Revert "Add -mno-shared to x86 assembler"
[deliverable/binutils-gdb.git] / gas / config / tc-i386.c
1 /* tc-i386.c -- Assemble code for the Intel 80386
2 Copyright (C) 1989-2015 Free Software Foundation, Inc.
3
4 This file is part of GAS, the GNU Assembler.
5
6 GAS is free software; you can redistribute it and/or modify
7 it under the terms of the GNU General Public License as published by
8 the Free Software Foundation; either version 3, or (at your option)
9 any later version.
10
11 GAS is distributed in the hope that it will be useful,
12 but WITHOUT ANY WARRANTY; without even the implied warranty of
13 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 GNU General Public License for more details.
15
16 You should have received a copy of the GNU General Public License
17 along with GAS; see the file COPYING. If not, write to the Free
18 Software Foundation, 51 Franklin Street - Fifth Floor, Boston, MA
19 02110-1301, USA. */
20
21 /* Intel 80386 machine specific gas.
22 Written by Eliot Dresselhaus (eliot@mgm.mit.edu).
23 x86_64 support by Jan Hubicka (jh@suse.cz)
24 VIA PadLock support by Michal Ludvig (mludvig@suse.cz)
25 Bugs & suggestions are completely welcome. This is free software.
26 Please help us make it better. */
27
28 #include "as.h"
29 #include "safe-ctype.h"
30 #include "subsegs.h"
31 #include "dwarf2dbg.h"
32 #include "dw2gencfi.h"
33 #include "elf/x86-64.h"
34 #include "opcodes/i386-init.h"
35
36 #ifdef TE_LINUX
37 /* Default to compress debug sections for Linux. */
38 enum compressed_debug_section_type flag_compress_debug
39 = COMPRESS_DEBUG_ZLIB;
40 #endif
41
42 #ifndef REGISTER_WARNINGS
43 #define REGISTER_WARNINGS 1
44 #endif
45
46 #ifndef INFER_ADDR_PREFIX
47 #define INFER_ADDR_PREFIX 1
48 #endif
49
50 #ifndef DEFAULT_ARCH
51 #define DEFAULT_ARCH "i386"
52 #endif
53
54 #ifndef INLINE
55 #if __GNUC__ >= 2
56 #define INLINE __inline__
57 #else
58 #define INLINE
59 #endif
60 #endif
61
62 /* Prefixes will be emitted in the order defined below.
63 WAIT_PREFIX must be the first prefix since FWAIT is really is an
64 instruction, and so must come before any prefixes.
65 The preferred prefix order is SEG_PREFIX, ADDR_PREFIX, DATA_PREFIX,
66 REP_PREFIX/HLE_PREFIX, LOCK_PREFIX. */
67 #define WAIT_PREFIX 0
68 #define SEG_PREFIX 1
69 #define ADDR_PREFIX 2
70 #define DATA_PREFIX 3
71 #define REP_PREFIX 4
72 #define HLE_PREFIX REP_PREFIX
73 #define BND_PREFIX REP_PREFIX
74 #define LOCK_PREFIX 5
75 #define REX_PREFIX 6 /* must come last. */
76 #define MAX_PREFIXES 7 /* max prefixes per opcode */
77
78 /* we define the syntax here (modulo base,index,scale syntax) */
79 #define REGISTER_PREFIX '%'
80 #define IMMEDIATE_PREFIX '$'
81 #define ABSOLUTE_PREFIX '*'
82
83 /* these are the instruction mnemonic suffixes in AT&T syntax or
84 memory operand size in Intel syntax. */
85 #define WORD_MNEM_SUFFIX 'w'
86 #define BYTE_MNEM_SUFFIX 'b'
87 #define SHORT_MNEM_SUFFIX 's'
88 #define LONG_MNEM_SUFFIX 'l'
89 #define QWORD_MNEM_SUFFIX 'q'
90 #define XMMWORD_MNEM_SUFFIX 'x'
91 #define YMMWORD_MNEM_SUFFIX 'y'
92 #define ZMMWORD_MNEM_SUFFIX 'z'
93 /* Intel Syntax. Use a non-ascii letter since since it never appears
94 in instructions. */
95 #define LONG_DOUBLE_MNEM_SUFFIX '\1'
96
97 #define END_OF_INSN '\0'
98
99 /*
100 'templates' is for grouping together 'template' structures for opcodes
101 of the same name. This is only used for storing the insns in the grand
102 ole hash table of insns.
103 The templates themselves start at START and range up to (but not including)
104 END.
105 */
106 typedef struct
107 {
108 const insn_template *start;
109 const insn_template *end;
110 }
111 templates;
112
113 /* 386 operand encoding bytes: see 386 book for details of this. */
114 typedef struct
115 {
116 unsigned int regmem; /* codes register or memory operand */
117 unsigned int reg; /* codes register operand (or extended opcode) */
118 unsigned int mode; /* how to interpret regmem & reg */
119 }
120 modrm_byte;
121
122 /* x86-64 extension prefix. */
123 typedef int rex_byte;
124
125 /* 386 opcode byte to code indirect addressing. */
126 typedef struct
127 {
128 unsigned base;
129 unsigned index;
130 unsigned scale;
131 }
132 sib_byte;
133
134 /* x86 arch names, types and features */
135 typedef struct
136 {
137 const char *name; /* arch name */
138 unsigned int len; /* arch string length */
139 enum processor_type type; /* arch type */
140 i386_cpu_flags flags; /* cpu feature flags */
141 unsigned int skip; /* show_arch should skip this. */
142 unsigned int negated; /* turn off indicated flags. */
143 }
144 arch_entry;
145
146 static void update_code_flag (int, int);
147 static void set_code_flag (int);
148 static void set_16bit_gcc_code_flag (int);
149 static void set_intel_syntax (int);
150 static void set_intel_mnemonic (int);
151 static void set_allow_index_reg (int);
152 static void set_check (int);
153 static void set_cpu_arch (int);
154 #ifdef TE_PE
155 static void pe_directive_secrel (int);
156 #endif
157 static void signed_cons (int);
158 static char *output_invalid (int c);
159 static int i386_finalize_immediate (segT, expressionS *, i386_operand_type,
160 const char *);
161 static int i386_finalize_displacement (segT, expressionS *, i386_operand_type,
162 const char *);
163 static int i386_att_operand (char *);
164 static int i386_intel_operand (char *, int);
165 static int i386_intel_simplify (expressionS *);
166 static int i386_intel_parse_name (const char *, expressionS *);
167 static const reg_entry *parse_register (char *, char **);
168 static char *parse_insn (char *, char *);
169 static char *parse_operands (char *, const char *);
170 static void swap_operands (void);
171 static void swap_2_operands (int, int);
172 static void optimize_imm (void);
173 static void optimize_disp (void);
174 static const insn_template *match_template (void);
175 static int check_string (void);
176 static int process_suffix (void);
177 static int check_byte_reg (void);
178 static int check_long_reg (void);
179 static int check_qword_reg (void);
180 static int check_word_reg (void);
181 static int finalize_imm (void);
182 static int process_operands (void);
183 static const seg_entry *build_modrm_byte (void);
184 static void output_insn (void);
185 static void output_imm (fragS *, offsetT);
186 static void output_disp (fragS *, offsetT);
187 #ifndef I386COFF
188 static void s_bss (int);
189 #endif
190 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
191 static void handle_large_common (int small ATTRIBUTE_UNUSED);
192 #endif
193
194 static const char *default_arch = DEFAULT_ARCH;
195
196 /* This struct describes rounding control and SAE in the instruction. */
197 struct RC_Operation
198 {
199 enum rc_type
200 {
201 rne = 0,
202 rd,
203 ru,
204 rz,
205 saeonly
206 } type;
207 int operand;
208 };
209
210 static struct RC_Operation rc_op;
211
212 /* The struct describes masking, applied to OPERAND in the instruction.
213 MASK is a pointer to the corresponding mask register. ZEROING tells
214 whether merging or zeroing mask is used. */
215 struct Mask_Operation
216 {
217 const reg_entry *mask;
218 unsigned int zeroing;
219 /* The operand where this operation is associated. */
220 int operand;
221 };
222
223 static struct Mask_Operation mask_op;
224
225 /* The struct describes broadcasting, applied to OPERAND. FACTOR is
226 broadcast factor. */
227 struct Broadcast_Operation
228 {
229 /* Type of broadcast: no broadcast, {1to8}, or {1to16}. */
230 int type;
231
232 /* Index of broadcasted operand. */
233 int operand;
234 };
235
236 static struct Broadcast_Operation broadcast_op;
237
238 /* VEX prefix. */
239 typedef struct
240 {
241 /* VEX prefix is either 2 byte or 3 byte. EVEX is 4 byte. */
242 unsigned char bytes[4];
243 unsigned int length;
244 /* Destination or source register specifier. */
245 const reg_entry *register_specifier;
246 } vex_prefix;
247
248 /* 'md_assemble ()' gathers together information and puts it into a
249 i386_insn. */
250
251 union i386_op
252 {
253 expressionS *disps;
254 expressionS *imms;
255 const reg_entry *regs;
256 };
257
258 enum i386_error
259 {
260 operand_size_mismatch,
261 operand_type_mismatch,
262 register_type_mismatch,
263 number_of_operands_mismatch,
264 invalid_instruction_suffix,
265 bad_imm4,
266 old_gcc_only,
267 unsupported_with_intel_mnemonic,
268 unsupported_syntax,
269 unsupported,
270 invalid_vsib_address,
271 invalid_vector_register_set,
272 unsupported_vector_index_register,
273 unsupported_broadcast,
274 broadcast_not_on_src_operand,
275 broadcast_needed,
276 unsupported_masking,
277 mask_not_on_destination,
278 no_default_mask,
279 unsupported_rc_sae,
280 rc_sae_operand_not_last_imm,
281 invalid_register_operand,
282 try_vector_disp8
283 };
284
285 struct _i386_insn
286 {
287 /* TM holds the template for the insn were currently assembling. */
288 insn_template tm;
289
290 /* SUFFIX holds the instruction size suffix for byte, word, dword
291 or qword, if given. */
292 char suffix;
293
294 /* OPERANDS gives the number of given operands. */
295 unsigned int operands;
296
297 /* REG_OPERANDS, DISP_OPERANDS, MEM_OPERANDS, IMM_OPERANDS give the number
298 of given register, displacement, memory operands and immediate
299 operands. */
300 unsigned int reg_operands, disp_operands, mem_operands, imm_operands;
301
302 /* TYPES [i] is the type (see above #defines) which tells us how to
303 use OP[i] for the corresponding operand. */
304 i386_operand_type types[MAX_OPERANDS];
305
306 /* Displacement expression, immediate expression, or register for each
307 operand. */
308 union i386_op op[MAX_OPERANDS];
309
310 /* Flags for operands. */
311 unsigned int flags[MAX_OPERANDS];
312 #define Operand_PCrel 1
313
314 /* Relocation type for operand */
315 enum bfd_reloc_code_real reloc[MAX_OPERANDS];
316
317 /* BASE_REG, INDEX_REG, and LOG2_SCALE_FACTOR are used to encode
318 the base index byte below. */
319 const reg_entry *base_reg;
320 const reg_entry *index_reg;
321 unsigned int log2_scale_factor;
322
323 /* SEG gives the seg_entries of this insn. They are zero unless
324 explicit segment overrides are given. */
325 const seg_entry *seg[2];
326
327 /* PREFIX holds all the given prefix opcodes (usually null).
328 PREFIXES is the number of prefix opcodes. */
329 unsigned int prefixes;
330 unsigned char prefix[MAX_PREFIXES];
331
332 /* RM and SIB are the modrm byte and the sib byte where the
333 addressing modes of this insn are encoded. */
334 modrm_byte rm;
335 rex_byte rex;
336 rex_byte vrex;
337 sib_byte sib;
338 vex_prefix vex;
339
340 /* Masking attributes. */
341 struct Mask_Operation *mask;
342
343 /* Rounding control and SAE attributes. */
344 struct RC_Operation *rounding;
345
346 /* Broadcasting attributes. */
347 struct Broadcast_Operation *broadcast;
348
349 /* Compressed disp8*N attribute. */
350 unsigned int memshift;
351
352 /* Swap operand in encoding. */
353 unsigned int swap_operand;
354
355 /* Prefer 8bit or 32bit displacement in encoding. */
356 enum
357 {
358 disp_encoding_default = 0,
359 disp_encoding_8bit,
360 disp_encoding_32bit
361 } disp_encoding;
362
363 /* REP prefix. */
364 const char *rep_prefix;
365
366 /* HLE prefix. */
367 const char *hle_prefix;
368
369 /* Have BND prefix. */
370 const char *bnd_prefix;
371
372 /* Need VREX to support upper 16 registers. */
373 int need_vrex;
374
375 /* Error message. */
376 enum i386_error error;
377 };
378
379 typedef struct _i386_insn i386_insn;
380
381 /* Link RC type with corresponding string, that'll be looked for in
382 asm. */
383 struct RC_name
384 {
385 enum rc_type type;
386 const char *name;
387 unsigned int len;
388 };
389
390 static const struct RC_name RC_NamesTable[] =
391 {
392 { rne, STRING_COMMA_LEN ("rn-sae") },
393 { rd, STRING_COMMA_LEN ("rd-sae") },
394 { ru, STRING_COMMA_LEN ("ru-sae") },
395 { rz, STRING_COMMA_LEN ("rz-sae") },
396 { saeonly, STRING_COMMA_LEN ("sae") },
397 };
398
399 /* List of chars besides those in app.c:symbol_chars that can start an
400 operand. Used to prevent the scrubber eating vital white-space. */
401 const char extra_symbol_chars[] = "*%-([{"
402 #ifdef LEX_AT
403 "@"
404 #endif
405 #ifdef LEX_QM
406 "?"
407 #endif
408 ;
409
410 #if (defined (TE_I386AIX) \
411 || ((defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)) \
412 && !defined (TE_GNU) \
413 && !defined (TE_LINUX) \
414 && !defined (TE_NACL) \
415 && !defined (TE_NETWARE) \
416 && !defined (TE_FreeBSD) \
417 && !defined (TE_DragonFly) \
418 && !defined (TE_NetBSD)))
419 /* This array holds the chars that always start a comment. If the
420 pre-processor is disabled, these aren't very useful. The option
421 --divide will remove '/' from this list. */
422 const char *i386_comment_chars = "#/";
423 #define SVR4_COMMENT_CHARS 1
424 #define PREFIX_SEPARATOR '\\'
425
426 #else
427 const char *i386_comment_chars = "#";
428 #define PREFIX_SEPARATOR '/'
429 #endif
430
431 /* This array holds the chars that only start a comment at the beginning of
432 a line. If the line seems to have the form '# 123 filename'
433 .line and .file directives will appear in the pre-processed output.
434 Note that input_file.c hand checks for '#' at the beginning of the
435 first line of the input file. This is because the compiler outputs
436 #NO_APP at the beginning of its output.
437 Also note that comments started like this one will always work if
438 '/' isn't otherwise defined. */
439 const char line_comment_chars[] = "#/";
440
441 const char line_separator_chars[] = ";";
442
443 /* Chars that can be used to separate mant from exp in floating point
444 nums. */
445 const char EXP_CHARS[] = "eE";
446
447 /* Chars that mean this number is a floating point constant
448 As in 0f12.456
449 or 0d1.2345e12. */
450 const char FLT_CHARS[] = "fFdDxX";
451
452 /* Tables for lexical analysis. */
453 static char mnemonic_chars[256];
454 static char register_chars[256];
455 static char operand_chars[256];
456 static char identifier_chars[256];
457 static char digit_chars[256];
458
459 /* Lexical macros. */
460 #define is_mnemonic_char(x) (mnemonic_chars[(unsigned char) x])
461 #define is_operand_char(x) (operand_chars[(unsigned char) x])
462 #define is_register_char(x) (register_chars[(unsigned char) x])
463 #define is_space_char(x) ((x) == ' ')
464 #define is_identifier_char(x) (identifier_chars[(unsigned char) x])
465 #define is_digit_char(x) (digit_chars[(unsigned char) x])
466
467 /* All non-digit non-letter characters that may occur in an operand. */
468 static char operand_special_chars[] = "%$-+(,)*._~/<>|&^!:[@]";
469
470 /* md_assemble() always leaves the strings it's passed unaltered. To
471 effect this we maintain a stack of saved characters that we've smashed
472 with '\0's (indicating end of strings for various sub-fields of the
473 assembler instruction). */
474 static char save_stack[32];
475 static char *save_stack_p;
476 #define END_STRING_AND_SAVE(s) \
477 do { *save_stack_p++ = *(s); *(s) = '\0'; } while (0)
478 #define RESTORE_END_STRING(s) \
479 do { *(s) = *--save_stack_p; } while (0)
480
481 /* The instruction we're assembling. */
482 static i386_insn i;
483
484 /* Possible templates for current insn. */
485 static const templates *current_templates;
486
487 /* Per instruction expressionS buffers: max displacements & immediates. */
488 static expressionS disp_expressions[MAX_MEMORY_OPERANDS];
489 static expressionS im_expressions[MAX_IMMEDIATE_OPERANDS];
490
491 /* Current operand we are working on. */
492 static int this_operand = -1;
493
494 /* We support four different modes. FLAG_CODE variable is used to distinguish
495 these. */
496
497 enum flag_code {
498 CODE_32BIT,
499 CODE_16BIT,
500 CODE_64BIT };
501
502 static enum flag_code flag_code;
503 static unsigned int object_64bit;
504 static unsigned int disallow_64bit_reloc;
505 static int use_rela_relocations = 0;
506
507 #if ((defined (OBJ_MAYBE_COFF) && defined (OBJ_MAYBE_AOUT)) \
508 || defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) \
509 || defined (TE_PE) || defined (TE_PEP) || defined (OBJ_MACH_O))
510
511 /* The ELF ABI to use. */
512 enum x86_elf_abi
513 {
514 I386_ABI,
515 X86_64_ABI,
516 X86_64_X32_ABI
517 };
518
519 static enum x86_elf_abi x86_elf_abi = I386_ABI;
520 #endif
521
522 #if defined (TE_PE) || defined (TE_PEP)
523 /* Use big object file format. */
524 static int use_big_obj = 0;
525 #endif
526
527 /* 1 for intel syntax,
528 0 if att syntax. */
529 static int intel_syntax = 0;
530
531 /* 1 for intel mnemonic,
532 0 if att mnemonic. */
533 static int intel_mnemonic = !SYSV386_COMPAT;
534
535 /* 1 if support old (<= 2.8.1) versions of gcc. */
536 static int old_gcc = OLDGCC_COMPAT;
537
538 /* 1 if pseudo registers are permitted. */
539 static int allow_pseudo_reg = 0;
540
541 /* 1 if register prefix % not required. */
542 static int allow_naked_reg = 0;
543
544 /* 1 if the assembler should add BND prefix for all control-tranferring
545 instructions supporting it, even if this prefix wasn't specified
546 explicitly. */
547 static int add_bnd_prefix = 0;
548
549 /* 1 if pseudo index register, eiz/riz, is allowed . */
550 static int allow_index_reg = 0;
551
552 /* 1 if the assembler should ignore LOCK prefix, even if it was
553 specified explicitly. */
554 static int omit_lock_prefix = 0;
555
556 static enum check_kind
557 {
558 check_none = 0,
559 check_warning,
560 check_error
561 }
562 sse_check, operand_check = check_warning;
563
564 /* Register prefix used for error message. */
565 static const char *register_prefix = "%";
566
567 /* Used in 16 bit gcc mode to add an l suffix to call, ret, enter,
568 leave, push, and pop instructions so that gcc has the same stack
569 frame as in 32 bit mode. */
570 static char stackop_size = '\0';
571
572 /* Non-zero to optimize code alignment. */
573 int optimize_align_code = 1;
574
575 /* Non-zero to quieten some warnings. */
576 static int quiet_warnings = 0;
577
578 /* CPU name. */
579 static const char *cpu_arch_name = NULL;
580 static char *cpu_sub_arch_name = NULL;
581
582 /* CPU feature flags. */
583 static i386_cpu_flags cpu_arch_flags = CPU_UNKNOWN_FLAGS;
584
585 /* If we have selected a cpu we are generating instructions for. */
586 static int cpu_arch_tune_set = 0;
587
588 /* Cpu we are generating instructions for. */
589 enum processor_type cpu_arch_tune = PROCESSOR_UNKNOWN;
590
591 /* CPU feature flags of cpu we are generating instructions for. */
592 static i386_cpu_flags cpu_arch_tune_flags;
593
594 /* CPU instruction set architecture used. */
595 enum processor_type cpu_arch_isa = PROCESSOR_UNKNOWN;
596
597 /* CPU feature flags of instruction set architecture used. */
598 i386_cpu_flags cpu_arch_isa_flags;
599
600 /* If set, conditional jumps are not automatically promoted to handle
601 larger than a byte offset. */
602 static unsigned int no_cond_jump_promotion = 0;
603
604 /* Encode SSE instructions with VEX prefix. */
605 static unsigned int sse2avx;
606
607 /* Encode scalar AVX instructions with specific vector length. */
608 static enum
609 {
610 vex128 = 0,
611 vex256
612 } avxscalar;
613
614 /* Encode scalar EVEX LIG instructions with specific vector length. */
615 static enum
616 {
617 evexl128 = 0,
618 evexl256,
619 evexl512
620 } evexlig;
621
622 /* Encode EVEX WIG instructions with specific evex.w. */
623 static enum
624 {
625 evexw0 = 0,
626 evexw1
627 } evexwig;
628
629 /* Value to encode in EVEX RC bits, for SAE-only instructions. */
630 static enum rc_type evexrcig = rne;
631
632 /* Pre-defined "_GLOBAL_OFFSET_TABLE_". */
633 static symbolS *GOT_symbol;
634
635 /* The dwarf2 return column, adjusted for 32 or 64 bit. */
636 unsigned int x86_dwarf2_return_column;
637
638 /* The dwarf2 data alignment, adjusted for 32 or 64 bit. */
639 int x86_cie_data_alignment;
640
641 /* Interface to relax_segment.
642 There are 3 major relax states for 386 jump insns because the
643 different types of jumps add different sizes to frags when we're
644 figuring out what sort of jump to choose to reach a given label. */
645
646 /* Types. */
647 #define UNCOND_JUMP 0
648 #define COND_JUMP 1
649 #define COND_JUMP86 2
650
651 /* Sizes. */
652 #define CODE16 1
653 #define SMALL 0
654 #define SMALL16 (SMALL | CODE16)
655 #define BIG 2
656 #define BIG16 (BIG | CODE16)
657
658 #ifndef INLINE
659 #ifdef __GNUC__
660 #define INLINE __inline__
661 #else
662 #define INLINE
663 #endif
664 #endif
665
666 #define ENCODE_RELAX_STATE(type, size) \
667 ((relax_substateT) (((type) << 2) | (size)))
668 #define TYPE_FROM_RELAX_STATE(s) \
669 ((s) >> 2)
670 #define DISP_SIZE_FROM_RELAX_STATE(s) \
671 ((((s) & 3) == BIG ? 4 : (((s) & 3) == BIG16 ? 2 : 1)))
672
673 /* This table is used by relax_frag to promote short jumps to long
674 ones where necessary. SMALL (short) jumps may be promoted to BIG
675 (32 bit long) ones, and SMALL16 jumps to BIG16 (16 bit long). We
676 don't allow a short jump in a 32 bit code segment to be promoted to
677 a 16 bit offset jump because it's slower (requires data size
678 prefix), and doesn't work, unless the destination is in the bottom
679 64k of the code segment (The top 16 bits of eip are zeroed). */
680
681 const relax_typeS md_relax_table[] =
682 {
683 /* The fields are:
684 1) most positive reach of this state,
685 2) most negative reach of this state,
686 3) how many bytes this mode will have in the variable part of the frag
687 4) which index into the table to try if we can't fit into this one. */
688
689 /* UNCOND_JUMP states. */
690 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (UNCOND_JUMP, BIG)},
691 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (UNCOND_JUMP, BIG16)},
692 /* dword jmp adds 4 bytes to frag:
693 0 extra opcode bytes, 4 displacement bytes. */
694 {0, 0, 4, 0},
695 /* word jmp adds 2 byte2 to frag:
696 0 extra opcode bytes, 2 displacement bytes. */
697 {0, 0, 2, 0},
698
699 /* COND_JUMP states. */
700 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (COND_JUMP, BIG)},
701 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (COND_JUMP, BIG16)},
702 /* dword conditionals adds 5 bytes to frag:
703 1 extra opcode byte, 4 displacement bytes. */
704 {0, 0, 5, 0},
705 /* word conditionals add 3 bytes to frag:
706 1 extra opcode byte, 2 displacement bytes. */
707 {0, 0, 3, 0},
708
709 /* COND_JUMP86 states. */
710 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (COND_JUMP86, BIG)},
711 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (COND_JUMP86, BIG16)},
712 /* dword conditionals adds 5 bytes to frag:
713 1 extra opcode byte, 4 displacement bytes. */
714 {0, 0, 5, 0},
715 /* word conditionals add 4 bytes to frag:
716 1 displacement byte and a 3 byte long branch insn. */
717 {0, 0, 4, 0}
718 };
719
720 static const arch_entry cpu_arch[] =
721 {
722 /* Do not replace the first two entries - i386_target_format()
723 relies on them being there in this order. */
724 { STRING_COMMA_LEN ("generic32"), PROCESSOR_GENERIC32,
725 CPU_GENERIC32_FLAGS, 0, 0 },
726 { STRING_COMMA_LEN ("generic64"), PROCESSOR_GENERIC64,
727 CPU_GENERIC64_FLAGS, 0, 0 },
728 { STRING_COMMA_LEN ("i8086"), PROCESSOR_UNKNOWN,
729 CPU_NONE_FLAGS, 0, 0 },
730 { STRING_COMMA_LEN ("i186"), PROCESSOR_UNKNOWN,
731 CPU_I186_FLAGS, 0, 0 },
732 { STRING_COMMA_LEN ("i286"), PROCESSOR_UNKNOWN,
733 CPU_I286_FLAGS, 0, 0 },
734 { STRING_COMMA_LEN ("i386"), PROCESSOR_I386,
735 CPU_I386_FLAGS, 0, 0 },
736 { STRING_COMMA_LEN ("i486"), PROCESSOR_I486,
737 CPU_I486_FLAGS, 0, 0 },
738 { STRING_COMMA_LEN ("i586"), PROCESSOR_PENTIUM,
739 CPU_I586_FLAGS, 0, 0 },
740 { STRING_COMMA_LEN ("i686"), PROCESSOR_PENTIUMPRO,
741 CPU_I686_FLAGS, 0, 0 },
742 { STRING_COMMA_LEN ("pentium"), PROCESSOR_PENTIUM,
743 CPU_I586_FLAGS, 0, 0 },
744 { STRING_COMMA_LEN ("pentiumpro"), PROCESSOR_PENTIUMPRO,
745 CPU_PENTIUMPRO_FLAGS, 0, 0 },
746 { STRING_COMMA_LEN ("pentiumii"), PROCESSOR_PENTIUMPRO,
747 CPU_P2_FLAGS, 0, 0 },
748 { STRING_COMMA_LEN ("pentiumiii"),PROCESSOR_PENTIUMPRO,
749 CPU_P3_FLAGS, 0, 0 },
750 { STRING_COMMA_LEN ("pentium4"), PROCESSOR_PENTIUM4,
751 CPU_P4_FLAGS, 0, 0 },
752 { STRING_COMMA_LEN ("prescott"), PROCESSOR_NOCONA,
753 CPU_CORE_FLAGS, 0, 0 },
754 { STRING_COMMA_LEN ("nocona"), PROCESSOR_NOCONA,
755 CPU_NOCONA_FLAGS, 0, 0 },
756 { STRING_COMMA_LEN ("yonah"), PROCESSOR_CORE,
757 CPU_CORE_FLAGS, 1, 0 },
758 { STRING_COMMA_LEN ("core"), PROCESSOR_CORE,
759 CPU_CORE_FLAGS, 0, 0 },
760 { STRING_COMMA_LEN ("merom"), PROCESSOR_CORE2,
761 CPU_CORE2_FLAGS, 1, 0 },
762 { STRING_COMMA_LEN ("core2"), PROCESSOR_CORE2,
763 CPU_CORE2_FLAGS, 0, 0 },
764 { STRING_COMMA_LEN ("corei7"), PROCESSOR_COREI7,
765 CPU_COREI7_FLAGS, 0, 0 },
766 { STRING_COMMA_LEN ("l1om"), PROCESSOR_L1OM,
767 CPU_L1OM_FLAGS, 0, 0 },
768 { STRING_COMMA_LEN ("k1om"), PROCESSOR_K1OM,
769 CPU_K1OM_FLAGS, 0, 0 },
770 { STRING_COMMA_LEN ("iamcu"), PROCESSOR_IAMCU,
771 CPU_IAMCU_FLAGS, 0, 0 },
772 { STRING_COMMA_LEN ("k6"), PROCESSOR_K6,
773 CPU_K6_FLAGS, 0, 0 },
774 { STRING_COMMA_LEN ("k6_2"), PROCESSOR_K6,
775 CPU_K6_2_FLAGS, 0, 0 },
776 { STRING_COMMA_LEN ("athlon"), PROCESSOR_ATHLON,
777 CPU_ATHLON_FLAGS, 0, 0 },
778 { STRING_COMMA_LEN ("sledgehammer"), PROCESSOR_K8,
779 CPU_K8_FLAGS, 1, 0 },
780 { STRING_COMMA_LEN ("opteron"), PROCESSOR_K8,
781 CPU_K8_FLAGS, 0, 0 },
782 { STRING_COMMA_LEN ("k8"), PROCESSOR_K8,
783 CPU_K8_FLAGS, 0, 0 },
784 { STRING_COMMA_LEN ("amdfam10"), PROCESSOR_AMDFAM10,
785 CPU_AMDFAM10_FLAGS, 0, 0 },
786 { STRING_COMMA_LEN ("bdver1"), PROCESSOR_BD,
787 CPU_BDVER1_FLAGS, 0, 0 },
788 { STRING_COMMA_LEN ("bdver2"), PROCESSOR_BD,
789 CPU_BDVER2_FLAGS, 0, 0 },
790 { STRING_COMMA_LEN ("bdver3"), PROCESSOR_BD,
791 CPU_BDVER3_FLAGS, 0, 0 },
792 { STRING_COMMA_LEN ("bdver4"), PROCESSOR_BD,
793 CPU_BDVER4_FLAGS, 0, 0 },
794 { STRING_COMMA_LEN ("znver1"), PROCESSOR_ZNVER,
795 CPU_ZNVER1_FLAGS, 0, 0 },
796 { STRING_COMMA_LEN ("btver1"), PROCESSOR_BT,
797 CPU_BTVER1_FLAGS, 0, 0 },
798 { STRING_COMMA_LEN ("btver2"), PROCESSOR_BT,
799 CPU_BTVER2_FLAGS, 0, 0 },
800 { STRING_COMMA_LEN (".8087"), PROCESSOR_UNKNOWN,
801 CPU_8087_FLAGS, 0, 0 },
802 { STRING_COMMA_LEN (".287"), PROCESSOR_UNKNOWN,
803 CPU_287_FLAGS, 0, 0 },
804 { STRING_COMMA_LEN (".387"), PROCESSOR_UNKNOWN,
805 CPU_387_FLAGS, 0, 0 },
806 { STRING_COMMA_LEN (".no87"), PROCESSOR_UNKNOWN,
807 CPU_ANY87_FLAGS, 0, 1 },
808 { STRING_COMMA_LEN (".mmx"), PROCESSOR_UNKNOWN,
809 CPU_MMX_FLAGS, 0, 0 },
810 { STRING_COMMA_LEN (".nommx"), PROCESSOR_UNKNOWN,
811 CPU_3DNOWA_FLAGS, 0, 1 },
812 { STRING_COMMA_LEN (".sse"), PROCESSOR_UNKNOWN,
813 CPU_SSE_FLAGS, 0, 0 },
814 { STRING_COMMA_LEN (".sse2"), PROCESSOR_UNKNOWN,
815 CPU_SSE2_FLAGS, 0, 0 },
816 { STRING_COMMA_LEN (".sse3"), PROCESSOR_UNKNOWN,
817 CPU_SSE3_FLAGS, 0, 0 },
818 { STRING_COMMA_LEN (".ssse3"), PROCESSOR_UNKNOWN,
819 CPU_SSSE3_FLAGS, 0, 0 },
820 { STRING_COMMA_LEN (".sse4.1"), PROCESSOR_UNKNOWN,
821 CPU_SSE4_1_FLAGS, 0, 0 },
822 { STRING_COMMA_LEN (".sse4.2"), PROCESSOR_UNKNOWN,
823 CPU_SSE4_2_FLAGS, 0, 0 },
824 { STRING_COMMA_LEN (".sse4"), PROCESSOR_UNKNOWN,
825 CPU_SSE4_2_FLAGS, 0, 0 },
826 { STRING_COMMA_LEN (".nosse"), PROCESSOR_UNKNOWN,
827 CPU_ANY_SSE_FLAGS, 0, 1 },
828 { STRING_COMMA_LEN (".avx"), PROCESSOR_UNKNOWN,
829 CPU_AVX_FLAGS, 0, 0 },
830 { STRING_COMMA_LEN (".avx2"), PROCESSOR_UNKNOWN,
831 CPU_AVX2_FLAGS, 0, 0 },
832 { STRING_COMMA_LEN (".avx512f"), PROCESSOR_UNKNOWN,
833 CPU_AVX512F_FLAGS, 0, 0 },
834 { STRING_COMMA_LEN (".avx512cd"), PROCESSOR_UNKNOWN,
835 CPU_AVX512CD_FLAGS, 0, 0 },
836 { STRING_COMMA_LEN (".avx512er"), PROCESSOR_UNKNOWN,
837 CPU_AVX512ER_FLAGS, 0, 0 },
838 { STRING_COMMA_LEN (".avx512pf"), PROCESSOR_UNKNOWN,
839 CPU_AVX512PF_FLAGS, 0, 0 },
840 { STRING_COMMA_LEN (".avx512dq"), PROCESSOR_UNKNOWN,
841 CPU_AVX512DQ_FLAGS, 0, 0 },
842 { STRING_COMMA_LEN (".avx512bw"), PROCESSOR_UNKNOWN,
843 CPU_AVX512BW_FLAGS, 0, 0 },
844 { STRING_COMMA_LEN (".avx512vl"), PROCESSOR_UNKNOWN,
845 CPU_AVX512VL_FLAGS, 0, 0 },
846 { STRING_COMMA_LEN (".noavx"), PROCESSOR_UNKNOWN,
847 CPU_ANY_AVX_FLAGS, 0, 1 },
848 { STRING_COMMA_LEN (".vmx"), PROCESSOR_UNKNOWN,
849 CPU_VMX_FLAGS, 0, 0 },
850 { STRING_COMMA_LEN (".vmfunc"), PROCESSOR_UNKNOWN,
851 CPU_VMFUNC_FLAGS, 0, 0 },
852 { STRING_COMMA_LEN (".smx"), PROCESSOR_UNKNOWN,
853 CPU_SMX_FLAGS, 0, 0 },
854 { STRING_COMMA_LEN (".xsave"), PROCESSOR_UNKNOWN,
855 CPU_XSAVE_FLAGS, 0, 0 },
856 { STRING_COMMA_LEN (".xsaveopt"), PROCESSOR_UNKNOWN,
857 CPU_XSAVEOPT_FLAGS, 0, 0 },
858 { STRING_COMMA_LEN (".xsavec"), PROCESSOR_UNKNOWN,
859 CPU_XSAVEC_FLAGS, 0, 0 },
860 { STRING_COMMA_LEN (".xsaves"), PROCESSOR_UNKNOWN,
861 CPU_XSAVES_FLAGS, 0, 0 },
862 { STRING_COMMA_LEN (".aes"), PROCESSOR_UNKNOWN,
863 CPU_AES_FLAGS, 0, 0 },
864 { STRING_COMMA_LEN (".pclmul"), PROCESSOR_UNKNOWN,
865 CPU_PCLMUL_FLAGS, 0, 0 },
866 { STRING_COMMA_LEN (".clmul"), PROCESSOR_UNKNOWN,
867 CPU_PCLMUL_FLAGS, 1, 0 },
868 { STRING_COMMA_LEN (".fsgsbase"), PROCESSOR_UNKNOWN,
869 CPU_FSGSBASE_FLAGS, 0, 0 },
870 { STRING_COMMA_LEN (".rdrnd"), PROCESSOR_UNKNOWN,
871 CPU_RDRND_FLAGS, 0, 0 },
872 { STRING_COMMA_LEN (".f16c"), PROCESSOR_UNKNOWN,
873 CPU_F16C_FLAGS, 0, 0 },
874 { STRING_COMMA_LEN (".bmi2"), PROCESSOR_UNKNOWN,
875 CPU_BMI2_FLAGS, 0, 0 },
876 { STRING_COMMA_LEN (".fma"), PROCESSOR_UNKNOWN,
877 CPU_FMA_FLAGS, 0, 0 },
878 { STRING_COMMA_LEN (".fma4"), PROCESSOR_UNKNOWN,
879 CPU_FMA4_FLAGS, 0, 0 },
880 { STRING_COMMA_LEN (".xop"), PROCESSOR_UNKNOWN,
881 CPU_XOP_FLAGS, 0, 0 },
882 { STRING_COMMA_LEN (".lwp"), PROCESSOR_UNKNOWN,
883 CPU_LWP_FLAGS, 0, 0 },
884 { STRING_COMMA_LEN (".movbe"), PROCESSOR_UNKNOWN,
885 CPU_MOVBE_FLAGS, 0, 0 },
886 { STRING_COMMA_LEN (".cx16"), PROCESSOR_UNKNOWN,
887 CPU_CX16_FLAGS, 0, 0 },
888 { STRING_COMMA_LEN (".ept"), PROCESSOR_UNKNOWN,
889 CPU_EPT_FLAGS, 0, 0 },
890 { STRING_COMMA_LEN (".lzcnt"), PROCESSOR_UNKNOWN,
891 CPU_LZCNT_FLAGS, 0, 0 },
892 { STRING_COMMA_LEN (".hle"), PROCESSOR_UNKNOWN,
893 CPU_HLE_FLAGS, 0, 0 },
894 { STRING_COMMA_LEN (".rtm"), PROCESSOR_UNKNOWN,
895 CPU_RTM_FLAGS, 0, 0 },
896 { STRING_COMMA_LEN (".invpcid"), PROCESSOR_UNKNOWN,
897 CPU_INVPCID_FLAGS, 0, 0 },
898 { STRING_COMMA_LEN (".clflush"), PROCESSOR_UNKNOWN,
899 CPU_CLFLUSH_FLAGS, 0, 0 },
900 { STRING_COMMA_LEN (".nop"), PROCESSOR_UNKNOWN,
901 CPU_NOP_FLAGS, 0, 0 },
902 { STRING_COMMA_LEN (".syscall"), PROCESSOR_UNKNOWN,
903 CPU_SYSCALL_FLAGS, 0, 0 },
904 { STRING_COMMA_LEN (".rdtscp"), PROCESSOR_UNKNOWN,
905 CPU_RDTSCP_FLAGS, 0, 0 },
906 { STRING_COMMA_LEN (".3dnow"), PROCESSOR_UNKNOWN,
907 CPU_3DNOW_FLAGS, 0, 0 },
908 { STRING_COMMA_LEN (".3dnowa"), PROCESSOR_UNKNOWN,
909 CPU_3DNOWA_FLAGS, 0, 0 },
910 { STRING_COMMA_LEN (".padlock"), PROCESSOR_UNKNOWN,
911 CPU_PADLOCK_FLAGS, 0, 0 },
912 { STRING_COMMA_LEN (".pacifica"), PROCESSOR_UNKNOWN,
913 CPU_SVME_FLAGS, 1, 0 },
914 { STRING_COMMA_LEN (".svme"), PROCESSOR_UNKNOWN,
915 CPU_SVME_FLAGS, 0, 0 },
916 { STRING_COMMA_LEN (".sse4a"), PROCESSOR_UNKNOWN,
917 CPU_SSE4A_FLAGS, 0, 0 },
918 { STRING_COMMA_LEN (".abm"), PROCESSOR_UNKNOWN,
919 CPU_ABM_FLAGS, 0, 0 },
920 { STRING_COMMA_LEN (".bmi"), PROCESSOR_UNKNOWN,
921 CPU_BMI_FLAGS, 0, 0 },
922 { STRING_COMMA_LEN (".tbm"), PROCESSOR_UNKNOWN,
923 CPU_TBM_FLAGS, 0, 0 },
924 { STRING_COMMA_LEN (".adx"), PROCESSOR_UNKNOWN,
925 CPU_ADX_FLAGS, 0, 0 },
926 { STRING_COMMA_LEN (".rdseed"), PROCESSOR_UNKNOWN,
927 CPU_RDSEED_FLAGS, 0, 0 },
928 { STRING_COMMA_LEN (".prfchw"), PROCESSOR_UNKNOWN,
929 CPU_PRFCHW_FLAGS, 0, 0 },
930 { STRING_COMMA_LEN (".smap"), PROCESSOR_UNKNOWN,
931 CPU_SMAP_FLAGS, 0, 0 },
932 { STRING_COMMA_LEN (".mpx"), PROCESSOR_UNKNOWN,
933 CPU_MPX_FLAGS, 0, 0 },
934 { STRING_COMMA_LEN (".sha"), PROCESSOR_UNKNOWN,
935 CPU_SHA_FLAGS, 0, 0 },
936 { STRING_COMMA_LEN (".clflushopt"), PROCESSOR_UNKNOWN,
937 CPU_CLFLUSHOPT_FLAGS, 0, 0 },
938 { STRING_COMMA_LEN (".prefetchwt1"), PROCESSOR_UNKNOWN,
939 CPU_PREFETCHWT1_FLAGS, 0, 0 },
940 { STRING_COMMA_LEN (".se1"), PROCESSOR_UNKNOWN,
941 CPU_SE1_FLAGS, 0, 0 },
942 { STRING_COMMA_LEN (".clwb"), PROCESSOR_UNKNOWN,
943 CPU_CLWB_FLAGS, 0, 0 },
944 { STRING_COMMA_LEN (".pcommit"), PROCESSOR_UNKNOWN,
945 CPU_PCOMMIT_FLAGS, 0, 0 },
946 { STRING_COMMA_LEN (".avx512ifma"), PROCESSOR_UNKNOWN,
947 CPU_AVX512IFMA_FLAGS, 0, 0 },
948 { STRING_COMMA_LEN (".avx512vbmi"), PROCESSOR_UNKNOWN,
949 CPU_AVX512VBMI_FLAGS, 0, 0 },
950 { STRING_COMMA_LEN (".clzero"), PROCESSOR_UNKNOWN,
951 CPU_CLZERO_FLAGS, 0, 0 },
952 };
953
954 #ifdef I386COFF
955 /* Like s_lcomm_internal in gas/read.c but the alignment string
956 is allowed to be optional. */
957
958 static symbolS *
959 pe_lcomm_internal (int needs_align, symbolS *symbolP, addressT size)
960 {
961 addressT align = 0;
962
963 SKIP_WHITESPACE ();
964
965 if (needs_align
966 && *input_line_pointer == ',')
967 {
968 align = parse_align (needs_align - 1);
969
970 if (align == (addressT) -1)
971 return NULL;
972 }
973 else
974 {
975 if (size >= 8)
976 align = 3;
977 else if (size >= 4)
978 align = 2;
979 else if (size >= 2)
980 align = 1;
981 else
982 align = 0;
983 }
984
985 bss_alloc (symbolP, size, align);
986 return symbolP;
987 }
988
989 static void
990 pe_lcomm (int needs_align)
991 {
992 s_comm_internal (needs_align * 2, pe_lcomm_internal);
993 }
994 #endif
995
996 const pseudo_typeS md_pseudo_table[] =
997 {
998 #if !defined(OBJ_AOUT) && !defined(USE_ALIGN_PTWO)
999 {"align", s_align_bytes, 0},
1000 #else
1001 {"align", s_align_ptwo, 0},
1002 #endif
1003 {"arch", set_cpu_arch, 0},
1004 #ifndef I386COFF
1005 {"bss", s_bss, 0},
1006 #else
1007 {"lcomm", pe_lcomm, 1},
1008 #endif
1009 {"ffloat", float_cons, 'f'},
1010 {"dfloat", float_cons, 'd'},
1011 {"tfloat", float_cons, 'x'},
1012 {"value", cons, 2},
1013 {"slong", signed_cons, 4},
1014 {"noopt", s_ignore, 0},
1015 {"optim", s_ignore, 0},
1016 {"code16gcc", set_16bit_gcc_code_flag, CODE_16BIT},
1017 {"code16", set_code_flag, CODE_16BIT},
1018 {"code32", set_code_flag, CODE_32BIT},
1019 {"code64", set_code_flag, CODE_64BIT},
1020 {"intel_syntax", set_intel_syntax, 1},
1021 {"att_syntax", set_intel_syntax, 0},
1022 {"intel_mnemonic", set_intel_mnemonic, 1},
1023 {"att_mnemonic", set_intel_mnemonic, 0},
1024 {"allow_index_reg", set_allow_index_reg, 1},
1025 {"disallow_index_reg", set_allow_index_reg, 0},
1026 {"sse_check", set_check, 0},
1027 {"operand_check", set_check, 1},
1028 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
1029 {"largecomm", handle_large_common, 0},
1030 #else
1031 {"file", (void (*) (int)) dwarf2_directive_file, 0},
1032 {"loc", dwarf2_directive_loc, 0},
1033 {"loc_mark_labels", dwarf2_directive_loc_mark_labels, 0},
1034 #endif
1035 #ifdef TE_PE
1036 {"secrel32", pe_directive_secrel, 0},
1037 #endif
1038 {0, 0, 0}
1039 };
1040
1041 /* For interface with expression (). */
1042 extern char *input_line_pointer;
1043
1044 /* Hash table for instruction mnemonic lookup. */
1045 static struct hash_control *op_hash;
1046
1047 /* Hash table for register lookup. */
1048 static struct hash_control *reg_hash;
1049 \f
1050 void
1051 i386_align_code (fragS *fragP, int count)
1052 {
1053 /* Various efficient no-op patterns for aligning code labels.
1054 Note: Don't try to assemble the instructions in the comments.
1055 0L and 0w are not legal. */
1056 static const char f32_1[] =
1057 {0x90}; /* nop */
1058 static const char f32_2[] =
1059 {0x66,0x90}; /* xchg %ax,%ax */
1060 static const char f32_3[] =
1061 {0x8d,0x76,0x00}; /* leal 0(%esi),%esi */
1062 static const char f32_4[] =
1063 {0x8d,0x74,0x26,0x00}; /* leal 0(%esi,1),%esi */
1064 static const char f32_5[] =
1065 {0x90, /* nop */
1066 0x8d,0x74,0x26,0x00}; /* leal 0(%esi,1),%esi */
1067 static const char f32_6[] =
1068 {0x8d,0xb6,0x00,0x00,0x00,0x00}; /* leal 0L(%esi),%esi */
1069 static const char f32_7[] =
1070 {0x8d,0xb4,0x26,0x00,0x00,0x00,0x00}; /* leal 0L(%esi,1),%esi */
1071 static const char f32_8[] =
1072 {0x90, /* nop */
1073 0x8d,0xb4,0x26,0x00,0x00,0x00,0x00}; /* leal 0L(%esi,1),%esi */
1074 static const char f32_9[] =
1075 {0x89,0xf6, /* movl %esi,%esi */
1076 0x8d,0xbc,0x27,0x00,0x00,0x00,0x00}; /* leal 0L(%edi,1),%edi */
1077 static const char f32_10[] =
1078 {0x8d,0x76,0x00, /* leal 0(%esi),%esi */
1079 0x8d,0xbc,0x27,0x00,0x00,0x00,0x00}; /* leal 0L(%edi,1),%edi */
1080 static const char f32_11[] =
1081 {0x8d,0x74,0x26,0x00, /* leal 0(%esi,1),%esi */
1082 0x8d,0xbc,0x27,0x00,0x00,0x00,0x00}; /* leal 0L(%edi,1),%edi */
1083 static const char f32_12[] =
1084 {0x8d,0xb6,0x00,0x00,0x00,0x00, /* leal 0L(%esi),%esi */
1085 0x8d,0xbf,0x00,0x00,0x00,0x00}; /* leal 0L(%edi),%edi */
1086 static const char f32_13[] =
1087 {0x8d,0xb6,0x00,0x00,0x00,0x00, /* leal 0L(%esi),%esi */
1088 0x8d,0xbc,0x27,0x00,0x00,0x00,0x00}; /* leal 0L(%edi,1),%edi */
1089 static const char f32_14[] =
1090 {0x8d,0xb4,0x26,0x00,0x00,0x00,0x00, /* leal 0L(%esi,1),%esi */
1091 0x8d,0xbc,0x27,0x00,0x00,0x00,0x00}; /* leal 0L(%edi,1),%edi */
1092 static const char f16_3[] =
1093 {0x8d,0x74,0x00}; /* lea 0(%esi),%esi */
1094 static const char f16_4[] =
1095 {0x8d,0xb4,0x00,0x00}; /* lea 0w(%si),%si */
1096 static const char f16_5[] =
1097 {0x90, /* nop */
1098 0x8d,0xb4,0x00,0x00}; /* lea 0w(%si),%si */
1099 static const char f16_6[] =
1100 {0x89,0xf6, /* mov %si,%si */
1101 0x8d,0xbd,0x00,0x00}; /* lea 0w(%di),%di */
1102 static const char f16_7[] =
1103 {0x8d,0x74,0x00, /* lea 0(%si),%si */
1104 0x8d,0xbd,0x00,0x00}; /* lea 0w(%di),%di */
1105 static const char f16_8[] =
1106 {0x8d,0xb4,0x00,0x00, /* lea 0w(%si),%si */
1107 0x8d,0xbd,0x00,0x00}; /* lea 0w(%di),%di */
1108 static const char jump_31[] =
1109 {0xeb,0x1d,0x90,0x90,0x90,0x90,0x90, /* jmp .+31; lotsa nops */
1110 0x90,0x90,0x90,0x90,0x90,0x90,0x90,0x90,
1111 0x90,0x90,0x90,0x90,0x90,0x90,0x90,0x90,
1112 0x90,0x90,0x90,0x90,0x90,0x90,0x90,0x90};
1113 static const char *const f32_patt[] = {
1114 f32_1, f32_2, f32_3, f32_4, f32_5, f32_6, f32_7, f32_8,
1115 f32_9, f32_10, f32_11, f32_12, f32_13, f32_14
1116 };
1117 static const char *const f16_patt[] = {
1118 f32_1, f32_2, f16_3, f16_4, f16_5, f16_6, f16_7, f16_8
1119 };
1120 /* nopl (%[re]ax) */
1121 static const char alt_3[] =
1122 {0x0f,0x1f,0x00};
1123 /* nopl 0(%[re]ax) */
1124 static const char alt_4[] =
1125 {0x0f,0x1f,0x40,0x00};
1126 /* nopl 0(%[re]ax,%[re]ax,1) */
1127 static const char alt_5[] =
1128 {0x0f,0x1f,0x44,0x00,0x00};
1129 /* nopw 0(%[re]ax,%[re]ax,1) */
1130 static const char alt_6[] =
1131 {0x66,0x0f,0x1f,0x44,0x00,0x00};
1132 /* nopl 0L(%[re]ax) */
1133 static const char alt_7[] =
1134 {0x0f,0x1f,0x80,0x00,0x00,0x00,0x00};
1135 /* nopl 0L(%[re]ax,%[re]ax,1) */
1136 static const char alt_8[] =
1137 {0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
1138 /* nopw 0L(%[re]ax,%[re]ax,1) */
1139 static const char alt_9[] =
1140 {0x66,0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
1141 /* nopw %cs:0L(%[re]ax,%[re]ax,1) */
1142 static const char alt_10[] =
1143 {0x66,0x2e,0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
1144 static const char *const alt_patt[] = {
1145 f32_1, f32_2, alt_3, alt_4, alt_5, alt_6, alt_7, alt_8,
1146 alt_9, alt_10
1147 };
1148
1149 /* Only align for at least a positive non-zero boundary. */
1150 if (count <= 0 || count > MAX_MEM_FOR_RS_ALIGN_CODE)
1151 return;
1152
1153 /* We need to decide which NOP sequence to use for 32bit and
1154 64bit. When -mtune= is used:
1155
1156 1. For PROCESSOR_I386, PROCESSOR_I486, PROCESSOR_PENTIUM and
1157 PROCESSOR_GENERIC32, f32_patt will be used.
1158 2. For the rest, alt_patt will be used.
1159
1160 When -mtune= isn't used, alt_patt will be used if
1161 cpu_arch_isa_flags has CpuNop. Otherwise, f32_patt will
1162 be used.
1163
1164 When -march= or .arch is used, we can't use anything beyond
1165 cpu_arch_isa_flags. */
1166
1167 if (flag_code == CODE_16BIT)
1168 {
1169 if (count > 8)
1170 {
1171 memcpy (fragP->fr_literal + fragP->fr_fix,
1172 jump_31, count);
1173 /* Adjust jump offset. */
1174 fragP->fr_literal[fragP->fr_fix + 1] = count - 2;
1175 }
1176 else
1177 memcpy (fragP->fr_literal + fragP->fr_fix,
1178 f16_patt[count - 1], count);
1179 }
1180 else
1181 {
1182 const char *const *patt = NULL;
1183
1184 if (fragP->tc_frag_data.isa == PROCESSOR_UNKNOWN)
1185 {
1186 /* PROCESSOR_UNKNOWN means that all ISAs may be used. */
1187 switch (cpu_arch_tune)
1188 {
1189 case PROCESSOR_UNKNOWN:
1190 /* We use cpu_arch_isa_flags to check if we SHOULD
1191 optimize with nops. */
1192 if (fragP->tc_frag_data.isa_flags.bitfield.cpunop)
1193 patt = alt_patt;
1194 else
1195 patt = f32_patt;
1196 break;
1197 case PROCESSOR_PENTIUM4:
1198 case PROCESSOR_NOCONA:
1199 case PROCESSOR_CORE:
1200 case PROCESSOR_CORE2:
1201 case PROCESSOR_COREI7:
1202 case PROCESSOR_L1OM:
1203 case PROCESSOR_K1OM:
1204 case PROCESSOR_GENERIC64:
1205 case PROCESSOR_K6:
1206 case PROCESSOR_ATHLON:
1207 case PROCESSOR_K8:
1208 case PROCESSOR_AMDFAM10:
1209 case PROCESSOR_BD:
1210 case PROCESSOR_ZNVER:
1211 case PROCESSOR_BT:
1212 patt = alt_patt;
1213 break;
1214 case PROCESSOR_I386:
1215 case PROCESSOR_I486:
1216 case PROCESSOR_PENTIUM:
1217 case PROCESSOR_PENTIUMPRO:
1218 case PROCESSOR_IAMCU:
1219 case PROCESSOR_GENERIC32:
1220 patt = f32_patt;
1221 break;
1222 }
1223 }
1224 else
1225 {
1226 switch (fragP->tc_frag_data.tune)
1227 {
1228 case PROCESSOR_UNKNOWN:
1229 /* When cpu_arch_isa is set, cpu_arch_tune shouldn't be
1230 PROCESSOR_UNKNOWN. */
1231 abort ();
1232 break;
1233
1234 case PROCESSOR_I386:
1235 case PROCESSOR_I486:
1236 case PROCESSOR_PENTIUM:
1237 case PROCESSOR_IAMCU:
1238 case PROCESSOR_K6:
1239 case PROCESSOR_ATHLON:
1240 case PROCESSOR_K8:
1241 case PROCESSOR_AMDFAM10:
1242 case PROCESSOR_BD:
1243 case PROCESSOR_ZNVER:
1244 case PROCESSOR_BT:
1245 case PROCESSOR_GENERIC32:
1246 /* We use cpu_arch_isa_flags to check if we CAN optimize
1247 with nops. */
1248 if (fragP->tc_frag_data.isa_flags.bitfield.cpunop)
1249 patt = alt_patt;
1250 else
1251 patt = f32_patt;
1252 break;
1253 case PROCESSOR_PENTIUMPRO:
1254 case PROCESSOR_PENTIUM4:
1255 case PROCESSOR_NOCONA:
1256 case PROCESSOR_CORE:
1257 case PROCESSOR_CORE2:
1258 case PROCESSOR_COREI7:
1259 case PROCESSOR_L1OM:
1260 case PROCESSOR_K1OM:
1261 if (fragP->tc_frag_data.isa_flags.bitfield.cpunop)
1262 patt = alt_patt;
1263 else
1264 patt = f32_patt;
1265 break;
1266 case PROCESSOR_GENERIC64:
1267 patt = alt_patt;
1268 break;
1269 }
1270 }
1271
1272 if (patt == f32_patt)
1273 {
1274 /* If the padding is less than 15 bytes, we use the normal
1275 ones. Otherwise, we use a jump instruction and adjust
1276 its offset. */
1277 int limit;
1278
1279 /* For 64bit, the limit is 3 bytes. */
1280 if (flag_code == CODE_64BIT
1281 && fragP->tc_frag_data.isa_flags.bitfield.cpulm)
1282 limit = 3;
1283 else
1284 limit = 15;
1285 if (count < limit)
1286 memcpy (fragP->fr_literal + fragP->fr_fix,
1287 patt[count - 1], count);
1288 else
1289 {
1290 memcpy (fragP->fr_literal + fragP->fr_fix,
1291 jump_31, count);
1292 /* Adjust jump offset. */
1293 fragP->fr_literal[fragP->fr_fix + 1] = count - 2;
1294 }
1295 }
1296 else
1297 {
1298 /* Maximum length of an instruction is 10 byte. If the
1299 padding is greater than 10 bytes and we don't use jump,
1300 we have to break it into smaller pieces. */
1301 int padding = count;
1302 while (padding > 10)
1303 {
1304 padding -= 10;
1305 memcpy (fragP->fr_literal + fragP->fr_fix + padding,
1306 patt [9], 10);
1307 }
1308
1309 if (padding)
1310 memcpy (fragP->fr_literal + fragP->fr_fix,
1311 patt [padding - 1], padding);
1312 }
1313 }
1314 fragP->fr_var = count;
1315 }
1316
1317 static INLINE int
1318 operand_type_all_zero (const union i386_operand_type *x)
1319 {
1320 switch (ARRAY_SIZE(x->array))
1321 {
1322 case 3:
1323 if (x->array[2])
1324 return 0;
1325 case 2:
1326 if (x->array[1])
1327 return 0;
1328 case 1:
1329 return !x->array[0];
1330 default:
1331 abort ();
1332 }
1333 }
1334
1335 static INLINE void
1336 operand_type_set (union i386_operand_type *x, unsigned int v)
1337 {
1338 switch (ARRAY_SIZE(x->array))
1339 {
1340 case 3:
1341 x->array[2] = v;
1342 case 2:
1343 x->array[1] = v;
1344 case 1:
1345 x->array[0] = v;
1346 break;
1347 default:
1348 abort ();
1349 }
1350 }
1351
1352 static INLINE int
1353 operand_type_equal (const union i386_operand_type *x,
1354 const union i386_operand_type *y)
1355 {
1356 switch (ARRAY_SIZE(x->array))
1357 {
1358 case 3:
1359 if (x->array[2] != y->array[2])
1360 return 0;
1361 case 2:
1362 if (x->array[1] != y->array[1])
1363 return 0;
1364 case 1:
1365 return x->array[0] == y->array[0];
1366 break;
1367 default:
1368 abort ();
1369 }
1370 }
1371
1372 static INLINE int
1373 cpu_flags_all_zero (const union i386_cpu_flags *x)
1374 {
1375 switch (ARRAY_SIZE(x->array))
1376 {
1377 case 3:
1378 if (x->array[2])
1379 return 0;
1380 case 2:
1381 if (x->array[1])
1382 return 0;
1383 case 1:
1384 return !x->array[0];
1385 default:
1386 abort ();
1387 }
1388 }
1389
1390 static INLINE int
1391 cpu_flags_equal (const union i386_cpu_flags *x,
1392 const union i386_cpu_flags *y)
1393 {
1394 switch (ARRAY_SIZE(x->array))
1395 {
1396 case 3:
1397 if (x->array[2] != y->array[2])
1398 return 0;
1399 case 2:
1400 if (x->array[1] != y->array[1])
1401 return 0;
1402 case 1:
1403 return x->array[0] == y->array[0];
1404 break;
1405 default:
1406 abort ();
1407 }
1408 }
1409
1410 static INLINE int
1411 cpu_flags_check_cpu64 (i386_cpu_flags f)
1412 {
1413 return !((flag_code == CODE_64BIT && f.bitfield.cpuno64)
1414 || (flag_code != CODE_64BIT && f.bitfield.cpu64));
1415 }
1416
1417 static INLINE i386_cpu_flags
1418 cpu_flags_and (i386_cpu_flags x, i386_cpu_flags y)
1419 {
1420 switch (ARRAY_SIZE (x.array))
1421 {
1422 case 3:
1423 x.array [2] &= y.array [2];
1424 case 2:
1425 x.array [1] &= y.array [1];
1426 case 1:
1427 x.array [0] &= y.array [0];
1428 break;
1429 default:
1430 abort ();
1431 }
1432 return x;
1433 }
1434
1435 static INLINE i386_cpu_flags
1436 cpu_flags_or (i386_cpu_flags x, i386_cpu_flags y)
1437 {
1438 switch (ARRAY_SIZE (x.array))
1439 {
1440 case 3:
1441 x.array [2] |= y.array [2];
1442 case 2:
1443 x.array [1] |= y.array [1];
1444 case 1:
1445 x.array [0] |= y.array [0];
1446 break;
1447 default:
1448 abort ();
1449 }
1450 return x;
1451 }
1452
1453 static INLINE i386_cpu_flags
1454 cpu_flags_and_not (i386_cpu_flags x, i386_cpu_flags y)
1455 {
1456 switch (ARRAY_SIZE (x.array))
1457 {
1458 case 3:
1459 x.array [2] &= ~y.array [2];
1460 case 2:
1461 x.array [1] &= ~y.array [1];
1462 case 1:
1463 x.array [0] &= ~y.array [0];
1464 break;
1465 default:
1466 abort ();
1467 }
1468 return x;
1469 }
1470
1471 static int
1472 valid_iamcu_cpu_flags (const i386_cpu_flags *flags)
1473 {
1474 if (cpu_arch_isa == PROCESSOR_IAMCU)
1475 {
1476 static const i386_cpu_flags iamcu_flags = CPU_IAMCU_COMPAT_FLAGS;
1477 i386_cpu_flags compat_flags;
1478 compat_flags = cpu_flags_and_not (*flags, iamcu_flags);
1479 return cpu_flags_all_zero (&compat_flags);
1480 }
1481 else
1482 return 1;
1483 }
1484
1485 #define CPU_FLAGS_ARCH_MATCH 0x1
1486 #define CPU_FLAGS_64BIT_MATCH 0x2
1487 #define CPU_FLAGS_AES_MATCH 0x4
1488 #define CPU_FLAGS_PCLMUL_MATCH 0x8
1489 #define CPU_FLAGS_AVX_MATCH 0x10
1490
1491 #define CPU_FLAGS_32BIT_MATCH \
1492 (CPU_FLAGS_ARCH_MATCH | CPU_FLAGS_AES_MATCH \
1493 | CPU_FLAGS_PCLMUL_MATCH | CPU_FLAGS_AVX_MATCH)
1494 #define CPU_FLAGS_PERFECT_MATCH \
1495 (CPU_FLAGS_32BIT_MATCH | CPU_FLAGS_64BIT_MATCH)
1496
1497 /* Return CPU flags match bits. */
1498
1499 static int
1500 cpu_flags_match (const insn_template *t)
1501 {
1502 i386_cpu_flags x = t->cpu_flags;
1503 int match = cpu_flags_check_cpu64 (x) ? CPU_FLAGS_64BIT_MATCH : 0;
1504
1505 x.bitfield.cpu64 = 0;
1506 x.bitfield.cpuno64 = 0;
1507
1508 if (cpu_flags_all_zero (&x))
1509 {
1510 /* This instruction is available on all archs. */
1511 match |= CPU_FLAGS_32BIT_MATCH;
1512 }
1513 else
1514 {
1515 /* This instruction is available only on some archs. */
1516 i386_cpu_flags cpu = cpu_arch_flags;
1517
1518 cpu.bitfield.cpu64 = 0;
1519 cpu.bitfield.cpuno64 = 0;
1520 cpu = cpu_flags_and (x, cpu);
1521 if (!cpu_flags_all_zero (&cpu))
1522 {
1523 if (x.bitfield.cpuavx)
1524 {
1525 /* We only need to check AES/PCLMUL/SSE2AVX with AVX. */
1526 if (cpu.bitfield.cpuavx)
1527 {
1528 /* Check SSE2AVX. */
1529 if (!t->opcode_modifier.sse2avx|| sse2avx)
1530 {
1531 match |= (CPU_FLAGS_ARCH_MATCH
1532 | CPU_FLAGS_AVX_MATCH);
1533 /* Check AES. */
1534 if (!x.bitfield.cpuaes || cpu.bitfield.cpuaes)
1535 match |= CPU_FLAGS_AES_MATCH;
1536 /* Check PCLMUL. */
1537 if (!x.bitfield.cpupclmul
1538 || cpu.bitfield.cpupclmul)
1539 match |= CPU_FLAGS_PCLMUL_MATCH;
1540 }
1541 }
1542 else
1543 match |= CPU_FLAGS_ARCH_MATCH;
1544 }
1545 else
1546 match |= CPU_FLAGS_32BIT_MATCH;
1547 }
1548 }
1549 return match;
1550 }
1551
1552 static INLINE i386_operand_type
1553 operand_type_and (i386_operand_type x, i386_operand_type y)
1554 {
1555 switch (ARRAY_SIZE (x.array))
1556 {
1557 case 3:
1558 x.array [2] &= y.array [2];
1559 case 2:
1560 x.array [1] &= y.array [1];
1561 case 1:
1562 x.array [0] &= y.array [0];
1563 break;
1564 default:
1565 abort ();
1566 }
1567 return x;
1568 }
1569
1570 static INLINE i386_operand_type
1571 operand_type_or (i386_operand_type x, i386_operand_type y)
1572 {
1573 switch (ARRAY_SIZE (x.array))
1574 {
1575 case 3:
1576 x.array [2] |= y.array [2];
1577 case 2:
1578 x.array [1] |= y.array [1];
1579 case 1:
1580 x.array [0] |= y.array [0];
1581 break;
1582 default:
1583 abort ();
1584 }
1585 return x;
1586 }
1587
1588 static INLINE i386_operand_type
1589 operand_type_xor (i386_operand_type x, i386_operand_type y)
1590 {
1591 switch (ARRAY_SIZE (x.array))
1592 {
1593 case 3:
1594 x.array [2] ^= y.array [2];
1595 case 2:
1596 x.array [1] ^= y.array [1];
1597 case 1:
1598 x.array [0] ^= y.array [0];
1599 break;
1600 default:
1601 abort ();
1602 }
1603 return x;
1604 }
1605
1606 static const i386_operand_type acc32 = OPERAND_TYPE_ACC32;
1607 static const i386_operand_type acc64 = OPERAND_TYPE_ACC64;
1608 static const i386_operand_type control = OPERAND_TYPE_CONTROL;
1609 static const i386_operand_type inoutportreg
1610 = OPERAND_TYPE_INOUTPORTREG;
1611 static const i386_operand_type reg16_inoutportreg
1612 = OPERAND_TYPE_REG16_INOUTPORTREG;
1613 static const i386_operand_type disp16 = OPERAND_TYPE_DISP16;
1614 static const i386_operand_type disp32 = OPERAND_TYPE_DISP32;
1615 static const i386_operand_type disp32s = OPERAND_TYPE_DISP32S;
1616 static const i386_operand_type disp16_32 = OPERAND_TYPE_DISP16_32;
1617 static const i386_operand_type anydisp
1618 = OPERAND_TYPE_ANYDISP;
1619 static const i386_operand_type regxmm = OPERAND_TYPE_REGXMM;
1620 static const i386_operand_type regymm = OPERAND_TYPE_REGYMM;
1621 static const i386_operand_type regzmm = OPERAND_TYPE_REGZMM;
1622 static const i386_operand_type regmask = OPERAND_TYPE_REGMASK;
1623 static const i386_operand_type imm8 = OPERAND_TYPE_IMM8;
1624 static const i386_operand_type imm8s = OPERAND_TYPE_IMM8S;
1625 static const i386_operand_type imm16 = OPERAND_TYPE_IMM16;
1626 static const i386_operand_type imm32 = OPERAND_TYPE_IMM32;
1627 static const i386_operand_type imm32s = OPERAND_TYPE_IMM32S;
1628 static const i386_operand_type imm64 = OPERAND_TYPE_IMM64;
1629 static const i386_operand_type imm16_32 = OPERAND_TYPE_IMM16_32;
1630 static const i386_operand_type imm16_32s = OPERAND_TYPE_IMM16_32S;
1631 static const i386_operand_type imm16_32_32s = OPERAND_TYPE_IMM16_32_32S;
1632 static const i386_operand_type vec_imm4 = OPERAND_TYPE_VEC_IMM4;
1633
1634 enum operand_type
1635 {
1636 reg,
1637 imm,
1638 disp,
1639 anymem
1640 };
1641
1642 static INLINE int
1643 operand_type_check (i386_operand_type t, enum operand_type c)
1644 {
1645 switch (c)
1646 {
1647 case reg:
1648 return (t.bitfield.reg8
1649 || t.bitfield.reg16
1650 || t.bitfield.reg32
1651 || t.bitfield.reg64);
1652
1653 case imm:
1654 return (t.bitfield.imm8
1655 || t.bitfield.imm8s
1656 || t.bitfield.imm16
1657 || t.bitfield.imm32
1658 || t.bitfield.imm32s
1659 || t.bitfield.imm64);
1660
1661 case disp:
1662 return (t.bitfield.disp8
1663 || t.bitfield.disp16
1664 || t.bitfield.disp32
1665 || t.bitfield.disp32s
1666 || t.bitfield.disp64);
1667
1668 case anymem:
1669 return (t.bitfield.disp8
1670 || t.bitfield.disp16
1671 || t.bitfield.disp32
1672 || t.bitfield.disp32s
1673 || t.bitfield.disp64
1674 || t.bitfield.baseindex);
1675
1676 default:
1677 abort ();
1678 }
1679
1680 return 0;
1681 }
1682
1683 /* Return 1 if there is no conflict in 8bit/16bit/32bit/64bit on
1684 operand J for instruction template T. */
1685
1686 static INLINE int
1687 match_reg_size (const insn_template *t, unsigned int j)
1688 {
1689 return !((i.types[j].bitfield.byte
1690 && !t->operand_types[j].bitfield.byte)
1691 || (i.types[j].bitfield.word
1692 && !t->operand_types[j].bitfield.word)
1693 || (i.types[j].bitfield.dword
1694 && !t->operand_types[j].bitfield.dword)
1695 || (i.types[j].bitfield.qword
1696 && !t->operand_types[j].bitfield.qword));
1697 }
1698
1699 /* Return 1 if there is no conflict in any size on operand J for
1700 instruction template T. */
1701
1702 static INLINE int
1703 match_mem_size (const insn_template *t, unsigned int j)
1704 {
1705 return (match_reg_size (t, j)
1706 && !((i.types[j].bitfield.unspecified
1707 && !i.broadcast
1708 && !t->operand_types[j].bitfield.unspecified)
1709 || (i.types[j].bitfield.fword
1710 && !t->operand_types[j].bitfield.fword)
1711 || (i.types[j].bitfield.tbyte
1712 && !t->operand_types[j].bitfield.tbyte)
1713 || (i.types[j].bitfield.xmmword
1714 && !t->operand_types[j].bitfield.xmmword)
1715 || (i.types[j].bitfield.ymmword
1716 && !t->operand_types[j].bitfield.ymmword)
1717 || (i.types[j].bitfield.zmmword
1718 && !t->operand_types[j].bitfield.zmmword)));
1719 }
1720
1721 /* Return 1 if there is no size conflict on any operands for
1722 instruction template T. */
1723
1724 static INLINE int
1725 operand_size_match (const insn_template *t)
1726 {
1727 unsigned int j;
1728 int match = 1;
1729
1730 /* Don't check jump instructions. */
1731 if (t->opcode_modifier.jump
1732 || t->opcode_modifier.jumpbyte
1733 || t->opcode_modifier.jumpdword
1734 || t->opcode_modifier.jumpintersegment)
1735 return match;
1736
1737 /* Check memory and accumulator operand size. */
1738 for (j = 0; j < i.operands; j++)
1739 {
1740 if (t->operand_types[j].bitfield.anysize)
1741 continue;
1742
1743 if (t->operand_types[j].bitfield.acc && !match_reg_size (t, j))
1744 {
1745 match = 0;
1746 break;
1747 }
1748
1749 if (i.types[j].bitfield.mem && !match_mem_size (t, j))
1750 {
1751 match = 0;
1752 break;
1753 }
1754 }
1755
1756 if (match)
1757 return match;
1758 else if (!t->opcode_modifier.d && !t->opcode_modifier.floatd)
1759 {
1760 mismatch:
1761 i.error = operand_size_mismatch;
1762 return 0;
1763 }
1764
1765 /* Check reverse. */
1766 gas_assert (i.operands == 2);
1767
1768 match = 1;
1769 for (j = 0; j < 2; j++)
1770 {
1771 if (t->operand_types[j].bitfield.acc
1772 && !match_reg_size (t, j ? 0 : 1))
1773 goto mismatch;
1774
1775 if (i.types[j].bitfield.mem
1776 && !match_mem_size (t, j ? 0 : 1))
1777 goto mismatch;
1778 }
1779
1780 return match;
1781 }
1782
1783 static INLINE int
1784 operand_type_match (i386_operand_type overlap,
1785 i386_operand_type given)
1786 {
1787 i386_operand_type temp = overlap;
1788
1789 temp.bitfield.jumpabsolute = 0;
1790 temp.bitfield.unspecified = 0;
1791 temp.bitfield.byte = 0;
1792 temp.bitfield.word = 0;
1793 temp.bitfield.dword = 0;
1794 temp.bitfield.fword = 0;
1795 temp.bitfield.qword = 0;
1796 temp.bitfield.tbyte = 0;
1797 temp.bitfield.xmmword = 0;
1798 temp.bitfield.ymmword = 0;
1799 temp.bitfield.zmmword = 0;
1800 if (operand_type_all_zero (&temp))
1801 goto mismatch;
1802
1803 if (given.bitfield.baseindex == overlap.bitfield.baseindex
1804 && given.bitfield.jumpabsolute == overlap.bitfield.jumpabsolute)
1805 return 1;
1806
1807 mismatch:
1808 i.error = operand_type_mismatch;
1809 return 0;
1810 }
1811
1812 /* If given types g0 and g1 are registers they must be of the same type
1813 unless the expected operand type register overlap is null.
1814 Note that Acc in a template matches every size of reg. */
1815
1816 static INLINE int
1817 operand_type_register_match (i386_operand_type m0,
1818 i386_operand_type g0,
1819 i386_operand_type t0,
1820 i386_operand_type m1,
1821 i386_operand_type g1,
1822 i386_operand_type t1)
1823 {
1824 if (!operand_type_check (g0, reg))
1825 return 1;
1826
1827 if (!operand_type_check (g1, reg))
1828 return 1;
1829
1830 if (g0.bitfield.reg8 == g1.bitfield.reg8
1831 && g0.bitfield.reg16 == g1.bitfield.reg16
1832 && g0.bitfield.reg32 == g1.bitfield.reg32
1833 && g0.bitfield.reg64 == g1.bitfield.reg64)
1834 return 1;
1835
1836 if (m0.bitfield.acc)
1837 {
1838 t0.bitfield.reg8 = 1;
1839 t0.bitfield.reg16 = 1;
1840 t0.bitfield.reg32 = 1;
1841 t0.bitfield.reg64 = 1;
1842 }
1843
1844 if (m1.bitfield.acc)
1845 {
1846 t1.bitfield.reg8 = 1;
1847 t1.bitfield.reg16 = 1;
1848 t1.bitfield.reg32 = 1;
1849 t1.bitfield.reg64 = 1;
1850 }
1851
1852 if (!(t0.bitfield.reg8 & t1.bitfield.reg8)
1853 && !(t0.bitfield.reg16 & t1.bitfield.reg16)
1854 && !(t0.bitfield.reg32 & t1.bitfield.reg32)
1855 && !(t0.bitfield.reg64 & t1.bitfield.reg64))
1856 return 1;
1857
1858 i.error = register_type_mismatch;
1859
1860 return 0;
1861 }
1862
1863 static INLINE unsigned int
1864 register_number (const reg_entry *r)
1865 {
1866 unsigned int nr = r->reg_num;
1867
1868 if (r->reg_flags & RegRex)
1869 nr += 8;
1870
1871 return nr;
1872 }
1873
1874 static INLINE unsigned int
1875 mode_from_disp_size (i386_operand_type t)
1876 {
1877 if (t.bitfield.disp8 || t.bitfield.vec_disp8)
1878 return 1;
1879 else if (t.bitfield.disp16
1880 || t.bitfield.disp32
1881 || t.bitfield.disp32s)
1882 return 2;
1883 else
1884 return 0;
1885 }
1886
1887 static INLINE int
1888 fits_in_signed_byte (addressT num)
1889 {
1890 return num + 0x80 <= 0xff;
1891 }
1892
1893 static INLINE int
1894 fits_in_unsigned_byte (addressT num)
1895 {
1896 return num <= 0xff;
1897 }
1898
1899 static INLINE int
1900 fits_in_unsigned_word (addressT num)
1901 {
1902 return num <= 0xffff;
1903 }
1904
1905 static INLINE int
1906 fits_in_signed_word (addressT num)
1907 {
1908 return num + 0x8000 <= 0xffff;
1909 }
1910
1911 static INLINE int
1912 fits_in_signed_long (addressT num ATTRIBUTE_UNUSED)
1913 {
1914 #ifndef BFD64
1915 return 1;
1916 #else
1917 return num + 0x80000000 <= 0xffffffff;
1918 #endif
1919 } /* fits_in_signed_long() */
1920
1921 static INLINE int
1922 fits_in_unsigned_long (addressT num ATTRIBUTE_UNUSED)
1923 {
1924 #ifndef BFD64
1925 return 1;
1926 #else
1927 return num <= 0xffffffff;
1928 #endif
1929 } /* fits_in_unsigned_long() */
1930
1931 static INLINE int
1932 fits_in_vec_disp8 (offsetT num)
1933 {
1934 int shift = i.memshift;
1935 unsigned int mask;
1936
1937 if (shift == -1)
1938 abort ();
1939
1940 mask = (1 << shift) - 1;
1941
1942 /* Return 0 if NUM isn't properly aligned. */
1943 if ((num & mask))
1944 return 0;
1945
1946 /* Check if NUM will fit in 8bit after shift. */
1947 return fits_in_signed_byte (num >> shift);
1948 }
1949
1950 static INLINE int
1951 fits_in_imm4 (offsetT num)
1952 {
1953 return (num & 0xf) == num;
1954 }
1955
1956 static i386_operand_type
1957 smallest_imm_type (offsetT num)
1958 {
1959 i386_operand_type t;
1960
1961 operand_type_set (&t, 0);
1962 t.bitfield.imm64 = 1;
1963
1964 if (cpu_arch_tune != PROCESSOR_I486 && num == 1)
1965 {
1966 /* This code is disabled on the 486 because all the Imm1 forms
1967 in the opcode table are slower on the i486. They're the
1968 versions with the implicitly specified single-position
1969 displacement, which has another syntax if you really want to
1970 use that form. */
1971 t.bitfield.imm1 = 1;
1972 t.bitfield.imm8 = 1;
1973 t.bitfield.imm8s = 1;
1974 t.bitfield.imm16 = 1;
1975 t.bitfield.imm32 = 1;
1976 t.bitfield.imm32s = 1;
1977 }
1978 else if (fits_in_signed_byte (num))
1979 {
1980 t.bitfield.imm8 = 1;
1981 t.bitfield.imm8s = 1;
1982 t.bitfield.imm16 = 1;
1983 t.bitfield.imm32 = 1;
1984 t.bitfield.imm32s = 1;
1985 }
1986 else if (fits_in_unsigned_byte (num))
1987 {
1988 t.bitfield.imm8 = 1;
1989 t.bitfield.imm16 = 1;
1990 t.bitfield.imm32 = 1;
1991 t.bitfield.imm32s = 1;
1992 }
1993 else if (fits_in_signed_word (num) || fits_in_unsigned_word (num))
1994 {
1995 t.bitfield.imm16 = 1;
1996 t.bitfield.imm32 = 1;
1997 t.bitfield.imm32s = 1;
1998 }
1999 else if (fits_in_signed_long (num))
2000 {
2001 t.bitfield.imm32 = 1;
2002 t.bitfield.imm32s = 1;
2003 }
2004 else if (fits_in_unsigned_long (num))
2005 t.bitfield.imm32 = 1;
2006
2007 return t;
2008 }
2009
2010 static offsetT
2011 offset_in_range (offsetT val, int size)
2012 {
2013 addressT mask;
2014
2015 switch (size)
2016 {
2017 case 1: mask = ((addressT) 1 << 8) - 1; break;
2018 case 2: mask = ((addressT) 1 << 16) - 1; break;
2019 case 4: mask = ((addressT) 2 << 31) - 1; break;
2020 #ifdef BFD64
2021 case 8: mask = ((addressT) 2 << 63) - 1; break;
2022 #endif
2023 default: abort ();
2024 }
2025
2026 #ifdef BFD64
2027 /* If BFD64, sign extend val for 32bit address mode. */
2028 if (flag_code != CODE_64BIT
2029 || i.prefix[ADDR_PREFIX])
2030 if ((val & ~(((addressT) 2 << 31) - 1)) == 0)
2031 val = (val ^ ((addressT) 1 << 31)) - ((addressT) 1 << 31);
2032 #endif
2033
2034 if ((val & ~mask) != 0 && (val & ~mask) != ~mask)
2035 {
2036 char buf1[40], buf2[40];
2037
2038 sprint_value (buf1, val);
2039 sprint_value (buf2, val & mask);
2040 as_warn (_("%s shortened to %s"), buf1, buf2);
2041 }
2042 return val & mask;
2043 }
2044
2045 enum PREFIX_GROUP
2046 {
2047 PREFIX_EXIST = 0,
2048 PREFIX_LOCK,
2049 PREFIX_REP,
2050 PREFIX_OTHER
2051 };
2052
2053 /* Returns
2054 a. PREFIX_EXIST if attempting to add a prefix where one from the
2055 same class already exists.
2056 b. PREFIX_LOCK if lock prefix is added.
2057 c. PREFIX_REP if rep/repne prefix is added.
2058 d. PREFIX_OTHER if other prefix is added.
2059 */
2060
2061 static enum PREFIX_GROUP
2062 add_prefix (unsigned int prefix)
2063 {
2064 enum PREFIX_GROUP ret = PREFIX_OTHER;
2065 unsigned int q;
2066
2067 if (prefix >= REX_OPCODE && prefix < REX_OPCODE + 16
2068 && flag_code == CODE_64BIT)
2069 {
2070 if ((i.prefix[REX_PREFIX] & prefix & REX_W)
2071 || ((i.prefix[REX_PREFIX] & (REX_R | REX_X | REX_B))
2072 && (prefix & (REX_R | REX_X | REX_B))))
2073 ret = PREFIX_EXIST;
2074 q = REX_PREFIX;
2075 }
2076 else
2077 {
2078 switch (prefix)
2079 {
2080 default:
2081 abort ();
2082
2083 case CS_PREFIX_OPCODE:
2084 case DS_PREFIX_OPCODE:
2085 case ES_PREFIX_OPCODE:
2086 case FS_PREFIX_OPCODE:
2087 case GS_PREFIX_OPCODE:
2088 case SS_PREFIX_OPCODE:
2089 q = SEG_PREFIX;
2090 break;
2091
2092 case REPNE_PREFIX_OPCODE:
2093 case REPE_PREFIX_OPCODE:
2094 q = REP_PREFIX;
2095 ret = PREFIX_REP;
2096 break;
2097
2098 case LOCK_PREFIX_OPCODE:
2099 q = LOCK_PREFIX;
2100 ret = PREFIX_LOCK;
2101 break;
2102
2103 case FWAIT_OPCODE:
2104 q = WAIT_PREFIX;
2105 break;
2106
2107 case ADDR_PREFIX_OPCODE:
2108 q = ADDR_PREFIX;
2109 break;
2110
2111 case DATA_PREFIX_OPCODE:
2112 q = DATA_PREFIX;
2113 break;
2114 }
2115 if (i.prefix[q] != 0)
2116 ret = PREFIX_EXIST;
2117 }
2118
2119 if (ret)
2120 {
2121 if (!i.prefix[q])
2122 ++i.prefixes;
2123 i.prefix[q] |= prefix;
2124 }
2125 else
2126 as_bad (_("same type of prefix used twice"));
2127
2128 return ret;
2129 }
2130
2131 static void
2132 update_code_flag (int value, int check)
2133 {
2134 PRINTF_LIKE ((*as_error));
2135
2136 flag_code = (enum flag_code) value;
2137 if (flag_code == CODE_64BIT)
2138 {
2139 cpu_arch_flags.bitfield.cpu64 = 1;
2140 cpu_arch_flags.bitfield.cpuno64 = 0;
2141 }
2142 else
2143 {
2144 cpu_arch_flags.bitfield.cpu64 = 0;
2145 cpu_arch_flags.bitfield.cpuno64 = 1;
2146 }
2147 if (value == CODE_64BIT && !cpu_arch_flags.bitfield.cpulm )
2148 {
2149 if (check)
2150 as_error = as_fatal;
2151 else
2152 as_error = as_bad;
2153 (*as_error) (_("64bit mode not supported on `%s'."),
2154 cpu_arch_name ? cpu_arch_name : default_arch);
2155 }
2156 if (value == CODE_32BIT && !cpu_arch_flags.bitfield.cpui386)
2157 {
2158 if (check)
2159 as_error = as_fatal;
2160 else
2161 as_error = as_bad;
2162 (*as_error) (_("32bit mode not supported on `%s'."),
2163 cpu_arch_name ? cpu_arch_name : default_arch);
2164 }
2165 stackop_size = '\0';
2166 }
2167
2168 static void
2169 set_code_flag (int value)
2170 {
2171 update_code_flag (value, 0);
2172 }
2173
2174 static void
2175 set_16bit_gcc_code_flag (int new_code_flag)
2176 {
2177 flag_code = (enum flag_code) new_code_flag;
2178 if (flag_code != CODE_16BIT)
2179 abort ();
2180 cpu_arch_flags.bitfield.cpu64 = 0;
2181 cpu_arch_flags.bitfield.cpuno64 = 1;
2182 stackop_size = LONG_MNEM_SUFFIX;
2183 }
2184
2185 static void
2186 set_intel_syntax (int syntax_flag)
2187 {
2188 /* Find out if register prefixing is specified. */
2189 int ask_naked_reg = 0;
2190
2191 SKIP_WHITESPACE ();
2192 if (!is_end_of_line[(unsigned char) *input_line_pointer])
2193 {
2194 char *string = input_line_pointer;
2195 int e = get_symbol_end ();
2196
2197 if (strcmp (string, "prefix") == 0)
2198 ask_naked_reg = 1;
2199 else if (strcmp (string, "noprefix") == 0)
2200 ask_naked_reg = -1;
2201 else
2202 as_bad (_("bad argument to syntax directive."));
2203 *input_line_pointer = e;
2204 }
2205 demand_empty_rest_of_line ();
2206
2207 intel_syntax = syntax_flag;
2208
2209 if (ask_naked_reg == 0)
2210 allow_naked_reg = (intel_syntax
2211 && (bfd_get_symbol_leading_char (stdoutput) != '\0'));
2212 else
2213 allow_naked_reg = (ask_naked_reg < 0);
2214
2215 expr_set_rank (O_full_ptr, syntax_flag ? 10 : 0);
2216
2217 identifier_chars['%'] = intel_syntax && allow_naked_reg ? '%' : 0;
2218 identifier_chars['$'] = intel_syntax ? '$' : 0;
2219 register_prefix = allow_naked_reg ? "" : "%";
2220 }
2221
2222 static void
2223 set_intel_mnemonic (int mnemonic_flag)
2224 {
2225 intel_mnemonic = mnemonic_flag;
2226 }
2227
2228 static void
2229 set_allow_index_reg (int flag)
2230 {
2231 allow_index_reg = flag;
2232 }
2233
2234 static void
2235 set_check (int what)
2236 {
2237 enum check_kind *kind;
2238 const char *str;
2239
2240 if (what)
2241 {
2242 kind = &operand_check;
2243 str = "operand";
2244 }
2245 else
2246 {
2247 kind = &sse_check;
2248 str = "sse";
2249 }
2250
2251 SKIP_WHITESPACE ();
2252
2253 if (!is_end_of_line[(unsigned char) *input_line_pointer])
2254 {
2255 char *string = input_line_pointer;
2256 int e = get_symbol_end ();
2257
2258 if (strcmp (string, "none") == 0)
2259 *kind = check_none;
2260 else if (strcmp (string, "warning") == 0)
2261 *kind = check_warning;
2262 else if (strcmp (string, "error") == 0)
2263 *kind = check_error;
2264 else
2265 as_bad (_("bad argument to %s_check directive."), str);
2266 *input_line_pointer = e;
2267 }
2268 else
2269 as_bad (_("missing argument for %s_check directive"), str);
2270
2271 demand_empty_rest_of_line ();
2272 }
2273
2274 static void
2275 check_cpu_arch_compatible (const char *name ATTRIBUTE_UNUSED,
2276 i386_cpu_flags new_flag ATTRIBUTE_UNUSED)
2277 {
2278 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
2279 static const char *arch;
2280
2281 /* Intel LIOM is only supported on ELF. */
2282 if (!IS_ELF)
2283 return;
2284
2285 if (!arch)
2286 {
2287 /* Use cpu_arch_name if it is set in md_parse_option. Otherwise
2288 use default_arch. */
2289 arch = cpu_arch_name;
2290 if (!arch)
2291 arch = default_arch;
2292 }
2293
2294 /* If we are targeting Intel MCU, we must enable it. */
2295 if (get_elf_backend_data (stdoutput)->elf_machine_code != EM_IAMCU
2296 || new_flag.bitfield.cpuiamcu)
2297 return;
2298
2299 /* If we are targeting Intel L1OM, we must enable it. */
2300 if (get_elf_backend_data (stdoutput)->elf_machine_code != EM_L1OM
2301 || new_flag.bitfield.cpul1om)
2302 return;
2303
2304 /* If we are targeting Intel K1OM, we must enable it. */
2305 if (get_elf_backend_data (stdoutput)->elf_machine_code != EM_K1OM
2306 || new_flag.bitfield.cpuk1om)
2307 return;
2308
2309 as_bad (_("`%s' is not supported on `%s'"), name, arch);
2310 #endif
2311 }
2312
2313 static void
2314 set_cpu_arch (int dummy ATTRIBUTE_UNUSED)
2315 {
2316 SKIP_WHITESPACE ();
2317
2318 if (!is_end_of_line[(unsigned char) *input_line_pointer])
2319 {
2320 char *string = input_line_pointer;
2321 int e = get_symbol_end ();
2322 unsigned int j;
2323 i386_cpu_flags flags;
2324
2325 for (j = 0; j < ARRAY_SIZE (cpu_arch); j++)
2326 {
2327 if (strcmp (string, cpu_arch[j].name) == 0)
2328 {
2329 check_cpu_arch_compatible (string, cpu_arch[j].flags);
2330
2331 if (*string != '.')
2332 {
2333 cpu_arch_name = cpu_arch[j].name;
2334 cpu_sub_arch_name = NULL;
2335 cpu_arch_flags = cpu_arch[j].flags;
2336 if (flag_code == CODE_64BIT)
2337 {
2338 cpu_arch_flags.bitfield.cpu64 = 1;
2339 cpu_arch_flags.bitfield.cpuno64 = 0;
2340 }
2341 else
2342 {
2343 cpu_arch_flags.bitfield.cpu64 = 0;
2344 cpu_arch_flags.bitfield.cpuno64 = 1;
2345 }
2346 cpu_arch_isa = cpu_arch[j].type;
2347 cpu_arch_isa_flags = cpu_arch[j].flags;
2348 if (!cpu_arch_tune_set)
2349 {
2350 cpu_arch_tune = cpu_arch_isa;
2351 cpu_arch_tune_flags = cpu_arch_isa_flags;
2352 }
2353 break;
2354 }
2355
2356 if (!cpu_arch[j].negated)
2357 flags = cpu_flags_or (cpu_arch_flags,
2358 cpu_arch[j].flags);
2359 else
2360 flags = cpu_flags_and_not (cpu_arch_flags,
2361 cpu_arch[j].flags);
2362
2363 if (!valid_iamcu_cpu_flags (&flags))
2364 as_fatal (_("`%s' isn't valid for Intel MCU"),
2365 cpu_arch[j].name);
2366 else if (!cpu_flags_equal (&flags, &cpu_arch_flags))
2367 {
2368 if (cpu_sub_arch_name)
2369 {
2370 char *name = cpu_sub_arch_name;
2371 cpu_sub_arch_name = concat (name,
2372 cpu_arch[j].name,
2373 (const char *) NULL);
2374 free (name);
2375 }
2376 else
2377 cpu_sub_arch_name = xstrdup (cpu_arch[j].name);
2378 cpu_arch_flags = flags;
2379 cpu_arch_isa_flags = flags;
2380 }
2381 *input_line_pointer = e;
2382 demand_empty_rest_of_line ();
2383 return;
2384 }
2385 }
2386 if (j >= ARRAY_SIZE (cpu_arch))
2387 as_bad (_("no such architecture: `%s'"), string);
2388
2389 *input_line_pointer = e;
2390 }
2391 else
2392 as_bad (_("missing cpu architecture"));
2393
2394 no_cond_jump_promotion = 0;
2395 if (*input_line_pointer == ','
2396 && !is_end_of_line[(unsigned char) input_line_pointer[1]])
2397 {
2398 char *string = ++input_line_pointer;
2399 int e = get_symbol_end ();
2400
2401 if (strcmp (string, "nojumps") == 0)
2402 no_cond_jump_promotion = 1;
2403 else if (strcmp (string, "jumps") == 0)
2404 ;
2405 else
2406 as_bad (_("no such architecture modifier: `%s'"), string);
2407
2408 *input_line_pointer = e;
2409 }
2410
2411 demand_empty_rest_of_line ();
2412 }
2413
2414 enum bfd_architecture
2415 i386_arch (void)
2416 {
2417 if (cpu_arch_isa == PROCESSOR_L1OM)
2418 {
2419 if (OUTPUT_FLAVOR != bfd_target_elf_flavour
2420 || flag_code != CODE_64BIT)
2421 as_fatal (_("Intel L1OM is 64bit ELF only"));
2422 return bfd_arch_l1om;
2423 }
2424 else if (cpu_arch_isa == PROCESSOR_K1OM)
2425 {
2426 if (OUTPUT_FLAVOR != bfd_target_elf_flavour
2427 || flag_code != CODE_64BIT)
2428 as_fatal (_("Intel K1OM is 64bit ELF only"));
2429 return bfd_arch_k1om;
2430 }
2431 else if (cpu_arch_isa == PROCESSOR_IAMCU)
2432 {
2433 if (OUTPUT_FLAVOR != bfd_target_elf_flavour
2434 || flag_code == CODE_64BIT)
2435 as_fatal (_("Intel MCU is 32bit ELF only"));
2436 return bfd_arch_iamcu;
2437 }
2438 else
2439 return bfd_arch_i386;
2440 }
2441
2442 unsigned long
2443 i386_mach (void)
2444 {
2445 if (!strncmp (default_arch, "x86_64", 6))
2446 {
2447 if (cpu_arch_isa == PROCESSOR_L1OM)
2448 {
2449 if (OUTPUT_FLAVOR != bfd_target_elf_flavour
2450 || default_arch[6] != '\0')
2451 as_fatal (_("Intel L1OM is 64bit ELF only"));
2452 return bfd_mach_l1om;
2453 }
2454 else if (cpu_arch_isa == PROCESSOR_K1OM)
2455 {
2456 if (OUTPUT_FLAVOR != bfd_target_elf_flavour
2457 || default_arch[6] != '\0')
2458 as_fatal (_("Intel K1OM is 64bit ELF only"));
2459 return bfd_mach_k1om;
2460 }
2461 else if (default_arch[6] == '\0')
2462 return bfd_mach_x86_64;
2463 else
2464 return bfd_mach_x64_32;
2465 }
2466 else if (!strcmp (default_arch, "i386")
2467 || !strcmp (default_arch, "iamcu"))
2468 {
2469 if (cpu_arch_isa == PROCESSOR_IAMCU)
2470 {
2471 if (OUTPUT_FLAVOR != bfd_target_elf_flavour)
2472 as_fatal (_("Intel MCU is 32bit ELF only"));
2473 return bfd_mach_i386_iamcu;
2474 }
2475 else
2476 return bfd_mach_i386_i386;
2477 }
2478 else
2479 as_fatal (_("unknown architecture"));
2480 }
2481 \f
2482 void
2483 md_begin (void)
2484 {
2485 const char *hash_err;
2486
2487 /* Initialize op_hash hash table. */
2488 op_hash = hash_new ();
2489
2490 {
2491 const insn_template *optab;
2492 templates *core_optab;
2493
2494 /* Setup for loop. */
2495 optab = i386_optab;
2496 core_optab = (templates *) xmalloc (sizeof (templates));
2497 core_optab->start = optab;
2498
2499 while (1)
2500 {
2501 ++optab;
2502 if (optab->name == NULL
2503 || strcmp (optab->name, (optab - 1)->name) != 0)
2504 {
2505 /* different name --> ship out current template list;
2506 add to hash table; & begin anew. */
2507 core_optab->end = optab;
2508 hash_err = hash_insert (op_hash,
2509 (optab - 1)->name,
2510 (void *) core_optab);
2511 if (hash_err)
2512 {
2513 as_fatal (_("can't hash %s: %s"),
2514 (optab - 1)->name,
2515 hash_err);
2516 }
2517 if (optab->name == NULL)
2518 break;
2519 core_optab = (templates *) xmalloc (sizeof (templates));
2520 core_optab->start = optab;
2521 }
2522 }
2523 }
2524
2525 /* Initialize reg_hash hash table. */
2526 reg_hash = hash_new ();
2527 {
2528 const reg_entry *regtab;
2529 unsigned int regtab_size = i386_regtab_size;
2530
2531 for (regtab = i386_regtab; regtab_size--; regtab++)
2532 {
2533 hash_err = hash_insert (reg_hash, regtab->reg_name, (void *) regtab);
2534 if (hash_err)
2535 as_fatal (_("can't hash %s: %s"),
2536 regtab->reg_name,
2537 hash_err);
2538 }
2539 }
2540
2541 /* Fill in lexical tables: mnemonic_chars, operand_chars. */
2542 {
2543 int c;
2544 char *p;
2545
2546 for (c = 0; c < 256; c++)
2547 {
2548 if (ISDIGIT (c))
2549 {
2550 digit_chars[c] = c;
2551 mnemonic_chars[c] = c;
2552 register_chars[c] = c;
2553 operand_chars[c] = c;
2554 }
2555 else if (ISLOWER (c))
2556 {
2557 mnemonic_chars[c] = c;
2558 register_chars[c] = c;
2559 operand_chars[c] = c;
2560 }
2561 else if (ISUPPER (c))
2562 {
2563 mnemonic_chars[c] = TOLOWER (c);
2564 register_chars[c] = mnemonic_chars[c];
2565 operand_chars[c] = c;
2566 }
2567 else if (c == '{' || c == '}')
2568 operand_chars[c] = c;
2569
2570 if (ISALPHA (c) || ISDIGIT (c))
2571 identifier_chars[c] = c;
2572 else if (c >= 128)
2573 {
2574 identifier_chars[c] = c;
2575 operand_chars[c] = c;
2576 }
2577 }
2578
2579 #ifdef LEX_AT
2580 identifier_chars['@'] = '@';
2581 #endif
2582 #ifdef LEX_QM
2583 identifier_chars['?'] = '?';
2584 operand_chars['?'] = '?';
2585 #endif
2586 digit_chars['-'] = '-';
2587 mnemonic_chars['_'] = '_';
2588 mnemonic_chars['-'] = '-';
2589 mnemonic_chars['.'] = '.';
2590 identifier_chars['_'] = '_';
2591 identifier_chars['.'] = '.';
2592
2593 for (p = operand_special_chars; *p != '\0'; p++)
2594 operand_chars[(unsigned char) *p] = *p;
2595 }
2596
2597 if (flag_code == CODE_64BIT)
2598 {
2599 #if defined (OBJ_COFF) && defined (TE_PE)
2600 x86_dwarf2_return_column = (OUTPUT_FLAVOR == bfd_target_coff_flavour
2601 ? 32 : 16);
2602 #else
2603 x86_dwarf2_return_column = 16;
2604 #endif
2605 x86_cie_data_alignment = -8;
2606 }
2607 else
2608 {
2609 x86_dwarf2_return_column = 8;
2610 x86_cie_data_alignment = -4;
2611 }
2612 }
2613
2614 void
2615 i386_print_statistics (FILE *file)
2616 {
2617 hash_print_statistics (file, "i386 opcode", op_hash);
2618 hash_print_statistics (file, "i386 register", reg_hash);
2619 }
2620 \f
2621 #ifdef DEBUG386
2622
2623 /* Debugging routines for md_assemble. */
2624 static void pte (insn_template *);
2625 static void pt (i386_operand_type);
2626 static void pe (expressionS *);
2627 static void ps (symbolS *);
2628
2629 static void
2630 pi (char *line, i386_insn *x)
2631 {
2632 unsigned int j;
2633
2634 fprintf (stdout, "%s: template ", line);
2635 pte (&x->tm);
2636 fprintf (stdout, " address: base %s index %s scale %x\n",
2637 x->base_reg ? x->base_reg->reg_name : "none",
2638 x->index_reg ? x->index_reg->reg_name : "none",
2639 x->log2_scale_factor);
2640 fprintf (stdout, " modrm: mode %x reg %x reg/mem %x\n",
2641 x->rm.mode, x->rm.reg, x->rm.regmem);
2642 fprintf (stdout, " sib: base %x index %x scale %x\n",
2643 x->sib.base, x->sib.index, x->sib.scale);
2644 fprintf (stdout, " rex: 64bit %x extX %x extY %x extZ %x\n",
2645 (x->rex & REX_W) != 0,
2646 (x->rex & REX_R) != 0,
2647 (x->rex & REX_X) != 0,
2648 (x->rex & REX_B) != 0);
2649 for (j = 0; j < x->operands; j++)
2650 {
2651 fprintf (stdout, " #%d: ", j + 1);
2652 pt (x->types[j]);
2653 fprintf (stdout, "\n");
2654 if (x->types[j].bitfield.reg8
2655 || x->types[j].bitfield.reg16
2656 || x->types[j].bitfield.reg32
2657 || x->types[j].bitfield.reg64
2658 || x->types[j].bitfield.regmmx
2659 || x->types[j].bitfield.regxmm
2660 || x->types[j].bitfield.regymm
2661 || x->types[j].bitfield.regzmm
2662 || x->types[j].bitfield.sreg2
2663 || x->types[j].bitfield.sreg3
2664 || x->types[j].bitfield.control
2665 || x->types[j].bitfield.debug
2666 || x->types[j].bitfield.test)
2667 fprintf (stdout, "%s\n", x->op[j].regs->reg_name);
2668 if (operand_type_check (x->types[j], imm))
2669 pe (x->op[j].imms);
2670 if (operand_type_check (x->types[j], disp))
2671 pe (x->op[j].disps);
2672 }
2673 }
2674
2675 static void
2676 pte (insn_template *t)
2677 {
2678 unsigned int j;
2679 fprintf (stdout, " %d operands ", t->operands);
2680 fprintf (stdout, "opcode %x ", t->base_opcode);
2681 if (t->extension_opcode != None)
2682 fprintf (stdout, "ext %x ", t->extension_opcode);
2683 if (t->opcode_modifier.d)
2684 fprintf (stdout, "D");
2685 if (t->opcode_modifier.w)
2686 fprintf (stdout, "W");
2687 fprintf (stdout, "\n");
2688 for (j = 0; j < t->operands; j++)
2689 {
2690 fprintf (stdout, " #%d type ", j + 1);
2691 pt (t->operand_types[j]);
2692 fprintf (stdout, "\n");
2693 }
2694 }
2695
2696 static void
2697 pe (expressionS *e)
2698 {
2699 fprintf (stdout, " operation %d\n", e->X_op);
2700 fprintf (stdout, " add_number %ld (%lx)\n",
2701 (long) e->X_add_number, (long) e->X_add_number);
2702 if (e->X_add_symbol)
2703 {
2704 fprintf (stdout, " add_symbol ");
2705 ps (e->X_add_symbol);
2706 fprintf (stdout, "\n");
2707 }
2708 if (e->X_op_symbol)
2709 {
2710 fprintf (stdout, " op_symbol ");
2711 ps (e->X_op_symbol);
2712 fprintf (stdout, "\n");
2713 }
2714 }
2715
2716 static void
2717 ps (symbolS *s)
2718 {
2719 fprintf (stdout, "%s type %s%s",
2720 S_GET_NAME (s),
2721 S_IS_EXTERNAL (s) ? "EXTERNAL " : "",
2722 segment_name (S_GET_SEGMENT (s)));
2723 }
2724
2725 static struct type_name
2726 {
2727 i386_operand_type mask;
2728 const char *name;
2729 }
2730 const type_names[] =
2731 {
2732 { OPERAND_TYPE_REG8, "r8" },
2733 { OPERAND_TYPE_REG16, "r16" },
2734 { OPERAND_TYPE_REG32, "r32" },
2735 { OPERAND_TYPE_REG64, "r64" },
2736 { OPERAND_TYPE_IMM8, "i8" },
2737 { OPERAND_TYPE_IMM8, "i8s" },
2738 { OPERAND_TYPE_IMM16, "i16" },
2739 { OPERAND_TYPE_IMM32, "i32" },
2740 { OPERAND_TYPE_IMM32S, "i32s" },
2741 { OPERAND_TYPE_IMM64, "i64" },
2742 { OPERAND_TYPE_IMM1, "i1" },
2743 { OPERAND_TYPE_BASEINDEX, "BaseIndex" },
2744 { OPERAND_TYPE_DISP8, "d8" },
2745 { OPERAND_TYPE_DISP16, "d16" },
2746 { OPERAND_TYPE_DISP32, "d32" },
2747 { OPERAND_TYPE_DISP32S, "d32s" },
2748 { OPERAND_TYPE_DISP64, "d64" },
2749 { OPERAND_TYPE_VEC_DISP8, "Vector d8" },
2750 { OPERAND_TYPE_INOUTPORTREG, "InOutPortReg" },
2751 { OPERAND_TYPE_SHIFTCOUNT, "ShiftCount" },
2752 { OPERAND_TYPE_CONTROL, "control reg" },
2753 { OPERAND_TYPE_TEST, "test reg" },
2754 { OPERAND_TYPE_DEBUG, "debug reg" },
2755 { OPERAND_TYPE_FLOATREG, "FReg" },
2756 { OPERAND_TYPE_FLOATACC, "FAcc" },
2757 { OPERAND_TYPE_SREG2, "SReg2" },
2758 { OPERAND_TYPE_SREG3, "SReg3" },
2759 { OPERAND_TYPE_ACC, "Acc" },
2760 { OPERAND_TYPE_JUMPABSOLUTE, "Jump Absolute" },
2761 { OPERAND_TYPE_REGMMX, "rMMX" },
2762 { OPERAND_TYPE_REGXMM, "rXMM" },
2763 { OPERAND_TYPE_REGYMM, "rYMM" },
2764 { OPERAND_TYPE_REGZMM, "rZMM" },
2765 { OPERAND_TYPE_REGMASK, "Mask reg" },
2766 { OPERAND_TYPE_ESSEG, "es" },
2767 };
2768
2769 static void
2770 pt (i386_operand_type t)
2771 {
2772 unsigned int j;
2773 i386_operand_type a;
2774
2775 for (j = 0; j < ARRAY_SIZE (type_names); j++)
2776 {
2777 a = operand_type_and (t, type_names[j].mask);
2778 if (!operand_type_all_zero (&a))
2779 fprintf (stdout, "%s, ", type_names[j].name);
2780 }
2781 fflush (stdout);
2782 }
2783
2784 #endif /* DEBUG386 */
2785 \f
2786 static bfd_reloc_code_real_type
2787 reloc (unsigned int size,
2788 int pcrel,
2789 int sign,
2790 bfd_reloc_code_real_type other)
2791 {
2792 if (other != NO_RELOC)
2793 {
2794 reloc_howto_type *rel;
2795
2796 if (size == 8)
2797 switch (other)
2798 {
2799 case BFD_RELOC_X86_64_GOT32:
2800 return BFD_RELOC_X86_64_GOT64;
2801 break;
2802 case BFD_RELOC_X86_64_GOTPLT64:
2803 return BFD_RELOC_X86_64_GOTPLT64;
2804 break;
2805 case BFD_RELOC_X86_64_PLTOFF64:
2806 return BFD_RELOC_X86_64_PLTOFF64;
2807 break;
2808 case BFD_RELOC_X86_64_GOTPC32:
2809 other = BFD_RELOC_X86_64_GOTPC64;
2810 break;
2811 case BFD_RELOC_X86_64_GOTPCREL:
2812 other = BFD_RELOC_X86_64_GOTPCREL64;
2813 break;
2814 case BFD_RELOC_X86_64_TPOFF32:
2815 other = BFD_RELOC_X86_64_TPOFF64;
2816 break;
2817 case BFD_RELOC_X86_64_DTPOFF32:
2818 other = BFD_RELOC_X86_64_DTPOFF64;
2819 break;
2820 default:
2821 break;
2822 }
2823
2824 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
2825 if (other == BFD_RELOC_SIZE32)
2826 {
2827 if (size == 8)
2828 other = BFD_RELOC_SIZE64;
2829 if (pcrel)
2830 {
2831 as_bad (_("there are no pc-relative size relocations"));
2832 return NO_RELOC;
2833 }
2834 }
2835 #endif
2836
2837 /* Sign-checking 4-byte relocations in 16-/32-bit code is pointless. */
2838 if (size == 4 && (flag_code != CODE_64BIT || disallow_64bit_reloc))
2839 sign = -1;
2840
2841 rel = bfd_reloc_type_lookup (stdoutput, other);
2842 if (!rel)
2843 as_bad (_("unknown relocation (%u)"), other);
2844 else if (size != bfd_get_reloc_size (rel))
2845 as_bad (_("%u-byte relocation cannot be applied to %u-byte field"),
2846 bfd_get_reloc_size (rel),
2847 size);
2848 else if (pcrel && !rel->pc_relative)
2849 as_bad (_("non-pc-relative relocation for pc-relative field"));
2850 else if ((rel->complain_on_overflow == complain_overflow_signed
2851 && !sign)
2852 || (rel->complain_on_overflow == complain_overflow_unsigned
2853 && sign > 0))
2854 as_bad (_("relocated field and relocation type differ in signedness"));
2855 else
2856 return other;
2857 return NO_RELOC;
2858 }
2859
2860 if (pcrel)
2861 {
2862 if (!sign)
2863 as_bad (_("there are no unsigned pc-relative relocations"));
2864 switch (size)
2865 {
2866 case 1: return BFD_RELOC_8_PCREL;
2867 case 2: return BFD_RELOC_16_PCREL;
2868 case 4: return BFD_RELOC_32_PCREL;
2869 case 8: return BFD_RELOC_64_PCREL;
2870 }
2871 as_bad (_("cannot do %u byte pc-relative relocation"), size);
2872 }
2873 else
2874 {
2875 if (sign > 0)
2876 switch (size)
2877 {
2878 case 4: return BFD_RELOC_X86_64_32S;
2879 }
2880 else
2881 switch (size)
2882 {
2883 case 1: return BFD_RELOC_8;
2884 case 2: return BFD_RELOC_16;
2885 case 4: return BFD_RELOC_32;
2886 case 8: return BFD_RELOC_64;
2887 }
2888 as_bad (_("cannot do %s %u byte relocation"),
2889 sign > 0 ? "signed" : "unsigned", size);
2890 }
2891
2892 return NO_RELOC;
2893 }
2894
2895 /* Here we decide which fixups can be adjusted to make them relative to
2896 the beginning of the section instead of the symbol. Basically we need
2897 to make sure that the dynamic relocations are done correctly, so in
2898 some cases we force the original symbol to be used. */
2899
2900 int
2901 tc_i386_fix_adjustable (fixS *fixP ATTRIBUTE_UNUSED)
2902 {
2903 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
2904 if (!IS_ELF)
2905 return 1;
2906
2907 /* Don't adjust pc-relative references to merge sections in 64-bit
2908 mode. */
2909 if (use_rela_relocations
2910 && (S_GET_SEGMENT (fixP->fx_addsy)->flags & SEC_MERGE) != 0
2911 && fixP->fx_pcrel)
2912 return 0;
2913
2914 /* The x86_64 GOTPCREL are represented as 32bit PCrel relocations
2915 and changed later by validate_fix. */
2916 if (GOT_symbol && fixP->fx_subsy == GOT_symbol
2917 && fixP->fx_r_type == BFD_RELOC_32_PCREL)
2918 return 0;
2919
2920 /* Adjust_reloc_syms doesn't know about the GOT. Need to keep symbol
2921 for size relocations. */
2922 if (fixP->fx_r_type == BFD_RELOC_SIZE32
2923 || fixP->fx_r_type == BFD_RELOC_SIZE64
2924 || fixP->fx_r_type == BFD_RELOC_386_GOTOFF
2925 || fixP->fx_r_type == BFD_RELOC_386_PLT32
2926 || fixP->fx_r_type == BFD_RELOC_386_GOT32
2927 || fixP->fx_r_type == BFD_RELOC_386_TLS_GD
2928 || fixP->fx_r_type == BFD_RELOC_386_TLS_LDM
2929 || fixP->fx_r_type == BFD_RELOC_386_TLS_LDO_32
2930 || fixP->fx_r_type == BFD_RELOC_386_TLS_IE_32
2931 || fixP->fx_r_type == BFD_RELOC_386_TLS_IE
2932 || fixP->fx_r_type == BFD_RELOC_386_TLS_GOTIE
2933 || fixP->fx_r_type == BFD_RELOC_386_TLS_LE_32
2934 || fixP->fx_r_type == BFD_RELOC_386_TLS_LE
2935 || fixP->fx_r_type == BFD_RELOC_386_TLS_GOTDESC
2936 || fixP->fx_r_type == BFD_RELOC_386_TLS_DESC_CALL
2937 || fixP->fx_r_type == BFD_RELOC_X86_64_PLT32
2938 || fixP->fx_r_type == BFD_RELOC_X86_64_GOT32
2939 || fixP->fx_r_type == BFD_RELOC_X86_64_GOTPCREL
2940 || fixP->fx_r_type == BFD_RELOC_X86_64_TLSGD
2941 || fixP->fx_r_type == BFD_RELOC_X86_64_TLSLD
2942 || fixP->fx_r_type == BFD_RELOC_X86_64_DTPOFF32
2943 || fixP->fx_r_type == BFD_RELOC_X86_64_DTPOFF64
2944 || fixP->fx_r_type == BFD_RELOC_X86_64_GOTTPOFF
2945 || fixP->fx_r_type == BFD_RELOC_X86_64_TPOFF32
2946 || fixP->fx_r_type == BFD_RELOC_X86_64_TPOFF64
2947 || fixP->fx_r_type == BFD_RELOC_X86_64_GOTOFF64
2948 || fixP->fx_r_type == BFD_RELOC_X86_64_GOTPC32_TLSDESC
2949 || fixP->fx_r_type == BFD_RELOC_X86_64_TLSDESC_CALL
2950 || fixP->fx_r_type == BFD_RELOC_VTABLE_INHERIT
2951 || fixP->fx_r_type == BFD_RELOC_VTABLE_ENTRY)
2952 return 0;
2953 #endif
2954 return 1;
2955 }
2956
2957 static int
2958 intel_float_operand (const char *mnemonic)
2959 {
2960 /* Note that the value returned is meaningful only for opcodes with (memory)
2961 operands, hence the code here is free to improperly handle opcodes that
2962 have no operands (for better performance and smaller code). */
2963
2964 if (mnemonic[0] != 'f')
2965 return 0; /* non-math */
2966
2967 switch (mnemonic[1])
2968 {
2969 /* fclex, fdecstp, fdisi, femms, feni, fincstp, finit, fsetpm, and
2970 the fs segment override prefix not currently handled because no
2971 call path can make opcodes without operands get here */
2972 case 'i':
2973 return 2 /* integer op */;
2974 case 'l':
2975 if (mnemonic[2] == 'd' && (mnemonic[3] == 'c' || mnemonic[3] == 'e'))
2976 return 3; /* fldcw/fldenv */
2977 break;
2978 case 'n':
2979 if (mnemonic[2] != 'o' /* fnop */)
2980 return 3; /* non-waiting control op */
2981 break;
2982 case 'r':
2983 if (mnemonic[2] == 's')
2984 return 3; /* frstor/frstpm */
2985 break;
2986 case 's':
2987 if (mnemonic[2] == 'a')
2988 return 3; /* fsave */
2989 if (mnemonic[2] == 't')
2990 {
2991 switch (mnemonic[3])
2992 {
2993 case 'c': /* fstcw */
2994 case 'd': /* fstdw */
2995 case 'e': /* fstenv */
2996 case 's': /* fsts[gw] */
2997 return 3;
2998 }
2999 }
3000 break;
3001 case 'x':
3002 if (mnemonic[2] == 'r' || mnemonic[2] == 's')
3003 return 0; /* fxsave/fxrstor are not really math ops */
3004 break;
3005 }
3006
3007 return 1;
3008 }
3009
3010 /* Build the VEX prefix. */
3011
3012 static void
3013 build_vex_prefix (const insn_template *t)
3014 {
3015 unsigned int register_specifier;
3016 unsigned int implied_prefix;
3017 unsigned int vector_length;
3018
3019 /* Check register specifier. */
3020 if (i.vex.register_specifier)
3021 {
3022 register_specifier =
3023 ~register_number (i.vex.register_specifier) & 0xf;
3024 gas_assert ((i.vex.register_specifier->reg_flags & RegVRex) == 0);
3025 }
3026 else
3027 register_specifier = 0xf;
3028
3029 /* Use 2-byte VEX prefix by swappping destination and source
3030 operand. */
3031 if (!i.swap_operand
3032 && i.operands == i.reg_operands
3033 && i.tm.opcode_modifier.vexopcode == VEX0F
3034 && i.tm.opcode_modifier.s
3035 && i.rex == REX_B)
3036 {
3037 unsigned int xchg = i.operands - 1;
3038 union i386_op temp_op;
3039 i386_operand_type temp_type;
3040
3041 temp_type = i.types[xchg];
3042 i.types[xchg] = i.types[0];
3043 i.types[0] = temp_type;
3044 temp_op = i.op[xchg];
3045 i.op[xchg] = i.op[0];
3046 i.op[0] = temp_op;
3047
3048 gas_assert (i.rm.mode == 3);
3049
3050 i.rex = REX_R;
3051 xchg = i.rm.regmem;
3052 i.rm.regmem = i.rm.reg;
3053 i.rm.reg = xchg;
3054
3055 /* Use the next insn. */
3056 i.tm = t[1];
3057 }
3058
3059 if (i.tm.opcode_modifier.vex == VEXScalar)
3060 vector_length = avxscalar;
3061 else
3062 vector_length = i.tm.opcode_modifier.vex == VEX256 ? 1 : 0;
3063
3064 switch ((i.tm.base_opcode >> 8) & 0xff)
3065 {
3066 case 0:
3067 implied_prefix = 0;
3068 break;
3069 case DATA_PREFIX_OPCODE:
3070 implied_prefix = 1;
3071 break;
3072 case REPE_PREFIX_OPCODE:
3073 implied_prefix = 2;
3074 break;
3075 case REPNE_PREFIX_OPCODE:
3076 implied_prefix = 3;
3077 break;
3078 default:
3079 abort ();
3080 }
3081
3082 /* Use 2-byte VEX prefix if possible. */
3083 if (i.tm.opcode_modifier.vexopcode == VEX0F
3084 && i.tm.opcode_modifier.vexw != VEXW1
3085 && (i.rex & (REX_W | REX_X | REX_B)) == 0)
3086 {
3087 /* 2-byte VEX prefix. */
3088 unsigned int r;
3089
3090 i.vex.length = 2;
3091 i.vex.bytes[0] = 0xc5;
3092
3093 /* Check the REX.R bit. */
3094 r = (i.rex & REX_R) ? 0 : 1;
3095 i.vex.bytes[1] = (r << 7
3096 | register_specifier << 3
3097 | vector_length << 2
3098 | implied_prefix);
3099 }
3100 else
3101 {
3102 /* 3-byte VEX prefix. */
3103 unsigned int m, w;
3104
3105 i.vex.length = 3;
3106
3107 switch (i.tm.opcode_modifier.vexopcode)
3108 {
3109 case VEX0F:
3110 m = 0x1;
3111 i.vex.bytes[0] = 0xc4;
3112 break;
3113 case VEX0F38:
3114 m = 0x2;
3115 i.vex.bytes[0] = 0xc4;
3116 break;
3117 case VEX0F3A:
3118 m = 0x3;
3119 i.vex.bytes[0] = 0xc4;
3120 break;
3121 case XOP08:
3122 m = 0x8;
3123 i.vex.bytes[0] = 0x8f;
3124 break;
3125 case XOP09:
3126 m = 0x9;
3127 i.vex.bytes[0] = 0x8f;
3128 break;
3129 case XOP0A:
3130 m = 0xa;
3131 i.vex.bytes[0] = 0x8f;
3132 break;
3133 default:
3134 abort ();
3135 }
3136
3137 /* The high 3 bits of the second VEX byte are 1's compliment
3138 of RXB bits from REX. */
3139 i.vex.bytes[1] = (~i.rex & 0x7) << 5 | m;
3140
3141 /* Check the REX.W bit. */
3142 w = (i.rex & REX_W) ? 1 : 0;
3143 if (i.tm.opcode_modifier.vexw == VEXW1)
3144 w = 1;
3145
3146 i.vex.bytes[2] = (w << 7
3147 | register_specifier << 3
3148 | vector_length << 2
3149 | implied_prefix);
3150 }
3151 }
3152
3153 /* Build the EVEX prefix. */
3154
3155 static void
3156 build_evex_prefix (void)
3157 {
3158 unsigned int register_specifier;
3159 unsigned int implied_prefix;
3160 unsigned int m, w;
3161 rex_byte vrex_used = 0;
3162
3163 /* Check register specifier. */
3164 if (i.vex.register_specifier)
3165 {
3166 gas_assert ((i.vrex & REX_X) == 0);
3167
3168 register_specifier = i.vex.register_specifier->reg_num;
3169 if ((i.vex.register_specifier->reg_flags & RegRex))
3170 register_specifier += 8;
3171 /* The upper 16 registers are encoded in the fourth byte of the
3172 EVEX prefix. */
3173 if (!(i.vex.register_specifier->reg_flags & RegVRex))
3174 i.vex.bytes[3] = 0x8;
3175 register_specifier = ~register_specifier & 0xf;
3176 }
3177 else
3178 {
3179 register_specifier = 0xf;
3180
3181 /* Encode upper 16 vector index register in the fourth byte of
3182 the EVEX prefix. */
3183 if (!(i.vrex & REX_X))
3184 i.vex.bytes[3] = 0x8;
3185 else
3186 vrex_used |= REX_X;
3187 }
3188
3189 switch ((i.tm.base_opcode >> 8) & 0xff)
3190 {
3191 case 0:
3192 implied_prefix = 0;
3193 break;
3194 case DATA_PREFIX_OPCODE:
3195 implied_prefix = 1;
3196 break;
3197 case REPE_PREFIX_OPCODE:
3198 implied_prefix = 2;
3199 break;
3200 case REPNE_PREFIX_OPCODE:
3201 implied_prefix = 3;
3202 break;
3203 default:
3204 abort ();
3205 }
3206
3207 /* 4 byte EVEX prefix. */
3208 i.vex.length = 4;
3209 i.vex.bytes[0] = 0x62;
3210
3211 /* mmmm bits. */
3212 switch (i.tm.opcode_modifier.vexopcode)
3213 {
3214 case VEX0F:
3215 m = 1;
3216 break;
3217 case VEX0F38:
3218 m = 2;
3219 break;
3220 case VEX0F3A:
3221 m = 3;
3222 break;
3223 default:
3224 abort ();
3225 break;
3226 }
3227
3228 /* The high 3 bits of the second EVEX byte are 1's compliment of RXB
3229 bits from REX. */
3230 i.vex.bytes[1] = (~i.rex & 0x7) << 5 | m;
3231
3232 /* The fifth bit of the second EVEX byte is 1's compliment of the
3233 REX_R bit in VREX. */
3234 if (!(i.vrex & REX_R))
3235 i.vex.bytes[1] |= 0x10;
3236 else
3237 vrex_used |= REX_R;
3238
3239 if ((i.reg_operands + i.imm_operands) == i.operands)
3240 {
3241 /* When all operands are registers, the REX_X bit in REX is not
3242 used. We reuse it to encode the upper 16 registers, which is
3243 indicated by the REX_B bit in VREX. The REX_X bit is encoded
3244 as 1's compliment. */
3245 if ((i.vrex & REX_B))
3246 {
3247 vrex_used |= REX_B;
3248 i.vex.bytes[1] &= ~0x40;
3249 }
3250 }
3251
3252 /* EVEX instructions shouldn't need the REX prefix. */
3253 i.vrex &= ~vrex_used;
3254 gas_assert (i.vrex == 0);
3255
3256 /* Check the REX.W bit. */
3257 w = (i.rex & REX_W) ? 1 : 0;
3258 if (i.tm.opcode_modifier.vexw)
3259 {
3260 if (i.tm.opcode_modifier.vexw == VEXW1)
3261 w = 1;
3262 }
3263 /* If w is not set it means we are dealing with WIG instruction. */
3264 else if (!w)
3265 {
3266 if (evexwig == evexw1)
3267 w = 1;
3268 }
3269
3270 /* Encode the U bit. */
3271 implied_prefix |= 0x4;
3272
3273 /* The third byte of the EVEX prefix. */
3274 i.vex.bytes[2] = (w << 7 | register_specifier << 3 | implied_prefix);
3275
3276 /* The fourth byte of the EVEX prefix. */
3277 /* The zeroing-masking bit. */
3278 if (i.mask && i.mask->zeroing)
3279 i.vex.bytes[3] |= 0x80;
3280
3281 /* Don't always set the broadcast bit if there is no RC. */
3282 if (!i.rounding)
3283 {
3284 /* Encode the vector length. */
3285 unsigned int vec_length;
3286
3287 switch (i.tm.opcode_modifier.evex)
3288 {
3289 case EVEXLIG: /* LL' is ignored */
3290 vec_length = evexlig << 5;
3291 break;
3292 case EVEX128:
3293 vec_length = 0 << 5;
3294 break;
3295 case EVEX256:
3296 vec_length = 1 << 5;
3297 break;
3298 case EVEX512:
3299 vec_length = 2 << 5;
3300 break;
3301 default:
3302 abort ();
3303 break;
3304 }
3305 i.vex.bytes[3] |= vec_length;
3306 /* Encode the broadcast bit. */
3307 if (i.broadcast)
3308 i.vex.bytes[3] |= 0x10;
3309 }
3310 else
3311 {
3312 if (i.rounding->type != saeonly)
3313 i.vex.bytes[3] |= 0x10 | (i.rounding->type << 5);
3314 else
3315 i.vex.bytes[3] |= 0x10 | (evexrcig << 5);
3316 }
3317
3318 if (i.mask && i.mask->mask)
3319 i.vex.bytes[3] |= i.mask->mask->reg_num;
3320 }
3321
3322 static void
3323 process_immext (void)
3324 {
3325 expressionS *exp;
3326
3327 if ((i.tm.cpu_flags.bitfield.cpusse3 || i.tm.cpu_flags.bitfield.cpusvme)
3328 && i.operands > 0)
3329 {
3330 /* MONITOR/MWAIT as well as SVME instructions have fixed operands
3331 with an opcode suffix which is coded in the same place as an
3332 8-bit immediate field would be.
3333 Here we check those operands and remove them afterwards. */
3334 unsigned int x;
3335
3336 for (x = 0; x < i.operands; x++)
3337 if (register_number (i.op[x].regs) != x)
3338 as_bad (_("can't use register '%s%s' as operand %d in '%s'."),
3339 register_prefix, i.op[x].regs->reg_name, x + 1,
3340 i.tm.name);
3341
3342 i.operands = 0;
3343 }
3344
3345 /* These AMD 3DNow! and SSE2 instructions have an opcode suffix
3346 which is coded in the same place as an 8-bit immediate field
3347 would be. Here we fake an 8-bit immediate operand from the
3348 opcode suffix stored in tm.extension_opcode.
3349
3350 AVX instructions also use this encoding, for some of
3351 3 argument instructions. */
3352
3353 gas_assert (i.imm_operands <= 1
3354 && (i.operands <= 2
3355 || ((i.tm.opcode_modifier.vex
3356 || i.tm.opcode_modifier.evex)
3357 && i.operands <= 4)));
3358
3359 exp = &im_expressions[i.imm_operands++];
3360 i.op[i.operands].imms = exp;
3361 i.types[i.operands] = imm8;
3362 i.operands++;
3363 exp->X_op = O_constant;
3364 exp->X_add_number = i.tm.extension_opcode;
3365 i.tm.extension_opcode = None;
3366 }
3367
3368
3369 static int
3370 check_hle (void)
3371 {
3372 switch (i.tm.opcode_modifier.hleprefixok)
3373 {
3374 default:
3375 abort ();
3376 case HLEPrefixNone:
3377 as_bad (_("invalid instruction `%s' after `%s'"),
3378 i.tm.name, i.hle_prefix);
3379 return 0;
3380 case HLEPrefixLock:
3381 if (i.prefix[LOCK_PREFIX])
3382 return 1;
3383 as_bad (_("missing `lock' with `%s'"), i.hle_prefix);
3384 return 0;
3385 case HLEPrefixAny:
3386 return 1;
3387 case HLEPrefixRelease:
3388 if (i.prefix[HLE_PREFIX] != XRELEASE_PREFIX_OPCODE)
3389 {
3390 as_bad (_("instruction `%s' after `xacquire' not allowed"),
3391 i.tm.name);
3392 return 0;
3393 }
3394 if (i.mem_operands == 0
3395 || !operand_type_check (i.types[i.operands - 1], anymem))
3396 {
3397 as_bad (_("memory destination needed for instruction `%s'"
3398 " after `xrelease'"), i.tm.name);
3399 return 0;
3400 }
3401 return 1;
3402 }
3403 }
3404
3405 /* This is the guts of the machine-dependent assembler. LINE points to a
3406 machine dependent instruction. This function is supposed to emit
3407 the frags/bytes it assembles to. */
3408
3409 void
3410 md_assemble (char *line)
3411 {
3412 unsigned int j;
3413 char mnemonic[MAX_MNEM_SIZE];
3414 const insn_template *t;
3415
3416 /* Initialize globals. */
3417 memset (&i, '\0', sizeof (i));
3418 for (j = 0; j < MAX_OPERANDS; j++)
3419 i.reloc[j] = NO_RELOC;
3420 memset (disp_expressions, '\0', sizeof (disp_expressions));
3421 memset (im_expressions, '\0', sizeof (im_expressions));
3422 save_stack_p = save_stack;
3423
3424 /* First parse an instruction mnemonic & call i386_operand for the operands.
3425 We assume that the scrubber has arranged it so that line[0] is the valid
3426 start of a (possibly prefixed) mnemonic. */
3427
3428 line = parse_insn (line, mnemonic);
3429 if (line == NULL)
3430 return;
3431
3432 line = parse_operands (line, mnemonic);
3433 this_operand = -1;
3434 if (line == NULL)
3435 return;
3436
3437 /* Now we've parsed the mnemonic into a set of templates, and have the
3438 operands at hand. */
3439
3440 /* All intel opcodes have reversed operands except for "bound" and
3441 "enter". We also don't reverse intersegment "jmp" and "call"
3442 instructions with 2 immediate operands so that the immediate segment
3443 precedes the offset, as it does when in AT&T mode. */
3444 if (intel_syntax
3445 && i.operands > 1
3446 && (strcmp (mnemonic, "bound") != 0)
3447 && (strcmp (mnemonic, "invlpga") != 0)
3448 && !(operand_type_check (i.types[0], imm)
3449 && operand_type_check (i.types[1], imm)))
3450 swap_operands ();
3451
3452 /* The order of the immediates should be reversed
3453 for 2 immediates extrq and insertq instructions */
3454 if (i.imm_operands == 2
3455 && (strcmp (mnemonic, "extrq") == 0
3456 || strcmp (mnemonic, "insertq") == 0))
3457 swap_2_operands (0, 1);
3458
3459 if (i.imm_operands)
3460 optimize_imm ();
3461
3462 /* Don't optimize displacement for movabs since it only takes 64bit
3463 displacement. */
3464 if (i.disp_operands
3465 && i.disp_encoding != disp_encoding_32bit
3466 && (flag_code != CODE_64BIT
3467 || strcmp (mnemonic, "movabs") != 0))
3468 optimize_disp ();
3469
3470 /* Next, we find a template that matches the given insn,
3471 making sure the overlap of the given operands types is consistent
3472 with the template operand types. */
3473
3474 if (!(t = match_template ()))
3475 return;
3476
3477 if (sse_check != check_none
3478 && !i.tm.opcode_modifier.noavx
3479 && (i.tm.cpu_flags.bitfield.cpusse
3480 || i.tm.cpu_flags.bitfield.cpusse2
3481 || i.tm.cpu_flags.bitfield.cpusse3
3482 || i.tm.cpu_flags.bitfield.cpussse3
3483 || i.tm.cpu_flags.bitfield.cpusse4_1
3484 || i.tm.cpu_flags.bitfield.cpusse4_2))
3485 {
3486 (sse_check == check_warning
3487 ? as_warn
3488 : as_bad) (_("SSE instruction `%s' is used"), i.tm.name);
3489 }
3490
3491 /* Zap movzx and movsx suffix. The suffix has been set from
3492 "word ptr" or "byte ptr" on the source operand in Intel syntax
3493 or extracted from mnemonic in AT&T syntax. But we'll use
3494 the destination register to choose the suffix for encoding. */
3495 if ((i.tm.base_opcode & ~9) == 0x0fb6)
3496 {
3497 /* In Intel syntax, there must be a suffix. In AT&T syntax, if
3498 there is no suffix, the default will be byte extension. */
3499 if (i.reg_operands != 2
3500 && !i.suffix
3501 && intel_syntax)
3502 as_bad (_("ambiguous operand size for `%s'"), i.tm.name);
3503
3504 i.suffix = 0;
3505 }
3506
3507 if (i.tm.opcode_modifier.fwait)
3508 if (!add_prefix (FWAIT_OPCODE))
3509 return;
3510
3511 /* Check if REP prefix is OK. */
3512 if (i.rep_prefix && !i.tm.opcode_modifier.repprefixok)
3513 {
3514 as_bad (_("invalid instruction `%s' after `%s'"),
3515 i.tm.name, i.rep_prefix);
3516 return;
3517 }
3518
3519 /* Check for lock without a lockable instruction. Destination operand
3520 must be memory unless it is xchg (0x86). */
3521 if (i.prefix[LOCK_PREFIX]
3522 && (!i.tm.opcode_modifier.islockable
3523 || i.mem_operands == 0
3524 || (i.tm.base_opcode != 0x86
3525 && !operand_type_check (i.types[i.operands - 1], anymem))))
3526 {
3527 as_bad (_("expecting lockable instruction after `lock'"));
3528 return;
3529 }
3530
3531 /* Check if HLE prefix is OK. */
3532 if (i.hle_prefix && !check_hle ())
3533 return;
3534
3535 /* Check BND prefix. */
3536 if (i.bnd_prefix && !i.tm.opcode_modifier.bndprefixok)
3537 as_bad (_("expecting valid branch instruction after `bnd'"));
3538
3539 if (i.tm.cpu_flags.bitfield.cpumpx
3540 && flag_code == CODE_64BIT
3541 && i.prefix[ADDR_PREFIX])
3542 as_bad (_("32-bit address isn't allowed in 64-bit MPX instructions."));
3543
3544 /* Insert BND prefix. */
3545 if (add_bnd_prefix
3546 && i.tm.opcode_modifier.bndprefixok
3547 && !i.prefix[BND_PREFIX])
3548 add_prefix (BND_PREFIX_OPCODE);
3549
3550 /* Check string instruction segment overrides. */
3551 if (i.tm.opcode_modifier.isstring && i.mem_operands != 0)
3552 {
3553 if (!check_string ())
3554 return;
3555 i.disp_operands = 0;
3556 }
3557
3558 if (!process_suffix ())
3559 return;
3560
3561 /* Update operand types. */
3562 for (j = 0; j < i.operands; j++)
3563 i.types[j] = operand_type_and (i.types[j], i.tm.operand_types[j]);
3564
3565 /* Make still unresolved immediate matches conform to size of immediate
3566 given in i.suffix. */
3567 if (!finalize_imm ())
3568 return;
3569
3570 if (i.types[0].bitfield.imm1)
3571 i.imm_operands = 0; /* kludge for shift insns. */
3572
3573 /* We only need to check those implicit registers for instructions
3574 with 3 operands or less. */
3575 if (i.operands <= 3)
3576 for (j = 0; j < i.operands; j++)
3577 if (i.types[j].bitfield.inoutportreg
3578 || i.types[j].bitfield.shiftcount
3579 || i.types[j].bitfield.acc
3580 || i.types[j].bitfield.floatacc)
3581 i.reg_operands--;
3582
3583 /* ImmExt should be processed after SSE2AVX. */
3584 if (!i.tm.opcode_modifier.sse2avx
3585 && i.tm.opcode_modifier.immext)
3586 process_immext ();
3587
3588 /* For insns with operands there are more diddles to do to the opcode. */
3589 if (i.operands)
3590 {
3591 if (!process_operands ())
3592 return;
3593 }
3594 else if (!quiet_warnings && i.tm.opcode_modifier.ugh)
3595 {
3596 /* UnixWare fsub no args is alias for fsubp, fadd -> faddp, etc. */
3597 as_warn (_("translating to `%sp'"), i.tm.name);
3598 }
3599
3600 if (i.tm.opcode_modifier.vex || i.tm.opcode_modifier.evex)
3601 {
3602 if (flag_code == CODE_16BIT)
3603 {
3604 as_bad (_("instruction `%s' isn't supported in 16-bit mode."),
3605 i.tm.name);
3606 return;
3607 }
3608
3609 if (i.tm.opcode_modifier.vex)
3610 build_vex_prefix (t);
3611 else
3612 build_evex_prefix ();
3613 }
3614
3615 /* Handle conversion of 'int $3' --> special int3 insn. XOP or FMA4
3616 instructions may define INT_OPCODE as well, so avoid this corner
3617 case for those instructions that use MODRM. */
3618 if (i.tm.base_opcode == INT_OPCODE
3619 && !i.tm.opcode_modifier.modrm
3620 && i.op[0].imms->X_add_number == 3)
3621 {
3622 i.tm.base_opcode = INT3_OPCODE;
3623 i.imm_operands = 0;
3624 }
3625
3626 if ((i.tm.opcode_modifier.jump
3627 || i.tm.opcode_modifier.jumpbyte
3628 || i.tm.opcode_modifier.jumpdword)
3629 && i.op[0].disps->X_op == O_constant)
3630 {
3631 /* Convert "jmp constant" (and "call constant") to a jump (call) to
3632 the absolute address given by the constant. Since ix86 jumps and
3633 calls are pc relative, we need to generate a reloc. */
3634 i.op[0].disps->X_add_symbol = &abs_symbol;
3635 i.op[0].disps->X_op = O_symbol;
3636 }
3637
3638 if (i.tm.opcode_modifier.rex64)
3639 i.rex |= REX_W;
3640
3641 /* For 8 bit registers we need an empty rex prefix. Also if the
3642 instruction already has a prefix, we need to convert old
3643 registers to new ones. */
3644
3645 if ((i.types[0].bitfield.reg8
3646 && (i.op[0].regs->reg_flags & RegRex64) != 0)
3647 || (i.types[1].bitfield.reg8
3648 && (i.op[1].regs->reg_flags & RegRex64) != 0)
3649 || ((i.types[0].bitfield.reg8
3650 || i.types[1].bitfield.reg8)
3651 && i.rex != 0))
3652 {
3653 int x;
3654
3655 i.rex |= REX_OPCODE;
3656 for (x = 0; x < 2; x++)
3657 {
3658 /* Look for 8 bit operand that uses old registers. */
3659 if (i.types[x].bitfield.reg8
3660 && (i.op[x].regs->reg_flags & RegRex64) == 0)
3661 {
3662 /* In case it is "hi" register, give up. */
3663 if (i.op[x].regs->reg_num > 3)
3664 as_bad (_("can't encode register '%s%s' in an "
3665 "instruction requiring REX prefix."),
3666 register_prefix, i.op[x].regs->reg_name);
3667
3668 /* Otherwise it is equivalent to the extended register.
3669 Since the encoding doesn't change this is merely
3670 cosmetic cleanup for debug output. */
3671
3672 i.op[x].regs = i.op[x].regs + 8;
3673 }
3674 }
3675 }
3676
3677 if (i.rex != 0)
3678 add_prefix (REX_OPCODE | i.rex);
3679
3680 /* We are ready to output the insn. */
3681 output_insn ();
3682 }
3683
3684 static char *
3685 parse_insn (char *line, char *mnemonic)
3686 {
3687 char *l = line;
3688 char *token_start = l;
3689 char *mnem_p;
3690 int supported;
3691 const insn_template *t;
3692 char *dot_p = NULL;
3693
3694 while (1)
3695 {
3696 mnem_p = mnemonic;
3697 while ((*mnem_p = mnemonic_chars[(unsigned char) *l]) != 0)
3698 {
3699 if (*mnem_p == '.')
3700 dot_p = mnem_p;
3701 mnem_p++;
3702 if (mnem_p >= mnemonic + MAX_MNEM_SIZE)
3703 {
3704 as_bad (_("no such instruction: `%s'"), token_start);
3705 return NULL;
3706 }
3707 l++;
3708 }
3709 if (!is_space_char (*l)
3710 && *l != END_OF_INSN
3711 && (intel_syntax
3712 || (*l != PREFIX_SEPARATOR
3713 && *l != ',')))
3714 {
3715 as_bad (_("invalid character %s in mnemonic"),
3716 output_invalid (*l));
3717 return NULL;
3718 }
3719 if (token_start == l)
3720 {
3721 if (!intel_syntax && *l == PREFIX_SEPARATOR)
3722 as_bad (_("expecting prefix; got nothing"));
3723 else
3724 as_bad (_("expecting mnemonic; got nothing"));
3725 return NULL;
3726 }
3727
3728 /* Look up instruction (or prefix) via hash table. */
3729 current_templates = (const templates *) hash_find (op_hash, mnemonic);
3730
3731 if (*l != END_OF_INSN
3732 && (!is_space_char (*l) || l[1] != END_OF_INSN)
3733 && current_templates
3734 && current_templates->start->opcode_modifier.isprefix)
3735 {
3736 if (!cpu_flags_check_cpu64 (current_templates->start->cpu_flags))
3737 {
3738 as_bad ((flag_code != CODE_64BIT
3739 ? _("`%s' is only supported in 64-bit mode")
3740 : _("`%s' is not supported in 64-bit mode")),
3741 current_templates->start->name);
3742 return NULL;
3743 }
3744 /* If we are in 16-bit mode, do not allow addr16 or data16.
3745 Similarly, in 32-bit mode, do not allow addr32 or data32. */
3746 if ((current_templates->start->opcode_modifier.size16
3747 || current_templates->start->opcode_modifier.size32)
3748 && flag_code != CODE_64BIT
3749 && (current_templates->start->opcode_modifier.size32
3750 ^ (flag_code == CODE_16BIT)))
3751 {
3752 as_bad (_("redundant %s prefix"),
3753 current_templates->start->name);
3754 return NULL;
3755 }
3756 /* Add prefix, checking for repeated prefixes. */
3757 switch (add_prefix (current_templates->start->base_opcode))
3758 {
3759 case PREFIX_EXIST:
3760 return NULL;
3761 case PREFIX_REP:
3762 if (current_templates->start->cpu_flags.bitfield.cpuhle)
3763 i.hle_prefix = current_templates->start->name;
3764 else if (current_templates->start->cpu_flags.bitfield.cpumpx)
3765 i.bnd_prefix = current_templates->start->name;
3766 else
3767 i.rep_prefix = current_templates->start->name;
3768 break;
3769 default:
3770 break;
3771 }
3772 /* Skip past PREFIX_SEPARATOR and reset token_start. */
3773 token_start = ++l;
3774 }
3775 else
3776 break;
3777 }
3778
3779 if (!current_templates)
3780 {
3781 /* Check if we should swap operand or force 32bit displacement in
3782 encoding. */
3783 if (mnem_p - 2 == dot_p && dot_p[1] == 's')
3784 i.swap_operand = 1;
3785 else if (mnem_p - 3 == dot_p
3786 && dot_p[1] == 'd'
3787 && dot_p[2] == '8')
3788 i.disp_encoding = disp_encoding_8bit;
3789 else if (mnem_p - 4 == dot_p
3790 && dot_p[1] == 'd'
3791 && dot_p[2] == '3'
3792 && dot_p[3] == '2')
3793 i.disp_encoding = disp_encoding_32bit;
3794 else
3795 goto check_suffix;
3796 mnem_p = dot_p;
3797 *dot_p = '\0';
3798 current_templates = (const templates *) hash_find (op_hash, mnemonic);
3799 }
3800
3801 if (!current_templates)
3802 {
3803 check_suffix:
3804 /* See if we can get a match by trimming off a suffix. */
3805 switch (mnem_p[-1])
3806 {
3807 case WORD_MNEM_SUFFIX:
3808 if (intel_syntax && (intel_float_operand (mnemonic) & 2))
3809 i.suffix = SHORT_MNEM_SUFFIX;
3810 else
3811 case BYTE_MNEM_SUFFIX:
3812 case QWORD_MNEM_SUFFIX:
3813 i.suffix = mnem_p[-1];
3814 mnem_p[-1] = '\0';
3815 current_templates = (const templates *) hash_find (op_hash,
3816 mnemonic);
3817 break;
3818 case SHORT_MNEM_SUFFIX:
3819 case LONG_MNEM_SUFFIX:
3820 if (!intel_syntax)
3821 {
3822 i.suffix = mnem_p[-1];
3823 mnem_p[-1] = '\0';
3824 current_templates = (const templates *) hash_find (op_hash,
3825 mnemonic);
3826 }
3827 break;
3828
3829 /* Intel Syntax. */
3830 case 'd':
3831 if (intel_syntax)
3832 {
3833 if (intel_float_operand (mnemonic) == 1)
3834 i.suffix = SHORT_MNEM_SUFFIX;
3835 else
3836 i.suffix = LONG_MNEM_SUFFIX;
3837 mnem_p[-1] = '\0';
3838 current_templates = (const templates *) hash_find (op_hash,
3839 mnemonic);
3840 }
3841 break;
3842 }
3843 if (!current_templates)
3844 {
3845 as_bad (_("no such instruction: `%s'"), token_start);
3846 return NULL;
3847 }
3848 }
3849
3850 if (current_templates->start->opcode_modifier.jump
3851 || current_templates->start->opcode_modifier.jumpbyte)
3852 {
3853 /* Check for a branch hint. We allow ",pt" and ",pn" for
3854 predict taken and predict not taken respectively.
3855 I'm not sure that branch hints actually do anything on loop
3856 and jcxz insns (JumpByte) for current Pentium4 chips. They
3857 may work in the future and it doesn't hurt to accept them
3858 now. */
3859 if (l[0] == ',' && l[1] == 'p')
3860 {
3861 if (l[2] == 't')
3862 {
3863 if (!add_prefix (DS_PREFIX_OPCODE))
3864 return NULL;
3865 l += 3;
3866 }
3867 else if (l[2] == 'n')
3868 {
3869 if (!add_prefix (CS_PREFIX_OPCODE))
3870 return NULL;
3871 l += 3;
3872 }
3873 }
3874 }
3875 /* Any other comma loses. */
3876 if (*l == ',')
3877 {
3878 as_bad (_("invalid character %s in mnemonic"),
3879 output_invalid (*l));
3880 return NULL;
3881 }
3882
3883 /* Check if instruction is supported on specified architecture. */
3884 supported = 0;
3885 for (t = current_templates->start; t < current_templates->end; ++t)
3886 {
3887 supported |= cpu_flags_match (t);
3888 if (supported == CPU_FLAGS_PERFECT_MATCH)
3889 goto skip;
3890 }
3891
3892 if (!(supported & CPU_FLAGS_64BIT_MATCH))
3893 {
3894 as_bad (flag_code == CODE_64BIT
3895 ? _("`%s' is not supported in 64-bit mode")
3896 : _("`%s' is only supported in 64-bit mode"),
3897 current_templates->start->name);
3898 return NULL;
3899 }
3900 if (supported != CPU_FLAGS_PERFECT_MATCH)
3901 {
3902 as_bad (_("`%s' is not supported on `%s%s'"),
3903 current_templates->start->name,
3904 cpu_arch_name ? cpu_arch_name : default_arch,
3905 cpu_sub_arch_name ? cpu_sub_arch_name : "");
3906 return NULL;
3907 }
3908
3909 skip:
3910 if (!cpu_arch_flags.bitfield.cpui386
3911 && (flag_code != CODE_16BIT))
3912 {
3913 as_warn (_("use .code16 to ensure correct addressing mode"));
3914 }
3915
3916 return l;
3917 }
3918
3919 static char *
3920 parse_operands (char *l, const char *mnemonic)
3921 {
3922 char *token_start;
3923
3924 /* 1 if operand is pending after ','. */
3925 unsigned int expecting_operand = 0;
3926
3927 /* Non-zero if operand parens not balanced. */
3928 unsigned int paren_not_balanced;
3929
3930 while (*l != END_OF_INSN)
3931 {
3932 /* Skip optional white space before operand. */
3933 if (is_space_char (*l))
3934 ++l;
3935 if (!is_operand_char (*l) && *l != END_OF_INSN)
3936 {
3937 as_bad (_("invalid character %s before operand %d"),
3938 output_invalid (*l),
3939 i.operands + 1);
3940 return NULL;
3941 }
3942 token_start = l; /* after white space */
3943 paren_not_balanced = 0;
3944 while (paren_not_balanced || *l != ',')
3945 {
3946 if (*l == END_OF_INSN)
3947 {
3948 if (paren_not_balanced)
3949 {
3950 if (!intel_syntax)
3951 as_bad (_("unbalanced parenthesis in operand %d."),
3952 i.operands + 1);
3953 else
3954 as_bad (_("unbalanced brackets in operand %d."),
3955 i.operands + 1);
3956 return NULL;
3957 }
3958 else
3959 break; /* we are done */
3960 }
3961 else if (!is_operand_char (*l) && !is_space_char (*l))
3962 {
3963 as_bad (_("invalid character %s in operand %d"),
3964 output_invalid (*l),
3965 i.operands + 1);
3966 return NULL;
3967 }
3968 if (!intel_syntax)
3969 {
3970 if (*l == '(')
3971 ++paren_not_balanced;
3972 if (*l == ')')
3973 --paren_not_balanced;
3974 }
3975 else
3976 {
3977 if (*l == '[')
3978 ++paren_not_balanced;
3979 if (*l == ']')
3980 --paren_not_balanced;
3981 }
3982 l++;
3983 }
3984 if (l != token_start)
3985 { /* Yes, we've read in another operand. */
3986 unsigned int operand_ok;
3987 this_operand = i.operands++;
3988 i.types[this_operand].bitfield.unspecified = 1;
3989 if (i.operands > MAX_OPERANDS)
3990 {
3991 as_bad (_("spurious operands; (%d operands/instruction max)"),
3992 MAX_OPERANDS);
3993 return NULL;
3994 }
3995 /* Now parse operand adding info to 'i' as we go along. */
3996 END_STRING_AND_SAVE (l);
3997
3998 if (intel_syntax)
3999 operand_ok =
4000 i386_intel_operand (token_start,
4001 intel_float_operand (mnemonic));
4002 else
4003 operand_ok = i386_att_operand (token_start);
4004
4005 RESTORE_END_STRING (l);
4006 if (!operand_ok)
4007 return NULL;
4008 }
4009 else
4010 {
4011 if (expecting_operand)
4012 {
4013 expecting_operand_after_comma:
4014 as_bad (_("expecting operand after ','; got nothing"));
4015 return NULL;
4016 }
4017 if (*l == ',')
4018 {
4019 as_bad (_("expecting operand before ','; got nothing"));
4020 return NULL;
4021 }
4022 }
4023
4024 /* Now *l must be either ',' or END_OF_INSN. */
4025 if (*l == ',')
4026 {
4027 if (*++l == END_OF_INSN)
4028 {
4029 /* Just skip it, if it's \n complain. */
4030 goto expecting_operand_after_comma;
4031 }
4032 expecting_operand = 1;
4033 }
4034 }
4035 return l;
4036 }
4037
4038 static void
4039 swap_2_operands (int xchg1, int xchg2)
4040 {
4041 union i386_op temp_op;
4042 i386_operand_type temp_type;
4043 enum bfd_reloc_code_real temp_reloc;
4044
4045 temp_type = i.types[xchg2];
4046 i.types[xchg2] = i.types[xchg1];
4047 i.types[xchg1] = temp_type;
4048 temp_op = i.op[xchg2];
4049 i.op[xchg2] = i.op[xchg1];
4050 i.op[xchg1] = temp_op;
4051 temp_reloc = i.reloc[xchg2];
4052 i.reloc[xchg2] = i.reloc[xchg1];
4053 i.reloc[xchg1] = temp_reloc;
4054
4055 if (i.mask)
4056 {
4057 if (i.mask->operand == xchg1)
4058 i.mask->operand = xchg2;
4059 else if (i.mask->operand == xchg2)
4060 i.mask->operand = xchg1;
4061 }
4062 if (i.broadcast)
4063 {
4064 if (i.broadcast->operand == xchg1)
4065 i.broadcast->operand = xchg2;
4066 else if (i.broadcast->operand == xchg2)
4067 i.broadcast->operand = xchg1;
4068 }
4069 if (i.rounding)
4070 {
4071 if (i.rounding->operand == xchg1)
4072 i.rounding->operand = xchg2;
4073 else if (i.rounding->operand == xchg2)
4074 i.rounding->operand = xchg1;
4075 }
4076 }
4077
4078 static void
4079 swap_operands (void)
4080 {
4081 switch (i.operands)
4082 {
4083 case 5:
4084 case 4:
4085 swap_2_operands (1, i.operands - 2);
4086 case 3:
4087 case 2:
4088 swap_2_operands (0, i.operands - 1);
4089 break;
4090 default:
4091 abort ();
4092 }
4093
4094 if (i.mem_operands == 2)
4095 {
4096 const seg_entry *temp_seg;
4097 temp_seg = i.seg[0];
4098 i.seg[0] = i.seg[1];
4099 i.seg[1] = temp_seg;
4100 }
4101 }
4102
4103 /* Try to ensure constant immediates are represented in the smallest
4104 opcode possible. */
4105 static void
4106 optimize_imm (void)
4107 {
4108 char guess_suffix = 0;
4109 int op;
4110
4111 if (i.suffix)
4112 guess_suffix = i.suffix;
4113 else if (i.reg_operands)
4114 {
4115 /* Figure out a suffix from the last register operand specified.
4116 We can't do this properly yet, ie. excluding InOutPortReg,
4117 but the following works for instructions with immediates.
4118 In any case, we can't set i.suffix yet. */
4119 for (op = i.operands; --op >= 0;)
4120 if (i.types[op].bitfield.reg8)
4121 {
4122 guess_suffix = BYTE_MNEM_SUFFIX;
4123 break;
4124 }
4125 else if (i.types[op].bitfield.reg16)
4126 {
4127 guess_suffix = WORD_MNEM_SUFFIX;
4128 break;
4129 }
4130 else if (i.types[op].bitfield.reg32)
4131 {
4132 guess_suffix = LONG_MNEM_SUFFIX;
4133 break;
4134 }
4135 else if (i.types[op].bitfield.reg64)
4136 {
4137 guess_suffix = QWORD_MNEM_SUFFIX;
4138 break;
4139 }
4140 }
4141 else if ((flag_code == CODE_16BIT) ^ (i.prefix[DATA_PREFIX] != 0))
4142 guess_suffix = WORD_MNEM_SUFFIX;
4143
4144 for (op = i.operands; --op >= 0;)
4145 if (operand_type_check (i.types[op], imm))
4146 {
4147 switch (i.op[op].imms->X_op)
4148 {
4149 case O_constant:
4150 /* If a suffix is given, this operand may be shortened. */
4151 switch (guess_suffix)
4152 {
4153 case LONG_MNEM_SUFFIX:
4154 i.types[op].bitfield.imm32 = 1;
4155 i.types[op].bitfield.imm64 = 1;
4156 break;
4157 case WORD_MNEM_SUFFIX:
4158 i.types[op].bitfield.imm16 = 1;
4159 i.types[op].bitfield.imm32 = 1;
4160 i.types[op].bitfield.imm32s = 1;
4161 i.types[op].bitfield.imm64 = 1;
4162 break;
4163 case BYTE_MNEM_SUFFIX:
4164 i.types[op].bitfield.imm8 = 1;
4165 i.types[op].bitfield.imm8s = 1;
4166 i.types[op].bitfield.imm16 = 1;
4167 i.types[op].bitfield.imm32 = 1;
4168 i.types[op].bitfield.imm32s = 1;
4169 i.types[op].bitfield.imm64 = 1;
4170 break;
4171 }
4172
4173 /* If this operand is at most 16 bits, convert it
4174 to a signed 16 bit number before trying to see
4175 whether it will fit in an even smaller size.
4176 This allows a 16-bit operand such as $0xffe0 to
4177 be recognised as within Imm8S range. */
4178 if ((i.types[op].bitfield.imm16)
4179 && (i.op[op].imms->X_add_number & ~(offsetT) 0xffff) == 0)
4180 {
4181 i.op[op].imms->X_add_number =
4182 (((i.op[op].imms->X_add_number & 0xffff) ^ 0x8000) - 0x8000);
4183 }
4184 if ((i.types[op].bitfield.imm32)
4185 && ((i.op[op].imms->X_add_number & ~(((offsetT) 2 << 31) - 1))
4186 == 0))
4187 {
4188 i.op[op].imms->X_add_number = ((i.op[op].imms->X_add_number
4189 ^ ((offsetT) 1 << 31))
4190 - ((offsetT) 1 << 31));
4191 }
4192 i.types[op]
4193 = operand_type_or (i.types[op],
4194 smallest_imm_type (i.op[op].imms->X_add_number));
4195
4196 /* We must avoid matching of Imm32 templates when 64bit
4197 only immediate is available. */
4198 if (guess_suffix == QWORD_MNEM_SUFFIX)
4199 i.types[op].bitfield.imm32 = 0;
4200 break;
4201
4202 case O_absent:
4203 case O_register:
4204 abort ();
4205
4206 /* Symbols and expressions. */
4207 default:
4208 /* Convert symbolic operand to proper sizes for matching, but don't
4209 prevent matching a set of insns that only supports sizes other
4210 than those matching the insn suffix. */
4211 {
4212 i386_operand_type mask, allowed;
4213 const insn_template *t;
4214
4215 operand_type_set (&mask, 0);
4216 operand_type_set (&allowed, 0);
4217
4218 for (t = current_templates->start;
4219 t < current_templates->end;
4220 ++t)
4221 allowed = operand_type_or (allowed,
4222 t->operand_types[op]);
4223 switch (guess_suffix)
4224 {
4225 case QWORD_MNEM_SUFFIX:
4226 mask.bitfield.imm64 = 1;
4227 mask.bitfield.imm32s = 1;
4228 break;
4229 case LONG_MNEM_SUFFIX:
4230 mask.bitfield.imm32 = 1;
4231 break;
4232 case WORD_MNEM_SUFFIX:
4233 mask.bitfield.imm16 = 1;
4234 break;
4235 case BYTE_MNEM_SUFFIX:
4236 mask.bitfield.imm8 = 1;
4237 break;
4238 default:
4239 break;
4240 }
4241 allowed = operand_type_and (mask, allowed);
4242 if (!operand_type_all_zero (&allowed))
4243 i.types[op] = operand_type_and (i.types[op], mask);
4244 }
4245 break;
4246 }
4247 }
4248 }
4249
4250 /* Try to use the smallest displacement type too. */
4251 static void
4252 optimize_disp (void)
4253 {
4254 int op;
4255
4256 for (op = i.operands; --op >= 0;)
4257 if (operand_type_check (i.types[op], disp))
4258 {
4259 if (i.op[op].disps->X_op == O_constant)
4260 {
4261 offsetT op_disp = i.op[op].disps->X_add_number;
4262
4263 if (i.types[op].bitfield.disp16
4264 && (op_disp & ~(offsetT) 0xffff) == 0)
4265 {
4266 /* If this operand is at most 16 bits, convert
4267 to a signed 16 bit number and don't use 64bit
4268 displacement. */
4269 op_disp = (((op_disp & 0xffff) ^ 0x8000) - 0x8000);
4270 i.types[op].bitfield.disp64 = 0;
4271 }
4272 if (i.types[op].bitfield.disp32
4273 && (op_disp & ~(((offsetT) 2 << 31) - 1)) == 0)
4274 {
4275 /* If this operand is at most 32 bits, convert
4276 to a signed 32 bit number and don't use 64bit
4277 displacement. */
4278 op_disp &= (((offsetT) 2 << 31) - 1);
4279 op_disp = (op_disp ^ ((offsetT) 1 << 31)) - ((addressT) 1 << 31);
4280 i.types[op].bitfield.disp64 = 0;
4281 }
4282 if (!op_disp && i.types[op].bitfield.baseindex)
4283 {
4284 i.types[op].bitfield.disp8 = 0;
4285 i.types[op].bitfield.disp16 = 0;
4286 i.types[op].bitfield.disp32 = 0;
4287 i.types[op].bitfield.disp32s = 0;
4288 i.types[op].bitfield.disp64 = 0;
4289 i.op[op].disps = 0;
4290 i.disp_operands--;
4291 }
4292 else if (flag_code == CODE_64BIT)
4293 {
4294 if (fits_in_signed_long (op_disp))
4295 {
4296 i.types[op].bitfield.disp64 = 0;
4297 i.types[op].bitfield.disp32s = 1;
4298 }
4299 if (i.prefix[ADDR_PREFIX]
4300 && fits_in_unsigned_long (op_disp))
4301 i.types[op].bitfield.disp32 = 1;
4302 }
4303 if ((i.types[op].bitfield.disp32
4304 || i.types[op].bitfield.disp32s
4305 || i.types[op].bitfield.disp16)
4306 && fits_in_signed_byte (op_disp))
4307 i.types[op].bitfield.disp8 = 1;
4308 }
4309 else if (i.reloc[op] == BFD_RELOC_386_TLS_DESC_CALL
4310 || i.reloc[op] == BFD_RELOC_X86_64_TLSDESC_CALL)
4311 {
4312 fix_new_exp (frag_now, frag_more (0) - frag_now->fr_literal, 0,
4313 i.op[op].disps, 0, i.reloc[op]);
4314 i.types[op].bitfield.disp8 = 0;
4315 i.types[op].bitfield.disp16 = 0;
4316 i.types[op].bitfield.disp32 = 0;
4317 i.types[op].bitfield.disp32s = 0;
4318 i.types[op].bitfield.disp64 = 0;
4319 }
4320 else
4321 /* We only support 64bit displacement on constants. */
4322 i.types[op].bitfield.disp64 = 0;
4323 }
4324 }
4325
4326 /* Check if operands are valid for the instruction. */
4327
4328 static int
4329 check_VecOperands (const insn_template *t)
4330 {
4331 unsigned int op;
4332
4333 /* Without VSIB byte, we can't have a vector register for index. */
4334 if (!t->opcode_modifier.vecsib
4335 && i.index_reg
4336 && (i.index_reg->reg_type.bitfield.regxmm
4337 || i.index_reg->reg_type.bitfield.regymm
4338 || i.index_reg->reg_type.bitfield.regzmm))
4339 {
4340 i.error = unsupported_vector_index_register;
4341 return 1;
4342 }
4343
4344 /* Check if default mask is allowed. */
4345 if (t->opcode_modifier.nodefmask
4346 && (!i.mask || i.mask->mask->reg_num == 0))
4347 {
4348 i.error = no_default_mask;
4349 return 1;
4350 }
4351
4352 /* For VSIB byte, we need a vector register for index, and all vector
4353 registers must be distinct. */
4354 if (t->opcode_modifier.vecsib)
4355 {
4356 if (!i.index_reg
4357 || !((t->opcode_modifier.vecsib == VecSIB128
4358 && i.index_reg->reg_type.bitfield.regxmm)
4359 || (t->opcode_modifier.vecsib == VecSIB256
4360 && i.index_reg->reg_type.bitfield.regymm)
4361 || (t->opcode_modifier.vecsib == VecSIB512
4362 && i.index_reg->reg_type.bitfield.regzmm)))
4363 {
4364 i.error = invalid_vsib_address;
4365 return 1;
4366 }
4367
4368 gas_assert (i.reg_operands == 2 || i.mask);
4369 if (i.reg_operands == 2 && !i.mask)
4370 {
4371 gas_assert (i.types[0].bitfield.regxmm
4372 || i.types[0].bitfield.regymm);
4373 gas_assert (i.types[2].bitfield.regxmm
4374 || i.types[2].bitfield.regymm);
4375 if (operand_check == check_none)
4376 return 0;
4377 if (register_number (i.op[0].regs)
4378 != register_number (i.index_reg)
4379 && register_number (i.op[2].regs)
4380 != register_number (i.index_reg)
4381 && register_number (i.op[0].regs)
4382 != register_number (i.op[2].regs))
4383 return 0;
4384 if (operand_check == check_error)
4385 {
4386 i.error = invalid_vector_register_set;
4387 return 1;
4388 }
4389 as_warn (_("mask, index, and destination registers should be distinct"));
4390 }
4391 else if (i.reg_operands == 1 && i.mask)
4392 {
4393 if ((i.types[1].bitfield.regymm
4394 || i.types[1].bitfield.regzmm)
4395 && (register_number (i.op[1].regs)
4396 == register_number (i.index_reg)))
4397 {
4398 if (operand_check == check_error)
4399 {
4400 i.error = invalid_vector_register_set;
4401 return 1;
4402 }
4403 if (operand_check != check_none)
4404 as_warn (_("index and destination registers should be distinct"));
4405 }
4406 }
4407 }
4408
4409 /* Check if broadcast is supported by the instruction and is applied
4410 to the memory operand. */
4411 if (i.broadcast)
4412 {
4413 int broadcasted_opnd_size;
4414
4415 /* Check if specified broadcast is supported in this instruction,
4416 and it's applied to memory operand of DWORD or QWORD type,
4417 depending on VecESize. */
4418 if (i.broadcast->type != t->opcode_modifier.broadcast
4419 || !i.types[i.broadcast->operand].bitfield.mem
4420 || (t->opcode_modifier.vecesize == 0
4421 && !i.types[i.broadcast->operand].bitfield.dword
4422 && !i.types[i.broadcast->operand].bitfield.unspecified)
4423 || (t->opcode_modifier.vecesize == 1
4424 && !i.types[i.broadcast->operand].bitfield.qword
4425 && !i.types[i.broadcast->operand].bitfield.unspecified))
4426 goto bad_broadcast;
4427
4428 broadcasted_opnd_size = t->opcode_modifier.vecesize ? 64 : 32;
4429 if (i.broadcast->type == BROADCAST_1TO16)
4430 broadcasted_opnd_size <<= 4; /* Broadcast 1to16. */
4431 else if (i.broadcast->type == BROADCAST_1TO8)
4432 broadcasted_opnd_size <<= 3; /* Broadcast 1to8. */
4433 else if (i.broadcast->type == BROADCAST_1TO4)
4434 broadcasted_opnd_size <<= 2; /* Broadcast 1to4. */
4435 else if (i.broadcast->type == BROADCAST_1TO2)
4436 broadcasted_opnd_size <<= 1; /* Broadcast 1to2. */
4437 else
4438 goto bad_broadcast;
4439
4440 if ((broadcasted_opnd_size == 256
4441 && !t->operand_types[i.broadcast->operand].bitfield.ymmword)
4442 || (broadcasted_opnd_size == 512
4443 && !t->operand_types[i.broadcast->operand].bitfield.zmmword))
4444 {
4445 bad_broadcast:
4446 i.error = unsupported_broadcast;
4447 return 1;
4448 }
4449 }
4450 /* If broadcast is supported in this instruction, we need to check if
4451 operand of one-element size isn't specified without broadcast. */
4452 else if (t->opcode_modifier.broadcast && i.mem_operands)
4453 {
4454 /* Find memory operand. */
4455 for (op = 0; op < i.operands; op++)
4456 if (operand_type_check (i.types[op], anymem))
4457 break;
4458 gas_assert (op < i.operands);
4459 /* Check size of the memory operand. */
4460 if ((t->opcode_modifier.vecesize == 0
4461 && i.types[op].bitfield.dword)
4462 || (t->opcode_modifier.vecesize == 1
4463 && i.types[op].bitfield.qword))
4464 {
4465 i.error = broadcast_needed;
4466 return 1;
4467 }
4468 }
4469
4470 /* Check if requested masking is supported. */
4471 if (i.mask
4472 && (!t->opcode_modifier.masking
4473 || (i.mask->zeroing
4474 && t->opcode_modifier.masking == MERGING_MASKING)))
4475 {
4476 i.error = unsupported_masking;
4477 return 1;
4478 }
4479
4480 /* Check if masking is applied to dest operand. */
4481 if (i.mask && (i.mask->operand != (int) (i.operands - 1)))
4482 {
4483 i.error = mask_not_on_destination;
4484 return 1;
4485 }
4486
4487 /* Check RC/SAE. */
4488 if (i.rounding)
4489 {
4490 if ((i.rounding->type != saeonly
4491 && !t->opcode_modifier.staticrounding)
4492 || (i.rounding->type == saeonly
4493 && (t->opcode_modifier.staticrounding
4494 || !t->opcode_modifier.sae)))
4495 {
4496 i.error = unsupported_rc_sae;
4497 return 1;
4498 }
4499 /* If the instruction has several immediate operands and one of
4500 them is rounding, the rounding operand should be the last
4501 immediate operand. */
4502 if (i.imm_operands > 1
4503 && i.rounding->operand != (int) (i.imm_operands - 1))
4504 {
4505 i.error = rc_sae_operand_not_last_imm;
4506 return 1;
4507 }
4508 }
4509
4510 /* Check vector Disp8 operand. */
4511 if (t->opcode_modifier.disp8memshift)
4512 {
4513 if (i.broadcast)
4514 i.memshift = t->opcode_modifier.vecesize ? 3 : 2;
4515 else
4516 i.memshift = t->opcode_modifier.disp8memshift;
4517
4518 for (op = 0; op < i.operands; op++)
4519 if (operand_type_check (i.types[op], disp)
4520 && i.op[op].disps->X_op == O_constant)
4521 {
4522 offsetT value = i.op[op].disps->X_add_number;
4523 int vec_disp8_ok = fits_in_vec_disp8 (value);
4524 if (t->operand_types [op].bitfield.vec_disp8)
4525 {
4526 if (vec_disp8_ok)
4527 i.types[op].bitfield.vec_disp8 = 1;
4528 else
4529 {
4530 /* Vector insn can only have Vec_Disp8/Disp32 in
4531 32/64bit modes, and Vec_Disp8/Disp16 in 16bit
4532 mode. */
4533 i.types[op].bitfield.disp8 = 0;
4534 if (flag_code != CODE_16BIT)
4535 i.types[op].bitfield.disp16 = 0;
4536 }
4537 }
4538 else if (flag_code != CODE_16BIT)
4539 {
4540 /* One form of this instruction supports vector Disp8.
4541 Try vector Disp8 if we need to use Disp32. */
4542 if (vec_disp8_ok && !fits_in_signed_byte (value))
4543 {
4544 i.error = try_vector_disp8;
4545 return 1;
4546 }
4547 }
4548 }
4549 }
4550 else
4551 i.memshift = -1;
4552
4553 return 0;
4554 }
4555
4556 /* Check if operands are valid for the instruction. Update VEX
4557 operand types. */
4558
4559 static int
4560 VEX_check_operands (const insn_template *t)
4561 {
4562 /* VREX is only valid with EVEX prefix. */
4563 if (i.need_vrex && !t->opcode_modifier.evex)
4564 {
4565 i.error = invalid_register_operand;
4566 return 1;
4567 }
4568
4569 if (!t->opcode_modifier.vex)
4570 return 0;
4571
4572 /* Only check VEX_Imm4, which must be the first operand. */
4573 if (t->operand_types[0].bitfield.vec_imm4)
4574 {
4575 if (i.op[0].imms->X_op != O_constant
4576 || !fits_in_imm4 (i.op[0].imms->X_add_number))
4577 {
4578 i.error = bad_imm4;
4579 return 1;
4580 }
4581
4582 /* Turn off Imm8 so that update_imm won't complain. */
4583 i.types[0] = vec_imm4;
4584 }
4585
4586 return 0;
4587 }
4588
4589 static const insn_template *
4590 match_template (void)
4591 {
4592 /* Points to template once we've found it. */
4593 const insn_template *t;
4594 i386_operand_type overlap0, overlap1, overlap2, overlap3;
4595 i386_operand_type overlap4;
4596 unsigned int found_reverse_match;
4597 i386_opcode_modifier suffix_check;
4598 i386_operand_type operand_types [MAX_OPERANDS];
4599 int addr_prefix_disp;
4600 unsigned int j;
4601 unsigned int found_cpu_match;
4602 unsigned int check_register;
4603 enum i386_error specific_error = 0;
4604
4605 #if MAX_OPERANDS != 5
4606 # error "MAX_OPERANDS must be 5."
4607 #endif
4608
4609 found_reverse_match = 0;
4610 addr_prefix_disp = -1;
4611
4612 memset (&suffix_check, 0, sizeof (suffix_check));
4613 if (i.suffix == BYTE_MNEM_SUFFIX)
4614 suffix_check.no_bsuf = 1;
4615 else if (i.suffix == WORD_MNEM_SUFFIX)
4616 suffix_check.no_wsuf = 1;
4617 else if (i.suffix == SHORT_MNEM_SUFFIX)
4618 suffix_check.no_ssuf = 1;
4619 else if (i.suffix == LONG_MNEM_SUFFIX)
4620 suffix_check.no_lsuf = 1;
4621 else if (i.suffix == QWORD_MNEM_SUFFIX)
4622 suffix_check.no_qsuf = 1;
4623 else if (i.suffix == LONG_DOUBLE_MNEM_SUFFIX)
4624 suffix_check.no_ldsuf = 1;
4625
4626 /* Must have right number of operands. */
4627 i.error = number_of_operands_mismatch;
4628
4629 for (t = current_templates->start; t < current_templates->end; t++)
4630 {
4631 addr_prefix_disp = -1;
4632
4633 if (i.operands != t->operands)
4634 continue;
4635
4636 /* Check processor support. */
4637 i.error = unsupported;
4638 found_cpu_match = (cpu_flags_match (t)
4639 == CPU_FLAGS_PERFECT_MATCH);
4640 if (!found_cpu_match)
4641 continue;
4642
4643 /* Check old gcc support. */
4644 i.error = old_gcc_only;
4645 if (!old_gcc && t->opcode_modifier.oldgcc)
4646 continue;
4647
4648 /* Check AT&T mnemonic. */
4649 i.error = unsupported_with_intel_mnemonic;
4650 if (intel_mnemonic && t->opcode_modifier.attmnemonic)
4651 continue;
4652
4653 /* Check AT&T/Intel syntax. */
4654 i.error = unsupported_syntax;
4655 if ((intel_syntax && t->opcode_modifier.attsyntax)
4656 || (!intel_syntax && t->opcode_modifier.intelsyntax))
4657 continue;
4658
4659 /* Check the suffix, except for some instructions in intel mode. */
4660 i.error = invalid_instruction_suffix;
4661 if ((!intel_syntax || !t->opcode_modifier.ignoresize)
4662 && ((t->opcode_modifier.no_bsuf && suffix_check.no_bsuf)
4663 || (t->opcode_modifier.no_wsuf && suffix_check.no_wsuf)
4664 || (t->opcode_modifier.no_lsuf && suffix_check.no_lsuf)
4665 || (t->opcode_modifier.no_ssuf && suffix_check.no_ssuf)
4666 || (t->opcode_modifier.no_qsuf && suffix_check.no_qsuf)
4667 || (t->opcode_modifier.no_ldsuf && suffix_check.no_ldsuf)))
4668 continue;
4669
4670 if (!operand_size_match (t))
4671 continue;
4672
4673 for (j = 0; j < MAX_OPERANDS; j++)
4674 operand_types[j] = t->operand_types[j];
4675
4676 /* In general, don't allow 64-bit operands in 32-bit mode. */
4677 if (i.suffix == QWORD_MNEM_SUFFIX
4678 && flag_code != CODE_64BIT
4679 && (intel_syntax
4680 ? (!t->opcode_modifier.ignoresize
4681 && !intel_float_operand (t->name))
4682 : intel_float_operand (t->name) != 2)
4683 && ((!operand_types[0].bitfield.regmmx
4684 && !operand_types[0].bitfield.regxmm
4685 && !operand_types[0].bitfield.regymm
4686 && !operand_types[0].bitfield.regzmm)
4687 || (!operand_types[t->operands > 1].bitfield.regmmx
4688 && operand_types[t->operands > 1].bitfield.regxmm
4689 && operand_types[t->operands > 1].bitfield.regymm
4690 && operand_types[t->operands > 1].bitfield.regzmm))
4691 && (t->base_opcode != 0x0fc7
4692 || t->extension_opcode != 1 /* cmpxchg8b */))
4693 continue;
4694
4695 /* In general, don't allow 32-bit operands on pre-386. */
4696 else if (i.suffix == LONG_MNEM_SUFFIX
4697 && !cpu_arch_flags.bitfield.cpui386
4698 && (intel_syntax
4699 ? (!t->opcode_modifier.ignoresize
4700 && !intel_float_operand (t->name))
4701 : intel_float_operand (t->name) != 2)
4702 && ((!operand_types[0].bitfield.regmmx
4703 && !operand_types[0].bitfield.regxmm)
4704 || (!operand_types[t->operands > 1].bitfield.regmmx
4705 && operand_types[t->operands > 1].bitfield.regxmm)))
4706 continue;
4707
4708 /* Do not verify operands when there are none. */
4709 else
4710 {
4711 if (!t->operands)
4712 /* We've found a match; break out of loop. */
4713 break;
4714 }
4715
4716 /* Address size prefix will turn Disp64/Disp32/Disp16 operand
4717 into Disp32/Disp16/Disp32 operand. */
4718 if (i.prefix[ADDR_PREFIX] != 0)
4719 {
4720 /* There should be only one Disp operand. */
4721 switch (flag_code)
4722 {
4723 case CODE_16BIT:
4724 for (j = 0; j < MAX_OPERANDS; j++)
4725 {
4726 if (operand_types[j].bitfield.disp16)
4727 {
4728 addr_prefix_disp = j;
4729 operand_types[j].bitfield.disp32 = 1;
4730 operand_types[j].bitfield.disp16 = 0;
4731 break;
4732 }
4733 }
4734 break;
4735 case CODE_32BIT:
4736 for (j = 0; j < MAX_OPERANDS; j++)
4737 {
4738 if (operand_types[j].bitfield.disp32)
4739 {
4740 addr_prefix_disp = j;
4741 operand_types[j].bitfield.disp32 = 0;
4742 operand_types[j].bitfield.disp16 = 1;
4743 break;
4744 }
4745 }
4746 break;
4747 case CODE_64BIT:
4748 for (j = 0; j < MAX_OPERANDS; j++)
4749 {
4750 if (operand_types[j].bitfield.disp64)
4751 {
4752 addr_prefix_disp = j;
4753 operand_types[j].bitfield.disp64 = 0;
4754 operand_types[j].bitfield.disp32 = 1;
4755 break;
4756 }
4757 }
4758 break;
4759 }
4760 }
4761
4762 /* We check register size if needed. */
4763 check_register = t->opcode_modifier.checkregsize;
4764 overlap0 = operand_type_and (i.types[0], operand_types[0]);
4765 switch (t->operands)
4766 {
4767 case 1:
4768 if (!operand_type_match (overlap0, i.types[0]))
4769 continue;
4770 break;
4771 case 2:
4772 /* xchg %eax, %eax is a special case. It is an aliase for nop
4773 only in 32bit mode and we can use opcode 0x90. In 64bit
4774 mode, we can't use 0x90 for xchg %eax, %eax since it should
4775 zero-extend %eax to %rax. */
4776 if (flag_code == CODE_64BIT
4777 && t->base_opcode == 0x90
4778 && operand_type_equal (&i.types [0], &acc32)
4779 && operand_type_equal (&i.types [1], &acc32))
4780 continue;
4781 if (i.swap_operand)
4782 {
4783 /* If we swap operand in encoding, we either match
4784 the next one or reverse direction of operands. */
4785 if (t->opcode_modifier.s)
4786 continue;
4787 else if (t->opcode_modifier.d)
4788 goto check_reverse;
4789 }
4790
4791 case 3:
4792 /* If we swap operand in encoding, we match the next one. */
4793 if (i.swap_operand && t->opcode_modifier.s)
4794 continue;
4795 case 4:
4796 case 5:
4797 overlap1 = operand_type_and (i.types[1], operand_types[1]);
4798 if (!operand_type_match (overlap0, i.types[0])
4799 || !operand_type_match (overlap1, i.types[1])
4800 || (check_register
4801 && !operand_type_register_match (overlap0, i.types[0],
4802 operand_types[0],
4803 overlap1, i.types[1],
4804 operand_types[1])))
4805 {
4806 /* Check if other direction is valid ... */
4807 if (!t->opcode_modifier.d && !t->opcode_modifier.floatd)
4808 continue;
4809
4810 check_reverse:
4811 /* Try reversing direction of operands. */
4812 overlap0 = operand_type_and (i.types[0], operand_types[1]);
4813 overlap1 = operand_type_and (i.types[1], operand_types[0]);
4814 if (!operand_type_match (overlap0, i.types[0])
4815 || !operand_type_match (overlap1, i.types[1])
4816 || (check_register
4817 && !operand_type_register_match (overlap0,
4818 i.types[0],
4819 operand_types[1],
4820 overlap1,
4821 i.types[1],
4822 operand_types[0])))
4823 {
4824 /* Does not match either direction. */
4825 continue;
4826 }
4827 /* found_reverse_match holds which of D or FloatDR
4828 we've found. */
4829 if (t->opcode_modifier.d)
4830 found_reverse_match = Opcode_D;
4831 else if (t->opcode_modifier.floatd)
4832 found_reverse_match = Opcode_FloatD;
4833 else
4834 found_reverse_match = 0;
4835 if (t->opcode_modifier.floatr)
4836 found_reverse_match |= Opcode_FloatR;
4837 }
4838 else
4839 {
4840 /* Found a forward 2 operand match here. */
4841 switch (t->operands)
4842 {
4843 case 5:
4844 overlap4 = operand_type_and (i.types[4],
4845 operand_types[4]);
4846 case 4:
4847 overlap3 = operand_type_and (i.types[3],
4848 operand_types[3]);
4849 case 3:
4850 overlap2 = operand_type_and (i.types[2],
4851 operand_types[2]);
4852 break;
4853 }
4854
4855 switch (t->operands)
4856 {
4857 case 5:
4858 if (!operand_type_match (overlap4, i.types[4])
4859 || !operand_type_register_match (overlap3,
4860 i.types[3],
4861 operand_types[3],
4862 overlap4,
4863 i.types[4],
4864 operand_types[4]))
4865 continue;
4866 case 4:
4867 if (!operand_type_match (overlap3, i.types[3])
4868 || (check_register
4869 && !operand_type_register_match (overlap2,
4870 i.types[2],
4871 operand_types[2],
4872 overlap3,
4873 i.types[3],
4874 operand_types[3])))
4875 continue;
4876 case 3:
4877 /* Here we make use of the fact that there are no
4878 reverse match 3 operand instructions, and all 3
4879 operand instructions only need to be checked for
4880 register consistency between operands 2 and 3. */
4881 if (!operand_type_match (overlap2, i.types[2])
4882 || (check_register
4883 && !operand_type_register_match (overlap1,
4884 i.types[1],
4885 operand_types[1],
4886 overlap2,
4887 i.types[2],
4888 operand_types[2])))
4889 continue;
4890 break;
4891 }
4892 }
4893 /* Found either forward/reverse 2, 3 or 4 operand match here:
4894 slip through to break. */
4895 }
4896 if (!found_cpu_match)
4897 {
4898 found_reverse_match = 0;
4899 continue;
4900 }
4901
4902 /* Check if vector and VEX operands are valid. */
4903 if (check_VecOperands (t) || VEX_check_operands (t))
4904 {
4905 specific_error = i.error;
4906 continue;
4907 }
4908
4909 /* We've found a match; break out of loop. */
4910 break;
4911 }
4912
4913 if (t == current_templates->end)
4914 {
4915 /* We found no match. */
4916 const char *err_msg;
4917 switch (specific_error ? specific_error : i.error)
4918 {
4919 default:
4920 abort ();
4921 case operand_size_mismatch:
4922 err_msg = _("operand size mismatch");
4923 break;
4924 case operand_type_mismatch:
4925 err_msg = _("operand type mismatch");
4926 break;
4927 case register_type_mismatch:
4928 err_msg = _("register type mismatch");
4929 break;
4930 case number_of_operands_mismatch:
4931 err_msg = _("number of operands mismatch");
4932 break;
4933 case invalid_instruction_suffix:
4934 err_msg = _("invalid instruction suffix");
4935 break;
4936 case bad_imm4:
4937 err_msg = _("constant doesn't fit in 4 bits");
4938 break;
4939 case old_gcc_only:
4940 err_msg = _("only supported with old gcc");
4941 break;
4942 case unsupported_with_intel_mnemonic:
4943 err_msg = _("unsupported with Intel mnemonic");
4944 break;
4945 case unsupported_syntax:
4946 err_msg = _("unsupported syntax");
4947 break;
4948 case unsupported:
4949 as_bad (_("unsupported instruction `%s'"),
4950 current_templates->start->name);
4951 return NULL;
4952 case invalid_vsib_address:
4953 err_msg = _("invalid VSIB address");
4954 break;
4955 case invalid_vector_register_set:
4956 err_msg = _("mask, index, and destination registers must be distinct");
4957 break;
4958 case unsupported_vector_index_register:
4959 err_msg = _("unsupported vector index register");
4960 break;
4961 case unsupported_broadcast:
4962 err_msg = _("unsupported broadcast");
4963 break;
4964 case broadcast_not_on_src_operand:
4965 err_msg = _("broadcast not on source memory operand");
4966 break;
4967 case broadcast_needed:
4968 err_msg = _("broadcast is needed for operand of such type");
4969 break;
4970 case unsupported_masking:
4971 err_msg = _("unsupported masking");
4972 break;
4973 case mask_not_on_destination:
4974 err_msg = _("mask not on destination operand");
4975 break;
4976 case no_default_mask:
4977 err_msg = _("default mask isn't allowed");
4978 break;
4979 case unsupported_rc_sae:
4980 err_msg = _("unsupported static rounding/sae");
4981 break;
4982 case rc_sae_operand_not_last_imm:
4983 if (intel_syntax)
4984 err_msg = _("RC/SAE operand must precede immediate operands");
4985 else
4986 err_msg = _("RC/SAE operand must follow immediate operands");
4987 break;
4988 case invalid_register_operand:
4989 err_msg = _("invalid register operand");
4990 break;
4991 }
4992 as_bad (_("%s for `%s'"), err_msg,
4993 current_templates->start->name);
4994 return NULL;
4995 }
4996
4997 if (!quiet_warnings)
4998 {
4999 if (!intel_syntax
5000 && (i.types[0].bitfield.jumpabsolute
5001 != operand_types[0].bitfield.jumpabsolute))
5002 {
5003 as_warn (_("indirect %s without `*'"), t->name);
5004 }
5005
5006 if (t->opcode_modifier.isprefix
5007 && t->opcode_modifier.ignoresize)
5008 {
5009 /* Warn them that a data or address size prefix doesn't
5010 affect assembly of the next line of code. */
5011 as_warn (_("stand-alone `%s' prefix"), t->name);
5012 }
5013 }
5014
5015 /* Copy the template we found. */
5016 i.tm = *t;
5017
5018 if (addr_prefix_disp != -1)
5019 i.tm.operand_types[addr_prefix_disp]
5020 = operand_types[addr_prefix_disp];
5021
5022 if (found_reverse_match)
5023 {
5024 /* If we found a reverse match we must alter the opcode
5025 direction bit. found_reverse_match holds bits to change
5026 (different for int & float insns). */
5027
5028 i.tm.base_opcode ^= found_reverse_match;
5029
5030 i.tm.operand_types[0] = operand_types[1];
5031 i.tm.operand_types[1] = operand_types[0];
5032 }
5033
5034 return t;
5035 }
5036
5037 static int
5038 check_string (void)
5039 {
5040 int mem_op = operand_type_check (i.types[0], anymem) ? 0 : 1;
5041 if (i.tm.operand_types[mem_op].bitfield.esseg)
5042 {
5043 if (i.seg[0] != NULL && i.seg[0] != &es)
5044 {
5045 as_bad (_("`%s' operand %d must use `%ses' segment"),
5046 i.tm.name,
5047 mem_op + 1,
5048 register_prefix);
5049 return 0;
5050 }
5051 /* There's only ever one segment override allowed per instruction.
5052 This instruction possibly has a legal segment override on the
5053 second operand, so copy the segment to where non-string
5054 instructions store it, allowing common code. */
5055 i.seg[0] = i.seg[1];
5056 }
5057 else if (i.tm.operand_types[mem_op + 1].bitfield.esseg)
5058 {
5059 if (i.seg[1] != NULL && i.seg[1] != &es)
5060 {
5061 as_bad (_("`%s' operand %d must use `%ses' segment"),
5062 i.tm.name,
5063 mem_op + 2,
5064 register_prefix);
5065 return 0;
5066 }
5067 }
5068 return 1;
5069 }
5070
5071 static int
5072 process_suffix (void)
5073 {
5074 /* If matched instruction specifies an explicit instruction mnemonic
5075 suffix, use it. */
5076 if (i.tm.opcode_modifier.size16)
5077 i.suffix = WORD_MNEM_SUFFIX;
5078 else if (i.tm.opcode_modifier.size32)
5079 i.suffix = LONG_MNEM_SUFFIX;
5080 else if (i.tm.opcode_modifier.size64)
5081 i.suffix = QWORD_MNEM_SUFFIX;
5082 else if (i.reg_operands)
5083 {
5084 /* If there's no instruction mnemonic suffix we try to invent one
5085 based on register operands. */
5086 if (!i.suffix)
5087 {
5088 /* We take i.suffix from the last register operand specified,
5089 Destination register type is more significant than source
5090 register type. crc32 in SSE4.2 prefers source register
5091 type. */
5092 if (i.tm.base_opcode == 0xf20f38f1)
5093 {
5094 if (i.types[0].bitfield.reg16)
5095 i.suffix = WORD_MNEM_SUFFIX;
5096 else if (i.types[0].bitfield.reg32)
5097 i.suffix = LONG_MNEM_SUFFIX;
5098 else if (i.types[0].bitfield.reg64)
5099 i.suffix = QWORD_MNEM_SUFFIX;
5100 }
5101 else if (i.tm.base_opcode == 0xf20f38f0)
5102 {
5103 if (i.types[0].bitfield.reg8)
5104 i.suffix = BYTE_MNEM_SUFFIX;
5105 }
5106
5107 if (!i.suffix)
5108 {
5109 int op;
5110
5111 if (i.tm.base_opcode == 0xf20f38f1
5112 || i.tm.base_opcode == 0xf20f38f0)
5113 {
5114 /* We have to know the operand size for crc32. */
5115 as_bad (_("ambiguous memory operand size for `%s`"),
5116 i.tm.name);
5117 return 0;
5118 }
5119
5120 for (op = i.operands; --op >= 0;)
5121 if (!i.tm.operand_types[op].bitfield.inoutportreg)
5122 {
5123 if (i.types[op].bitfield.reg8)
5124 {
5125 i.suffix = BYTE_MNEM_SUFFIX;
5126 break;
5127 }
5128 else if (i.types[op].bitfield.reg16)
5129 {
5130 i.suffix = WORD_MNEM_SUFFIX;
5131 break;
5132 }
5133 else if (i.types[op].bitfield.reg32)
5134 {
5135 i.suffix = LONG_MNEM_SUFFIX;
5136 break;
5137 }
5138 else if (i.types[op].bitfield.reg64)
5139 {
5140 i.suffix = QWORD_MNEM_SUFFIX;
5141 break;
5142 }
5143 }
5144 }
5145 }
5146 else if (i.suffix == BYTE_MNEM_SUFFIX)
5147 {
5148 if (intel_syntax
5149 && i.tm.opcode_modifier.ignoresize
5150 && i.tm.opcode_modifier.no_bsuf)
5151 i.suffix = 0;
5152 else if (!check_byte_reg ())
5153 return 0;
5154 }
5155 else if (i.suffix == LONG_MNEM_SUFFIX)
5156 {
5157 if (intel_syntax
5158 && i.tm.opcode_modifier.ignoresize
5159 && i.tm.opcode_modifier.no_lsuf)
5160 i.suffix = 0;
5161 else if (!check_long_reg ())
5162 return 0;
5163 }
5164 else if (i.suffix == QWORD_MNEM_SUFFIX)
5165 {
5166 if (intel_syntax
5167 && i.tm.opcode_modifier.ignoresize
5168 && i.tm.opcode_modifier.no_qsuf)
5169 i.suffix = 0;
5170 else if (!check_qword_reg ())
5171 return 0;
5172 }
5173 else if (i.suffix == WORD_MNEM_SUFFIX)
5174 {
5175 if (intel_syntax
5176 && i.tm.opcode_modifier.ignoresize
5177 && i.tm.opcode_modifier.no_wsuf)
5178 i.suffix = 0;
5179 else if (!check_word_reg ())
5180 return 0;
5181 }
5182 else if (i.suffix == XMMWORD_MNEM_SUFFIX
5183 || i.suffix == YMMWORD_MNEM_SUFFIX
5184 || i.suffix == ZMMWORD_MNEM_SUFFIX)
5185 {
5186 /* Skip if the instruction has x/y/z suffix. match_template
5187 should check if it is a valid suffix. */
5188 }
5189 else if (intel_syntax && i.tm.opcode_modifier.ignoresize)
5190 /* Do nothing if the instruction is going to ignore the prefix. */
5191 ;
5192 else
5193 abort ();
5194 }
5195 else if (i.tm.opcode_modifier.defaultsize
5196 && !i.suffix
5197 /* exclude fldenv/frstor/fsave/fstenv */
5198 && i.tm.opcode_modifier.no_ssuf)
5199 {
5200 i.suffix = stackop_size;
5201 }
5202 else if (intel_syntax
5203 && !i.suffix
5204 && (i.tm.operand_types[0].bitfield.jumpabsolute
5205 || i.tm.opcode_modifier.jumpbyte
5206 || i.tm.opcode_modifier.jumpintersegment
5207 || (i.tm.base_opcode == 0x0f01 /* [ls][gi]dt */
5208 && i.tm.extension_opcode <= 3)))
5209 {
5210 switch (flag_code)
5211 {
5212 case CODE_64BIT:
5213 if (!i.tm.opcode_modifier.no_qsuf)
5214 {
5215 i.suffix = QWORD_MNEM_SUFFIX;
5216 break;
5217 }
5218 case CODE_32BIT:
5219 if (!i.tm.opcode_modifier.no_lsuf)
5220 i.suffix = LONG_MNEM_SUFFIX;
5221 break;
5222 case CODE_16BIT:
5223 if (!i.tm.opcode_modifier.no_wsuf)
5224 i.suffix = WORD_MNEM_SUFFIX;
5225 break;
5226 }
5227 }
5228
5229 if (!i.suffix)
5230 {
5231 if (!intel_syntax)
5232 {
5233 if (i.tm.opcode_modifier.w)
5234 {
5235 as_bad (_("no instruction mnemonic suffix given and "
5236 "no register operands; can't size instruction"));
5237 return 0;
5238 }
5239 }
5240 else
5241 {
5242 unsigned int suffixes;
5243
5244 suffixes = !i.tm.opcode_modifier.no_bsuf;
5245 if (!i.tm.opcode_modifier.no_wsuf)
5246 suffixes |= 1 << 1;
5247 if (!i.tm.opcode_modifier.no_lsuf)
5248 suffixes |= 1 << 2;
5249 if (!i.tm.opcode_modifier.no_ldsuf)
5250 suffixes |= 1 << 3;
5251 if (!i.tm.opcode_modifier.no_ssuf)
5252 suffixes |= 1 << 4;
5253 if (!i.tm.opcode_modifier.no_qsuf)
5254 suffixes |= 1 << 5;
5255
5256 /* There are more than suffix matches. */
5257 if (i.tm.opcode_modifier.w
5258 || ((suffixes & (suffixes - 1))
5259 && !i.tm.opcode_modifier.defaultsize
5260 && !i.tm.opcode_modifier.ignoresize))
5261 {
5262 as_bad (_("ambiguous operand size for `%s'"), i.tm.name);
5263 return 0;
5264 }
5265 }
5266 }
5267
5268 /* Change the opcode based on the operand size given by i.suffix;
5269 We don't need to change things for byte insns. */
5270
5271 if (i.suffix
5272 && i.suffix != BYTE_MNEM_SUFFIX
5273 && i.suffix != XMMWORD_MNEM_SUFFIX
5274 && i.suffix != YMMWORD_MNEM_SUFFIX
5275 && i.suffix != ZMMWORD_MNEM_SUFFIX)
5276 {
5277 /* It's not a byte, select word/dword operation. */
5278 if (i.tm.opcode_modifier.w)
5279 {
5280 if (i.tm.opcode_modifier.shortform)
5281 i.tm.base_opcode |= 8;
5282 else
5283 i.tm.base_opcode |= 1;
5284 }
5285
5286 /* Now select between word & dword operations via the operand
5287 size prefix, except for instructions that will ignore this
5288 prefix anyway. */
5289 if (i.tm.opcode_modifier.addrprefixop0)
5290 {
5291 /* The address size override prefix changes the size of the
5292 first operand. */
5293 if ((flag_code == CODE_32BIT
5294 && i.op->regs[0].reg_type.bitfield.reg16)
5295 || (flag_code != CODE_32BIT
5296 && i.op->regs[0].reg_type.bitfield.reg32))
5297 if (!add_prefix (ADDR_PREFIX_OPCODE))
5298 return 0;
5299 }
5300 else if (i.suffix != QWORD_MNEM_SUFFIX
5301 && i.suffix != LONG_DOUBLE_MNEM_SUFFIX
5302 && !i.tm.opcode_modifier.ignoresize
5303 && !i.tm.opcode_modifier.floatmf
5304 && ((i.suffix == LONG_MNEM_SUFFIX) == (flag_code == CODE_16BIT)
5305 || (flag_code == CODE_64BIT
5306 && i.tm.opcode_modifier.jumpbyte)))
5307 {
5308 unsigned int prefix = DATA_PREFIX_OPCODE;
5309
5310 if (i.tm.opcode_modifier.jumpbyte) /* jcxz, loop */
5311 prefix = ADDR_PREFIX_OPCODE;
5312
5313 if (!add_prefix (prefix))
5314 return 0;
5315 }
5316
5317 /* Set mode64 for an operand. */
5318 if (i.suffix == QWORD_MNEM_SUFFIX
5319 && flag_code == CODE_64BIT
5320 && !i.tm.opcode_modifier.norex64)
5321 {
5322 /* Special case for xchg %rax,%rax. It is NOP and doesn't
5323 need rex64. cmpxchg8b is also a special case. */
5324 if (! (i.operands == 2
5325 && i.tm.base_opcode == 0x90
5326 && i.tm.extension_opcode == None
5327 && operand_type_equal (&i.types [0], &acc64)
5328 && operand_type_equal (&i.types [1], &acc64))
5329 && ! (i.operands == 1
5330 && i.tm.base_opcode == 0xfc7
5331 && i.tm.extension_opcode == 1
5332 && !operand_type_check (i.types [0], reg)
5333 && operand_type_check (i.types [0], anymem)))
5334 i.rex |= REX_W;
5335 }
5336
5337 /* Size floating point instruction. */
5338 if (i.suffix == LONG_MNEM_SUFFIX)
5339 if (i.tm.opcode_modifier.floatmf)
5340 i.tm.base_opcode ^= 4;
5341 }
5342
5343 return 1;
5344 }
5345
5346 static int
5347 check_byte_reg (void)
5348 {
5349 int op;
5350
5351 for (op = i.operands; --op >= 0;)
5352 {
5353 /* If this is an eight bit register, it's OK. If it's the 16 or
5354 32 bit version of an eight bit register, we will just use the
5355 low portion, and that's OK too. */
5356 if (i.types[op].bitfield.reg8)
5357 continue;
5358
5359 /* I/O port address operands are OK too. */
5360 if (i.tm.operand_types[op].bitfield.inoutportreg)
5361 continue;
5362
5363 /* crc32 doesn't generate this warning. */
5364 if (i.tm.base_opcode == 0xf20f38f0)
5365 continue;
5366
5367 if ((i.types[op].bitfield.reg16
5368 || i.types[op].bitfield.reg32
5369 || i.types[op].bitfield.reg64)
5370 && i.op[op].regs->reg_num < 4
5371 /* Prohibit these changes in 64bit mode, since the lowering
5372 would be more complicated. */
5373 && flag_code != CODE_64BIT)
5374 {
5375 #if REGISTER_WARNINGS
5376 if (!quiet_warnings)
5377 as_warn (_("using `%s%s' instead of `%s%s' due to `%c' suffix"),
5378 register_prefix,
5379 (i.op[op].regs + (i.types[op].bitfield.reg16
5380 ? REGNAM_AL - REGNAM_AX
5381 : REGNAM_AL - REGNAM_EAX))->reg_name,
5382 register_prefix,
5383 i.op[op].regs->reg_name,
5384 i.suffix);
5385 #endif
5386 continue;
5387 }
5388 /* Any other register is bad. */
5389 if (i.types[op].bitfield.reg16
5390 || i.types[op].bitfield.reg32
5391 || i.types[op].bitfield.reg64
5392 || i.types[op].bitfield.regmmx
5393 || i.types[op].bitfield.regxmm
5394 || i.types[op].bitfield.regymm
5395 || i.types[op].bitfield.regzmm
5396 || i.types[op].bitfield.sreg2
5397 || i.types[op].bitfield.sreg3
5398 || i.types[op].bitfield.control
5399 || i.types[op].bitfield.debug
5400 || i.types[op].bitfield.test
5401 || i.types[op].bitfield.floatreg
5402 || i.types[op].bitfield.floatacc)
5403 {
5404 as_bad (_("`%s%s' not allowed with `%s%c'"),
5405 register_prefix,
5406 i.op[op].regs->reg_name,
5407 i.tm.name,
5408 i.suffix);
5409 return 0;
5410 }
5411 }
5412 return 1;
5413 }
5414
5415 static int
5416 check_long_reg (void)
5417 {
5418 int op;
5419
5420 for (op = i.operands; --op >= 0;)
5421 /* Reject eight bit registers, except where the template requires
5422 them. (eg. movzb) */
5423 if (i.types[op].bitfield.reg8
5424 && (i.tm.operand_types[op].bitfield.reg16
5425 || i.tm.operand_types[op].bitfield.reg32
5426 || i.tm.operand_types[op].bitfield.acc))
5427 {
5428 as_bad (_("`%s%s' not allowed with `%s%c'"),
5429 register_prefix,
5430 i.op[op].regs->reg_name,
5431 i.tm.name,
5432 i.suffix);
5433 return 0;
5434 }
5435 /* Warn if the e prefix on a general reg is missing. */
5436 else if ((!quiet_warnings || flag_code == CODE_64BIT)
5437 && i.types[op].bitfield.reg16
5438 && (i.tm.operand_types[op].bitfield.reg32
5439 || i.tm.operand_types[op].bitfield.acc))
5440 {
5441 /* Prohibit these changes in the 64bit mode, since the
5442 lowering is more complicated. */
5443 if (flag_code == CODE_64BIT)
5444 {
5445 as_bad (_("incorrect register `%s%s' used with `%c' suffix"),
5446 register_prefix, i.op[op].regs->reg_name,
5447 i.suffix);
5448 return 0;
5449 }
5450 #if REGISTER_WARNINGS
5451 as_warn (_("using `%s%s' instead of `%s%s' due to `%c' suffix"),
5452 register_prefix,
5453 (i.op[op].regs + REGNAM_EAX - REGNAM_AX)->reg_name,
5454 register_prefix, i.op[op].regs->reg_name, i.suffix);
5455 #endif
5456 }
5457 /* Warn if the r prefix on a general reg is present. */
5458 else if (i.types[op].bitfield.reg64
5459 && (i.tm.operand_types[op].bitfield.reg32
5460 || i.tm.operand_types[op].bitfield.acc))
5461 {
5462 if (intel_syntax
5463 && i.tm.opcode_modifier.toqword
5464 && !i.types[0].bitfield.regxmm)
5465 {
5466 /* Convert to QWORD. We want REX byte. */
5467 i.suffix = QWORD_MNEM_SUFFIX;
5468 }
5469 else
5470 {
5471 as_bad (_("incorrect register `%s%s' used with `%c' suffix"),
5472 register_prefix, i.op[op].regs->reg_name,
5473 i.suffix);
5474 return 0;
5475 }
5476 }
5477 return 1;
5478 }
5479
5480 static int
5481 check_qword_reg (void)
5482 {
5483 int op;
5484
5485 for (op = i.operands; --op >= 0; )
5486 /* Reject eight bit registers, except where the template requires
5487 them. (eg. movzb) */
5488 if (i.types[op].bitfield.reg8
5489 && (i.tm.operand_types[op].bitfield.reg16
5490 || i.tm.operand_types[op].bitfield.reg32
5491 || i.tm.operand_types[op].bitfield.acc))
5492 {
5493 as_bad (_("`%s%s' not allowed with `%s%c'"),
5494 register_prefix,
5495 i.op[op].regs->reg_name,
5496 i.tm.name,
5497 i.suffix);
5498 return 0;
5499 }
5500 /* Warn if the r prefix on a general reg is missing. */
5501 else if ((i.types[op].bitfield.reg16
5502 || i.types[op].bitfield.reg32)
5503 && (i.tm.operand_types[op].bitfield.reg32
5504 || i.tm.operand_types[op].bitfield.acc))
5505 {
5506 /* Prohibit these changes in the 64bit mode, since the
5507 lowering is more complicated. */
5508 if (intel_syntax
5509 && i.tm.opcode_modifier.todword
5510 && !i.types[0].bitfield.regxmm)
5511 {
5512 /* Convert to DWORD. We don't want REX byte. */
5513 i.suffix = LONG_MNEM_SUFFIX;
5514 }
5515 else
5516 {
5517 as_bad (_("incorrect register `%s%s' used with `%c' suffix"),
5518 register_prefix, i.op[op].regs->reg_name,
5519 i.suffix);
5520 return 0;
5521 }
5522 }
5523 return 1;
5524 }
5525
5526 static int
5527 check_word_reg (void)
5528 {
5529 int op;
5530 for (op = i.operands; --op >= 0;)
5531 /* Reject eight bit registers, except where the template requires
5532 them. (eg. movzb) */
5533 if (i.types[op].bitfield.reg8
5534 && (i.tm.operand_types[op].bitfield.reg16
5535 || i.tm.operand_types[op].bitfield.reg32
5536 || i.tm.operand_types[op].bitfield.acc))
5537 {
5538 as_bad (_("`%s%s' not allowed with `%s%c'"),
5539 register_prefix,
5540 i.op[op].regs->reg_name,
5541 i.tm.name,
5542 i.suffix);
5543 return 0;
5544 }
5545 /* Warn if the e or r prefix on a general reg is present. */
5546 else if ((!quiet_warnings || flag_code == CODE_64BIT)
5547 && (i.types[op].bitfield.reg32
5548 || i.types[op].bitfield.reg64)
5549 && (i.tm.operand_types[op].bitfield.reg16
5550 || i.tm.operand_types[op].bitfield.acc))
5551 {
5552 /* Prohibit these changes in the 64bit mode, since the
5553 lowering is more complicated. */
5554 if (flag_code == CODE_64BIT)
5555 {
5556 as_bad (_("incorrect register `%s%s' used with `%c' suffix"),
5557 register_prefix, i.op[op].regs->reg_name,
5558 i.suffix);
5559 return 0;
5560 }
5561 #if REGISTER_WARNINGS
5562 as_warn (_("using `%s%s' instead of `%s%s' due to `%c' suffix"),
5563 register_prefix,
5564 (i.op[op].regs + REGNAM_AX - REGNAM_EAX)->reg_name,
5565 register_prefix, i.op[op].regs->reg_name, i.suffix);
5566 #endif
5567 }
5568 return 1;
5569 }
5570
5571 static int
5572 update_imm (unsigned int j)
5573 {
5574 i386_operand_type overlap = i.types[j];
5575 if ((overlap.bitfield.imm8
5576 || overlap.bitfield.imm8s
5577 || overlap.bitfield.imm16
5578 || overlap.bitfield.imm32
5579 || overlap.bitfield.imm32s
5580 || overlap.bitfield.imm64)
5581 && !operand_type_equal (&overlap, &imm8)
5582 && !operand_type_equal (&overlap, &imm8s)
5583 && !operand_type_equal (&overlap, &imm16)
5584 && !operand_type_equal (&overlap, &imm32)
5585 && !operand_type_equal (&overlap, &imm32s)
5586 && !operand_type_equal (&overlap, &imm64))
5587 {
5588 if (i.suffix)
5589 {
5590 i386_operand_type temp;
5591
5592 operand_type_set (&temp, 0);
5593 if (i.suffix == BYTE_MNEM_SUFFIX)
5594 {
5595 temp.bitfield.imm8 = overlap.bitfield.imm8;
5596 temp.bitfield.imm8s = overlap.bitfield.imm8s;
5597 }
5598 else if (i.suffix == WORD_MNEM_SUFFIX)
5599 temp.bitfield.imm16 = overlap.bitfield.imm16;
5600 else if (i.suffix == QWORD_MNEM_SUFFIX)
5601 {
5602 temp.bitfield.imm64 = overlap.bitfield.imm64;
5603 temp.bitfield.imm32s = overlap.bitfield.imm32s;
5604 }
5605 else
5606 temp.bitfield.imm32 = overlap.bitfield.imm32;
5607 overlap = temp;
5608 }
5609 else if (operand_type_equal (&overlap, &imm16_32_32s)
5610 || operand_type_equal (&overlap, &imm16_32)
5611 || operand_type_equal (&overlap, &imm16_32s))
5612 {
5613 if ((flag_code == CODE_16BIT) ^ (i.prefix[DATA_PREFIX] != 0))
5614 overlap = imm16;
5615 else
5616 overlap = imm32s;
5617 }
5618 if (!operand_type_equal (&overlap, &imm8)
5619 && !operand_type_equal (&overlap, &imm8s)
5620 && !operand_type_equal (&overlap, &imm16)
5621 && !operand_type_equal (&overlap, &imm32)
5622 && !operand_type_equal (&overlap, &imm32s)
5623 && !operand_type_equal (&overlap, &imm64))
5624 {
5625 as_bad (_("no instruction mnemonic suffix given; "
5626 "can't determine immediate size"));
5627 return 0;
5628 }
5629 }
5630 i.types[j] = overlap;
5631
5632 return 1;
5633 }
5634
5635 static int
5636 finalize_imm (void)
5637 {
5638 unsigned int j, n;
5639
5640 /* Update the first 2 immediate operands. */
5641 n = i.operands > 2 ? 2 : i.operands;
5642 if (n)
5643 {
5644 for (j = 0; j < n; j++)
5645 if (update_imm (j) == 0)
5646 return 0;
5647
5648 /* The 3rd operand can't be immediate operand. */
5649 gas_assert (operand_type_check (i.types[2], imm) == 0);
5650 }
5651
5652 return 1;
5653 }
5654
5655 static int
5656 bad_implicit_operand (int xmm)
5657 {
5658 const char *ireg = xmm ? "xmm0" : "ymm0";
5659
5660 if (intel_syntax)
5661 as_bad (_("the last operand of `%s' must be `%s%s'"),
5662 i.tm.name, register_prefix, ireg);
5663 else
5664 as_bad (_("the first operand of `%s' must be `%s%s'"),
5665 i.tm.name, register_prefix, ireg);
5666 return 0;
5667 }
5668
5669 static int
5670 process_operands (void)
5671 {
5672 /* Default segment register this instruction will use for memory
5673 accesses. 0 means unknown. This is only for optimizing out
5674 unnecessary segment overrides. */
5675 const seg_entry *default_seg = 0;
5676
5677 if (i.tm.opcode_modifier.sse2avx && i.tm.opcode_modifier.vexvvvv)
5678 {
5679 unsigned int dupl = i.operands;
5680 unsigned int dest = dupl - 1;
5681 unsigned int j;
5682
5683 /* The destination must be an xmm register. */
5684 gas_assert (i.reg_operands
5685 && MAX_OPERANDS > dupl
5686 && operand_type_equal (&i.types[dest], &regxmm));
5687
5688 if (i.tm.opcode_modifier.firstxmm0)
5689 {
5690 /* The first operand is implicit and must be xmm0. */
5691 gas_assert (operand_type_equal (&i.types[0], &regxmm));
5692 if (register_number (i.op[0].regs) != 0)
5693 return bad_implicit_operand (1);
5694
5695 if (i.tm.opcode_modifier.vexsources == VEX3SOURCES)
5696 {
5697 /* Keep xmm0 for instructions with VEX prefix and 3
5698 sources. */
5699 goto duplicate;
5700 }
5701 else
5702 {
5703 /* We remove the first xmm0 and keep the number of
5704 operands unchanged, which in fact duplicates the
5705 destination. */
5706 for (j = 1; j < i.operands; j++)
5707 {
5708 i.op[j - 1] = i.op[j];
5709 i.types[j - 1] = i.types[j];
5710 i.tm.operand_types[j - 1] = i.tm.operand_types[j];
5711 }
5712 }
5713 }
5714 else if (i.tm.opcode_modifier.implicit1stxmm0)
5715 {
5716 gas_assert ((MAX_OPERANDS - 1) > dupl
5717 && (i.tm.opcode_modifier.vexsources
5718 == VEX3SOURCES));
5719
5720 /* Add the implicit xmm0 for instructions with VEX prefix
5721 and 3 sources. */
5722 for (j = i.operands; j > 0; j--)
5723 {
5724 i.op[j] = i.op[j - 1];
5725 i.types[j] = i.types[j - 1];
5726 i.tm.operand_types[j] = i.tm.operand_types[j - 1];
5727 }
5728 i.op[0].regs
5729 = (const reg_entry *) hash_find (reg_hash, "xmm0");
5730 i.types[0] = regxmm;
5731 i.tm.operand_types[0] = regxmm;
5732
5733 i.operands += 2;
5734 i.reg_operands += 2;
5735 i.tm.operands += 2;
5736
5737 dupl++;
5738 dest++;
5739 i.op[dupl] = i.op[dest];
5740 i.types[dupl] = i.types[dest];
5741 i.tm.operand_types[dupl] = i.tm.operand_types[dest];
5742 }
5743 else
5744 {
5745 duplicate:
5746 i.operands++;
5747 i.reg_operands++;
5748 i.tm.operands++;
5749
5750 i.op[dupl] = i.op[dest];
5751 i.types[dupl] = i.types[dest];
5752 i.tm.operand_types[dupl] = i.tm.operand_types[dest];
5753 }
5754
5755 if (i.tm.opcode_modifier.immext)
5756 process_immext ();
5757 }
5758 else if (i.tm.opcode_modifier.firstxmm0)
5759 {
5760 unsigned int j;
5761
5762 /* The first operand is implicit and must be xmm0/ymm0/zmm0. */
5763 gas_assert (i.reg_operands
5764 && (operand_type_equal (&i.types[0], &regxmm)
5765 || operand_type_equal (&i.types[0], &regymm)
5766 || operand_type_equal (&i.types[0], &regzmm)));
5767 if (register_number (i.op[0].regs) != 0)
5768 return bad_implicit_operand (i.types[0].bitfield.regxmm);
5769
5770 for (j = 1; j < i.operands; j++)
5771 {
5772 i.op[j - 1] = i.op[j];
5773 i.types[j - 1] = i.types[j];
5774
5775 /* We need to adjust fields in i.tm since they are used by
5776 build_modrm_byte. */
5777 i.tm.operand_types [j - 1] = i.tm.operand_types [j];
5778 }
5779
5780 i.operands--;
5781 i.reg_operands--;
5782 i.tm.operands--;
5783 }
5784 else if (i.tm.opcode_modifier.regkludge)
5785 {
5786 /* The imul $imm, %reg instruction is converted into
5787 imul $imm, %reg, %reg, and the clr %reg instruction
5788 is converted into xor %reg, %reg. */
5789
5790 unsigned int first_reg_op;
5791
5792 if (operand_type_check (i.types[0], reg))
5793 first_reg_op = 0;
5794 else
5795 first_reg_op = 1;
5796 /* Pretend we saw the extra register operand. */
5797 gas_assert (i.reg_operands == 1
5798 && i.op[first_reg_op + 1].regs == 0);
5799 i.op[first_reg_op + 1].regs = i.op[first_reg_op].regs;
5800 i.types[first_reg_op + 1] = i.types[first_reg_op];
5801 i.operands++;
5802 i.reg_operands++;
5803 }
5804
5805 if (i.tm.opcode_modifier.shortform)
5806 {
5807 if (i.types[0].bitfield.sreg2
5808 || i.types[0].bitfield.sreg3)
5809 {
5810 if (i.tm.base_opcode == POP_SEG_SHORT
5811 && i.op[0].regs->reg_num == 1)
5812 {
5813 as_bad (_("you can't `pop %scs'"), register_prefix);
5814 return 0;
5815 }
5816 i.tm.base_opcode |= (i.op[0].regs->reg_num << 3);
5817 if ((i.op[0].regs->reg_flags & RegRex) != 0)
5818 i.rex |= REX_B;
5819 }
5820 else
5821 {
5822 /* The register or float register operand is in operand
5823 0 or 1. */
5824 unsigned int op;
5825
5826 if (i.types[0].bitfield.floatreg
5827 || operand_type_check (i.types[0], reg))
5828 op = 0;
5829 else
5830 op = 1;
5831 /* Register goes in low 3 bits of opcode. */
5832 i.tm.base_opcode |= i.op[op].regs->reg_num;
5833 if ((i.op[op].regs->reg_flags & RegRex) != 0)
5834 i.rex |= REX_B;
5835 if (!quiet_warnings && i.tm.opcode_modifier.ugh)
5836 {
5837 /* Warn about some common errors, but press on regardless.
5838 The first case can be generated by gcc (<= 2.8.1). */
5839 if (i.operands == 2)
5840 {
5841 /* Reversed arguments on faddp, fsubp, etc. */
5842 as_warn (_("translating to `%s %s%s,%s%s'"), i.tm.name,
5843 register_prefix, i.op[!intel_syntax].regs->reg_name,
5844 register_prefix, i.op[intel_syntax].regs->reg_name);
5845 }
5846 else
5847 {
5848 /* Extraneous `l' suffix on fp insn. */
5849 as_warn (_("translating to `%s %s%s'"), i.tm.name,
5850 register_prefix, i.op[0].regs->reg_name);
5851 }
5852 }
5853 }
5854 }
5855 else if (i.tm.opcode_modifier.modrm)
5856 {
5857 /* The opcode is completed (modulo i.tm.extension_opcode which
5858 must be put into the modrm byte). Now, we make the modrm and
5859 index base bytes based on all the info we've collected. */
5860
5861 default_seg = build_modrm_byte ();
5862 }
5863 else if ((i.tm.base_opcode & ~0x3) == MOV_AX_DISP32)
5864 {
5865 default_seg = &ds;
5866 }
5867 else if (i.tm.opcode_modifier.isstring)
5868 {
5869 /* For the string instructions that allow a segment override
5870 on one of their operands, the default segment is ds. */
5871 default_seg = &ds;
5872 }
5873
5874 if (i.tm.base_opcode == 0x8d /* lea */
5875 && i.seg[0]
5876 && !quiet_warnings)
5877 as_warn (_("segment override on `%s' is ineffectual"), i.tm.name);
5878
5879 /* If a segment was explicitly specified, and the specified segment
5880 is not the default, use an opcode prefix to select it. If we
5881 never figured out what the default segment is, then default_seg
5882 will be zero at this point, and the specified segment prefix will
5883 always be used. */
5884 if ((i.seg[0]) && (i.seg[0] != default_seg))
5885 {
5886 if (!add_prefix (i.seg[0]->seg_prefix))
5887 return 0;
5888 }
5889 return 1;
5890 }
5891
5892 static const seg_entry *
5893 build_modrm_byte (void)
5894 {
5895 const seg_entry *default_seg = 0;
5896 unsigned int source, dest;
5897 int vex_3_sources;
5898
5899 /* The first operand of instructions with VEX prefix and 3 sources
5900 must be VEX_Imm4. */
5901 vex_3_sources = i.tm.opcode_modifier.vexsources == VEX3SOURCES;
5902 if (vex_3_sources)
5903 {
5904 unsigned int nds, reg_slot;
5905 expressionS *exp;
5906
5907 if (i.tm.opcode_modifier.veximmext
5908 && i.tm.opcode_modifier.immext)
5909 {
5910 dest = i.operands - 2;
5911 gas_assert (dest == 3);
5912 }
5913 else
5914 dest = i.operands - 1;
5915 nds = dest - 1;
5916
5917 /* There are 2 kinds of instructions:
5918 1. 5 operands: 4 register operands or 3 register operands
5919 plus 1 memory operand plus one Vec_Imm4 operand, VexXDS, and
5920 VexW0 or VexW1. The destination must be either XMM, YMM or
5921 ZMM register.
5922 2. 4 operands: 4 register operands or 3 register operands
5923 plus 1 memory operand, VexXDS, and VexImmExt */
5924 gas_assert ((i.reg_operands == 4
5925 || (i.reg_operands == 3 && i.mem_operands == 1))
5926 && i.tm.opcode_modifier.vexvvvv == VEXXDS
5927 && (i.tm.opcode_modifier.veximmext
5928 || (i.imm_operands == 1
5929 && i.types[0].bitfield.vec_imm4
5930 && (i.tm.opcode_modifier.vexw == VEXW0
5931 || i.tm.opcode_modifier.vexw == VEXW1)
5932 && (operand_type_equal (&i.tm.operand_types[dest], &regxmm)
5933 || operand_type_equal (&i.tm.operand_types[dest], &regymm)
5934 || operand_type_equal (&i.tm.operand_types[dest], &regzmm)))));
5935
5936 if (i.imm_operands == 0)
5937 {
5938 /* When there is no immediate operand, generate an 8bit
5939 immediate operand to encode the first operand. */
5940 exp = &im_expressions[i.imm_operands++];
5941 i.op[i.operands].imms = exp;
5942 i.types[i.operands] = imm8;
5943 i.operands++;
5944 /* If VexW1 is set, the first operand is the source and
5945 the second operand is encoded in the immediate operand. */
5946 if (i.tm.opcode_modifier.vexw == VEXW1)
5947 {
5948 source = 0;
5949 reg_slot = 1;
5950 }
5951 else
5952 {
5953 source = 1;
5954 reg_slot = 0;
5955 }
5956
5957 /* FMA swaps REG and NDS. */
5958 if (i.tm.cpu_flags.bitfield.cpufma)
5959 {
5960 unsigned int tmp;
5961 tmp = reg_slot;
5962 reg_slot = nds;
5963 nds = tmp;
5964 }
5965
5966 gas_assert (operand_type_equal (&i.tm.operand_types[reg_slot],
5967 &regxmm)
5968 || operand_type_equal (&i.tm.operand_types[reg_slot],
5969 &regymm)
5970 || operand_type_equal (&i.tm.operand_types[reg_slot],
5971 &regzmm));
5972 exp->X_op = O_constant;
5973 exp->X_add_number = register_number (i.op[reg_slot].regs) << 4;
5974 gas_assert ((i.op[reg_slot].regs->reg_flags & RegVRex) == 0);
5975 }
5976 else
5977 {
5978 unsigned int imm_slot;
5979
5980 if (i.tm.opcode_modifier.vexw == VEXW0)
5981 {
5982 /* If VexW0 is set, the third operand is the source and
5983 the second operand is encoded in the immediate
5984 operand. */
5985 source = 2;
5986 reg_slot = 1;
5987 }
5988 else
5989 {
5990 /* VexW1 is set, the second operand is the source and
5991 the third operand is encoded in the immediate
5992 operand. */
5993 source = 1;
5994 reg_slot = 2;
5995 }
5996
5997 if (i.tm.opcode_modifier.immext)
5998 {
5999 /* When ImmExt is set, the immdiate byte is the last
6000 operand. */
6001 imm_slot = i.operands - 1;
6002 source--;
6003 reg_slot--;
6004 }
6005 else
6006 {
6007 imm_slot = 0;
6008
6009 /* Turn on Imm8 so that output_imm will generate it. */
6010 i.types[imm_slot].bitfield.imm8 = 1;
6011 }
6012
6013 gas_assert (operand_type_equal (&i.tm.operand_types[reg_slot],
6014 &regxmm)
6015 || operand_type_equal (&i.tm.operand_types[reg_slot],
6016 &regymm)
6017 || operand_type_equal (&i.tm.operand_types[reg_slot],
6018 &regzmm));
6019 i.op[imm_slot].imms->X_add_number
6020 |= register_number (i.op[reg_slot].regs) << 4;
6021 gas_assert ((i.op[reg_slot].regs->reg_flags & RegVRex) == 0);
6022 }
6023
6024 gas_assert (operand_type_equal (&i.tm.operand_types[nds], &regxmm)
6025 || operand_type_equal (&i.tm.operand_types[nds],
6026 &regymm)
6027 || operand_type_equal (&i.tm.operand_types[nds],
6028 &regzmm));
6029 i.vex.register_specifier = i.op[nds].regs;
6030 }
6031 else
6032 source = dest = 0;
6033
6034 /* i.reg_operands MUST be the number of real register operands;
6035 implicit registers do not count. If there are 3 register
6036 operands, it must be a instruction with VexNDS. For a
6037 instruction with VexNDD, the destination register is encoded
6038 in VEX prefix. If there are 4 register operands, it must be
6039 a instruction with VEX prefix and 3 sources. */
6040 if (i.mem_operands == 0
6041 && ((i.reg_operands == 2
6042 && i.tm.opcode_modifier.vexvvvv <= VEXXDS)
6043 || (i.reg_operands == 3
6044 && i.tm.opcode_modifier.vexvvvv == VEXXDS)
6045 || (i.reg_operands == 4 && vex_3_sources)))
6046 {
6047 switch (i.operands)
6048 {
6049 case 2:
6050 source = 0;
6051 break;
6052 case 3:
6053 /* When there are 3 operands, one of them may be immediate,
6054 which may be the first or the last operand. Otherwise,
6055 the first operand must be shift count register (cl) or it
6056 is an instruction with VexNDS. */
6057 gas_assert (i.imm_operands == 1
6058 || (i.imm_operands == 0
6059 && (i.tm.opcode_modifier.vexvvvv == VEXXDS
6060 || i.types[0].bitfield.shiftcount)));
6061 if (operand_type_check (i.types[0], imm)
6062 || i.types[0].bitfield.shiftcount)
6063 source = 1;
6064 else
6065 source = 0;
6066 break;
6067 case 4:
6068 /* When there are 4 operands, the first two must be 8bit
6069 immediate operands. The source operand will be the 3rd
6070 one.
6071
6072 For instructions with VexNDS, if the first operand
6073 an imm8, the source operand is the 2nd one. If the last
6074 operand is imm8, the source operand is the first one. */
6075 gas_assert ((i.imm_operands == 2
6076 && i.types[0].bitfield.imm8
6077 && i.types[1].bitfield.imm8)
6078 || (i.tm.opcode_modifier.vexvvvv == VEXXDS
6079 && i.imm_operands == 1
6080 && (i.types[0].bitfield.imm8
6081 || i.types[i.operands - 1].bitfield.imm8
6082 || i.rounding)));
6083 if (i.imm_operands == 2)
6084 source = 2;
6085 else
6086 {
6087 if (i.types[0].bitfield.imm8)
6088 source = 1;
6089 else
6090 source = 0;
6091 }
6092 break;
6093 case 5:
6094 if (i.tm.opcode_modifier.evex)
6095 {
6096 /* For EVEX instructions, when there are 5 operands, the
6097 first one must be immediate operand. If the second one
6098 is immediate operand, the source operand is the 3th
6099 one. If the last one is immediate operand, the source
6100 operand is the 2nd one. */
6101 gas_assert (i.imm_operands == 2
6102 && i.tm.opcode_modifier.sae
6103 && operand_type_check (i.types[0], imm));
6104 if (operand_type_check (i.types[1], imm))
6105 source = 2;
6106 else if (operand_type_check (i.types[4], imm))
6107 source = 1;
6108 else
6109 abort ();
6110 }
6111 break;
6112 default:
6113 abort ();
6114 }
6115
6116 if (!vex_3_sources)
6117 {
6118 dest = source + 1;
6119
6120 /* RC/SAE operand could be between DEST and SRC. That happens
6121 when one operand is GPR and the other one is XMM/YMM/ZMM
6122 register. */
6123 if (i.rounding && i.rounding->operand == (int) dest)
6124 dest++;
6125
6126 if (i.tm.opcode_modifier.vexvvvv == VEXXDS)
6127 {
6128 /* For instructions with VexNDS, the register-only source
6129 operand must be 32/64bit integer, XMM, YMM or ZMM
6130 register. It is encoded in VEX prefix. We need to
6131 clear RegMem bit before calling operand_type_equal. */
6132
6133 i386_operand_type op;
6134 unsigned int vvvv;
6135
6136 /* Check register-only source operand when two source
6137 operands are swapped. */
6138 if (!i.tm.operand_types[source].bitfield.baseindex
6139 && i.tm.operand_types[dest].bitfield.baseindex)
6140 {
6141 vvvv = source;
6142 source = dest;
6143 }
6144 else
6145 vvvv = dest;
6146
6147 op = i.tm.operand_types[vvvv];
6148 op.bitfield.regmem = 0;
6149 if ((dest + 1) >= i.operands
6150 || (!op.bitfield.reg32
6151 && op.bitfield.reg64
6152 && !operand_type_equal (&op, &regxmm)
6153 && !operand_type_equal (&op, &regymm)
6154 && !operand_type_equal (&op, &regzmm)
6155 && !operand_type_equal (&op, &regmask)))
6156 abort ();
6157 i.vex.register_specifier = i.op[vvvv].regs;
6158 dest++;
6159 }
6160 }
6161
6162 i.rm.mode = 3;
6163 /* One of the register operands will be encoded in the i.tm.reg
6164 field, the other in the combined i.tm.mode and i.tm.regmem
6165 fields. If no form of this instruction supports a memory
6166 destination operand, then we assume the source operand may
6167 sometimes be a memory operand and so we need to store the
6168 destination in the i.rm.reg field. */
6169 if (!i.tm.operand_types[dest].bitfield.regmem
6170 && operand_type_check (i.tm.operand_types[dest], anymem) == 0)
6171 {
6172 i.rm.reg = i.op[dest].regs->reg_num;
6173 i.rm.regmem = i.op[source].regs->reg_num;
6174 if ((i.op[dest].regs->reg_flags & RegRex) != 0)
6175 i.rex |= REX_R;
6176 if ((i.op[dest].regs->reg_flags & RegVRex) != 0)
6177 i.vrex |= REX_R;
6178 if ((i.op[source].regs->reg_flags & RegRex) != 0)
6179 i.rex |= REX_B;
6180 if ((i.op[source].regs->reg_flags & RegVRex) != 0)
6181 i.vrex |= REX_B;
6182 }
6183 else
6184 {
6185 i.rm.reg = i.op[source].regs->reg_num;
6186 i.rm.regmem = i.op[dest].regs->reg_num;
6187 if ((i.op[dest].regs->reg_flags & RegRex) != 0)
6188 i.rex |= REX_B;
6189 if ((i.op[dest].regs->reg_flags & RegVRex) != 0)
6190 i.vrex |= REX_B;
6191 if ((i.op[source].regs->reg_flags & RegRex) != 0)
6192 i.rex |= REX_R;
6193 if ((i.op[source].regs->reg_flags & RegVRex) != 0)
6194 i.vrex |= REX_R;
6195 }
6196 if (flag_code != CODE_64BIT && (i.rex & (REX_R | REX_B)))
6197 {
6198 if (!i.types[0].bitfield.control
6199 && !i.types[1].bitfield.control)
6200 abort ();
6201 i.rex &= ~(REX_R | REX_B);
6202 add_prefix (LOCK_PREFIX_OPCODE);
6203 }
6204 }
6205 else
6206 { /* If it's not 2 reg operands... */
6207 unsigned int mem;
6208
6209 if (i.mem_operands)
6210 {
6211 unsigned int fake_zero_displacement = 0;
6212 unsigned int op;
6213
6214 for (op = 0; op < i.operands; op++)
6215 if (operand_type_check (i.types[op], anymem))
6216 break;
6217 gas_assert (op < i.operands);
6218
6219 if (i.tm.opcode_modifier.vecsib)
6220 {
6221 if (i.index_reg->reg_num == RegEiz
6222 || i.index_reg->reg_num == RegRiz)
6223 abort ();
6224
6225 i.rm.regmem = ESCAPE_TO_TWO_BYTE_ADDRESSING;
6226 if (!i.base_reg)
6227 {
6228 i.sib.base = NO_BASE_REGISTER;
6229 i.sib.scale = i.log2_scale_factor;
6230 /* No Vec_Disp8 if there is no base. */
6231 i.types[op].bitfield.vec_disp8 = 0;
6232 i.types[op].bitfield.disp8 = 0;
6233 i.types[op].bitfield.disp16 = 0;
6234 i.types[op].bitfield.disp64 = 0;
6235 if (flag_code != CODE_64BIT)
6236 {
6237 /* Must be 32 bit */
6238 i.types[op].bitfield.disp32 = 1;
6239 i.types[op].bitfield.disp32s = 0;
6240 }
6241 else
6242 {
6243 i.types[op].bitfield.disp32 = 0;
6244 i.types[op].bitfield.disp32s = 1;
6245 }
6246 }
6247 i.sib.index = i.index_reg->reg_num;
6248 if ((i.index_reg->reg_flags & RegRex) != 0)
6249 i.rex |= REX_X;
6250 if ((i.index_reg->reg_flags & RegVRex) != 0)
6251 i.vrex |= REX_X;
6252 }
6253
6254 default_seg = &ds;
6255
6256 if (i.base_reg == 0)
6257 {
6258 i.rm.mode = 0;
6259 if (!i.disp_operands)
6260 {
6261 fake_zero_displacement = 1;
6262 /* Instructions with VSIB byte need 32bit displacement
6263 if there is no base register. */
6264 if (i.tm.opcode_modifier.vecsib)
6265 i.types[op].bitfield.disp32 = 1;
6266 }
6267 if (i.index_reg == 0)
6268 {
6269 gas_assert (!i.tm.opcode_modifier.vecsib);
6270 /* Operand is just <disp> */
6271 if (flag_code == CODE_64BIT)
6272 {
6273 /* 64bit mode overwrites the 32bit absolute
6274 addressing by RIP relative addressing and
6275 absolute addressing is encoded by one of the
6276 redundant SIB forms. */
6277 i.rm.regmem = ESCAPE_TO_TWO_BYTE_ADDRESSING;
6278 i.sib.base = NO_BASE_REGISTER;
6279 i.sib.index = NO_INDEX_REGISTER;
6280 i.types[op] = ((i.prefix[ADDR_PREFIX] == 0)
6281 ? disp32s : disp32);
6282 }
6283 else if ((flag_code == CODE_16BIT)
6284 ^ (i.prefix[ADDR_PREFIX] != 0))
6285 {
6286 i.rm.regmem = NO_BASE_REGISTER_16;
6287 i.types[op] = disp16;
6288 }
6289 else
6290 {
6291 i.rm.regmem = NO_BASE_REGISTER;
6292 i.types[op] = disp32;
6293 }
6294 }
6295 else if (!i.tm.opcode_modifier.vecsib)
6296 {
6297 /* !i.base_reg && i.index_reg */
6298 if (i.index_reg->reg_num == RegEiz
6299 || i.index_reg->reg_num == RegRiz)
6300 i.sib.index = NO_INDEX_REGISTER;
6301 else
6302 i.sib.index = i.index_reg->reg_num;
6303 i.sib.base = NO_BASE_REGISTER;
6304 i.sib.scale = i.log2_scale_factor;
6305 i.rm.regmem = ESCAPE_TO_TWO_BYTE_ADDRESSING;
6306 /* No Vec_Disp8 if there is no base. */
6307 i.types[op].bitfield.vec_disp8 = 0;
6308 i.types[op].bitfield.disp8 = 0;
6309 i.types[op].bitfield.disp16 = 0;
6310 i.types[op].bitfield.disp64 = 0;
6311 if (flag_code != CODE_64BIT)
6312 {
6313 /* Must be 32 bit */
6314 i.types[op].bitfield.disp32 = 1;
6315 i.types[op].bitfield.disp32s = 0;
6316 }
6317 else
6318 {
6319 i.types[op].bitfield.disp32 = 0;
6320 i.types[op].bitfield.disp32s = 1;
6321 }
6322 if ((i.index_reg->reg_flags & RegRex) != 0)
6323 i.rex |= REX_X;
6324 }
6325 }
6326 /* RIP addressing for 64bit mode. */
6327 else if (i.base_reg->reg_num == RegRip ||
6328 i.base_reg->reg_num == RegEip)
6329 {
6330 gas_assert (!i.tm.opcode_modifier.vecsib);
6331 i.rm.regmem = NO_BASE_REGISTER;
6332 i.types[op].bitfield.disp8 = 0;
6333 i.types[op].bitfield.disp16 = 0;
6334 i.types[op].bitfield.disp32 = 0;
6335 i.types[op].bitfield.disp32s = 1;
6336 i.types[op].bitfield.disp64 = 0;
6337 i.types[op].bitfield.vec_disp8 = 0;
6338 i.flags[op] |= Operand_PCrel;
6339 if (! i.disp_operands)
6340 fake_zero_displacement = 1;
6341 }
6342 else if (i.base_reg->reg_type.bitfield.reg16)
6343 {
6344 gas_assert (!i.tm.opcode_modifier.vecsib);
6345 switch (i.base_reg->reg_num)
6346 {
6347 case 3: /* (%bx) */
6348 if (i.index_reg == 0)
6349 i.rm.regmem = 7;
6350 else /* (%bx,%si) -> 0, or (%bx,%di) -> 1 */
6351 i.rm.regmem = i.index_reg->reg_num - 6;
6352 break;
6353 case 5: /* (%bp) */
6354 default_seg = &ss;
6355 if (i.index_reg == 0)
6356 {
6357 i.rm.regmem = 6;
6358 if (operand_type_check (i.types[op], disp) == 0)
6359 {
6360 /* fake (%bp) into 0(%bp) */
6361 if (i.tm.operand_types[op].bitfield.vec_disp8)
6362 i.types[op].bitfield.vec_disp8 = 1;
6363 else
6364 i.types[op].bitfield.disp8 = 1;
6365 fake_zero_displacement = 1;
6366 }
6367 }
6368 else /* (%bp,%si) -> 2, or (%bp,%di) -> 3 */
6369 i.rm.regmem = i.index_reg->reg_num - 6 + 2;
6370 break;
6371 default: /* (%si) -> 4 or (%di) -> 5 */
6372 i.rm.regmem = i.base_reg->reg_num - 6 + 4;
6373 }
6374 i.rm.mode = mode_from_disp_size (i.types[op]);
6375 }
6376 else /* i.base_reg and 32/64 bit mode */
6377 {
6378 if (flag_code == CODE_64BIT
6379 && operand_type_check (i.types[op], disp))
6380 {
6381 i386_operand_type temp;
6382 operand_type_set (&temp, 0);
6383 temp.bitfield.disp8 = i.types[op].bitfield.disp8;
6384 temp.bitfield.vec_disp8
6385 = i.types[op].bitfield.vec_disp8;
6386 i.types[op] = temp;
6387 if (i.prefix[ADDR_PREFIX] == 0)
6388 i.types[op].bitfield.disp32s = 1;
6389 else
6390 i.types[op].bitfield.disp32 = 1;
6391 }
6392
6393 if (!i.tm.opcode_modifier.vecsib)
6394 i.rm.regmem = i.base_reg->reg_num;
6395 if ((i.base_reg->reg_flags & RegRex) != 0)
6396 i.rex |= REX_B;
6397 i.sib.base = i.base_reg->reg_num;
6398 /* x86-64 ignores REX prefix bit here to avoid decoder
6399 complications. */
6400 if (!(i.base_reg->reg_flags & RegRex)
6401 && (i.base_reg->reg_num == EBP_REG_NUM
6402 || i.base_reg->reg_num == ESP_REG_NUM))
6403 default_seg = &ss;
6404 if (i.base_reg->reg_num == 5 && i.disp_operands == 0)
6405 {
6406 fake_zero_displacement = 1;
6407 if (i.tm.operand_types [op].bitfield.vec_disp8)
6408 i.types[op].bitfield.vec_disp8 = 1;
6409 else
6410 i.types[op].bitfield.disp8 = 1;
6411 }
6412 i.sib.scale = i.log2_scale_factor;
6413 if (i.index_reg == 0)
6414 {
6415 gas_assert (!i.tm.opcode_modifier.vecsib);
6416 /* <disp>(%esp) becomes two byte modrm with no index
6417 register. We've already stored the code for esp
6418 in i.rm.regmem ie. ESCAPE_TO_TWO_BYTE_ADDRESSING.
6419 Any base register besides %esp will not use the
6420 extra modrm byte. */
6421 i.sib.index = NO_INDEX_REGISTER;
6422 }
6423 else if (!i.tm.opcode_modifier.vecsib)
6424 {
6425 if (i.index_reg->reg_num == RegEiz
6426 || i.index_reg->reg_num == RegRiz)
6427 i.sib.index = NO_INDEX_REGISTER;
6428 else
6429 i.sib.index = i.index_reg->reg_num;
6430 i.rm.regmem = ESCAPE_TO_TWO_BYTE_ADDRESSING;
6431 if ((i.index_reg->reg_flags & RegRex) != 0)
6432 i.rex |= REX_X;
6433 }
6434
6435 if (i.disp_operands
6436 && (i.reloc[op] == BFD_RELOC_386_TLS_DESC_CALL
6437 || i.reloc[op] == BFD_RELOC_X86_64_TLSDESC_CALL))
6438 i.rm.mode = 0;
6439 else
6440 {
6441 if (!fake_zero_displacement
6442 && !i.disp_operands
6443 && i.disp_encoding)
6444 {
6445 fake_zero_displacement = 1;
6446 if (i.disp_encoding == disp_encoding_8bit)
6447 i.types[op].bitfield.disp8 = 1;
6448 else
6449 i.types[op].bitfield.disp32 = 1;
6450 }
6451 i.rm.mode = mode_from_disp_size (i.types[op]);
6452 }
6453 }
6454
6455 if (fake_zero_displacement)
6456 {
6457 /* Fakes a zero displacement assuming that i.types[op]
6458 holds the correct displacement size. */
6459 expressionS *exp;
6460
6461 gas_assert (i.op[op].disps == 0);
6462 exp = &disp_expressions[i.disp_operands++];
6463 i.op[op].disps = exp;
6464 exp->X_op = O_constant;
6465 exp->X_add_number = 0;
6466 exp->X_add_symbol = (symbolS *) 0;
6467 exp->X_op_symbol = (symbolS *) 0;
6468 }
6469
6470 mem = op;
6471 }
6472 else
6473 mem = ~0;
6474
6475 if (i.tm.opcode_modifier.vexsources == XOP2SOURCES)
6476 {
6477 if (operand_type_check (i.types[0], imm))
6478 i.vex.register_specifier = NULL;
6479 else
6480 {
6481 /* VEX.vvvv encodes one of the sources when the first
6482 operand is not an immediate. */
6483 if (i.tm.opcode_modifier.vexw == VEXW0)
6484 i.vex.register_specifier = i.op[0].regs;
6485 else
6486 i.vex.register_specifier = i.op[1].regs;
6487 }
6488
6489 /* Destination is a XMM register encoded in the ModRM.reg
6490 and VEX.R bit. */
6491 i.rm.reg = i.op[2].regs->reg_num;
6492 if ((i.op[2].regs->reg_flags & RegRex) != 0)
6493 i.rex |= REX_R;
6494
6495 /* ModRM.rm and VEX.B encodes the other source. */
6496 if (!i.mem_operands)
6497 {
6498 i.rm.mode = 3;
6499
6500 if (i.tm.opcode_modifier.vexw == VEXW0)
6501 i.rm.regmem = i.op[1].regs->reg_num;
6502 else
6503 i.rm.regmem = i.op[0].regs->reg_num;
6504
6505 if ((i.op[1].regs->reg_flags & RegRex) != 0)
6506 i.rex |= REX_B;
6507 }
6508 }
6509 else if (i.tm.opcode_modifier.vexvvvv == VEXLWP)
6510 {
6511 i.vex.register_specifier = i.op[2].regs;
6512 if (!i.mem_operands)
6513 {
6514 i.rm.mode = 3;
6515 i.rm.regmem = i.op[1].regs->reg_num;
6516 if ((i.op[1].regs->reg_flags & RegRex) != 0)
6517 i.rex |= REX_B;
6518 }
6519 }
6520 /* Fill in i.rm.reg or i.rm.regmem field with register operand
6521 (if any) based on i.tm.extension_opcode. Again, we must be
6522 careful to make sure that segment/control/debug/test/MMX
6523 registers are coded into the i.rm.reg field. */
6524 else if (i.reg_operands)
6525 {
6526 unsigned int op;
6527 unsigned int vex_reg = ~0;
6528
6529 for (op = 0; op < i.operands; op++)
6530 if (i.types[op].bitfield.reg8
6531 || i.types[op].bitfield.reg16
6532 || i.types[op].bitfield.reg32
6533 || i.types[op].bitfield.reg64
6534 || i.types[op].bitfield.regmmx
6535 || i.types[op].bitfield.regxmm
6536 || i.types[op].bitfield.regymm
6537 || i.types[op].bitfield.regbnd
6538 || i.types[op].bitfield.regzmm
6539 || i.types[op].bitfield.regmask
6540 || i.types[op].bitfield.sreg2
6541 || i.types[op].bitfield.sreg3
6542 || i.types[op].bitfield.control
6543 || i.types[op].bitfield.debug
6544 || i.types[op].bitfield.test)
6545 break;
6546
6547 if (vex_3_sources)
6548 op = dest;
6549 else if (i.tm.opcode_modifier.vexvvvv == VEXXDS)
6550 {
6551 /* For instructions with VexNDS, the register-only
6552 source operand is encoded in VEX prefix. */
6553 gas_assert (mem != (unsigned int) ~0);
6554
6555 if (op > mem)
6556 {
6557 vex_reg = op++;
6558 gas_assert (op < i.operands);
6559 }
6560 else
6561 {
6562 /* Check register-only source operand when two source
6563 operands are swapped. */
6564 if (!i.tm.operand_types[op].bitfield.baseindex
6565 && i.tm.operand_types[op + 1].bitfield.baseindex)
6566 {
6567 vex_reg = op;
6568 op += 2;
6569 gas_assert (mem == (vex_reg + 1)
6570 && op < i.operands);
6571 }
6572 else
6573 {
6574 vex_reg = op + 1;
6575 gas_assert (vex_reg < i.operands);
6576 }
6577 }
6578 }
6579 else if (i.tm.opcode_modifier.vexvvvv == VEXNDD)
6580 {
6581 /* For instructions with VexNDD, the register destination
6582 is encoded in VEX prefix. */
6583 if (i.mem_operands == 0)
6584 {
6585 /* There is no memory operand. */
6586 gas_assert ((op + 2) == i.operands);
6587 vex_reg = op + 1;
6588 }
6589 else
6590 {
6591 /* There are only 2 operands. */
6592 gas_assert (op < 2 && i.operands == 2);
6593 vex_reg = 1;
6594 }
6595 }
6596 else
6597 gas_assert (op < i.operands);
6598
6599 if (vex_reg != (unsigned int) ~0)
6600 {
6601 i386_operand_type *type = &i.tm.operand_types[vex_reg];
6602
6603 if (type->bitfield.reg32 != 1
6604 && type->bitfield.reg64 != 1
6605 && !operand_type_equal (type, &regxmm)
6606 && !operand_type_equal (type, &regymm)
6607 && !operand_type_equal (type, &regzmm)
6608 && !operand_type_equal (type, &regmask))
6609 abort ();
6610
6611 i.vex.register_specifier = i.op[vex_reg].regs;
6612 }
6613
6614 /* Don't set OP operand twice. */
6615 if (vex_reg != op)
6616 {
6617 /* If there is an extension opcode to put here, the
6618 register number must be put into the regmem field. */
6619 if (i.tm.extension_opcode != None)
6620 {
6621 i.rm.regmem = i.op[op].regs->reg_num;
6622 if ((i.op[op].regs->reg_flags & RegRex) != 0)
6623 i.rex |= REX_B;
6624 if ((i.op[op].regs->reg_flags & RegVRex) != 0)
6625 i.vrex |= REX_B;
6626 }
6627 else
6628 {
6629 i.rm.reg = i.op[op].regs->reg_num;
6630 if ((i.op[op].regs->reg_flags & RegRex) != 0)
6631 i.rex |= REX_R;
6632 if ((i.op[op].regs->reg_flags & RegVRex) != 0)
6633 i.vrex |= REX_R;
6634 }
6635 }
6636
6637 /* Now, if no memory operand has set i.rm.mode = 0, 1, 2 we
6638 must set it to 3 to indicate this is a register operand
6639 in the regmem field. */
6640 if (!i.mem_operands)
6641 i.rm.mode = 3;
6642 }
6643
6644 /* Fill in i.rm.reg field with extension opcode (if any). */
6645 if (i.tm.extension_opcode != None)
6646 i.rm.reg = i.tm.extension_opcode;
6647 }
6648 return default_seg;
6649 }
6650
6651 static void
6652 output_branch (void)
6653 {
6654 char *p;
6655 int size;
6656 int code16;
6657 int prefix;
6658 relax_substateT subtype;
6659 symbolS *sym;
6660 offsetT off;
6661
6662 code16 = flag_code == CODE_16BIT ? CODE16 : 0;
6663 size = i.disp_encoding == disp_encoding_32bit ? BIG : SMALL;
6664
6665 prefix = 0;
6666 if (i.prefix[DATA_PREFIX] != 0)
6667 {
6668 prefix = 1;
6669 i.prefixes -= 1;
6670 code16 ^= CODE16;
6671 }
6672 /* Pentium4 branch hints. */
6673 if (i.prefix[SEG_PREFIX] == CS_PREFIX_OPCODE /* not taken */
6674 || i.prefix[SEG_PREFIX] == DS_PREFIX_OPCODE /* taken */)
6675 {
6676 prefix++;
6677 i.prefixes--;
6678 }
6679 if (i.prefix[REX_PREFIX] != 0)
6680 {
6681 prefix++;
6682 i.prefixes--;
6683 }
6684
6685 /* BND prefixed jump. */
6686 if (i.prefix[BND_PREFIX] != 0)
6687 {
6688 FRAG_APPEND_1_CHAR (i.prefix[BND_PREFIX]);
6689 i.prefixes -= 1;
6690 }
6691
6692 if (i.prefixes != 0 && !intel_syntax)
6693 as_warn (_("skipping prefixes on this instruction"));
6694
6695 /* It's always a symbol; End frag & setup for relax.
6696 Make sure there is enough room in this frag for the largest
6697 instruction we may generate in md_convert_frag. This is 2
6698 bytes for the opcode and room for the prefix and largest
6699 displacement. */
6700 frag_grow (prefix + 2 + 4);
6701 /* Prefix and 1 opcode byte go in fr_fix. */
6702 p = frag_more (prefix + 1);
6703 if (i.prefix[DATA_PREFIX] != 0)
6704 *p++ = DATA_PREFIX_OPCODE;
6705 if (i.prefix[SEG_PREFIX] == CS_PREFIX_OPCODE
6706 || i.prefix[SEG_PREFIX] == DS_PREFIX_OPCODE)
6707 *p++ = i.prefix[SEG_PREFIX];
6708 if (i.prefix[REX_PREFIX] != 0)
6709 *p++ = i.prefix[REX_PREFIX];
6710 *p = i.tm.base_opcode;
6711
6712 if ((unsigned char) *p == JUMP_PC_RELATIVE)
6713 subtype = ENCODE_RELAX_STATE (UNCOND_JUMP, size);
6714 else if (cpu_arch_flags.bitfield.cpui386)
6715 subtype = ENCODE_RELAX_STATE (COND_JUMP, size);
6716 else
6717 subtype = ENCODE_RELAX_STATE (COND_JUMP86, size);
6718 subtype |= code16;
6719
6720 sym = i.op[0].disps->X_add_symbol;
6721 off = i.op[0].disps->X_add_number;
6722
6723 if (i.op[0].disps->X_op != O_constant
6724 && i.op[0].disps->X_op != O_symbol)
6725 {
6726 /* Handle complex expressions. */
6727 sym = make_expr_symbol (i.op[0].disps);
6728 off = 0;
6729 }
6730
6731 /* 1 possible extra opcode + 4 byte displacement go in var part.
6732 Pass reloc in fr_var. */
6733 frag_var (rs_machine_dependent, 5, i.reloc[0], subtype, sym, off, p);
6734 }
6735
6736 static void
6737 output_jump (void)
6738 {
6739 char *p;
6740 int size;
6741 fixS *fixP;
6742
6743 if (i.tm.opcode_modifier.jumpbyte)
6744 {
6745 /* This is a loop or jecxz type instruction. */
6746 size = 1;
6747 if (i.prefix[ADDR_PREFIX] != 0)
6748 {
6749 FRAG_APPEND_1_CHAR (ADDR_PREFIX_OPCODE);
6750 i.prefixes -= 1;
6751 }
6752 /* Pentium4 branch hints. */
6753 if (i.prefix[SEG_PREFIX] == CS_PREFIX_OPCODE /* not taken */
6754 || i.prefix[SEG_PREFIX] == DS_PREFIX_OPCODE /* taken */)
6755 {
6756 FRAG_APPEND_1_CHAR (i.prefix[SEG_PREFIX]);
6757 i.prefixes--;
6758 }
6759 }
6760 else
6761 {
6762 int code16;
6763
6764 code16 = 0;
6765 if (flag_code == CODE_16BIT)
6766 code16 = CODE16;
6767
6768 if (i.prefix[DATA_PREFIX] != 0)
6769 {
6770 FRAG_APPEND_1_CHAR (DATA_PREFIX_OPCODE);
6771 i.prefixes -= 1;
6772 code16 ^= CODE16;
6773 }
6774
6775 size = 4;
6776 if (code16)
6777 size = 2;
6778 }
6779
6780 if (i.prefix[REX_PREFIX] != 0)
6781 {
6782 FRAG_APPEND_1_CHAR (i.prefix[REX_PREFIX]);
6783 i.prefixes -= 1;
6784 }
6785
6786 /* BND prefixed jump. */
6787 if (i.prefix[BND_PREFIX] != 0)
6788 {
6789 FRAG_APPEND_1_CHAR (i.prefix[BND_PREFIX]);
6790 i.prefixes -= 1;
6791 }
6792
6793 if (i.prefixes != 0 && !intel_syntax)
6794 as_warn (_("skipping prefixes on this instruction"));
6795
6796 p = frag_more (i.tm.opcode_length + size);
6797 switch (i.tm.opcode_length)
6798 {
6799 case 2:
6800 *p++ = i.tm.base_opcode >> 8;
6801 case 1:
6802 *p++ = i.tm.base_opcode;
6803 break;
6804 default:
6805 abort ();
6806 }
6807
6808 fixP = fix_new_exp (frag_now, p - frag_now->fr_literal, size,
6809 i.op[0].disps, 1, reloc (size, 1, 1, i.reloc[0]));
6810
6811 /* All jumps handled here are signed, but don't use a signed limit
6812 check for 32 and 16 bit jumps as we want to allow wrap around at
6813 4G and 64k respectively. */
6814 if (size == 1)
6815 fixP->fx_signed = 1;
6816 }
6817
6818 static void
6819 output_interseg_jump (void)
6820 {
6821 char *p;
6822 int size;
6823 int prefix;
6824 int code16;
6825
6826 code16 = 0;
6827 if (flag_code == CODE_16BIT)
6828 code16 = CODE16;
6829
6830 prefix = 0;
6831 if (i.prefix[DATA_PREFIX] != 0)
6832 {
6833 prefix = 1;
6834 i.prefixes -= 1;
6835 code16 ^= CODE16;
6836 }
6837 if (i.prefix[REX_PREFIX] != 0)
6838 {
6839 prefix++;
6840 i.prefixes -= 1;
6841 }
6842
6843 size = 4;
6844 if (code16)
6845 size = 2;
6846
6847 if (i.prefixes != 0 && !intel_syntax)
6848 as_warn (_("skipping prefixes on this instruction"));
6849
6850 /* 1 opcode; 2 segment; offset */
6851 p = frag_more (prefix + 1 + 2 + size);
6852
6853 if (i.prefix[DATA_PREFIX] != 0)
6854 *p++ = DATA_PREFIX_OPCODE;
6855
6856 if (i.prefix[REX_PREFIX] != 0)
6857 *p++ = i.prefix[REX_PREFIX];
6858
6859 *p++ = i.tm.base_opcode;
6860 if (i.op[1].imms->X_op == O_constant)
6861 {
6862 offsetT n = i.op[1].imms->X_add_number;
6863
6864 if (size == 2
6865 && !fits_in_unsigned_word (n)
6866 && !fits_in_signed_word (n))
6867 {
6868 as_bad (_("16-bit jump out of range"));
6869 return;
6870 }
6871 md_number_to_chars (p, n, size);
6872 }
6873 else
6874 fix_new_exp (frag_now, p - frag_now->fr_literal, size,
6875 i.op[1].imms, 0, reloc (size, 0, 0, i.reloc[1]));
6876 if (i.op[0].imms->X_op != O_constant)
6877 as_bad (_("can't handle non absolute segment in `%s'"),
6878 i.tm.name);
6879 md_number_to_chars (p + size, (valueT) i.op[0].imms->X_add_number, 2);
6880 }
6881
6882 static void
6883 output_insn (void)
6884 {
6885 fragS *insn_start_frag;
6886 offsetT insn_start_off;
6887
6888 /* Tie dwarf2 debug info to the address at the start of the insn.
6889 We can't do this after the insn has been output as the current
6890 frag may have been closed off. eg. by frag_var. */
6891 dwarf2_emit_insn (0);
6892
6893 insn_start_frag = frag_now;
6894 insn_start_off = frag_now_fix ();
6895
6896 /* Output jumps. */
6897 if (i.tm.opcode_modifier.jump)
6898 output_branch ();
6899 else if (i.tm.opcode_modifier.jumpbyte
6900 || i.tm.opcode_modifier.jumpdword)
6901 output_jump ();
6902 else if (i.tm.opcode_modifier.jumpintersegment)
6903 output_interseg_jump ();
6904 else
6905 {
6906 /* Output normal instructions here. */
6907 char *p;
6908 unsigned char *q;
6909 unsigned int j;
6910 unsigned int prefix;
6911
6912 /* Some processors fail on LOCK prefix. This options makes
6913 assembler ignore LOCK prefix and serves as a workaround. */
6914 if (omit_lock_prefix)
6915 {
6916 if (i.tm.base_opcode == LOCK_PREFIX_OPCODE)
6917 return;
6918 i.prefix[LOCK_PREFIX] = 0;
6919 }
6920
6921 /* Since the VEX/EVEX prefix contains the implicit prefix, we
6922 don't need the explicit prefix. */
6923 if (!i.tm.opcode_modifier.vex && !i.tm.opcode_modifier.evex)
6924 {
6925 switch (i.tm.opcode_length)
6926 {
6927 case 3:
6928 if (i.tm.base_opcode & 0xff000000)
6929 {
6930 prefix = (i.tm.base_opcode >> 24) & 0xff;
6931 goto check_prefix;
6932 }
6933 break;
6934 case 2:
6935 if ((i.tm.base_opcode & 0xff0000) != 0)
6936 {
6937 prefix = (i.tm.base_opcode >> 16) & 0xff;
6938 if (i.tm.cpu_flags.bitfield.cpupadlock)
6939 {
6940 check_prefix:
6941 if (prefix != REPE_PREFIX_OPCODE
6942 || (i.prefix[REP_PREFIX]
6943 != REPE_PREFIX_OPCODE))
6944 add_prefix (prefix);
6945 }
6946 else
6947 add_prefix (prefix);
6948 }
6949 break;
6950 case 1:
6951 break;
6952 default:
6953 abort ();
6954 }
6955
6956 #if defined (OBJ_MAYBE_ELF) || defined (OBJ_ELF)
6957 /* For x32, add a dummy REX_OPCODE prefix for mov/add with
6958 R_X86_64_GOTTPOFF relocation so that linker can safely
6959 perform IE->LE optimization. */
6960 if (x86_elf_abi == X86_64_X32_ABI
6961 && i.operands == 2
6962 && i.reloc[0] == BFD_RELOC_X86_64_GOTTPOFF
6963 && i.prefix[REX_PREFIX] == 0)
6964 add_prefix (REX_OPCODE);
6965 #endif
6966
6967 /* The prefix bytes. */
6968 for (j = ARRAY_SIZE (i.prefix), q = i.prefix; j > 0; j--, q++)
6969 if (*q)
6970 FRAG_APPEND_1_CHAR (*q);
6971 }
6972 else
6973 {
6974 for (j = 0, q = i.prefix; j < ARRAY_SIZE (i.prefix); j++, q++)
6975 if (*q)
6976 switch (j)
6977 {
6978 case REX_PREFIX:
6979 /* REX byte is encoded in VEX prefix. */
6980 break;
6981 case SEG_PREFIX:
6982 case ADDR_PREFIX:
6983 FRAG_APPEND_1_CHAR (*q);
6984 break;
6985 default:
6986 /* There should be no other prefixes for instructions
6987 with VEX prefix. */
6988 abort ();
6989 }
6990
6991 /* For EVEX instructions i.vrex should become 0 after
6992 build_evex_prefix. For VEX instructions upper 16 registers
6993 aren't available, so VREX should be 0. */
6994 if (i.vrex)
6995 abort ();
6996 /* Now the VEX prefix. */
6997 p = frag_more (i.vex.length);
6998 for (j = 0; j < i.vex.length; j++)
6999 p[j] = i.vex.bytes[j];
7000 }
7001
7002 /* Now the opcode; be careful about word order here! */
7003 if (i.tm.opcode_length == 1)
7004 {
7005 FRAG_APPEND_1_CHAR (i.tm.base_opcode);
7006 }
7007 else
7008 {
7009 switch (i.tm.opcode_length)
7010 {
7011 case 4:
7012 p = frag_more (4);
7013 *p++ = (i.tm.base_opcode >> 24) & 0xff;
7014 *p++ = (i.tm.base_opcode >> 16) & 0xff;
7015 break;
7016 case 3:
7017 p = frag_more (3);
7018 *p++ = (i.tm.base_opcode >> 16) & 0xff;
7019 break;
7020 case 2:
7021 p = frag_more (2);
7022 break;
7023 default:
7024 abort ();
7025 break;
7026 }
7027
7028 /* Put out high byte first: can't use md_number_to_chars! */
7029 *p++ = (i.tm.base_opcode >> 8) & 0xff;
7030 *p = i.tm.base_opcode & 0xff;
7031 }
7032
7033 /* Now the modrm byte and sib byte (if present). */
7034 if (i.tm.opcode_modifier.modrm)
7035 {
7036 FRAG_APPEND_1_CHAR ((i.rm.regmem << 0
7037 | i.rm.reg << 3
7038 | i.rm.mode << 6));
7039 /* If i.rm.regmem == ESP (4)
7040 && i.rm.mode != (Register mode)
7041 && not 16 bit
7042 ==> need second modrm byte. */
7043 if (i.rm.regmem == ESCAPE_TO_TWO_BYTE_ADDRESSING
7044 && i.rm.mode != 3
7045 && !(i.base_reg && i.base_reg->reg_type.bitfield.reg16))
7046 FRAG_APPEND_1_CHAR ((i.sib.base << 0
7047 | i.sib.index << 3
7048 | i.sib.scale << 6));
7049 }
7050
7051 if (i.disp_operands)
7052 output_disp (insn_start_frag, insn_start_off);
7053
7054 if (i.imm_operands)
7055 output_imm (insn_start_frag, insn_start_off);
7056 }
7057
7058 #ifdef DEBUG386
7059 if (flag_debug)
7060 {
7061 pi ("" /*line*/, &i);
7062 }
7063 #endif /* DEBUG386 */
7064 }
7065
7066 /* Return the size of the displacement operand N. */
7067
7068 static int
7069 disp_size (unsigned int n)
7070 {
7071 int size = 4;
7072
7073 /* Vec_Disp8 has to be 8bit. */
7074 if (i.types[n].bitfield.vec_disp8)
7075 size = 1;
7076 else if (i.types[n].bitfield.disp64)
7077 size = 8;
7078 else if (i.types[n].bitfield.disp8)
7079 size = 1;
7080 else if (i.types[n].bitfield.disp16)
7081 size = 2;
7082 return size;
7083 }
7084
7085 /* Return the size of the immediate operand N. */
7086
7087 static int
7088 imm_size (unsigned int n)
7089 {
7090 int size = 4;
7091 if (i.types[n].bitfield.imm64)
7092 size = 8;
7093 else if (i.types[n].bitfield.imm8 || i.types[n].bitfield.imm8s)
7094 size = 1;
7095 else if (i.types[n].bitfield.imm16)
7096 size = 2;
7097 return size;
7098 }
7099
7100 static void
7101 output_disp (fragS *insn_start_frag, offsetT insn_start_off)
7102 {
7103 char *p;
7104 unsigned int n;
7105
7106 for (n = 0; n < i.operands; n++)
7107 {
7108 if (i.types[n].bitfield.vec_disp8
7109 || operand_type_check (i.types[n], disp))
7110 {
7111 if (i.op[n].disps->X_op == O_constant)
7112 {
7113 int size = disp_size (n);
7114 offsetT val = i.op[n].disps->X_add_number;
7115
7116 if (i.types[n].bitfield.vec_disp8)
7117 val >>= i.memshift;
7118 val = offset_in_range (val, size);
7119 p = frag_more (size);
7120 md_number_to_chars (p, val, size);
7121 }
7122 else
7123 {
7124 enum bfd_reloc_code_real reloc_type;
7125 int size = disp_size (n);
7126 int sign = i.types[n].bitfield.disp32s;
7127 int pcrel = (i.flags[n] & Operand_PCrel) != 0;
7128
7129 /* We can't have 8 bit displacement here. */
7130 gas_assert (!i.types[n].bitfield.disp8);
7131
7132 /* The PC relative address is computed relative
7133 to the instruction boundary, so in case immediate
7134 fields follows, we need to adjust the value. */
7135 if (pcrel && i.imm_operands)
7136 {
7137 unsigned int n1;
7138 int sz = 0;
7139
7140 for (n1 = 0; n1 < i.operands; n1++)
7141 if (operand_type_check (i.types[n1], imm))
7142 {
7143 /* Only one immediate is allowed for PC
7144 relative address. */
7145 gas_assert (sz == 0);
7146 sz = imm_size (n1);
7147 i.op[n].disps->X_add_number -= sz;
7148 }
7149 /* We should find the immediate. */
7150 gas_assert (sz != 0);
7151 }
7152
7153 p = frag_more (size);
7154 reloc_type = reloc (size, pcrel, sign, i.reloc[n]);
7155 if (GOT_symbol
7156 && GOT_symbol == i.op[n].disps->X_add_symbol
7157 && (((reloc_type == BFD_RELOC_32
7158 || reloc_type == BFD_RELOC_X86_64_32S
7159 || (reloc_type == BFD_RELOC_64
7160 && object_64bit))
7161 && (i.op[n].disps->X_op == O_symbol
7162 || (i.op[n].disps->X_op == O_add
7163 && ((symbol_get_value_expression
7164 (i.op[n].disps->X_op_symbol)->X_op)
7165 == O_subtract))))
7166 || reloc_type == BFD_RELOC_32_PCREL))
7167 {
7168 offsetT add;
7169
7170 if (insn_start_frag == frag_now)
7171 add = (p - frag_now->fr_literal) - insn_start_off;
7172 else
7173 {
7174 fragS *fr;
7175
7176 add = insn_start_frag->fr_fix - insn_start_off;
7177 for (fr = insn_start_frag->fr_next;
7178 fr && fr != frag_now; fr = fr->fr_next)
7179 add += fr->fr_fix;
7180 add += p - frag_now->fr_literal;
7181 }
7182
7183 if (!object_64bit)
7184 {
7185 reloc_type = BFD_RELOC_386_GOTPC;
7186 i.op[n].imms->X_add_number += add;
7187 }
7188 else if (reloc_type == BFD_RELOC_64)
7189 reloc_type = BFD_RELOC_X86_64_GOTPC64;
7190 else
7191 /* Don't do the adjustment for x86-64, as there
7192 the pcrel addressing is relative to the _next_
7193 insn, and that is taken care of in other code. */
7194 reloc_type = BFD_RELOC_X86_64_GOTPC32;
7195 }
7196 fix_new_exp (frag_now, p - frag_now->fr_literal, size,
7197 i.op[n].disps, pcrel, reloc_type);
7198 }
7199 }
7200 }
7201 }
7202
7203 static void
7204 output_imm (fragS *insn_start_frag, offsetT insn_start_off)
7205 {
7206 char *p;
7207 unsigned int n;
7208
7209 for (n = 0; n < i.operands; n++)
7210 {
7211 /* Skip SAE/RC Imm operand in EVEX. They are already handled. */
7212 if (i.rounding && (int) n == i.rounding->operand)
7213 continue;
7214
7215 if (operand_type_check (i.types[n], imm))
7216 {
7217 if (i.op[n].imms->X_op == O_constant)
7218 {
7219 int size = imm_size (n);
7220 offsetT val;
7221
7222 val = offset_in_range (i.op[n].imms->X_add_number,
7223 size);
7224 p = frag_more (size);
7225 md_number_to_chars (p, val, size);
7226 }
7227 else
7228 {
7229 /* Not absolute_section.
7230 Need a 32-bit fixup (don't support 8bit
7231 non-absolute imms). Try to support other
7232 sizes ... */
7233 enum bfd_reloc_code_real reloc_type;
7234 int size = imm_size (n);
7235 int sign;
7236
7237 if (i.types[n].bitfield.imm32s
7238 && (i.suffix == QWORD_MNEM_SUFFIX
7239 || (!i.suffix && i.tm.opcode_modifier.no_lsuf)))
7240 sign = 1;
7241 else
7242 sign = 0;
7243
7244 p = frag_more (size);
7245 reloc_type = reloc (size, 0, sign, i.reloc[n]);
7246
7247 /* This is tough to explain. We end up with this one if we
7248 * have operands that look like
7249 * "_GLOBAL_OFFSET_TABLE_+[.-.L284]". The goal here is to
7250 * obtain the absolute address of the GOT, and it is strongly
7251 * preferable from a performance point of view to avoid using
7252 * a runtime relocation for this. The actual sequence of
7253 * instructions often look something like:
7254 *
7255 * call .L66
7256 * .L66:
7257 * popl %ebx
7258 * addl $_GLOBAL_OFFSET_TABLE_+[.-.L66],%ebx
7259 *
7260 * The call and pop essentially return the absolute address
7261 * of the label .L66 and store it in %ebx. The linker itself
7262 * will ultimately change the first operand of the addl so
7263 * that %ebx points to the GOT, but to keep things simple, the
7264 * .o file must have this operand set so that it generates not
7265 * the absolute address of .L66, but the absolute address of
7266 * itself. This allows the linker itself simply treat a GOTPC
7267 * relocation as asking for a pcrel offset to the GOT to be
7268 * added in, and the addend of the relocation is stored in the
7269 * operand field for the instruction itself.
7270 *
7271 * Our job here is to fix the operand so that it would add
7272 * the correct offset so that %ebx would point to itself. The
7273 * thing that is tricky is that .-.L66 will point to the
7274 * beginning of the instruction, so we need to further modify
7275 * the operand so that it will point to itself. There are
7276 * other cases where you have something like:
7277 *
7278 * .long $_GLOBAL_OFFSET_TABLE_+[.-.L66]
7279 *
7280 * and here no correction would be required. Internally in
7281 * the assembler we treat operands of this form as not being
7282 * pcrel since the '.' is explicitly mentioned, and I wonder
7283 * whether it would simplify matters to do it this way. Who
7284 * knows. In earlier versions of the PIC patches, the
7285 * pcrel_adjust field was used to store the correction, but
7286 * since the expression is not pcrel, I felt it would be
7287 * confusing to do it this way. */
7288
7289 if ((reloc_type == BFD_RELOC_32
7290 || reloc_type == BFD_RELOC_X86_64_32S
7291 || reloc_type == BFD_RELOC_64)
7292 && GOT_symbol
7293 && GOT_symbol == i.op[n].imms->X_add_symbol
7294 && (i.op[n].imms->X_op == O_symbol
7295 || (i.op[n].imms->X_op == O_add
7296 && ((symbol_get_value_expression
7297 (i.op[n].imms->X_op_symbol)->X_op)
7298 == O_subtract))))
7299 {
7300 offsetT add;
7301
7302 if (insn_start_frag == frag_now)
7303 add = (p - frag_now->fr_literal) - insn_start_off;
7304 else
7305 {
7306 fragS *fr;
7307
7308 add = insn_start_frag->fr_fix - insn_start_off;
7309 for (fr = insn_start_frag->fr_next;
7310 fr && fr != frag_now; fr = fr->fr_next)
7311 add += fr->fr_fix;
7312 add += p - frag_now->fr_literal;
7313 }
7314
7315 if (!object_64bit)
7316 reloc_type = BFD_RELOC_386_GOTPC;
7317 else if (size == 4)
7318 reloc_type = BFD_RELOC_X86_64_GOTPC32;
7319 else if (size == 8)
7320 reloc_type = BFD_RELOC_X86_64_GOTPC64;
7321 i.op[n].imms->X_add_number += add;
7322 }
7323 fix_new_exp (frag_now, p - frag_now->fr_literal, size,
7324 i.op[n].imms, 0, reloc_type);
7325 }
7326 }
7327 }
7328 }
7329 \f
7330 /* x86_cons_fix_new is called via the expression parsing code when a
7331 reloc is needed. We use this hook to get the correct .got reloc. */
7332 static int cons_sign = -1;
7333
7334 void
7335 x86_cons_fix_new (fragS *frag, unsigned int off, unsigned int len,
7336 expressionS *exp, bfd_reloc_code_real_type r)
7337 {
7338 r = reloc (len, 0, cons_sign, r);
7339
7340 #ifdef TE_PE
7341 if (exp->X_op == O_secrel)
7342 {
7343 exp->X_op = O_symbol;
7344 r = BFD_RELOC_32_SECREL;
7345 }
7346 #endif
7347
7348 fix_new_exp (frag, off, len, exp, 0, r);
7349 }
7350
7351 /* Export the ABI address size for use by TC_ADDRESS_BYTES for the
7352 purpose of the `.dc.a' internal pseudo-op. */
7353
7354 int
7355 x86_address_bytes (void)
7356 {
7357 if ((stdoutput->arch_info->mach & bfd_mach_x64_32))
7358 return 4;
7359 return stdoutput->arch_info->bits_per_address / 8;
7360 }
7361
7362 #if !(defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) || defined (OBJ_MACH_O)) \
7363 || defined (LEX_AT)
7364 # define lex_got(reloc, adjust, types) NULL
7365 #else
7366 /* Parse operands of the form
7367 <symbol>@GOTOFF+<nnn>
7368 and similar .plt or .got references.
7369
7370 If we find one, set up the correct relocation in RELOC and copy the
7371 input string, minus the `@GOTOFF' into a malloc'd buffer for
7372 parsing by the calling routine. Return this buffer, and if ADJUST
7373 is non-null set it to the length of the string we removed from the
7374 input line. Otherwise return NULL. */
7375 static char *
7376 lex_got (enum bfd_reloc_code_real *rel,
7377 int *adjust,
7378 i386_operand_type *types)
7379 {
7380 /* Some of the relocations depend on the size of what field is to
7381 be relocated. But in our callers i386_immediate and i386_displacement
7382 we don't yet know the operand size (this will be set by insn
7383 matching). Hence we record the word32 relocation here,
7384 and adjust the reloc according to the real size in reloc(). */
7385 static const struct {
7386 const char *str;
7387 int len;
7388 const enum bfd_reloc_code_real rel[2];
7389 const i386_operand_type types64;
7390 } gotrel[] = {
7391 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
7392 { STRING_COMMA_LEN ("SIZE"), { BFD_RELOC_SIZE32,
7393 BFD_RELOC_SIZE32 },
7394 OPERAND_TYPE_IMM32_64 },
7395 #endif
7396 { STRING_COMMA_LEN ("PLTOFF"), { _dummy_first_bfd_reloc_code_real,
7397 BFD_RELOC_X86_64_PLTOFF64 },
7398 OPERAND_TYPE_IMM64 },
7399 { STRING_COMMA_LEN ("PLT"), { BFD_RELOC_386_PLT32,
7400 BFD_RELOC_X86_64_PLT32 },
7401 OPERAND_TYPE_IMM32_32S_DISP32 },
7402 { STRING_COMMA_LEN ("GOTPLT"), { _dummy_first_bfd_reloc_code_real,
7403 BFD_RELOC_X86_64_GOTPLT64 },
7404 OPERAND_TYPE_IMM64_DISP64 },
7405 { STRING_COMMA_LEN ("GOTOFF"), { BFD_RELOC_386_GOTOFF,
7406 BFD_RELOC_X86_64_GOTOFF64 },
7407 OPERAND_TYPE_IMM64_DISP64 },
7408 { STRING_COMMA_LEN ("GOTPCREL"), { _dummy_first_bfd_reloc_code_real,
7409 BFD_RELOC_X86_64_GOTPCREL },
7410 OPERAND_TYPE_IMM32_32S_DISP32 },
7411 { STRING_COMMA_LEN ("TLSGD"), { BFD_RELOC_386_TLS_GD,
7412 BFD_RELOC_X86_64_TLSGD },
7413 OPERAND_TYPE_IMM32_32S_DISP32 },
7414 { STRING_COMMA_LEN ("TLSLDM"), { BFD_RELOC_386_TLS_LDM,
7415 _dummy_first_bfd_reloc_code_real },
7416 OPERAND_TYPE_NONE },
7417 { STRING_COMMA_LEN ("TLSLD"), { _dummy_first_bfd_reloc_code_real,
7418 BFD_RELOC_X86_64_TLSLD },
7419 OPERAND_TYPE_IMM32_32S_DISP32 },
7420 { STRING_COMMA_LEN ("GOTTPOFF"), { BFD_RELOC_386_TLS_IE_32,
7421 BFD_RELOC_X86_64_GOTTPOFF },
7422 OPERAND_TYPE_IMM32_32S_DISP32 },
7423 { STRING_COMMA_LEN ("TPOFF"), { BFD_RELOC_386_TLS_LE_32,
7424 BFD_RELOC_X86_64_TPOFF32 },
7425 OPERAND_TYPE_IMM32_32S_64_DISP32_64 },
7426 { STRING_COMMA_LEN ("NTPOFF"), { BFD_RELOC_386_TLS_LE,
7427 _dummy_first_bfd_reloc_code_real },
7428 OPERAND_TYPE_NONE },
7429 { STRING_COMMA_LEN ("DTPOFF"), { BFD_RELOC_386_TLS_LDO_32,
7430 BFD_RELOC_X86_64_DTPOFF32 },
7431 OPERAND_TYPE_IMM32_32S_64_DISP32_64 },
7432 { STRING_COMMA_LEN ("GOTNTPOFF"),{ BFD_RELOC_386_TLS_GOTIE,
7433 _dummy_first_bfd_reloc_code_real },
7434 OPERAND_TYPE_NONE },
7435 { STRING_COMMA_LEN ("INDNTPOFF"),{ BFD_RELOC_386_TLS_IE,
7436 _dummy_first_bfd_reloc_code_real },
7437 OPERAND_TYPE_NONE },
7438 { STRING_COMMA_LEN ("GOT"), { BFD_RELOC_386_GOT32,
7439 BFD_RELOC_X86_64_GOT32 },
7440 OPERAND_TYPE_IMM32_32S_64_DISP32 },
7441 { STRING_COMMA_LEN ("TLSDESC"), { BFD_RELOC_386_TLS_GOTDESC,
7442 BFD_RELOC_X86_64_GOTPC32_TLSDESC },
7443 OPERAND_TYPE_IMM32_32S_DISP32 },
7444 { STRING_COMMA_LEN ("TLSCALL"), { BFD_RELOC_386_TLS_DESC_CALL,
7445 BFD_RELOC_X86_64_TLSDESC_CALL },
7446 OPERAND_TYPE_IMM32_32S_DISP32 },
7447 };
7448 char *cp;
7449 unsigned int j;
7450
7451 #if defined (OBJ_MAYBE_ELF)
7452 if (!IS_ELF)
7453 return NULL;
7454 #endif
7455
7456 for (cp = input_line_pointer; *cp != '@'; cp++)
7457 if (is_end_of_line[(unsigned char) *cp] || *cp == ',')
7458 return NULL;
7459
7460 for (j = 0; j < ARRAY_SIZE (gotrel); j++)
7461 {
7462 int len = gotrel[j].len;
7463 if (strncasecmp (cp + 1, gotrel[j].str, len) == 0)
7464 {
7465 if (gotrel[j].rel[object_64bit] != 0)
7466 {
7467 int first, second;
7468 char *tmpbuf, *past_reloc;
7469
7470 *rel = gotrel[j].rel[object_64bit];
7471
7472 if (types)
7473 {
7474 if (flag_code != CODE_64BIT)
7475 {
7476 types->bitfield.imm32 = 1;
7477 types->bitfield.disp32 = 1;
7478 }
7479 else
7480 *types = gotrel[j].types64;
7481 }
7482
7483 if (j != 0 && GOT_symbol == NULL)
7484 GOT_symbol = symbol_find_or_make (GLOBAL_OFFSET_TABLE_NAME);
7485
7486 /* The length of the first part of our input line. */
7487 first = cp - input_line_pointer;
7488
7489 /* The second part goes from after the reloc token until
7490 (and including) an end_of_line char or comma. */
7491 past_reloc = cp + 1 + len;
7492 cp = past_reloc;
7493 while (!is_end_of_line[(unsigned char) *cp] && *cp != ',')
7494 ++cp;
7495 second = cp + 1 - past_reloc;
7496
7497 /* Allocate and copy string. The trailing NUL shouldn't
7498 be necessary, but be safe. */
7499 tmpbuf = (char *) xmalloc (first + second + 2);
7500 memcpy (tmpbuf, input_line_pointer, first);
7501 if (second != 0 && *past_reloc != ' ')
7502 /* Replace the relocation token with ' ', so that
7503 errors like foo@GOTOFF1 will be detected. */
7504 tmpbuf[first++] = ' ';
7505 else
7506 /* Increment length by 1 if the relocation token is
7507 removed. */
7508 len++;
7509 if (adjust)
7510 *adjust = len;
7511 memcpy (tmpbuf + first, past_reloc, second);
7512 tmpbuf[first + second] = '\0';
7513 return tmpbuf;
7514 }
7515
7516 as_bad (_("@%s reloc is not supported with %d-bit output format"),
7517 gotrel[j].str, 1 << (5 + object_64bit));
7518 return NULL;
7519 }
7520 }
7521
7522 /* Might be a symbol version string. Don't as_bad here. */
7523 return NULL;
7524 }
7525 #endif
7526
7527 #ifdef TE_PE
7528 #ifdef lex_got
7529 #undef lex_got
7530 #endif
7531 /* Parse operands of the form
7532 <symbol>@SECREL32+<nnn>
7533
7534 If we find one, set up the correct relocation in RELOC and copy the
7535 input string, minus the `@SECREL32' into a malloc'd buffer for
7536 parsing by the calling routine. Return this buffer, and if ADJUST
7537 is non-null set it to the length of the string we removed from the
7538 input line. Otherwise return NULL.
7539
7540 This function is copied from the ELF version above adjusted for PE targets. */
7541
7542 static char *
7543 lex_got (enum bfd_reloc_code_real *rel ATTRIBUTE_UNUSED,
7544 int *adjust ATTRIBUTE_UNUSED,
7545 i386_operand_type *types)
7546 {
7547 static const struct
7548 {
7549 const char *str;
7550 int len;
7551 const enum bfd_reloc_code_real rel[2];
7552 const i386_operand_type types64;
7553 }
7554 gotrel[] =
7555 {
7556 { STRING_COMMA_LEN ("SECREL32"), { BFD_RELOC_32_SECREL,
7557 BFD_RELOC_32_SECREL },
7558 OPERAND_TYPE_IMM32_32S_64_DISP32_64 },
7559 };
7560
7561 char *cp;
7562 unsigned j;
7563
7564 for (cp = input_line_pointer; *cp != '@'; cp++)
7565 if (is_end_of_line[(unsigned char) *cp] || *cp == ',')
7566 return NULL;
7567
7568 for (j = 0; j < ARRAY_SIZE (gotrel); j++)
7569 {
7570 int len = gotrel[j].len;
7571
7572 if (strncasecmp (cp + 1, gotrel[j].str, len) == 0)
7573 {
7574 if (gotrel[j].rel[object_64bit] != 0)
7575 {
7576 int first, second;
7577 char *tmpbuf, *past_reloc;
7578
7579 *rel = gotrel[j].rel[object_64bit];
7580 if (adjust)
7581 *adjust = len;
7582
7583 if (types)
7584 {
7585 if (flag_code != CODE_64BIT)
7586 {
7587 types->bitfield.imm32 = 1;
7588 types->bitfield.disp32 = 1;
7589 }
7590 else
7591 *types = gotrel[j].types64;
7592 }
7593
7594 /* The length of the first part of our input line. */
7595 first = cp - input_line_pointer;
7596
7597 /* The second part goes from after the reloc token until
7598 (and including) an end_of_line char or comma. */
7599 past_reloc = cp + 1 + len;
7600 cp = past_reloc;
7601 while (!is_end_of_line[(unsigned char) *cp] && *cp != ',')
7602 ++cp;
7603 second = cp + 1 - past_reloc;
7604
7605 /* Allocate and copy string. The trailing NUL shouldn't
7606 be necessary, but be safe. */
7607 tmpbuf = (char *) xmalloc (first + second + 2);
7608 memcpy (tmpbuf, input_line_pointer, first);
7609 if (second != 0 && *past_reloc != ' ')
7610 /* Replace the relocation token with ' ', so that
7611 errors like foo@SECLREL321 will be detected. */
7612 tmpbuf[first++] = ' ';
7613 memcpy (tmpbuf + first, past_reloc, second);
7614 tmpbuf[first + second] = '\0';
7615 return tmpbuf;
7616 }
7617
7618 as_bad (_("@%s reloc is not supported with %d-bit output format"),
7619 gotrel[j].str, 1 << (5 + object_64bit));
7620 return NULL;
7621 }
7622 }
7623
7624 /* Might be a symbol version string. Don't as_bad here. */
7625 return NULL;
7626 }
7627
7628 #endif /* TE_PE */
7629
7630 bfd_reloc_code_real_type
7631 x86_cons (expressionS *exp, int size)
7632 {
7633 bfd_reloc_code_real_type got_reloc = NO_RELOC;
7634
7635 intel_syntax = -intel_syntax;
7636
7637 exp->X_md = 0;
7638 if (size == 4 || (object_64bit && size == 8))
7639 {
7640 /* Handle @GOTOFF and the like in an expression. */
7641 char *save;
7642 char *gotfree_input_line;
7643 int adjust = 0;
7644
7645 save = input_line_pointer;
7646 gotfree_input_line = lex_got (&got_reloc, &adjust, NULL);
7647 if (gotfree_input_line)
7648 input_line_pointer = gotfree_input_line;
7649
7650 expression (exp);
7651
7652 if (gotfree_input_line)
7653 {
7654 /* expression () has merrily parsed up to the end of line,
7655 or a comma - in the wrong buffer. Transfer how far
7656 input_line_pointer has moved to the right buffer. */
7657 input_line_pointer = (save
7658 + (input_line_pointer - gotfree_input_line)
7659 + adjust);
7660 free (gotfree_input_line);
7661 if (exp->X_op == O_constant
7662 || exp->X_op == O_absent
7663 || exp->X_op == O_illegal
7664 || exp->X_op == O_register
7665 || exp->X_op == O_big)
7666 {
7667 char c = *input_line_pointer;
7668 *input_line_pointer = 0;
7669 as_bad (_("missing or invalid expression `%s'"), save);
7670 *input_line_pointer = c;
7671 }
7672 }
7673 }
7674 else
7675 expression (exp);
7676
7677 intel_syntax = -intel_syntax;
7678
7679 if (intel_syntax)
7680 i386_intel_simplify (exp);
7681
7682 return got_reloc;
7683 }
7684
7685 static void
7686 signed_cons (int size)
7687 {
7688 if (flag_code == CODE_64BIT)
7689 cons_sign = 1;
7690 cons (size);
7691 cons_sign = -1;
7692 }
7693
7694 #ifdef TE_PE
7695 static void
7696 pe_directive_secrel (int dummy ATTRIBUTE_UNUSED)
7697 {
7698 expressionS exp;
7699
7700 do
7701 {
7702 expression (&exp);
7703 if (exp.X_op == O_symbol)
7704 exp.X_op = O_secrel;
7705
7706 emit_expr (&exp, 4);
7707 }
7708 while (*input_line_pointer++ == ',');
7709
7710 input_line_pointer--;
7711 demand_empty_rest_of_line ();
7712 }
7713 #endif
7714
7715 /* Handle Vector operations. */
7716
7717 static char *
7718 check_VecOperations (char *op_string, char *op_end)
7719 {
7720 const reg_entry *mask;
7721 const char *saved;
7722 char *end_op;
7723
7724 while (*op_string
7725 && (op_end == NULL || op_string < op_end))
7726 {
7727 saved = op_string;
7728 if (*op_string == '{')
7729 {
7730 op_string++;
7731
7732 /* Check broadcasts. */
7733 if (strncmp (op_string, "1to", 3) == 0)
7734 {
7735 int bcst_type;
7736
7737 if (i.broadcast)
7738 goto duplicated_vec_op;
7739
7740 op_string += 3;
7741 if (*op_string == '8')
7742 bcst_type = BROADCAST_1TO8;
7743 else if (*op_string == '4')
7744 bcst_type = BROADCAST_1TO4;
7745 else if (*op_string == '2')
7746 bcst_type = BROADCAST_1TO2;
7747 else if (*op_string == '1'
7748 && *(op_string+1) == '6')
7749 {
7750 bcst_type = BROADCAST_1TO16;
7751 op_string++;
7752 }
7753 else
7754 {
7755 as_bad (_("Unsupported broadcast: `%s'"), saved);
7756 return NULL;
7757 }
7758 op_string++;
7759
7760 broadcast_op.type = bcst_type;
7761 broadcast_op.operand = this_operand;
7762 i.broadcast = &broadcast_op;
7763 }
7764 /* Check masking operation. */
7765 else if ((mask = parse_register (op_string, &end_op)) != NULL)
7766 {
7767 /* k0 can't be used for write mask. */
7768 if (mask->reg_num == 0)
7769 {
7770 as_bad (_("`%s' can't be used for write mask"),
7771 op_string);
7772 return NULL;
7773 }
7774
7775 if (!i.mask)
7776 {
7777 mask_op.mask = mask;
7778 mask_op.zeroing = 0;
7779 mask_op.operand = this_operand;
7780 i.mask = &mask_op;
7781 }
7782 else
7783 {
7784 if (i.mask->mask)
7785 goto duplicated_vec_op;
7786
7787 i.mask->mask = mask;
7788
7789 /* Only "{z}" is allowed here. No need to check
7790 zeroing mask explicitly. */
7791 if (i.mask->operand != this_operand)
7792 {
7793 as_bad (_("invalid write mask `%s'"), saved);
7794 return NULL;
7795 }
7796 }
7797
7798 op_string = end_op;
7799 }
7800 /* Check zeroing-flag for masking operation. */
7801 else if (*op_string == 'z')
7802 {
7803 if (!i.mask)
7804 {
7805 mask_op.mask = NULL;
7806 mask_op.zeroing = 1;
7807 mask_op.operand = this_operand;
7808 i.mask = &mask_op;
7809 }
7810 else
7811 {
7812 if (i.mask->zeroing)
7813 {
7814 duplicated_vec_op:
7815 as_bad (_("duplicated `%s'"), saved);
7816 return NULL;
7817 }
7818
7819 i.mask->zeroing = 1;
7820
7821 /* Only "{%k}" is allowed here. No need to check mask
7822 register explicitly. */
7823 if (i.mask->operand != this_operand)
7824 {
7825 as_bad (_("invalid zeroing-masking `%s'"),
7826 saved);
7827 return NULL;
7828 }
7829 }
7830
7831 op_string++;
7832 }
7833 else
7834 goto unknown_vec_op;
7835
7836 if (*op_string != '}')
7837 {
7838 as_bad (_("missing `}' in `%s'"), saved);
7839 return NULL;
7840 }
7841 op_string++;
7842 continue;
7843 }
7844 unknown_vec_op:
7845 /* We don't know this one. */
7846 as_bad (_("unknown vector operation: `%s'"), saved);
7847 return NULL;
7848 }
7849
7850 return op_string;
7851 }
7852
7853 static int
7854 i386_immediate (char *imm_start)
7855 {
7856 char *save_input_line_pointer;
7857 char *gotfree_input_line;
7858 segT exp_seg = 0;
7859 expressionS *exp;
7860 i386_operand_type types;
7861
7862 operand_type_set (&types, ~0);
7863
7864 if (i.imm_operands == MAX_IMMEDIATE_OPERANDS)
7865 {
7866 as_bad (_("at most %d immediate operands are allowed"),
7867 MAX_IMMEDIATE_OPERANDS);
7868 return 0;
7869 }
7870
7871 exp = &im_expressions[i.imm_operands++];
7872 i.op[this_operand].imms = exp;
7873
7874 if (is_space_char (*imm_start))
7875 ++imm_start;
7876
7877 save_input_line_pointer = input_line_pointer;
7878 input_line_pointer = imm_start;
7879
7880 gotfree_input_line = lex_got (&i.reloc[this_operand], NULL, &types);
7881 if (gotfree_input_line)
7882 input_line_pointer = gotfree_input_line;
7883
7884 exp_seg = expression (exp);
7885
7886 SKIP_WHITESPACE ();
7887
7888 /* Handle vector operations. */
7889 if (*input_line_pointer == '{')
7890 {
7891 input_line_pointer = check_VecOperations (input_line_pointer,
7892 NULL);
7893 if (input_line_pointer == NULL)
7894 return 0;
7895 }
7896
7897 if (*input_line_pointer)
7898 as_bad (_("junk `%s' after expression"), input_line_pointer);
7899
7900 input_line_pointer = save_input_line_pointer;
7901 if (gotfree_input_line)
7902 {
7903 free (gotfree_input_line);
7904
7905 if (exp->X_op == O_constant || exp->X_op == O_register)
7906 exp->X_op = O_illegal;
7907 }
7908
7909 return i386_finalize_immediate (exp_seg, exp, types, imm_start);
7910 }
7911
7912 static int
7913 i386_finalize_immediate (segT exp_seg ATTRIBUTE_UNUSED, expressionS *exp,
7914 i386_operand_type types, const char *imm_start)
7915 {
7916 if (exp->X_op == O_absent || exp->X_op == O_illegal || exp->X_op == O_big)
7917 {
7918 if (imm_start)
7919 as_bad (_("missing or invalid immediate expression `%s'"),
7920 imm_start);
7921 return 0;
7922 }
7923 else if (exp->X_op == O_constant)
7924 {
7925 /* Size it properly later. */
7926 i.types[this_operand].bitfield.imm64 = 1;
7927 /* If not 64bit, sign extend val. */
7928 if (flag_code != CODE_64BIT
7929 && (exp->X_add_number & ~(((addressT) 2 << 31) - 1)) == 0)
7930 exp->X_add_number
7931 = (exp->X_add_number ^ ((addressT) 1 << 31)) - ((addressT) 1 << 31);
7932 }
7933 #if (defined (OBJ_AOUT) || defined (OBJ_MAYBE_AOUT))
7934 else if (OUTPUT_FLAVOR == bfd_target_aout_flavour
7935 && exp_seg != absolute_section
7936 && exp_seg != text_section
7937 && exp_seg != data_section
7938 && exp_seg != bss_section
7939 && exp_seg != undefined_section
7940 && !bfd_is_com_section (exp_seg))
7941 {
7942 as_bad (_("unimplemented segment %s in operand"), exp_seg->name);
7943 return 0;
7944 }
7945 #endif
7946 else if (!intel_syntax && exp_seg == reg_section)
7947 {
7948 if (imm_start)
7949 as_bad (_("illegal immediate register operand %s"), imm_start);
7950 return 0;
7951 }
7952 else
7953 {
7954 /* This is an address. The size of the address will be
7955 determined later, depending on destination register,
7956 suffix, or the default for the section. */
7957 i.types[this_operand].bitfield.imm8 = 1;
7958 i.types[this_operand].bitfield.imm16 = 1;
7959 i.types[this_operand].bitfield.imm32 = 1;
7960 i.types[this_operand].bitfield.imm32s = 1;
7961 i.types[this_operand].bitfield.imm64 = 1;
7962 i.types[this_operand] = operand_type_and (i.types[this_operand],
7963 types);
7964 }
7965
7966 return 1;
7967 }
7968
7969 static char *
7970 i386_scale (char *scale)
7971 {
7972 offsetT val;
7973 char *save = input_line_pointer;
7974
7975 input_line_pointer = scale;
7976 val = get_absolute_expression ();
7977
7978 switch (val)
7979 {
7980 case 1:
7981 i.log2_scale_factor = 0;
7982 break;
7983 case 2:
7984 i.log2_scale_factor = 1;
7985 break;
7986 case 4:
7987 i.log2_scale_factor = 2;
7988 break;
7989 case 8:
7990 i.log2_scale_factor = 3;
7991 break;
7992 default:
7993 {
7994 char sep = *input_line_pointer;
7995
7996 *input_line_pointer = '\0';
7997 as_bad (_("expecting scale factor of 1, 2, 4, or 8: got `%s'"),
7998 scale);
7999 *input_line_pointer = sep;
8000 input_line_pointer = save;
8001 return NULL;
8002 }
8003 }
8004 if (i.log2_scale_factor != 0 && i.index_reg == 0)
8005 {
8006 as_warn (_("scale factor of %d without an index register"),
8007 1 << i.log2_scale_factor);
8008 i.log2_scale_factor = 0;
8009 }
8010 scale = input_line_pointer;
8011 input_line_pointer = save;
8012 return scale;
8013 }
8014
8015 static int
8016 i386_displacement (char *disp_start, char *disp_end)
8017 {
8018 expressionS *exp;
8019 segT exp_seg = 0;
8020 char *save_input_line_pointer;
8021 char *gotfree_input_line;
8022 int override;
8023 i386_operand_type bigdisp, types = anydisp;
8024 int ret;
8025
8026 if (i.disp_operands == MAX_MEMORY_OPERANDS)
8027 {
8028 as_bad (_("at most %d displacement operands are allowed"),
8029 MAX_MEMORY_OPERANDS);
8030 return 0;
8031 }
8032
8033 operand_type_set (&bigdisp, 0);
8034 if ((i.types[this_operand].bitfield.jumpabsolute)
8035 || (!current_templates->start->opcode_modifier.jump
8036 && !current_templates->start->opcode_modifier.jumpdword))
8037 {
8038 bigdisp.bitfield.disp32 = 1;
8039 override = (i.prefix[ADDR_PREFIX] != 0);
8040 if (flag_code == CODE_64BIT)
8041 {
8042 if (!override)
8043 {
8044 bigdisp.bitfield.disp32s = 1;
8045 bigdisp.bitfield.disp64 = 1;
8046 }
8047 }
8048 else if ((flag_code == CODE_16BIT) ^ override)
8049 {
8050 bigdisp.bitfield.disp32 = 0;
8051 bigdisp.bitfield.disp16 = 1;
8052 }
8053 }
8054 else
8055 {
8056 /* For PC-relative branches, the width of the displacement
8057 is dependent upon data size, not address size. */
8058 override = (i.prefix[DATA_PREFIX] != 0);
8059 if (flag_code == CODE_64BIT)
8060 {
8061 if (override || i.suffix == WORD_MNEM_SUFFIX)
8062 bigdisp.bitfield.disp16 = 1;
8063 else
8064 {
8065 bigdisp.bitfield.disp32 = 1;
8066 bigdisp.bitfield.disp32s = 1;
8067 }
8068 }
8069 else
8070 {
8071 if (!override)
8072 override = (i.suffix == (flag_code != CODE_16BIT
8073 ? WORD_MNEM_SUFFIX
8074 : LONG_MNEM_SUFFIX));
8075 bigdisp.bitfield.disp32 = 1;
8076 if ((flag_code == CODE_16BIT) ^ override)
8077 {
8078 bigdisp.bitfield.disp32 = 0;
8079 bigdisp.bitfield.disp16 = 1;
8080 }
8081 }
8082 }
8083 i.types[this_operand] = operand_type_or (i.types[this_operand],
8084 bigdisp);
8085
8086 exp = &disp_expressions[i.disp_operands];
8087 i.op[this_operand].disps = exp;
8088 i.disp_operands++;
8089 save_input_line_pointer = input_line_pointer;
8090 input_line_pointer = disp_start;
8091 END_STRING_AND_SAVE (disp_end);
8092
8093 #ifndef GCC_ASM_O_HACK
8094 #define GCC_ASM_O_HACK 0
8095 #endif
8096 #if GCC_ASM_O_HACK
8097 END_STRING_AND_SAVE (disp_end + 1);
8098 if (i.types[this_operand].bitfield.baseIndex
8099 && displacement_string_end[-1] == '+')
8100 {
8101 /* This hack is to avoid a warning when using the "o"
8102 constraint within gcc asm statements.
8103 For instance:
8104
8105 #define _set_tssldt_desc(n,addr,limit,type) \
8106 __asm__ __volatile__ ( \
8107 "movw %w2,%0\n\t" \
8108 "movw %w1,2+%0\n\t" \
8109 "rorl $16,%1\n\t" \
8110 "movb %b1,4+%0\n\t" \
8111 "movb %4,5+%0\n\t" \
8112 "movb $0,6+%0\n\t" \
8113 "movb %h1,7+%0\n\t" \
8114 "rorl $16,%1" \
8115 : "=o"(*(n)) : "q" (addr), "ri"(limit), "i"(type))
8116
8117 This works great except that the output assembler ends
8118 up looking a bit weird if it turns out that there is
8119 no offset. You end up producing code that looks like:
8120
8121 #APP
8122 movw $235,(%eax)
8123 movw %dx,2+(%eax)
8124 rorl $16,%edx
8125 movb %dl,4+(%eax)
8126 movb $137,5+(%eax)
8127 movb $0,6+(%eax)
8128 movb %dh,7+(%eax)
8129 rorl $16,%edx
8130 #NO_APP
8131
8132 So here we provide the missing zero. */
8133
8134 *displacement_string_end = '0';
8135 }
8136 #endif
8137 gotfree_input_line = lex_got (&i.reloc[this_operand], NULL, &types);
8138 if (gotfree_input_line)
8139 input_line_pointer = gotfree_input_line;
8140
8141 exp_seg = expression (exp);
8142
8143 SKIP_WHITESPACE ();
8144 if (*input_line_pointer)
8145 as_bad (_("junk `%s' after expression"), input_line_pointer);
8146 #if GCC_ASM_O_HACK
8147 RESTORE_END_STRING (disp_end + 1);
8148 #endif
8149 input_line_pointer = save_input_line_pointer;
8150 if (gotfree_input_line)
8151 {
8152 free (gotfree_input_line);
8153
8154 if (exp->X_op == O_constant || exp->X_op == O_register)
8155 exp->X_op = O_illegal;
8156 }
8157
8158 ret = i386_finalize_displacement (exp_seg, exp, types, disp_start);
8159
8160 RESTORE_END_STRING (disp_end);
8161
8162 return ret;
8163 }
8164
8165 static int
8166 i386_finalize_displacement (segT exp_seg ATTRIBUTE_UNUSED, expressionS *exp,
8167 i386_operand_type types, const char *disp_start)
8168 {
8169 i386_operand_type bigdisp;
8170 int ret = 1;
8171
8172 /* We do this to make sure that the section symbol is in
8173 the symbol table. We will ultimately change the relocation
8174 to be relative to the beginning of the section. */
8175 if (i.reloc[this_operand] == BFD_RELOC_386_GOTOFF
8176 || i.reloc[this_operand] == BFD_RELOC_X86_64_GOTPCREL
8177 || i.reloc[this_operand] == BFD_RELOC_X86_64_GOTOFF64)
8178 {
8179 if (exp->X_op != O_symbol)
8180 goto inv_disp;
8181
8182 if (S_IS_LOCAL (exp->X_add_symbol)
8183 && S_GET_SEGMENT (exp->X_add_symbol) != undefined_section
8184 && S_GET_SEGMENT (exp->X_add_symbol) != expr_section)
8185 section_symbol (S_GET_SEGMENT (exp->X_add_symbol));
8186 exp->X_op = O_subtract;
8187 exp->X_op_symbol = GOT_symbol;
8188 if (i.reloc[this_operand] == BFD_RELOC_X86_64_GOTPCREL)
8189 i.reloc[this_operand] = BFD_RELOC_32_PCREL;
8190 else if (i.reloc[this_operand] == BFD_RELOC_X86_64_GOTOFF64)
8191 i.reloc[this_operand] = BFD_RELOC_64;
8192 else
8193 i.reloc[this_operand] = BFD_RELOC_32;
8194 }
8195
8196 else if (exp->X_op == O_absent
8197 || exp->X_op == O_illegal
8198 || exp->X_op == O_big)
8199 {
8200 inv_disp:
8201 as_bad (_("missing or invalid displacement expression `%s'"),
8202 disp_start);
8203 ret = 0;
8204 }
8205
8206 else if (flag_code == CODE_64BIT
8207 && !i.prefix[ADDR_PREFIX]
8208 && exp->X_op == O_constant)
8209 {
8210 /* Since displacement is signed extended to 64bit, don't allow
8211 disp32 and turn off disp32s if they are out of range. */
8212 i.types[this_operand].bitfield.disp32 = 0;
8213 if (!fits_in_signed_long (exp->X_add_number))
8214 {
8215 i.types[this_operand].bitfield.disp32s = 0;
8216 if (i.types[this_operand].bitfield.baseindex)
8217 {
8218 as_bad (_("0x%lx out range of signed 32bit displacement"),
8219 (long) exp->X_add_number);
8220 ret = 0;
8221 }
8222 }
8223 }
8224
8225 #if (defined (OBJ_AOUT) || defined (OBJ_MAYBE_AOUT))
8226 else if (exp->X_op != O_constant
8227 && OUTPUT_FLAVOR == bfd_target_aout_flavour
8228 && exp_seg != absolute_section
8229 && exp_seg != text_section
8230 && exp_seg != data_section
8231 && exp_seg != bss_section
8232 && exp_seg != undefined_section
8233 && !bfd_is_com_section (exp_seg))
8234 {
8235 as_bad (_("unimplemented segment %s in operand"), exp_seg->name);
8236 ret = 0;
8237 }
8238 #endif
8239
8240 /* Check if this is a displacement only operand. */
8241 bigdisp = i.types[this_operand];
8242 bigdisp.bitfield.disp8 = 0;
8243 bigdisp.bitfield.disp16 = 0;
8244 bigdisp.bitfield.disp32 = 0;
8245 bigdisp.bitfield.disp32s = 0;
8246 bigdisp.bitfield.disp64 = 0;
8247 if (operand_type_all_zero (&bigdisp))
8248 i.types[this_operand] = operand_type_and (i.types[this_operand],
8249 types);
8250
8251 return ret;
8252 }
8253
8254 /* Make sure the memory operand we've been dealt is valid.
8255 Return 1 on success, 0 on a failure. */
8256
8257 static int
8258 i386_index_check (const char *operand_string)
8259 {
8260 const char *kind = "base/index";
8261 enum flag_code addr_mode;
8262
8263 if (i.prefix[ADDR_PREFIX])
8264 addr_mode = flag_code == CODE_32BIT ? CODE_16BIT : CODE_32BIT;
8265 else
8266 {
8267 addr_mode = flag_code;
8268
8269 #if INFER_ADDR_PREFIX
8270 if (i.mem_operands == 0)
8271 {
8272 /* Infer address prefix from the first memory operand. */
8273 const reg_entry *addr_reg = i.base_reg;
8274
8275 if (addr_reg == NULL)
8276 addr_reg = i.index_reg;
8277
8278 if (addr_reg)
8279 {
8280 if (addr_reg->reg_num == RegEip
8281 || addr_reg->reg_num == RegEiz
8282 || addr_reg->reg_type.bitfield.reg32)
8283 addr_mode = CODE_32BIT;
8284 else if (flag_code != CODE_64BIT
8285 && addr_reg->reg_type.bitfield.reg16)
8286 addr_mode = CODE_16BIT;
8287
8288 if (addr_mode != flag_code)
8289 {
8290 i.prefix[ADDR_PREFIX] = ADDR_PREFIX_OPCODE;
8291 i.prefixes += 1;
8292 /* Change the size of any displacement too. At most one
8293 of Disp16 or Disp32 is set.
8294 FIXME. There doesn't seem to be any real need for
8295 separate Disp16 and Disp32 flags. The same goes for
8296 Imm16 and Imm32. Removing them would probably clean
8297 up the code quite a lot. */
8298 if (flag_code != CODE_64BIT
8299 && (i.types[this_operand].bitfield.disp16
8300 || i.types[this_operand].bitfield.disp32))
8301 i.types[this_operand]
8302 = operand_type_xor (i.types[this_operand], disp16_32);
8303 }
8304 }
8305 }
8306 #endif
8307 }
8308
8309 if (current_templates->start->opcode_modifier.isstring
8310 && !current_templates->start->opcode_modifier.immext
8311 && (current_templates->end[-1].opcode_modifier.isstring
8312 || i.mem_operands))
8313 {
8314 /* Memory operands of string insns are special in that they only allow
8315 a single register (rDI, rSI, or rBX) as their memory address. */
8316 const reg_entry *expected_reg;
8317 static const char *di_si[][2] =
8318 {
8319 { "esi", "edi" },
8320 { "si", "di" },
8321 { "rsi", "rdi" }
8322 };
8323 static const char *bx[] = { "ebx", "bx", "rbx" };
8324
8325 kind = "string address";
8326
8327 if (current_templates->start->opcode_modifier.w)
8328 {
8329 i386_operand_type type = current_templates->end[-1].operand_types[0];
8330
8331 if (!type.bitfield.baseindex
8332 || ((!i.mem_operands != !intel_syntax)
8333 && current_templates->end[-1].operand_types[1]
8334 .bitfield.baseindex))
8335 type = current_templates->end[-1].operand_types[1];
8336 expected_reg = hash_find (reg_hash,
8337 di_si[addr_mode][type.bitfield.esseg]);
8338
8339 }
8340 else
8341 expected_reg = hash_find (reg_hash, bx[addr_mode]);
8342
8343 if (i.base_reg != expected_reg
8344 || i.index_reg
8345 || operand_type_check (i.types[this_operand], disp))
8346 {
8347 /* The second memory operand must have the same size as
8348 the first one. */
8349 if (i.mem_operands
8350 && i.base_reg
8351 && !((addr_mode == CODE_64BIT
8352 && i.base_reg->reg_type.bitfield.reg64)
8353 || (addr_mode == CODE_32BIT
8354 ? i.base_reg->reg_type.bitfield.reg32
8355 : i.base_reg->reg_type.bitfield.reg16)))
8356 goto bad_address;
8357
8358 as_warn (_("`%s' is not valid here (expected `%c%s%s%c')"),
8359 operand_string,
8360 intel_syntax ? '[' : '(',
8361 register_prefix,
8362 expected_reg->reg_name,
8363 intel_syntax ? ']' : ')');
8364 return 1;
8365 }
8366 else
8367 return 1;
8368
8369 bad_address:
8370 as_bad (_("`%s' is not a valid %s expression"),
8371 operand_string, kind);
8372 return 0;
8373 }
8374 else
8375 {
8376 if (addr_mode != CODE_16BIT)
8377 {
8378 /* 32-bit/64-bit checks. */
8379 if ((i.base_reg
8380 && (addr_mode == CODE_64BIT
8381 ? !i.base_reg->reg_type.bitfield.reg64
8382 : !i.base_reg->reg_type.bitfield.reg32)
8383 && (i.index_reg
8384 || (i.base_reg->reg_num
8385 != (addr_mode == CODE_64BIT ? RegRip : RegEip))))
8386 || (i.index_reg
8387 && !i.index_reg->reg_type.bitfield.regxmm
8388 && !i.index_reg->reg_type.bitfield.regymm
8389 && !i.index_reg->reg_type.bitfield.regzmm
8390 && ((addr_mode == CODE_64BIT
8391 ? !(i.index_reg->reg_type.bitfield.reg64
8392 || i.index_reg->reg_num == RegRiz)
8393 : !(i.index_reg->reg_type.bitfield.reg32
8394 || i.index_reg->reg_num == RegEiz))
8395 || !i.index_reg->reg_type.bitfield.baseindex)))
8396 goto bad_address;
8397 }
8398 else
8399 {
8400 /* 16-bit checks. */
8401 if ((i.base_reg
8402 && (!i.base_reg->reg_type.bitfield.reg16
8403 || !i.base_reg->reg_type.bitfield.baseindex))
8404 || (i.index_reg
8405 && (!i.index_reg->reg_type.bitfield.reg16
8406 || !i.index_reg->reg_type.bitfield.baseindex
8407 || !(i.base_reg
8408 && i.base_reg->reg_num < 6
8409 && i.index_reg->reg_num >= 6
8410 && i.log2_scale_factor == 0))))
8411 goto bad_address;
8412 }
8413 }
8414 return 1;
8415 }
8416
8417 /* Handle vector immediates. */
8418
8419 static int
8420 RC_SAE_immediate (const char *imm_start)
8421 {
8422 unsigned int match_found, j;
8423 const char *pstr = imm_start;
8424 expressionS *exp;
8425
8426 if (*pstr != '{')
8427 return 0;
8428
8429 pstr++;
8430 match_found = 0;
8431 for (j = 0; j < ARRAY_SIZE (RC_NamesTable); j++)
8432 {
8433 if (!strncmp (pstr, RC_NamesTable[j].name, RC_NamesTable[j].len))
8434 {
8435 if (!i.rounding)
8436 {
8437 rc_op.type = RC_NamesTable[j].type;
8438 rc_op.operand = this_operand;
8439 i.rounding = &rc_op;
8440 }
8441 else
8442 {
8443 as_bad (_("duplicated `%s'"), imm_start);
8444 return 0;
8445 }
8446 pstr += RC_NamesTable[j].len;
8447 match_found = 1;
8448 break;
8449 }
8450 }
8451 if (!match_found)
8452 return 0;
8453
8454 if (*pstr++ != '}')
8455 {
8456 as_bad (_("Missing '}': '%s'"), imm_start);
8457 return 0;
8458 }
8459 /* RC/SAE immediate string should contain nothing more. */;
8460 if (*pstr != 0)
8461 {
8462 as_bad (_("Junk after '}': '%s'"), imm_start);
8463 return 0;
8464 }
8465
8466 exp = &im_expressions[i.imm_operands++];
8467 i.op[this_operand].imms = exp;
8468
8469 exp->X_op = O_constant;
8470 exp->X_add_number = 0;
8471 exp->X_add_symbol = (symbolS *) 0;
8472 exp->X_op_symbol = (symbolS *) 0;
8473
8474 i.types[this_operand].bitfield.imm8 = 1;
8475 return 1;
8476 }
8477
8478 /* Parse OPERAND_STRING into the i386_insn structure I. Returns zero
8479 on error. */
8480
8481 static int
8482 i386_att_operand (char *operand_string)
8483 {
8484 const reg_entry *r;
8485 char *end_op;
8486 char *op_string = operand_string;
8487
8488 if (is_space_char (*op_string))
8489 ++op_string;
8490
8491 /* We check for an absolute prefix (differentiating,
8492 for example, 'jmp pc_relative_label' from 'jmp *absolute_label'. */
8493 if (*op_string == ABSOLUTE_PREFIX)
8494 {
8495 ++op_string;
8496 if (is_space_char (*op_string))
8497 ++op_string;
8498 i.types[this_operand].bitfield.jumpabsolute = 1;
8499 }
8500
8501 /* Check if operand is a register. */
8502 if ((r = parse_register (op_string, &end_op)) != NULL)
8503 {
8504 i386_operand_type temp;
8505
8506 /* Check for a segment override by searching for ':' after a
8507 segment register. */
8508 op_string = end_op;
8509 if (is_space_char (*op_string))
8510 ++op_string;
8511 if (*op_string == ':'
8512 && (r->reg_type.bitfield.sreg2
8513 || r->reg_type.bitfield.sreg3))
8514 {
8515 switch (r->reg_num)
8516 {
8517 case 0:
8518 i.seg[i.mem_operands] = &es;
8519 break;
8520 case 1:
8521 i.seg[i.mem_operands] = &cs;
8522 break;
8523 case 2:
8524 i.seg[i.mem_operands] = &ss;
8525 break;
8526 case 3:
8527 i.seg[i.mem_operands] = &ds;
8528 break;
8529 case 4:
8530 i.seg[i.mem_operands] = &fs;
8531 break;
8532 case 5:
8533 i.seg[i.mem_operands] = &gs;
8534 break;
8535 }
8536
8537 /* Skip the ':' and whitespace. */
8538 ++op_string;
8539 if (is_space_char (*op_string))
8540 ++op_string;
8541
8542 if (!is_digit_char (*op_string)
8543 && !is_identifier_char (*op_string)
8544 && *op_string != '('
8545 && *op_string != ABSOLUTE_PREFIX)
8546 {
8547 as_bad (_("bad memory operand `%s'"), op_string);
8548 return 0;
8549 }
8550 /* Handle case of %es:*foo. */
8551 if (*op_string == ABSOLUTE_PREFIX)
8552 {
8553 ++op_string;
8554 if (is_space_char (*op_string))
8555 ++op_string;
8556 i.types[this_operand].bitfield.jumpabsolute = 1;
8557 }
8558 goto do_memory_reference;
8559 }
8560
8561 /* Handle vector operations. */
8562 if (*op_string == '{')
8563 {
8564 op_string = check_VecOperations (op_string, NULL);
8565 if (op_string == NULL)
8566 return 0;
8567 }
8568
8569 if (*op_string)
8570 {
8571 as_bad (_("junk `%s' after register"), op_string);
8572 return 0;
8573 }
8574 temp = r->reg_type;
8575 temp.bitfield.baseindex = 0;
8576 i.types[this_operand] = operand_type_or (i.types[this_operand],
8577 temp);
8578 i.types[this_operand].bitfield.unspecified = 0;
8579 i.op[this_operand].regs = r;
8580 i.reg_operands++;
8581 }
8582 else if (*op_string == REGISTER_PREFIX)
8583 {
8584 as_bad (_("bad register name `%s'"), op_string);
8585 return 0;
8586 }
8587 else if (*op_string == IMMEDIATE_PREFIX)
8588 {
8589 ++op_string;
8590 if (i.types[this_operand].bitfield.jumpabsolute)
8591 {
8592 as_bad (_("immediate operand illegal with absolute jump"));
8593 return 0;
8594 }
8595 if (!i386_immediate (op_string))
8596 return 0;
8597 }
8598 else if (RC_SAE_immediate (operand_string))
8599 {
8600 /* If it is a RC or SAE immediate, do nothing. */
8601 ;
8602 }
8603 else if (is_digit_char (*op_string)
8604 || is_identifier_char (*op_string)
8605 || *op_string == '(')
8606 {
8607 /* This is a memory reference of some sort. */
8608 char *base_string;
8609
8610 /* Start and end of displacement string expression (if found). */
8611 char *displacement_string_start;
8612 char *displacement_string_end;
8613 char *vop_start;
8614
8615 do_memory_reference:
8616 if ((i.mem_operands == 1
8617 && !current_templates->start->opcode_modifier.isstring)
8618 || i.mem_operands == 2)
8619 {
8620 as_bad (_("too many memory references for `%s'"),
8621 current_templates->start->name);
8622 return 0;
8623 }
8624
8625 /* Check for base index form. We detect the base index form by
8626 looking for an ')' at the end of the operand, searching
8627 for the '(' matching it, and finding a REGISTER_PREFIX or ','
8628 after the '('. */
8629 base_string = op_string + strlen (op_string);
8630
8631 /* Handle vector operations. */
8632 vop_start = strchr (op_string, '{');
8633 if (vop_start && vop_start < base_string)
8634 {
8635 if (check_VecOperations (vop_start, base_string) == NULL)
8636 return 0;
8637 base_string = vop_start;
8638 }
8639
8640 --base_string;
8641 if (is_space_char (*base_string))
8642 --base_string;
8643
8644 /* If we only have a displacement, set-up for it to be parsed later. */
8645 displacement_string_start = op_string;
8646 displacement_string_end = base_string + 1;
8647
8648 if (*base_string == ')')
8649 {
8650 char *temp_string;
8651 unsigned int parens_balanced = 1;
8652 /* We've already checked that the number of left & right ()'s are
8653 equal, so this loop will not be infinite. */
8654 do
8655 {
8656 base_string--;
8657 if (*base_string == ')')
8658 parens_balanced++;
8659 if (*base_string == '(')
8660 parens_balanced--;
8661 }
8662 while (parens_balanced);
8663
8664 temp_string = base_string;
8665
8666 /* Skip past '(' and whitespace. */
8667 ++base_string;
8668 if (is_space_char (*base_string))
8669 ++base_string;
8670
8671 if (*base_string == ','
8672 || ((i.base_reg = parse_register (base_string, &end_op))
8673 != NULL))
8674 {
8675 displacement_string_end = temp_string;
8676
8677 i.types[this_operand].bitfield.baseindex = 1;
8678
8679 if (i.base_reg)
8680 {
8681 base_string = end_op;
8682 if (is_space_char (*base_string))
8683 ++base_string;
8684 }
8685
8686 /* There may be an index reg or scale factor here. */
8687 if (*base_string == ',')
8688 {
8689 ++base_string;
8690 if (is_space_char (*base_string))
8691 ++base_string;
8692
8693 if ((i.index_reg = parse_register (base_string, &end_op))
8694 != NULL)
8695 {
8696 base_string = end_op;
8697 if (is_space_char (*base_string))
8698 ++base_string;
8699 if (*base_string == ',')
8700 {
8701 ++base_string;
8702 if (is_space_char (*base_string))
8703 ++base_string;
8704 }
8705 else if (*base_string != ')')
8706 {
8707 as_bad (_("expecting `,' or `)' "
8708 "after index register in `%s'"),
8709 operand_string);
8710 return 0;
8711 }
8712 }
8713 else if (*base_string == REGISTER_PREFIX)
8714 {
8715 end_op = strchr (base_string, ',');
8716 if (end_op)
8717 *end_op = '\0';
8718 as_bad (_("bad register name `%s'"), base_string);
8719 return 0;
8720 }
8721
8722 /* Check for scale factor. */
8723 if (*base_string != ')')
8724 {
8725 char *end_scale = i386_scale (base_string);
8726
8727 if (!end_scale)
8728 return 0;
8729
8730 base_string = end_scale;
8731 if (is_space_char (*base_string))
8732 ++base_string;
8733 if (*base_string != ')')
8734 {
8735 as_bad (_("expecting `)' "
8736 "after scale factor in `%s'"),
8737 operand_string);
8738 return 0;
8739 }
8740 }
8741 else if (!i.index_reg)
8742 {
8743 as_bad (_("expecting index register or scale factor "
8744 "after `,'; got '%c'"),
8745 *base_string);
8746 return 0;
8747 }
8748 }
8749 else if (*base_string != ')')
8750 {
8751 as_bad (_("expecting `,' or `)' "
8752 "after base register in `%s'"),
8753 operand_string);
8754 return 0;
8755 }
8756 }
8757 else if (*base_string == REGISTER_PREFIX)
8758 {
8759 end_op = strchr (base_string, ',');
8760 if (end_op)
8761 *end_op = '\0';
8762 as_bad (_("bad register name `%s'"), base_string);
8763 return 0;
8764 }
8765 }
8766
8767 /* If there's an expression beginning the operand, parse it,
8768 assuming displacement_string_start and
8769 displacement_string_end are meaningful. */
8770 if (displacement_string_start != displacement_string_end)
8771 {
8772 if (!i386_displacement (displacement_string_start,
8773 displacement_string_end))
8774 return 0;
8775 }
8776
8777 /* Special case for (%dx) while doing input/output op. */
8778 if (i.base_reg
8779 && operand_type_equal (&i.base_reg->reg_type,
8780 &reg16_inoutportreg)
8781 && i.index_reg == 0
8782 && i.log2_scale_factor == 0
8783 && i.seg[i.mem_operands] == 0
8784 && !operand_type_check (i.types[this_operand], disp))
8785 {
8786 i.types[this_operand] = inoutportreg;
8787 return 1;
8788 }
8789
8790 if (i386_index_check (operand_string) == 0)
8791 return 0;
8792 i.types[this_operand].bitfield.mem = 1;
8793 i.mem_operands++;
8794 }
8795 else
8796 {
8797 /* It's not a memory operand; argh! */
8798 as_bad (_("invalid char %s beginning operand %d `%s'"),
8799 output_invalid (*op_string),
8800 this_operand + 1,
8801 op_string);
8802 return 0;
8803 }
8804 return 1; /* Normal return. */
8805 }
8806 \f
8807 /* Calculate the maximum variable size (i.e., excluding fr_fix)
8808 that an rs_machine_dependent frag may reach. */
8809
8810 unsigned int
8811 i386_frag_max_var (fragS *frag)
8812 {
8813 /* The only relaxable frags are for jumps.
8814 Unconditional jumps can grow by 4 bytes and others by 5 bytes. */
8815 gas_assert (frag->fr_type == rs_machine_dependent);
8816 return TYPE_FROM_RELAX_STATE (frag->fr_subtype) == UNCOND_JUMP ? 4 : 5;
8817 }
8818
8819 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
8820 static int
8821 elf_symbol_resolved_in_segment_p (symbolS *fr_symbol)
8822 {
8823 /* STT_GNU_IFUNC symbol must go through PLT. */
8824 if ((symbol_get_bfdsym (fr_symbol)->flags
8825 & BSF_GNU_INDIRECT_FUNCTION) != 0)
8826 return 0;
8827
8828 if (!S_IS_EXTERNAL (fr_symbol))
8829 /* Symbol may be weak or local. */
8830 return !S_IS_WEAK (fr_symbol);
8831
8832 /* Global symbols with default visibility in a shared library may be
8833 preempted by another definition. */
8834 return ELF_ST_VISIBILITY (S_GET_OTHER (fr_symbol)) != STV_DEFAULT;
8835 }
8836 #endif
8837
8838 /* md_estimate_size_before_relax()
8839
8840 Called just before relax() for rs_machine_dependent frags. The x86
8841 assembler uses these frags to handle variable size jump
8842 instructions.
8843
8844 Any symbol that is now undefined will not become defined.
8845 Return the correct fr_subtype in the frag.
8846 Return the initial "guess for variable size of frag" to caller.
8847 The guess is actually the growth beyond the fixed part. Whatever
8848 we do to grow the fixed or variable part contributes to our
8849 returned value. */
8850
8851 int
8852 md_estimate_size_before_relax (fragS *fragP, segT segment)
8853 {
8854 /* We've already got fragP->fr_subtype right; all we have to do is
8855 check for un-relaxable symbols. On an ELF system, we can't relax
8856 an externally visible symbol, because it may be overridden by a
8857 shared library. */
8858 if (S_GET_SEGMENT (fragP->fr_symbol) != segment
8859 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
8860 || (IS_ELF
8861 && !elf_symbol_resolved_in_segment_p (fragP->fr_symbol))
8862 #endif
8863 #if defined (OBJ_COFF) && defined (TE_PE)
8864 || (OUTPUT_FLAVOR == bfd_target_coff_flavour
8865 && S_IS_WEAK (fragP->fr_symbol))
8866 #endif
8867 )
8868 {
8869 /* Symbol is undefined in this segment, or we need to keep a
8870 reloc so that weak symbols can be overridden. */
8871 int size = (fragP->fr_subtype & CODE16) ? 2 : 4;
8872 enum bfd_reloc_code_real reloc_type;
8873 unsigned char *opcode;
8874 int old_fr_fix;
8875
8876 if (fragP->fr_var != NO_RELOC)
8877 reloc_type = (enum bfd_reloc_code_real) fragP->fr_var;
8878 else if (size == 2)
8879 reloc_type = BFD_RELOC_16_PCREL;
8880 else
8881 reloc_type = BFD_RELOC_32_PCREL;
8882
8883 old_fr_fix = fragP->fr_fix;
8884 opcode = (unsigned char *) fragP->fr_opcode;
8885
8886 switch (TYPE_FROM_RELAX_STATE (fragP->fr_subtype))
8887 {
8888 case UNCOND_JUMP:
8889 /* Make jmp (0xeb) a (d)word displacement jump. */
8890 opcode[0] = 0xe9;
8891 fragP->fr_fix += size;
8892 fix_new (fragP, old_fr_fix, size,
8893 fragP->fr_symbol,
8894 fragP->fr_offset, 1,
8895 reloc_type);
8896 break;
8897
8898 case COND_JUMP86:
8899 if (size == 2
8900 && (!no_cond_jump_promotion || fragP->fr_var != NO_RELOC))
8901 {
8902 /* Negate the condition, and branch past an
8903 unconditional jump. */
8904 opcode[0] ^= 1;
8905 opcode[1] = 3;
8906 /* Insert an unconditional jump. */
8907 opcode[2] = 0xe9;
8908 /* We added two extra opcode bytes, and have a two byte
8909 offset. */
8910 fragP->fr_fix += 2 + 2;
8911 fix_new (fragP, old_fr_fix + 2, 2,
8912 fragP->fr_symbol,
8913 fragP->fr_offset, 1,
8914 reloc_type);
8915 break;
8916 }
8917 /* Fall through. */
8918
8919 case COND_JUMP:
8920 if (no_cond_jump_promotion && fragP->fr_var == NO_RELOC)
8921 {
8922 fixS *fixP;
8923
8924 fragP->fr_fix += 1;
8925 fixP = fix_new (fragP, old_fr_fix, 1,
8926 fragP->fr_symbol,
8927 fragP->fr_offset, 1,
8928 BFD_RELOC_8_PCREL);
8929 fixP->fx_signed = 1;
8930 break;
8931 }
8932
8933 /* This changes the byte-displacement jump 0x7N
8934 to the (d)word-displacement jump 0x0f,0x8N. */
8935 opcode[1] = opcode[0] + 0x10;
8936 opcode[0] = TWO_BYTE_OPCODE_ESCAPE;
8937 /* We've added an opcode byte. */
8938 fragP->fr_fix += 1 + size;
8939 fix_new (fragP, old_fr_fix + 1, size,
8940 fragP->fr_symbol,
8941 fragP->fr_offset, 1,
8942 reloc_type);
8943 break;
8944
8945 default:
8946 BAD_CASE (fragP->fr_subtype);
8947 break;
8948 }
8949 frag_wane (fragP);
8950 return fragP->fr_fix - old_fr_fix;
8951 }
8952
8953 /* Guess size depending on current relax state. Initially the relax
8954 state will correspond to a short jump and we return 1, because
8955 the variable part of the frag (the branch offset) is one byte
8956 long. However, we can relax a section more than once and in that
8957 case we must either set fr_subtype back to the unrelaxed state,
8958 or return the value for the appropriate branch. */
8959 return md_relax_table[fragP->fr_subtype].rlx_length;
8960 }
8961
8962 /* Called after relax() is finished.
8963
8964 In: Address of frag.
8965 fr_type == rs_machine_dependent.
8966 fr_subtype is what the address relaxed to.
8967
8968 Out: Any fixSs and constants are set up.
8969 Caller will turn frag into a ".space 0". */
8970
8971 void
8972 md_convert_frag (bfd *abfd ATTRIBUTE_UNUSED, segT sec ATTRIBUTE_UNUSED,
8973 fragS *fragP)
8974 {
8975 unsigned char *opcode;
8976 unsigned char *where_to_put_displacement = NULL;
8977 offsetT target_address;
8978 offsetT opcode_address;
8979 unsigned int extension = 0;
8980 offsetT displacement_from_opcode_start;
8981
8982 opcode = (unsigned char *) fragP->fr_opcode;
8983
8984 /* Address we want to reach in file space. */
8985 target_address = S_GET_VALUE (fragP->fr_symbol) + fragP->fr_offset;
8986
8987 /* Address opcode resides at in file space. */
8988 opcode_address = fragP->fr_address + fragP->fr_fix;
8989
8990 /* Displacement from opcode start to fill into instruction. */
8991 displacement_from_opcode_start = target_address - opcode_address;
8992
8993 if ((fragP->fr_subtype & BIG) == 0)
8994 {
8995 /* Don't have to change opcode. */
8996 extension = 1; /* 1 opcode + 1 displacement */
8997 where_to_put_displacement = &opcode[1];
8998 }
8999 else
9000 {
9001 if (no_cond_jump_promotion
9002 && TYPE_FROM_RELAX_STATE (fragP->fr_subtype) != UNCOND_JUMP)
9003 as_warn_where (fragP->fr_file, fragP->fr_line,
9004 _("long jump required"));
9005
9006 switch (fragP->fr_subtype)
9007 {
9008 case ENCODE_RELAX_STATE (UNCOND_JUMP, BIG):
9009 extension = 4; /* 1 opcode + 4 displacement */
9010 opcode[0] = 0xe9;
9011 where_to_put_displacement = &opcode[1];
9012 break;
9013
9014 case ENCODE_RELAX_STATE (UNCOND_JUMP, BIG16):
9015 extension = 2; /* 1 opcode + 2 displacement */
9016 opcode[0] = 0xe9;
9017 where_to_put_displacement = &opcode[1];
9018 break;
9019
9020 case ENCODE_RELAX_STATE (COND_JUMP, BIG):
9021 case ENCODE_RELAX_STATE (COND_JUMP86, BIG):
9022 extension = 5; /* 2 opcode + 4 displacement */
9023 opcode[1] = opcode[0] + 0x10;
9024 opcode[0] = TWO_BYTE_OPCODE_ESCAPE;
9025 where_to_put_displacement = &opcode[2];
9026 break;
9027
9028 case ENCODE_RELAX_STATE (COND_JUMP, BIG16):
9029 extension = 3; /* 2 opcode + 2 displacement */
9030 opcode[1] = opcode[0] + 0x10;
9031 opcode[0] = TWO_BYTE_OPCODE_ESCAPE;
9032 where_to_put_displacement = &opcode[2];
9033 break;
9034
9035 case ENCODE_RELAX_STATE (COND_JUMP86, BIG16):
9036 extension = 4;
9037 opcode[0] ^= 1;
9038 opcode[1] = 3;
9039 opcode[2] = 0xe9;
9040 where_to_put_displacement = &opcode[3];
9041 break;
9042
9043 default:
9044 BAD_CASE (fragP->fr_subtype);
9045 break;
9046 }
9047 }
9048
9049 /* If size if less then four we are sure that the operand fits,
9050 but if it's 4, then it could be that the displacement is larger
9051 then -/+ 2GB. */
9052 if (DISP_SIZE_FROM_RELAX_STATE (fragP->fr_subtype) == 4
9053 && object_64bit
9054 && ((addressT) (displacement_from_opcode_start - extension
9055 + ((addressT) 1 << 31))
9056 > (((addressT) 2 << 31) - 1)))
9057 {
9058 as_bad_where (fragP->fr_file, fragP->fr_line,
9059 _("jump target out of range"));
9060 /* Make us emit 0. */
9061 displacement_from_opcode_start = extension;
9062 }
9063 /* Now put displacement after opcode. */
9064 md_number_to_chars ((char *) where_to_put_displacement,
9065 (valueT) (displacement_from_opcode_start - extension),
9066 DISP_SIZE_FROM_RELAX_STATE (fragP->fr_subtype));
9067 fragP->fr_fix += extension;
9068 }
9069 \f
9070 /* Apply a fixup (fixP) to segment data, once it has been determined
9071 by our caller that we have all the info we need to fix it up.
9072
9073 Parameter valP is the pointer to the value of the bits.
9074
9075 On the 386, immediates, displacements, and data pointers are all in
9076 the same (little-endian) format, so we don't need to care about which
9077 we are handling. */
9078
9079 void
9080 md_apply_fix (fixS *fixP, valueT *valP, segT seg ATTRIBUTE_UNUSED)
9081 {
9082 char *p = fixP->fx_where + fixP->fx_frag->fr_literal;
9083 valueT value = *valP;
9084
9085 #if !defined (TE_Mach)
9086 if (fixP->fx_pcrel)
9087 {
9088 switch (fixP->fx_r_type)
9089 {
9090 default:
9091 break;
9092
9093 case BFD_RELOC_64:
9094 fixP->fx_r_type = BFD_RELOC_64_PCREL;
9095 break;
9096 case BFD_RELOC_32:
9097 case BFD_RELOC_X86_64_32S:
9098 fixP->fx_r_type = BFD_RELOC_32_PCREL;
9099 break;
9100 case BFD_RELOC_16:
9101 fixP->fx_r_type = BFD_RELOC_16_PCREL;
9102 break;
9103 case BFD_RELOC_8:
9104 fixP->fx_r_type = BFD_RELOC_8_PCREL;
9105 break;
9106 }
9107 }
9108
9109 if (fixP->fx_addsy != NULL
9110 && (fixP->fx_r_type == BFD_RELOC_32_PCREL
9111 || fixP->fx_r_type == BFD_RELOC_64_PCREL
9112 || fixP->fx_r_type == BFD_RELOC_16_PCREL
9113 || fixP->fx_r_type == BFD_RELOC_8_PCREL)
9114 && !use_rela_relocations)
9115 {
9116 /* This is a hack. There should be a better way to handle this.
9117 This covers for the fact that bfd_install_relocation will
9118 subtract the current location (for partial_inplace, PC relative
9119 relocations); see more below. */
9120 #ifndef OBJ_AOUT
9121 if (IS_ELF
9122 #ifdef TE_PE
9123 || OUTPUT_FLAVOR == bfd_target_coff_flavour
9124 #endif
9125 )
9126 value += fixP->fx_where + fixP->fx_frag->fr_address;
9127 #endif
9128 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
9129 if (IS_ELF)
9130 {
9131 segT sym_seg = S_GET_SEGMENT (fixP->fx_addsy);
9132
9133 if ((sym_seg == seg
9134 || (symbol_section_p (fixP->fx_addsy)
9135 && sym_seg != absolute_section))
9136 && !generic_force_reloc (fixP))
9137 {
9138 /* Yes, we add the values in twice. This is because
9139 bfd_install_relocation subtracts them out again. I think
9140 bfd_install_relocation is broken, but I don't dare change
9141 it. FIXME. */
9142 value += fixP->fx_where + fixP->fx_frag->fr_address;
9143 }
9144 }
9145 #endif
9146 #if defined (OBJ_COFF) && defined (TE_PE)
9147 /* For some reason, the PE format does not store a
9148 section address offset for a PC relative symbol. */
9149 if (S_GET_SEGMENT (fixP->fx_addsy) != seg
9150 || S_IS_WEAK (fixP->fx_addsy))
9151 value += md_pcrel_from (fixP);
9152 #endif
9153 }
9154 #if defined (OBJ_COFF) && defined (TE_PE)
9155 if (fixP->fx_addsy != NULL
9156 && S_IS_WEAK (fixP->fx_addsy)
9157 /* PR 16858: Do not modify weak function references. */
9158 && ! fixP->fx_pcrel)
9159 {
9160 #if !defined (TE_PEP)
9161 /* For x86 PE weak function symbols are neither PC-relative
9162 nor do they set S_IS_FUNCTION. So the only reliable way
9163 to detect them is to check the flags of their containing
9164 section. */
9165 if (S_GET_SEGMENT (fixP->fx_addsy) != NULL
9166 && S_GET_SEGMENT (fixP->fx_addsy)->flags & SEC_CODE)
9167 ;
9168 else
9169 #endif
9170 value -= S_GET_VALUE (fixP->fx_addsy);
9171 }
9172 #endif
9173
9174 /* Fix a few things - the dynamic linker expects certain values here,
9175 and we must not disappoint it. */
9176 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
9177 if (IS_ELF && fixP->fx_addsy)
9178 switch (fixP->fx_r_type)
9179 {
9180 case BFD_RELOC_386_PLT32:
9181 case BFD_RELOC_X86_64_PLT32:
9182 /* Make the jump instruction point to the address of the operand. At
9183 runtime we merely add the offset to the actual PLT entry. */
9184 value = -4;
9185 break;
9186
9187 case BFD_RELOC_386_TLS_GD:
9188 case BFD_RELOC_386_TLS_LDM:
9189 case BFD_RELOC_386_TLS_IE_32:
9190 case BFD_RELOC_386_TLS_IE:
9191 case BFD_RELOC_386_TLS_GOTIE:
9192 case BFD_RELOC_386_TLS_GOTDESC:
9193 case BFD_RELOC_X86_64_TLSGD:
9194 case BFD_RELOC_X86_64_TLSLD:
9195 case BFD_RELOC_X86_64_GOTTPOFF:
9196 case BFD_RELOC_X86_64_GOTPC32_TLSDESC:
9197 value = 0; /* Fully resolved at runtime. No addend. */
9198 /* Fallthrough */
9199 case BFD_RELOC_386_TLS_LE:
9200 case BFD_RELOC_386_TLS_LDO_32:
9201 case BFD_RELOC_386_TLS_LE_32:
9202 case BFD_RELOC_X86_64_DTPOFF32:
9203 case BFD_RELOC_X86_64_DTPOFF64:
9204 case BFD_RELOC_X86_64_TPOFF32:
9205 case BFD_RELOC_X86_64_TPOFF64:
9206 S_SET_THREAD_LOCAL (fixP->fx_addsy);
9207 break;
9208
9209 case BFD_RELOC_386_TLS_DESC_CALL:
9210 case BFD_RELOC_X86_64_TLSDESC_CALL:
9211 value = 0; /* Fully resolved at runtime. No addend. */
9212 S_SET_THREAD_LOCAL (fixP->fx_addsy);
9213 fixP->fx_done = 0;
9214 return;
9215
9216 case BFD_RELOC_386_GOT32:
9217 case BFD_RELOC_X86_64_GOT32:
9218 value = 0; /* Fully resolved at runtime. No addend. */
9219 break;
9220
9221 case BFD_RELOC_VTABLE_INHERIT:
9222 case BFD_RELOC_VTABLE_ENTRY:
9223 fixP->fx_done = 0;
9224 return;
9225
9226 default:
9227 break;
9228 }
9229 #endif /* defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) */
9230 *valP = value;
9231 #endif /* !defined (TE_Mach) */
9232
9233 /* Are we finished with this relocation now? */
9234 if (fixP->fx_addsy == NULL)
9235 fixP->fx_done = 1;
9236 #if defined (OBJ_COFF) && defined (TE_PE)
9237 else if (fixP->fx_addsy != NULL && S_IS_WEAK (fixP->fx_addsy))
9238 {
9239 fixP->fx_done = 0;
9240 /* Remember value for tc_gen_reloc. */
9241 fixP->fx_addnumber = value;
9242 /* Clear out the frag for now. */
9243 value = 0;
9244 }
9245 #endif
9246 else if (use_rela_relocations)
9247 {
9248 fixP->fx_no_overflow = 1;
9249 /* Remember value for tc_gen_reloc. */
9250 fixP->fx_addnumber = value;
9251 value = 0;
9252 }
9253
9254 md_number_to_chars (p, value, fixP->fx_size);
9255 }
9256 \f
9257 char *
9258 md_atof (int type, char *litP, int *sizeP)
9259 {
9260 /* This outputs the LITTLENUMs in REVERSE order;
9261 in accord with the bigendian 386. */
9262 return ieee_md_atof (type, litP, sizeP, FALSE);
9263 }
9264 \f
9265 static char output_invalid_buf[sizeof (unsigned char) * 2 + 6];
9266
9267 static char *
9268 output_invalid (int c)
9269 {
9270 if (ISPRINT (c))
9271 snprintf (output_invalid_buf, sizeof (output_invalid_buf),
9272 "'%c'", c);
9273 else
9274 snprintf (output_invalid_buf, sizeof (output_invalid_buf),
9275 "(0x%x)", (unsigned char) c);
9276 return output_invalid_buf;
9277 }
9278
9279 /* REG_STRING starts *before* REGISTER_PREFIX. */
9280
9281 static const reg_entry *
9282 parse_real_register (char *reg_string, char **end_op)
9283 {
9284 char *s = reg_string;
9285 char *p;
9286 char reg_name_given[MAX_REG_NAME_SIZE + 1];
9287 const reg_entry *r;
9288
9289 /* Skip possible REGISTER_PREFIX and possible whitespace. */
9290 if (*s == REGISTER_PREFIX)
9291 ++s;
9292
9293 if (is_space_char (*s))
9294 ++s;
9295
9296 p = reg_name_given;
9297 while ((*p++ = register_chars[(unsigned char) *s]) != '\0')
9298 {
9299 if (p >= reg_name_given + MAX_REG_NAME_SIZE)
9300 return (const reg_entry *) NULL;
9301 s++;
9302 }
9303
9304 /* For naked regs, make sure that we are not dealing with an identifier.
9305 This prevents confusing an identifier like `eax_var' with register
9306 `eax'. */
9307 if (allow_naked_reg && identifier_chars[(unsigned char) *s])
9308 return (const reg_entry *) NULL;
9309
9310 *end_op = s;
9311
9312 r = (const reg_entry *) hash_find (reg_hash, reg_name_given);
9313
9314 /* Handle floating point regs, allowing spaces in the (i) part. */
9315 if (r == i386_regtab /* %st is first entry of table */)
9316 {
9317 if (is_space_char (*s))
9318 ++s;
9319 if (*s == '(')
9320 {
9321 ++s;
9322 if (is_space_char (*s))
9323 ++s;
9324 if (*s >= '0' && *s <= '7')
9325 {
9326 int fpr = *s - '0';
9327 ++s;
9328 if (is_space_char (*s))
9329 ++s;
9330 if (*s == ')')
9331 {
9332 *end_op = s + 1;
9333 r = (const reg_entry *) hash_find (reg_hash, "st(0)");
9334 know (r);
9335 return r + fpr;
9336 }
9337 }
9338 /* We have "%st(" then garbage. */
9339 return (const reg_entry *) NULL;
9340 }
9341 }
9342
9343 if (r == NULL || allow_pseudo_reg)
9344 return r;
9345
9346 if (operand_type_all_zero (&r->reg_type))
9347 return (const reg_entry *) NULL;
9348
9349 if ((r->reg_type.bitfield.reg32
9350 || r->reg_type.bitfield.sreg3
9351 || r->reg_type.bitfield.control
9352 || r->reg_type.bitfield.debug
9353 || r->reg_type.bitfield.test)
9354 && !cpu_arch_flags.bitfield.cpui386)
9355 return (const reg_entry *) NULL;
9356
9357 if (r->reg_type.bitfield.floatreg
9358 && !cpu_arch_flags.bitfield.cpu8087
9359 && !cpu_arch_flags.bitfield.cpu287
9360 && !cpu_arch_flags.bitfield.cpu387)
9361 return (const reg_entry *) NULL;
9362
9363 if (r->reg_type.bitfield.regmmx && !cpu_arch_flags.bitfield.cpummx)
9364 return (const reg_entry *) NULL;
9365
9366 if (r->reg_type.bitfield.regxmm && !cpu_arch_flags.bitfield.cpusse)
9367 return (const reg_entry *) NULL;
9368
9369 if (r->reg_type.bitfield.regymm && !cpu_arch_flags.bitfield.cpuavx)
9370 return (const reg_entry *) NULL;
9371
9372 if ((r->reg_type.bitfield.regzmm || r->reg_type.bitfield.regmask)
9373 && !cpu_arch_flags.bitfield.cpuavx512f)
9374 return (const reg_entry *) NULL;
9375
9376 /* Don't allow fake index register unless allow_index_reg isn't 0. */
9377 if (!allow_index_reg
9378 && (r->reg_num == RegEiz || r->reg_num == RegRiz))
9379 return (const reg_entry *) NULL;
9380
9381 /* Upper 16 vector register is only available with VREX in 64bit
9382 mode. */
9383 if ((r->reg_flags & RegVRex))
9384 {
9385 if (!cpu_arch_flags.bitfield.cpuvrex
9386 || flag_code != CODE_64BIT)
9387 return (const reg_entry *) NULL;
9388
9389 i.need_vrex = 1;
9390 }
9391
9392 if (((r->reg_flags & (RegRex64 | RegRex))
9393 || r->reg_type.bitfield.reg64)
9394 && (!cpu_arch_flags.bitfield.cpulm
9395 || !operand_type_equal (&r->reg_type, &control))
9396 && flag_code != CODE_64BIT)
9397 return (const reg_entry *) NULL;
9398
9399 if (r->reg_type.bitfield.sreg3 && r->reg_num == RegFlat && !intel_syntax)
9400 return (const reg_entry *) NULL;
9401
9402 return r;
9403 }
9404
9405 /* REG_STRING starts *before* REGISTER_PREFIX. */
9406
9407 static const reg_entry *
9408 parse_register (char *reg_string, char **end_op)
9409 {
9410 const reg_entry *r;
9411
9412 if (*reg_string == REGISTER_PREFIX || allow_naked_reg)
9413 r = parse_real_register (reg_string, end_op);
9414 else
9415 r = NULL;
9416 if (!r)
9417 {
9418 char *save = input_line_pointer;
9419 char c;
9420 symbolS *symbolP;
9421
9422 input_line_pointer = reg_string;
9423 c = get_symbol_end ();
9424 symbolP = symbol_find (reg_string);
9425 if (symbolP && S_GET_SEGMENT (symbolP) == reg_section)
9426 {
9427 const expressionS *e = symbol_get_value_expression (symbolP);
9428
9429 know (e->X_op == O_register);
9430 know (e->X_add_number >= 0
9431 && (valueT) e->X_add_number < i386_regtab_size);
9432 r = i386_regtab + e->X_add_number;
9433 if ((r->reg_flags & RegVRex))
9434 i.need_vrex = 1;
9435 *end_op = input_line_pointer;
9436 }
9437 *input_line_pointer = c;
9438 input_line_pointer = save;
9439 }
9440 return r;
9441 }
9442
9443 int
9444 i386_parse_name (char *name, expressionS *e, char *nextcharP)
9445 {
9446 const reg_entry *r;
9447 char *end = input_line_pointer;
9448
9449 *end = *nextcharP;
9450 r = parse_register (name, &input_line_pointer);
9451 if (r && end <= input_line_pointer)
9452 {
9453 *nextcharP = *input_line_pointer;
9454 *input_line_pointer = 0;
9455 e->X_op = O_register;
9456 e->X_add_number = r - i386_regtab;
9457 return 1;
9458 }
9459 input_line_pointer = end;
9460 *end = 0;
9461 return intel_syntax ? i386_intel_parse_name (name, e) : 0;
9462 }
9463
9464 void
9465 md_operand (expressionS *e)
9466 {
9467 char *end;
9468 const reg_entry *r;
9469
9470 switch (*input_line_pointer)
9471 {
9472 case REGISTER_PREFIX:
9473 r = parse_real_register (input_line_pointer, &end);
9474 if (r)
9475 {
9476 e->X_op = O_register;
9477 e->X_add_number = r - i386_regtab;
9478 input_line_pointer = end;
9479 }
9480 break;
9481
9482 case '[':
9483 gas_assert (intel_syntax);
9484 end = input_line_pointer++;
9485 expression (e);
9486 if (*input_line_pointer == ']')
9487 {
9488 ++input_line_pointer;
9489 e->X_op_symbol = make_expr_symbol (e);
9490 e->X_add_symbol = NULL;
9491 e->X_add_number = 0;
9492 e->X_op = O_index;
9493 }
9494 else
9495 {
9496 e->X_op = O_absent;
9497 input_line_pointer = end;
9498 }
9499 break;
9500 }
9501 }
9502
9503 \f
9504 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
9505 const char *md_shortopts = "kVQ:sqn";
9506 #else
9507 const char *md_shortopts = "qn";
9508 #endif
9509
9510 #define OPTION_32 (OPTION_MD_BASE + 0)
9511 #define OPTION_64 (OPTION_MD_BASE + 1)
9512 #define OPTION_DIVIDE (OPTION_MD_BASE + 2)
9513 #define OPTION_MARCH (OPTION_MD_BASE + 3)
9514 #define OPTION_MTUNE (OPTION_MD_BASE + 4)
9515 #define OPTION_MMNEMONIC (OPTION_MD_BASE + 5)
9516 #define OPTION_MSYNTAX (OPTION_MD_BASE + 6)
9517 #define OPTION_MINDEX_REG (OPTION_MD_BASE + 7)
9518 #define OPTION_MNAKED_REG (OPTION_MD_BASE + 8)
9519 #define OPTION_MOLD_GCC (OPTION_MD_BASE + 9)
9520 #define OPTION_MSSE2AVX (OPTION_MD_BASE + 10)
9521 #define OPTION_MSSE_CHECK (OPTION_MD_BASE + 11)
9522 #define OPTION_MOPERAND_CHECK (OPTION_MD_BASE + 12)
9523 #define OPTION_MAVXSCALAR (OPTION_MD_BASE + 13)
9524 #define OPTION_X32 (OPTION_MD_BASE + 14)
9525 #define OPTION_MADD_BND_PREFIX (OPTION_MD_BASE + 15)
9526 #define OPTION_MEVEXLIG (OPTION_MD_BASE + 16)
9527 #define OPTION_MEVEXWIG (OPTION_MD_BASE + 17)
9528 #define OPTION_MBIG_OBJ (OPTION_MD_BASE + 18)
9529 #define OPTION_OMIT_LOCK_PREFIX (OPTION_MD_BASE + 19)
9530 #define OPTION_MEVEXRCIG (OPTION_MD_BASE + 20)
9531
9532 struct option md_longopts[] =
9533 {
9534 {"32", no_argument, NULL, OPTION_32},
9535 #if (defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) \
9536 || defined (TE_PE) || defined (TE_PEP) || defined (OBJ_MACH_O))
9537 {"64", no_argument, NULL, OPTION_64},
9538 #endif
9539 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
9540 {"x32", no_argument, NULL, OPTION_X32},
9541 #endif
9542 {"divide", no_argument, NULL, OPTION_DIVIDE},
9543 {"march", required_argument, NULL, OPTION_MARCH},
9544 {"mtune", required_argument, NULL, OPTION_MTUNE},
9545 {"mmnemonic", required_argument, NULL, OPTION_MMNEMONIC},
9546 {"msyntax", required_argument, NULL, OPTION_MSYNTAX},
9547 {"mindex-reg", no_argument, NULL, OPTION_MINDEX_REG},
9548 {"mnaked-reg", no_argument, NULL, OPTION_MNAKED_REG},
9549 {"mold-gcc", no_argument, NULL, OPTION_MOLD_GCC},
9550 {"msse2avx", no_argument, NULL, OPTION_MSSE2AVX},
9551 {"msse-check", required_argument, NULL, OPTION_MSSE_CHECK},
9552 {"moperand-check", required_argument, NULL, OPTION_MOPERAND_CHECK},
9553 {"mavxscalar", required_argument, NULL, OPTION_MAVXSCALAR},
9554 {"madd-bnd-prefix", no_argument, NULL, OPTION_MADD_BND_PREFIX},
9555 {"mevexlig", required_argument, NULL, OPTION_MEVEXLIG},
9556 {"mevexwig", required_argument, NULL, OPTION_MEVEXWIG},
9557 # if defined (TE_PE) || defined (TE_PEP)
9558 {"mbig-obj", no_argument, NULL, OPTION_MBIG_OBJ},
9559 #endif
9560 {"momit-lock-prefix", required_argument, NULL, OPTION_OMIT_LOCK_PREFIX},
9561 {"mevexrcig", required_argument, NULL, OPTION_MEVEXRCIG},
9562 {NULL, no_argument, NULL, 0}
9563 };
9564 size_t md_longopts_size = sizeof (md_longopts);
9565
9566 int
9567 md_parse_option (int c, char *arg)
9568 {
9569 unsigned int j;
9570 char *arch, *next;
9571
9572 switch (c)
9573 {
9574 case 'n':
9575 optimize_align_code = 0;
9576 break;
9577
9578 case 'q':
9579 quiet_warnings = 1;
9580 break;
9581
9582 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
9583 /* -Qy, -Qn: SVR4 arguments controlling whether a .comment section
9584 should be emitted or not. FIXME: Not implemented. */
9585 case 'Q':
9586 break;
9587
9588 /* -V: SVR4 argument to print version ID. */
9589 case 'V':
9590 print_version_id ();
9591 break;
9592
9593 /* -k: Ignore for FreeBSD compatibility. */
9594 case 'k':
9595 break;
9596
9597 case 's':
9598 /* -s: On i386 Solaris, this tells the native assembler to use
9599 .stab instead of .stab.excl. We always use .stab anyhow. */
9600 break;
9601 #endif
9602 #if (defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) \
9603 || defined (TE_PE) || defined (TE_PEP) || defined (OBJ_MACH_O))
9604 case OPTION_64:
9605 {
9606 const char **list, **l;
9607
9608 list = bfd_target_list ();
9609 for (l = list; *l != NULL; l++)
9610 if (CONST_STRNEQ (*l, "elf64-x86-64")
9611 || strcmp (*l, "coff-x86-64") == 0
9612 || strcmp (*l, "pe-x86-64") == 0
9613 || strcmp (*l, "pei-x86-64") == 0
9614 || strcmp (*l, "mach-o-x86-64") == 0)
9615 {
9616 default_arch = "x86_64";
9617 break;
9618 }
9619 if (*l == NULL)
9620 as_fatal (_("no compiled in support for x86_64"));
9621 free (list);
9622 }
9623 break;
9624 #endif
9625
9626 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
9627 case OPTION_X32:
9628 if (IS_ELF)
9629 {
9630 const char **list, **l;
9631
9632 list = bfd_target_list ();
9633 for (l = list; *l != NULL; l++)
9634 if (CONST_STRNEQ (*l, "elf32-x86-64"))
9635 {
9636 default_arch = "x86_64:32";
9637 break;
9638 }
9639 if (*l == NULL)
9640 as_fatal (_("no compiled in support for 32bit x86_64"));
9641 free (list);
9642 }
9643 else
9644 as_fatal (_("32bit x86_64 is only supported for ELF"));
9645 break;
9646 #endif
9647
9648 case OPTION_32:
9649 default_arch = "i386";
9650 break;
9651
9652 case OPTION_DIVIDE:
9653 #ifdef SVR4_COMMENT_CHARS
9654 {
9655 char *n, *t;
9656 const char *s;
9657
9658 n = (char *) xmalloc (strlen (i386_comment_chars) + 1);
9659 t = n;
9660 for (s = i386_comment_chars; *s != '\0'; s++)
9661 if (*s != '/')
9662 *t++ = *s;
9663 *t = '\0';
9664 i386_comment_chars = n;
9665 }
9666 #endif
9667 break;
9668
9669 case OPTION_MARCH:
9670 arch = xstrdup (arg);
9671 do
9672 {
9673 if (*arch == '.')
9674 as_fatal (_("invalid -march= option: `%s'"), arg);
9675 next = strchr (arch, '+');
9676 if (next)
9677 *next++ = '\0';
9678 for (j = 0; j < ARRAY_SIZE (cpu_arch); j++)
9679 {
9680 if (strcmp (arch, cpu_arch [j].name) == 0)
9681 {
9682 /* Processor. */
9683 if (! cpu_arch[j].flags.bitfield.cpui386)
9684 continue;
9685
9686 cpu_arch_name = cpu_arch[j].name;
9687 cpu_sub_arch_name = NULL;
9688 cpu_arch_flags = cpu_arch[j].flags;
9689 cpu_arch_isa = cpu_arch[j].type;
9690 cpu_arch_isa_flags = cpu_arch[j].flags;
9691 if (!cpu_arch_tune_set)
9692 {
9693 cpu_arch_tune = cpu_arch_isa;
9694 cpu_arch_tune_flags = cpu_arch_isa_flags;
9695 }
9696 break;
9697 }
9698 else if (*cpu_arch [j].name == '.'
9699 && strcmp (arch, cpu_arch [j].name + 1) == 0)
9700 {
9701 /* ISA entension. */
9702 i386_cpu_flags flags;
9703
9704 if (!cpu_arch[j].negated)
9705 flags = cpu_flags_or (cpu_arch_flags,
9706 cpu_arch[j].flags);
9707 else
9708 flags = cpu_flags_and_not (cpu_arch_flags,
9709 cpu_arch[j].flags);
9710
9711 if (!valid_iamcu_cpu_flags (&flags))
9712 as_fatal (_("`%s' isn't valid for Intel MCU"), arch);
9713 else if (!cpu_flags_equal (&flags, &cpu_arch_flags))
9714 {
9715 if (cpu_sub_arch_name)
9716 {
9717 char *name = cpu_sub_arch_name;
9718 cpu_sub_arch_name = concat (name,
9719 cpu_arch[j].name,
9720 (const char *) NULL);
9721 free (name);
9722 }
9723 else
9724 cpu_sub_arch_name = xstrdup (cpu_arch[j].name);
9725 cpu_arch_flags = flags;
9726 cpu_arch_isa_flags = flags;
9727 }
9728 break;
9729 }
9730 }
9731
9732 if (j >= ARRAY_SIZE (cpu_arch))
9733 as_fatal (_("invalid -march= option: `%s'"), arg);
9734
9735 arch = next;
9736 }
9737 while (next != NULL );
9738 break;
9739
9740 case OPTION_MTUNE:
9741 if (*arg == '.')
9742 as_fatal (_("invalid -mtune= option: `%s'"), arg);
9743 for (j = 0; j < ARRAY_SIZE (cpu_arch); j++)
9744 {
9745 if (strcmp (arg, cpu_arch [j].name) == 0)
9746 {
9747 cpu_arch_tune_set = 1;
9748 cpu_arch_tune = cpu_arch [j].type;
9749 cpu_arch_tune_flags = cpu_arch[j].flags;
9750 break;
9751 }
9752 }
9753 if (j >= ARRAY_SIZE (cpu_arch))
9754 as_fatal (_("invalid -mtune= option: `%s'"), arg);
9755 break;
9756
9757 case OPTION_MMNEMONIC:
9758 if (strcasecmp (arg, "att") == 0)
9759 intel_mnemonic = 0;
9760 else if (strcasecmp (arg, "intel") == 0)
9761 intel_mnemonic = 1;
9762 else
9763 as_fatal (_("invalid -mmnemonic= option: `%s'"), arg);
9764 break;
9765
9766 case OPTION_MSYNTAX:
9767 if (strcasecmp (arg, "att") == 0)
9768 intel_syntax = 0;
9769 else if (strcasecmp (arg, "intel") == 0)
9770 intel_syntax = 1;
9771 else
9772 as_fatal (_("invalid -msyntax= option: `%s'"), arg);
9773 break;
9774
9775 case OPTION_MINDEX_REG:
9776 allow_index_reg = 1;
9777 break;
9778
9779 case OPTION_MNAKED_REG:
9780 allow_naked_reg = 1;
9781 break;
9782
9783 case OPTION_MOLD_GCC:
9784 old_gcc = 1;
9785 break;
9786
9787 case OPTION_MSSE2AVX:
9788 sse2avx = 1;
9789 break;
9790
9791 case OPTION_MSSE_CHECK:
9792 if (strcasecmp (arg, "error") == 0)
9793 sse_check = check_error;
9794 else if (strcasecmp (arg, "warning") == 0)
9795 sse_check = check_warning;
9796 else if (strcasecmp (arg, "none") == 0)
9797 sse_check = check_none;
9798 else
9799 as_fatal (_("invalid -msse-check= option: `%s'"), arg);
9800 break;
9801
9802 case OPTION_MOPERAND_CHECK:
9803 if (strcasecmp (arg, "error") == 0)
9804 operand_check = check_error;
9805 else if (strcasecmp (arg, "warning") == 0)
9806 operand_check = check_warning;
9807 else if (strcasecmp (arg, "none") == 0)
9808 operand_check = check_none;
9809 else
9810 as_fatal (_("invalid -moperand-check= option: `%s'"), arg);
9811 break;
9812
9813 case OPTION_MAVXSCALAR:
9814 if (strcasecmp (arg, "128") == 0)
9815 avxscalar = vex128;
9816 else if (strcasecmp (arg, "256") == 0)
9817 avxscalar = vex256;
9818 else
9819 as_fatal (_("invalid -mavxscalar= option: `%s'"), arg);
9820 break;
9821
9822 case OPTION_MADD_BND_PREFIX:
9823 add_bnd_prefix = 1;
9824 break;
9825
9826 case OPTION_MEVEXLIG:
9827 if (strcmp (arg, "128") == 0)
9828 evexlig = evexl128;
9829 else if (strcmp (arg, "256") == 0)
9830 evexlig = evexl256;
9831 else if (strcmp (arg, "512") == 0)
9832 evexlig = evexl512;
9833 else
9834 as_fatal (_("invalid -mevexlig= option: `%s'"), arg);
9835 break;
9836
9837 case OPTION_MEVEXRCIG:
9838 if (strcmp (arg, "rne") == 0)
9839 evexrcig = rne;
9840 else if (strcmp (arg, "rd") == 0)
9841 evexrcig = rd;
9842 else if (strcmp (arg, "ru") == 0)
9843 evexrcig = ru;
9844 else if (strcmp (arg, "rz") == 0)
9845 evexrcig = rz;
9846 else
9847 as_fatal (_("invalid -mevexrcig= option: `%s'"), arg);
9848 break;
9849
9850 case OPTION_MEVEXWIG:
9851 if (strcmp (arg, "0") == 0)
9852 evexwig = evexw0;
9853 else if (strcmp (arg, "1") == 0)
9854 evexwig = evexw1;
9855 else
9856 as_fatal (_("invalid -mevexwig= option: `%s'"), arg);
9857 break;
9858
9859 # if defined (TE_PE) || defined (TE_PEP)
9860 case OPTION_MBIG_OBJ:
9861 use_big_obj = 1;
9862 break;
9863 #endif
9864
9865 case OPTION_OMIT_LOCK_PREFIX:
9866 if (strcasecmp (arg, "yes") == 0)
9867 omit_lock_prefix = 1;
9868 else if (strcasecmp (arg, "no") == 0)
9869 omit_lock_prefix = 0;
9870 else
9871 as_fatal (_("invalid -momit-lock-prefix= option: `%s'"), arg);
9872 break;
9873
9874 default:
9875 return 0;
9876 }
9877 return 1;
9878 }
9879
9880 #define MESSAGE_TEMPLATE \
9881 " "
9882
9883 static void
9884 show_arch (FILE *stream, int ext, int check)
9885 {
9886 static char message[] = MESSAGE_TEMPLATE;
9887 char *start = message + 27;
9888 char *p;
9889 int size = sizeof (MESSAGE_TEMPLATE);
9890 int left;
9891 const char *name;
9892 int len;
9893 unsigned int j;
9894
9895 p = start;
9896 left = size - (start - message);
9897 for (j = 0; j < ARRAY_SIZE (cpu_arch); j++)
9898 {
9899 /* Should it be skipped? */
9900 if (cpu_arch [j].skip)
9901 continue;
9902
9903 name = cpu_arch [j].name;
9904 len = cpu_arch [j].len;
9905 if (*name == '.')
9906 {
9907 /* It is an extension. Skip if we aren't asked to show it. */
9908 if (ext)
9909 {
9910 name++;
9911 len--;
9912 }
9913 else
9914 continue;
9915 }
9916 else if (ext)
9917 {
9918 /* It is an processor. Skip if we show only extension. */
9919 continue;
9920 }
9921 else if (check && ! cpu_arch[j].flags.bitfield.cpui386)
9922 {
9923 /* It is an impossible processor - skip. */
9924 continue;
9925 }
9926
9927 /* Reserve 2 spaces for ", " or ",\0" */
9928 left -= len + 2;
9929
9930 /* Check if there is any room. */
9931 if (left >= 0)
9932 {
9933 if (p != start)
9934 {
9935 *p++ = ',';
9936 *p++ = ' ';
9937 }
9938 p = mempcpy (p, name, len);
9939 }
9940 else
9941 {
9942 /* Output the current message now and start a new one. */
9943 *p++ = ',';
9944 *p = '\0';
9945 fprintf (stream, "%s\n", message);
9946 p = start;
9947 left = size - (start - message) - len - 2;
9948
9949 gas_assert (left >= 0);
9950
9951 p = mempcpy (p, name, len);
9952 }
9953 }
9954
9955 *p = '\0';
9956 fprintf (stream, "%s\n", message);
9957 }
9958
9959 void
9960 md_show_usage (FILE *stream)
9961 {
9962 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
9963 fprintf (stream, _("\
9964 -Q ignored\n\
9965 -V print assembler version number\n\
9966 -k ignored\n"));
9967 #endif
9968 fprintf (stream, _("\
9969 -n Do not optimize code alignment\n\
9970 -q quieten some warnings\n"));
9971 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
9972 fprintf (stream, _("\
9973 -s ignored\n"));
9974 #endif
9975 #if (defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) \
9976 || defined (TE_PE) || defined (TE_PEP))
9977 fprintf (stream, _("\
9978 --32/--64/--x32 generate 32bit/64bit/x32 code\n"));
9979 #endif
9980 #ifdef SVR4_COMMENT_CHARS
9981 fprintf (stream, _("\
9982 --divide do not treat `/' as a comment character\n"));
9983 #else
9984 fprintf (stream, _("\
9985 --divide ignored\n"));
9986 #endif
9987 fprintf (stream, _("\
9988 -march=CPU[,+EXTENSION...]\n\
9989 generate code for CPU and EXTENSION, CPU is one of:\n"));
9990 show_arch (stream, 0, 1);
9991 fprintf (stream, _("\
9992 EXTENSION is combination of:\n"));
9993 show_arch (stream, 1, 0);
9994 fprintf (stream, _("\
9995 -mtune=CPU optimize for CPU, CPU is one of:\n"));
9996 show_arch (stream, 0, 0);
9997 fprintf (stream, _("\
9998 -msse2avx encode SSE instructions with VEX prefix\n"));
9999 fprintf (stream, _("\
10000 -msse-check=[none|error|warning]\n\
10001 check SSE instructions\n"));
10002 fprintf (stream, _("\
10003 -moperand-check=[none|error|warning]\n\
10004 check operand combinations for validity\n"));
10005 fprintf (stream, _("\
10006 -mavxscalar=[128|256] encode scalar AVX instructions with specific vector\n\
10007 length\n"));
10008 fprintf (stream, _("\
10009 -mevexlig=[128|256|512] encode scalar EVEX instructions with specific vector\n\
10010 length\n"));
10011 fprintf (stream, _("\
10012 -mevexwig=[0|1] encode EVEX instructions with specific EVEX.W value\n\
10013 for EVEX.W bit ignored instructions\n"));
10014 fprintf (stream, _("\
10015 -mevexrcig=[rne|rd|ru|rz]\n\
10016 encode EVEX instructions with specific EVEX.RC value\n\
10017 for SAE-only ignored instructions\n"));
10018 fprintf (stream, _("\
10019 -mmnemonic=[att|intel] use AT&T/Intel mnemonic\n"));
10020 fprintf (stream, _("\
10021 -msyntax=[att|intel] use AT&T/Intel syntax\n"));
10022 fprintf (stream, _("\
10023 -mindex-reg support pseudo index registers\n"));
10024 fprintf (stream, _("\
10025 -mnaked-reg don't require `%%' prefix for registers\n"));
10026 fprintf (stream, _("\
10027 -mold-gcc support old (<= 2.8.1) versions of gcc\n"));
10028 fprintf (stream, _("\
10029 -madd-bnd-prefix add BND prefix for all valid branches\n"));
10030 # if defined (TE_PE) || defined (TE_PEP)
10031 fprintf (stream, _("\
10032 -mbig-obj generate big object files\n"));
10033 #endif
10034 fprintf (stream, _("\
10035 -momit-lock-prefix=[no|yes]\n\
10036 strip all lock prefixes\n"));
10037 }
10038
10039 #if ((defined (OBJ_MAYBE_COFF) && defined (OBJ_MAYBE_AOUT)) \
10040 || defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) \
10041 || defined (TE_PE) || defined (TE_PEP) || defined (OBJ_MACH_O))
10042
10043 /* Pick the target format to use. */
10044
10045 const char *
10046 i386_target_format (void)
10047 {
10048 if (!strncmp (default_arch, "x86_64", 6))
10049 {
10050 update_code_flag (CODE_64BIT, 1);
10051 if (default_arch[6] == '\0')
10052 x86_elf_abi = X86_64_ABI;
10053 else
10054 x86_elf_abi = X86_64_X32_ABI;
10055 }
10056 else if (!strcmp (default_arch, "i386"))
10057 update_code_flag (CODE_32BIT, 1);
10058 else if (!strcmp (default_arch, "iamcu"))
10059 {
10060 update_code_flag (CODE_32BIT, 1);
10061 if (cpu_arch_isa == PROCESSOR_UNKNOWN)
10062 {
10063 static const i386_cpu_flags iamcu_flags = CPU_IAMCU_FLAGS;
10064 cpu_arch_name = "iamcu";
10065 cpu_sub_arch_name = NULL;
10066 cpu_arch_flags = iamcu_flags;
10067 cpu_arch_isa = PROCESSOR_IAMCU;
10068 cpu_arch_isa_flags = iamcu_flags;
10069 if (!cpu_arch_tune_set)
10070 {
10071 cpu_arch_tune = cpu_arch_isa;
10072 cpu_arch_tune_flags = cpu_arch_isa_flags;
10073 }
10074 }
10075 else
10076 as_fatal (_("Intel MCU doesn't support `%s' architecture"),
10077 cpu_arch_name);
10078 }
10079 else
10080 as_fatal (_("unknown architecture"));
10081
10082 if (cpu_flags_all_zero (&cpu_arch_isa_flags))
10083 cpu_arch_isa_flags = cpu_arch[flag_code == CODE_64BIT].flags;
10084 if (cpu_flags_all_zero (&cpu_arch_tune_flags))
10085 cpu_arch_tune_flags = cpu_arch[flag_code == CODE_64BIT].flags;
10086
10087 switch (OUTPUT_FLAVOR)
10088 {
10089 #if defined (OBJ_MAYBE_AOUT) || defined (OBJ_AOUT)
10090 case bfd_target_aout_flavour:
10091 return AOUT_TARGET_FORMAT;
10092 #endif
10093 #if defined (OBJ_MAYBE_COFF) || defined (OBJ_COFF)
10094 # if defined (TE_PE) || defined (TE_PEP)
10095 case bfd_target_coff_flavour:
10096 if (flag_code == CODE_64BIT)
10097 return use_big_obj ? "pe-bigobj-x86-64" : "pe-x86-64";
10098 else
10099 return "pe-i386";
10100 # elif defined (TE_GO32)
10101 case bfd_target_coff_flavour:
10102 return "coff-go32";
10103 # else
10104 case bfd_target_coff_flavour:
10105 return "coff-i386";
10106 # endif
10107 #endif
10108 #if defined (OBJ_MAYBE_ELF) || defined (OBJ_ELF)
10109 case bfd_target_elf_flavour:
10110 {
10111 const char *format;
10112
10113 switch (x86_elf_abi)
10114 {
10115 default:
10116 format = ELF_TARGET_FORMAT;
10117 break;
10118 case X86_64_ABI:
10119 use_rela_relocations = 1;
10120 object_64bit = 1;
10121 format = ELF_TARGET_FORMAT64;
10122 break;
10123 case X86_64_X32_ABI:
10124 use_rela_relocations = 1;
10125 object_64bit = 1;
10126 disallow_64bit_reloc = 1;
10127 format = ELF_TARGET_FORMAT32;
10128 break;
10129 }
10130 if (cpu_arch_isa == PROCESSOR_L1OM)
10131 {
10132 if (x86_elf_abi != X86_64_ABI)
10133 as_fatal (_("Intel L1OM is 64bit only"));
10134 return ELF_TARGET_L1OM_FORMAT;
10135 }
10136 else if (cpu_arch_isa == PROCESSOR_K1OM)
10137 {
10138 if (x86_elf_abi != X86_64_ABI)
10139 as_fatal (_("Intel K1OM is 64bit only"));
10140 return ELF_TARGET_K1OM_FORMAT;
10141 }
10142 else if (cpu_arch_isa == PROCESSOR_IAMCU)
10143 {
10144 if (x86_elf_abi != I386_ABI)
10145 as_fatal (_("Intel MCU is 32bit only"));
10146 return ELF_TARGET_IAMCU_FORMAT;
10147 }
10148 else
10149 return format;
10150 }
10151 #endif
10152 #if defined (OBJ_MACH_O)
10153 case bfd_target_mach_o_flavour:
10154 if (flag_code == CODE_64BIT)
10155 {
10156 use_rela_relocations = 1;
10157 object_64bit = 1;
10158 return "mach-o-x86-64";
10159 }
10160 else
10161 return "mach-o-i386";
10162 #endif
10163 default:
10164 abort ();
10165 return NULL;
10166 }
10167 }
10168
10169 #endif /* OBJ_MAYBE_ more than one */
10170 \f
10171 symbolS *
10172 md_undefined_symbol (char *name)
10173 {
10174 if (name[0] == GLOBAL_OFFSET_TABLE_NAME[0]
10175 && name[1] == GLOBAL_OFFSET_TABLE_NAME[1]
10176 && name[2] == GLOBAL_OFFSET_TABLE_NAME[2]
10177 && strcmp (name, GLOBAL_OFFSET_TABLE_NAME) == 0)
10178 {
10179 if (!GOT_symbol)
10180 {
10181 if (symbol_find (name))
10182 as_bad (_("GOT already in symbol table"));
10183 GOT_symbol = symbol_new (name, undefined_section,
10184 (valueT) 0, &zero_address_frag);
10185 };
10186 return GOT_symbol;
10187 }
10188 return 0;
10189 }
10190
10191 /* Round up a section size to the appropriate boundary. */
10192
10193 valueT
10194 md_section_align (segT segment ATTRIBUTE_UNUSED, valueT size)
10195 {
10196 #if (defined (OBJ_AOUT) || defined (OBJ_MAYBE_AOUT))
10197 if (OUTPUT_FLAVOR == bfd_target_aout_flavour)
10198 {
10199 /* For a.out, force the section size to be aligned. If we don't do
10200 this, BFD will align it for us, but it will not write out the
10201 final bytes of the section. This may be a bug in BFD, but it is
10202 easier to fix it here since that is how the other a.out targets
10203 work. */
10204 int align;
10205
10206 align = bfd_get_section_alignment (stdoutput, segment);
10207 size = ((size + (1 << align) - 1) & ((valueT) -1 << align));
10208 }
10209 #endif
10210
10211 return size;
10212 }
10213
10214 /* On the i386, PC-relative offsets are relative to the start of the
10215 next instruction. That is, the address of the offset, plus its
10216 size, since the offset is always the last part of the insn. */
10217
10218 long
10219 md_pcrel_from (fixS *fixP)
10220 {
10221 return fixP->fx_size + fixP->fx_where + fixP->fx_frag->fr_address;
10222 }
10223
10224 #ifndef I386COFF
10225
10226 static void
10227 s_bss (int ignore ATTRIBUTE_UNUSED)
10228 {
10229 int temp;
10230
10231 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
10232 if (IS_ELF)
10233 obj_elf_section_change_hook ();
10234 #endif
10235 temp = get_absolute_expression ();
10236 subseg_set (bss_section, (subsegT) temp);
10237 demand_empty_rest_of_line ();
10238 }
10239
10240 #endif
10241
10242 void
10243 i386_validate_fix (fixS *fixp)
10244 {
10245 if (fixp->fx_subsy && fixp->fx_subsy == GOT_symbol)
10246 {
10247 if (fixp->fx_r_type == BFD_RELOC_32_PCREL)
10248 {
10249 if (!object_64bit)
10250 abort ();
10251 fixp->fx_r_type = BFD_RELOC_X86_64_GOTPCREL;
10252 }
10253 else
10254 {
10255 if (!object_64bit)
10256 fixp->fx_r_type = BFD_RELOC_386_GOTOFF;
10257 else
10258 fixp->fx_r_type = BFD_RELOC_X86_64_GOTOFF64;
10259 }
10260 fixp->fx_subsy = 0;
10261 }
10262 }
10263
10264 arelent *
10265 tc_gen_reloc (asection *section ATTRIBUTE_UNUSED, fixS *fixp)
10266 {
10267 arelent *rel;
10268 bfd_reloc_code_real_type code;
10269
10270 switch (fixp->fx_r_type)
10271 {
10272 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
10273 case BFD_RELOC_SIZE32:
10274 case BFD_RELOC_SIZE64:
10275 if (S_IS_DEFINED (fixp->fx_addsy)
10276 && !S_IS_EXTERNAL (fixp->fx_addsy))
10277 {
10278 /* Resolve size relocation against local symbol to size of
10279 the symbol plus addend. */
10280 valueT value = S_GET_SIZE (fixp->fx_addsy) + fixp->fx_offset;
10281 if (fixp->fx_r_type == BFD_RELOC_SIZE32
10282 && !fits_in_unsigned_long (value))
10283 as_bad_where (fixp->fx_file, fixp->fx_line,
10284 _("symbol size computation overflow"));
10285 fixp->fx_addsy = NULL;
10286 fixp->fx_subsy = NULL;
10287 md_apply_fix (fixp, (valueT *) &value, NULL);
10288 return NULL;
10289 }
10290 #endif
10291
10292 case BFD_RELOC_X86_64_PLT32:
10293 case BFD_RELOC_X86_64_GOT32:
10294 case BFD_RELOC_X86_64_GOTPCREL:
10295 case BFD_RELOC_386_PLT32:
10296 case BFD_RELOC_386_GOT32:
10297 case BFD_RELOC_386_GOTOFF:
10298 case BFD_RELOC_386_GOTPC:
10299 case BFD_RELOC_386_TLS_GD:
10300 case BFD_RELOC_386_TLS_LDM:
10301 case BFD_RELOC_386_TLS_LDO_32:
10302 case BFD_RELOC_386_TLS_IE_32:
10303 case BFD_RELOC_386_TLS_IE:
10304 case BFD_RELOC_386_TLS_GOTIE:
10305 case BFD_RELOC_386_TLS_LE_32:
10306 case BFD_RELOC_386_TLS_LE:
10307 case BFD_RELOC_386_TLS_GOTDESC:
10308 case BFD_RELOC_386_TLS_DESC_CALL:
10309 case BFD_RELOC_X86_64_TLSGD:
10310 case BFD_RELOC_X86_64_TLSLD:
10311 case BFD_RELOC_X86_64_DTPOFF32:
10312 case BFD_RELOC_X86_64_DTPOFF64:
10313 case BFD_RELOC_X86_64_GOTTPOFF:
10314 case BFD_RELOC_X86_64_TPOFF32:
10315 case BFD_RELOC_X86_64_TPOFF64:
10316 case BFD_RELOC_X86_64_GOTOFF64:
10317 case BFD_RELOC_X86_64_GOTPC32:
10318 case BFD_RELOC_X86_64_GOT64:
10319 case BFD_RELOC_X86_64_GOTPCREL64:
10320 case BFD_RELOC_X86_64_GOTPC64:
10321 case BFD_RELOC_X86_64_GOTPLT64:
10322 case BFD_RELOC_X86_64_PLTOFF64:
10323 case BFD_RELOC_X86_64_GOTPC32_TLSDESC:
10324 case BFD_RELOC_X86_64_TLSDESC_CALL:
10325 case BFD_RELOC_RVA:
10326 case BFD_RELOC_VTABLE_ENTRY:
10327 case BFD_RELOC_VTABLE_INHERIT:
10328 #ifdef TE_PE
10329 case BFD_RELOC_32_SECREL:
10330 #endif
10331 code = fixp->fx_r_type;
10332 break;
10333 case BFD_RELOC_X86_64_32S:
10334 if (!fixp->fx_pcrel)
10335 {
10336 /* Don't turn BFD_RELOC_X86_64_32S into BFD_RELOC_32. */
10337 code = fixp->fx_r_type;
10338 break;
10339 }
10340 default:
10341 if (fixp->fx_pcrel)
10342 {
10343 switch (fixp->fx_size)
10344 {
10345 default:
10346 as_bad_where (fixp->fx_file, fixp->fx_line,
10347 _("can not do %d byte pc-relative relocation"),
10348 fixp->fx_size);
10349 code = BFD_RELOC_32_PCREL;
10350 break;
10351 case 1: code = BFD_RELOC_8_PCREL; break;
10352 case 2: code = BFD_RELOC_16_PCREL; break;
10353 case 4: code = BFD_RELOC_32_PCREL; break;
10354 #ifdef BFD64
10355 case 8: code = BFD_RELOC_64_PCREL; break;
10356 #endif
10357 }
10358 }
10359 else
10360 {
10361 switch (fixp->fx_size)
10362 {
10363 default:
10364 as_bad_where (fixp->fx_file, fixp->fx_line,
10365 _("can not do %d byte relocation"),
10366 fixp->fx_size);
10367 code = BFD_RELOC_32;
10368 break;
10369 case 1: code = BFD_RELOC_8; break;
10370 case 2: code = BFD_RELOC_16; break;
10371 case 4: code = BFD_RELOC_32; break;
10372 #ifdef BFD64
10373 case 8: code = BFD_RELOC_64; break;
10374 #endif
10375 }
10376 }
10377 break;
10378 }
10379
10380 if ((code == BFD_RELOC_32
10381 || code == BFD_RELOC_32_PCREL
10382 || code == BFD_RELOC_X86_64_32S)
10383 && GOT_symbol
10384 && fixp->fx_addsy == GOT_symbol)
10385 {
10386 if (!object_64bit)
10387 code = BFD_RELOC_386_GOTPC;
10388 else
10389 code = BFD_RELOC_X86_64_GOTPC32;
10390 }
10391 if ((code == BFD_RELOC_64 || code == BFD_RELOC_64_PCREL)
10392 && GOT_symbol
10393 && fixp->fx_addsy == GOT_symbol)
10394 {
10395 code = BFD_RELOC_X86_64_GOTPC64;
10396 }
10397
10398 rel = (arelent *) xmalloc (sizeof (arelent));
10399 rel->sym_ptr_ptr = (asymbol **) xmalloc (sizeof (asymbol *));
10400 *rel->sym_ptr_ptr = symbol_get_bfdsym (fixp->fx_addsy);
10401
10402 rel->address = fixp->fx_frag->fr_address + fixp->fx_where;
10403
10404 if (!use_rela_relocations)
10405 {
10406 /* HACK: Since i386 ELF uses Rel instead of Rela, encode the
10407 vtable entry to be used in the relocation's section offset. */
10408 if (fixp->fx_r_type == BFD_RELOC_VTABLE_ENTRY)
10409 rel->address = fixp->fx_offset;
10410 #if defined (OBJ_COFF) && defined (TE_PE)
10411 else if (fixp->fx_addsy && S_IS_WEAK (fixp->fx_addsy))
10412 rel->addend = fixp->fx_addnumber - (S_GET_VALUE (fixp->fx_addsy) * 2);
10413 else
10414 #endif
10415 rel->addend = 0;
10416 }
10417 /* Use the rela in 64bit mode. */
10418 else
10419 {
10420 if (disallow_64bit_reloc)
10421 switch (code)
10422 {
10423 case BFD_RELOC_X86_64_DTPOFF64:
10424 case BFD_RELOC_X86_64_TPOFF64:
10425 case BFD_RELOC_64_PCREL:
10426 case BFD_RELOC_X86_64_GOTOFF64:
10427 case BFD_RELOC_X86_64_GOT64:
10428 case BFD_RELOC_X86_64_GOTPCREL64:
10429 case BFD_RELOC_X86_64_GOTPC64:
10430 case BFD_RELOC_X86_64_GOTPLT64:
10431 case BFD_RELOC_X86_64_PLTOFF64:
10432 as_bad_where (fixp->fx_file, fixp->fx_line,
10433 _("cannot represent relocation type %s in x32 mode"),
10434 bfd_get_reloc_code_name (code));
10435 break;
10436 default:
10437 break;
10438 }
10439
10440 if (!fixp->fx_pcrel)
10441 rel->addend = fixp->fx_offset;
10442 else
10443 switch (code)
10444 {
10445 case BFD_RELOC_X86_64_PLT32:
10446 case BFD_RELOC_X86_64_GOT32:
10447 case BFD_RELOC_X86_64_GOTPCREL:
10448 case BFD_RELOC_X86_64_TLSGD:
10449 case BFD_RELOC_X86_64_TLSLD:
10450 case BFD_RELOC_X86_64_GOTTPOFF:
10451 case BFD_RELOC_X86_64_GOTPC32_TLSDESC:
10452 case BFD_RELOC_X86_64_TLSDESC_CALL:
10453 rel->addend = fixp->fx_offset - fixp->fx_size;
10454 break;
10455 default:
10456 rel->addend = (section->vma
10457 - fixp->fx_size
10458 + fixp->fx_addnumber
10459 + md_pcrel_from (fixp));
10460 break;
10461 }
10462 }
10463
10464 rel->howto = bfd_reloc_type_lookup (stdoutput, code);
10465 if (rel->howto == NULL)
10466 {
10467 as_bad_where (fixp->fx_file, fixp->fx_line,
10468 _("cannot represent relocation type %s"),
10469 bfd_get_reloc_code_name (code));
10470 /* Set howto to a garbage value so that we can keep going. */
10471 rel->howto = bfd_reloc_type_lookup (stdoutput, BFD_RELOC_32);
10472 gas_assert (rel->howto != NULL);
10473 }
10474
10475 return rel;
10476 }
10477
10478 #include "tc-i386-intel.c"
10479
10480 void
10481 tc_x86_parse_to_dw2regnum (expressionS *exp)
10482 {
10483 int saved_naked_reg;
10484 char saved_register_dot;
10485
10486 saved_naked_reg = allow_naked_reg;
10487 allow_naked_reg = 1;
10488 saved_register_dot = register_chars['.'];
10489 register_chars['.'] = '.';
10490 allow_pseudo_reg = 1;
10491 expression_and_evaluate (exp);
10492 allow_pseudo_reg = 0;
10493 register_chars['.'] = saved_register_dot;
10494 allow_naked_reg = saved_naked_reg;
10495
10496 if (exp->X_op == O_register && exp->X_add_number >= 0)
10497 {
10498 if ((addressT) exp->X_add_number < i386_regtab_size)
10499 {
10500 exp->X_op = O_constant;
10501 exp->X_add_number = i386_regtab[exp->X_add_number]
10502 .dw2_regnum[flag_code >> 1];
10503 }
10504 else
10505 exp->X_op = O_illegal;
10506 }
10507 }
10508
10509 void
10510 tc_x86_frame_initial_instructions (void)
10511 {
10512 static unsigned int sp_regno[2];
10513
10514 if (!sp_regno[flag_code >> 1])
10515 {
10516 char *saved_input = input_line_pointer;
10517 char sp[][4] = {"esp", "rsp"};
10518 expressionS exp;
10519
10520 input_line_pointer = sp[flag_code >> 1];
10521 tc_x86_parse_to_dw2regnum (&exp);
10522 gas_assert (exp.X_op == O_constant);
10523 sp_regno[flag_code >> 1] = exp.X_add_number;
10524 input_line_pointer = saved_input;
10525 }
10526
10527 cfi_add_CFA_def_cfa (sp_regno[flag_code >> 1], -x86_cie_data_alignment);
10528 cfi_add_CFA_offset (x86_dwarf2_return_column, x86_cie_data_alignment);
10529 }
10530
10531 int
10532 x86_dwarf2_addr_size (void)
10533 {
10534 #if defined (OBJ_MAYBE_ELF) || defined (OBJ_ELF)
10535 if (x86_elf_abi == X86_64_X32_ABI)
10536 return 4;
10537 #endif
10538 return bfd_arch_bits_per_address (stdoutput) / 8;
10539 }
10540
10541 int
10542 i386_elf_section_type (const char *str, size_t len)
10543 {
10544 if (flag_code == CODE_64BIT
10545 && len == sizeof ("unwind") - 1
10546 && strncmp (str, "unwind", 6) == 0)
10547 return SHT_X86_64_UNWIND;
10548
10549 return -1;
10550 }
10551
10552 #ifdef TE_SOLARIS
10553 void
10554 i386_solaris_fix_up_eh_frame (segT sec)
10555 {
10556 if (flag_code == CODE_64BIT)
10557 elf_section_type (sec) = SHT_X86_64_UNWIND;
10558 }
10559 #endif
10560
10561 #ifdef TE_PE
10562 void
10563 tc_pe_dwarf2_emit_offset (symbolS *symbol, unsigned int size)
10564 {
10565 expressionS exp;
10566
10567 exp.X_op = O_secrel;
10568 exp.X_add_symbol = symbol;
10569 exp.X_add_number = 0;
10570 emit_expr (&exp, size);
10571 }
10572 #endif
10573
10574 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
10575 /* For ELF on x86-64, add support for SHF_X86_64_LARGE. */
10576
10577 bfd_vma
10578 x86_64_section_letter (int letter, char **ptr_msg)
10579 {
10580 if (flag_code == CODE_64BIT)
10581 {
10582 if (letter == 'l')
10583 return SHF_X86_64_LARGE;
10584
10585 *ptr_msg = _("bad .section directive: want a,l,w,x,M,S,G,T in string");
10586 }
10587 else
10588 *ptr_msg = _("bad .section directive: want a,w,x,M,S,G,T in string");
10589 return -1;
10590 }
10591
10592 bfd_vma
10593 x86_64_section_word (char *str, size_t len)
10594 {
10595 if (len == 5 && flag_code == CODE_64BIT && CONST_STRNEQ (str, "large"))
10596 return SHF_X86_64_LARGE;
10597
10598 return -1;
10599 }
10600
10601 static void
10602 handle_large_common (int small ATTRIBUTE_UNUSED)
10603 {
10604 if (flag_code != CODE_64BIT)
10605 {
10606 s_comm_internal (0, elf_common_parse);
10607 as_warn (_(".largecomm supported only in 64bit mode, producing .comm"));
10608 }
10609 else
10610 {
10611 static segT lbss_section;
10612 asection *saved_com_section_ptr = elf_com_section_ptr;
10613 asection *saved_bss_section = bss_section;
10614
10615 if (lbss_section == NULL)
10616 {
10617 flagword applicable;
10618 segT seg = now_seg;
10619 subsegT subseg = now_subseg;
10620
10621 /* The .lbss section is for local .largecomm symbols. */
10622 lbss_section = subseg_new (".lbss", 0);
10623 applicable = bfd_applicable_section_flags (stdoutput);
10624 bfd_set_section_flags (stdoutput, lbss_section,
10625 applicable & SEC_ALLOC);
10626 seg_info (lbss_section)->bss = 1;
10627
10628 subseg_set (seg, subseg);
10629 }
10630
10631 elf_com_section_ptr = &_bfd_elf_large_com_section;
10632 bss_section = lbss_section;
10633
10634 s_comm_internal (0, elf_common_parse);
10635
10636 elf_com_section_ptr = saved_com_section_ptr;
10637 bss_section = saved_bss_section;
10638 }
10639 }
10640 #endif /* OBJ_ELF || OBJ_MAYBE_ELF */
This page took 0.235588 seconds and 5 git commands to generate.