| 1 | /* tc-arm.c -- Assemble for the ARM |
| 2 | Copyright (C) 1994-2017 Free Software Foundation, Inc. |
| 3 | Contributed by Richard Earnshaw (rwe@pegasus.esprit.ec.org) |
| 4 | Modified by David Taylor (dtaylor@armltd.co.uk) |
| 5 | Cirrus coprocessor mods by Aldy Hernandez (aldyh@redhat.com) |
| 6 | Cirrus coprocessor fixes by Petko Manolov (petkan@nucleusys.com) |
| 7 | Cirrus coprocessor fixes by Vladimir Ivanov (vladitx@nucleusys.com) |
| 8 | |
| 9 | This file is part of GAS, the GNU Assembler. |
| 10 | |
| 11 | GAS is free software; you can redistribute it and/or modify |
| 12 | it under the terms of the GNU General Public License as published by |
| 13 | the Free Software Foundation; either version 3, or (at your option) |
| 14 | any later version. |
| 15 | |
| 16 | GAS is distributed in the hope that it will be useful, |
| 17 | but WITHOUT ANY WARRANTY; without even the implied warranty of |
| 18 | MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
| 19 | GNU General Public License for more details. |
| 20 | |
| 21 | You should have received a copy of the GNU General Public License |
| 22 | along with GAS; see the file COPYING. If not, write to the Free |
| 23 | Software Foundation, 51 Franklin Street - Fifth Floor, Boston, MA |
| 24 | 02110-1301, USA. */ |
| 25 | |
| 26 | #include "as.h" |
| 27 | #include <limits.h> |
| 28 | #include <stdarg.h> |
| 29 | #define NO_RELOC 0 |
| 30 | #include "safe-ctype.h" |
| 31 | #include "subsegs.h" |
| 32 | #include "obstack.h" |
| 33 | #include "libiberty.h" |
| 34 | #include "opcode/arm.h" |
| 35 | |
| 36 | #ifdef OBJ_ELF |
| 37 | #include "elf/arm.h" |
| 38 | #include "dw2gencfi.h" |
| 39 | #endif |
| 40 | |
| 41 | #include "dwarf2dbg.h" |
| 42 | |
| 43 | #ifdef OBJ_ELF |
| 44 | /* Must be at least the size of the largest unwind opcode (currently two). */ |
| 45 | #define ARM_OPCODE_CHUNK_SIZE 8 |
| 46 | |
| 47 | /* This structure holds the unwinding state. */ |
| 48 | |
| 49 | static struct |
| 50 | { |
| 51 | symbolS * proc_start; |
| 52 | symbolS * table_entry; |
| 53 | symbolS * personality_routine; |
| 54 | int personality_index; |
| 55 | /* The segment containing the function. */ |
| 56 | segT saved_seg; |
| 57 | subsegT saved_subseg; |
| 58 | /* Opcodes generated from this function. */ |
| 59 | unsigned char * opcodes; |
| 60 | int opcode_count; |
| 61 | int opcode_alloc; |
| 62 | /* The number of bytes pushed to the stack. */ |
| 63 | offsetT frame_size; |
| 64 | /* We don't add stack adjustment opcodes immediately so that we can merge |
| 65 | multiple adjustments. We can also omit the final adjustment |
| 66 | when using a frame pointer. */ |
| 67 | offsetT pending_offset; |
| 68 | /* These two fields are set by both unwind_movsp and unwind_setfp. They |
| 69 | hold the reg+offset to use when restoring sp from a frame pointer. */ |
| 70 | offsetT fp_offset; |
| 71 | int fp_reg; |
| 72 | /* Nonzero if an unwind_setfp directive has been seen. */ |
| 73 | unsigned fp_used:1; |
| 74 | /* Nonzero if the last opcode restores sp from fp_reg. */ |
| 75 | unsigned sp_restored:1; |
| 76 | } unwind; |
| 77 | |
| 78 | #endif /* OBJ_ELF */ |
| 79 | |
| 80 | /* Results from operand parsing worker functions. */ |
| 81 | |
| 82 | typedef enum |
| 83 | { |
| 84 | PARSE_OPERAND_SUCCESS, |
| 85 | PARSE_OPERAND_FAIL, |
| 86 | PARSE_OPERAND_FAIL_NO_BACKTRACK |
| 87 | } parse_operand_result; |
| 88 | |
| 89 | enum arm_float_abi |
| 90 | { |
| 91 | ARM_FLOAT_ABI_HARD, |
| 92 | ARM_FLOAT_ABI_SOFTFP, |
| 93 | ARM_FLOAT_ABI_SOFT |
| 94 | }; |
| 95 | |
| 96 | /* Types of processor to assemble for. */ |
| 97 | #ifndef CPU_DEFAULT |
| 98 | /* The code that was here used to select a default CPU depending on compiler |
| 99 | pre-defines which were only present when doing native builds, thus |
| 100 | changing gas' default behaviour depending upon the build host. |
| 101 | |
| 102 | If you have a target that requires a default CPU option then the you |
| 103 | should define CPU_DEFAULT here. */ |
| 104 | #endif |
| 105 | |
| 106 | #ifndef FPU_DEFAULT |
| 107 | # ifdef TE_LINUX |
| 108 | # define FPU_DEFAULT FPU_ARCH_FPA |
| 109 | # elif defined (TE_NetBSD) |
| 110 | # ifdef OBJ_ELF |
| 111 | # define FPU_DEFAULT FPU_ARCH_VFP /* Soft-float, but VFP order. */ |
| 112 | # else |
| 113 | /* Legacy a.out format. */ |
| 114 | # define FPU_DEFAULT FPU_ARCH_FPA /* Soft-float, but FPA order. */ |
| 115 | # endif |
| 116 | # elif defined (TE_VXWORKS) |
| 117 | # define FPU_DEFAULT FPU_ARCH_VFP /* Soft-float, VFP order. */ |
| 118 | # else |
| 119 | /* For backwards compatibility, default to FPA. */ |
| 120 | # define FPU_DEFAULT FPU_ARCH_FPA |
| 121 | # endif |
| 122 | #endif /* ifndef FPU_DEFAULT */ |
| 123 | |
| 124 | #define streq(a, b) (strcmp (a, b) == 0) |
| 125 | |
| 126 | static arm_feature_set cpu_variant; |
| 127 | static arm_feature_set arm_arch_used; |
| 128 | static arm_feature_set thumb_arch_used; |
| 129 | |
| 130 | /* Flags stored in private area of BFD structure. */ |
| 131 | static int uses_apcs_26 = FALSE; |
| 132 | static int atpcs = FALSE; |
| 133 | static int support_interwork = FALSE; |
| 134 | static int uses_apcs_float = FALSE; |
| 135 | static int pic_code = FALSE; |
| 136 | static int fix_v4bx = FALSE; |
| 137 | /* Warn on using deprecated features. */ |
| 138 | static int warn_on_deprecated = TRUE; |
| 139 | |
| 140 | /* Understand CodeComposer Studio assembly syntax. */ |
| 141 | bfd_boolean codecomposer_syntax = FALSE; |
| 142 | |
| 143 | /* Variables that we set while parsing command-line options. Once all |
| 144 | options have been read we re-process these values to set the real |
| 145 | assembly flags. */ |
| 146 | static const arm_feature_set *legacy_cpu = NULL; |
| 147 | static const arm_feature_set *legacy_fpu = NULL; |
| 148 | |
| 149 | static const arm_feature_set *mcpu_cpu_opt = NULL; |
| 150 | static arm_feature_set *dyn_mcpu_ext_opt = NULL; |
| 151 | static const arm_feature_set *mcpu_fpu_opt = NULL; |
| 152 | static const arm_feature_set *march_cpu_opt = NULL; |
| 153 | static arm_feature_set *dyn_march_ext_opt = NULL; |
| 154 | static const arm_feature_set *march_fpu_opt = NULL; |
| 155 | static const arm_feature_set *mfpu_opt = NULL; |
| 156 | static const arm_feature_set *object_arch = NULL; |
| 157 | |
| 158 | /* Constants for known architecture features. */ |
| 159 | static const arm_feature_set fpu_default = FPU_DEFAULT; |
| 160 | static const arm_feature_set fpu_arch_vfp_v1 ATTRIBUTE_UNUSED = FPU_ARCH_VFP_V1; |
| 161 | static const arm_feature_set fpu_arch_vfp_v2 = FPU_ARCH_VFP_V2; |
| 162 | static const arm_feature_set fpu_arch_vfp_v3 ATTRIBUTE_UNUSED = FPU_ARCH_VFP_V3; |
| 163 | static const arm_feature_set fpu_arch_neon_v1 ATTRIBUTE_UNUSED = FPU_ARCH_NEON_V1; |
| 164 | static const arm_feature_set fpu_arch_fpa = FPU_ARCH_FPA; |
| 165 | static const arm_feature_set fpu_any_hard = FPU_ANY_HARD; |
| 166 | #ifdef OBJ_ELF |
| 167 | static const arm_feature_set fpu_arch_maverick = FPU_ARCH_MAVERICK; |
| 168 | #endif |
| 169 | static const arm_feature_set fpu_endian_pure = FPU_ARCH_ENDIAN_PURE; |
| 170 | |
| 171 | #ifdef CPU_DEFAULT |
| 172 | static const arm_feature_set cpu_default = CPU_DEFAULT; |
| 173 | #endif |
| 174 | |
| 175 | static const arm_feature_set arm_ext_v1 = ARM_FEATURE_CORE_LOW (ARM_EXT_V1); |
| 176 | static const arm_feature_set arm_ext_v2 = ARM_FEATURE_CORE_LOW (ARM_EXT_V2); |
| 177 | static const arm_feature_set arm_ext_v2s = ARM_FEATURE_CORE_LOW (ARM_EXT_V2S); |
| 178 | static const arm_feature_set arm_ext_v3 = ARM_FEATURE_CORE_LOW (ARM_EXT_V3); |
| 179 | static const arm_feature_set arm_ext_v3m = ARM_FEATURE_CORE_LOW (ARM_EXT_V3M); |
| 180 | static const arm_feature_set arm_ext_v4 = ARM_FEATURE_CORE_LOW (ARM_EXT_V4); |
| 181 | static const arm_feature_set arm_ext_v4t = ARM_FEATURE_CORE_LOW (ARM_EXT_V4T); |
| 182 | static const arm_feature_set arm_ext_v5 = ARM_FEATURE_CORE_LOW (ARM_EXT_V5); |
| 183 | static const arm_feature_set arm_ext_v4t_5 = |
| 184 | ARM_FEATURE_CORE_LOW (ARM_EXT_V4T | ARM_EXT_V5); |
| 185 | static const arm_feature_set arm_ext_v5t = ARM_FEATURE_CORE_LOW (ARM_EXT_V5T); |
| 186 | static const arm_feature_set arm_ext_v5e = ARM_FEATURE_CORE_LOW (ARM_EXT_V5E); |
| 187 | static const arm_feature_set arm_ext_v5exp = ARM_FEATURE_CORE_LOW (ARM_EXT_V5ExP); |
| 188 | static const arm_feature_set arm_ext_v5j = ARM_FEATURE_CORE_LOW (ARM_EXT_V5J); |
| 189 | static const arm_feature_set arm_ext_v6 = ARM_FEATURE_CORE_LOW (ARM_EXT_V6); |
| 190 | static const arm_feature_set arm_ext_v6k = ARM_FEATURE_CORE_LOW (ARM_EXT_V6K); |
| 191 | static const arm_feature_set arm_ext_v6t2 = ARM_FEATURE_CORE_LOW (ARM_EXT_V6T2); |
| 192 | static const arm_feature_set arm_ext_v6_notm = |
| 193 | ARM_FEATURE_CORE_LOW (ARM_EXT_V6_NOTM); |
| 194 | static const arm_feature_set arm_ext_v6_dsp = |
| 195 | ARM_FEATURE_CORE_LOW (ARM_EXT_V6_DSP); |
| 196 | static const arm_feature_set arm_ext_barrier = |
| 197 | ARM_FEATURE_CORE_LOW (ARM_EXT_BARRIER); |
| 198 | static const arm_feature_set arm_ext_msr = |
| 199 | ARM_FEATURE_CORE_LOW (ARM_EXT_THUMB_MSR); |
| 200 | static const arm_feature_set arm_ext_div = ARM_FEATURE_CORE_LOW (ARM_EXT_DIV); |
| 201 | static const arm_feature_set arm_ext_v7 = ARM_FEATURE_CORE_LOW (ARM_EXT_V7); |
| 202 | static const arm_feature_set arm_ext_v7a = ARM_FEATURE_CORE_LOW (ARM_EXT_V7A); |
| 203 | static const arm_feature_set arm_ext_v7r = ARM_FEATURE_CORE_LOW (ARM_EXT_V7R); |
| 204 | #ifdef OBJ_ELF |
| 205 | static const arm_feature_set ATTRIBUTE_UNUSED arm_ext_v7m = ARM_FEATURE_CORE_LOW (ARM_EXT_V7M); |
| 206 | #endif |
| 207 | static const arm_feature_set arm_ext_v8 = ARM_FEATURE_CORE_LOW (ARM_EXT_V8); |
| 208 | static const arm_feature_set arm_ext_m = |
| 209 | ARM_FEATURE_CORE (ARM_EXT_V6M | ARM_EXT_V7M, |
| 210 | ARM_EXT2_V8M | ARM_EXT2_V8M_MAIN); |
| 211 | static const arm_feature_set arm_ext_mp = ARM_FEATURE_CORE_LOW (ARM_EXT_MP); |
| 212 | static const arm_feature_set arm_ext_sec = ARM_FEATURE_CORE_LOW (ARM_EXT_SEC); |
| 213 | static const arm_feature_set arm_ext_os = ARM_FEATURE_CORE_LOW (ARM_EXT_OS); |
| 214 | static const arm_feature_set arm_ext_adiv = ARM_FEATURE_CORE_LOW (ARM_EXT_ADIV); |
| 215 | static const arm_feature_set arm_ext_virt = ARM_FEATURE_CORE_LOW (ARM_EXT_VIRT); |
| 216 | static const arm_feature_set arm_ext_pan = ARM_FEATURE_CORE_HIGH (ARM_EXT2_PAN); |
| 217 | static const arm_feature_set arm_ext_v8m = ARM_FEATURE_CORE_HIGH (ARM_EXT2_V8M); |
| 218 | static const arm_feature_set arm_ext_v8m_main = |
| 219 | ARM_FEATURE_CORE_HIGH (ARM_EXT2_V8M_MAIN); |
| 220 | /* Instructions in ARMv8-M only found in M profile architectures. */ |
| 221 | static const arm_feature_set arm_ext_v8m_m_only = |
| 222 | ARM_FEATURE_CORE_HIGH (ARM_EXT2_V8M | ARM_EXT2_V8M_MAIN); |
| 223 | static const arm_feature_set arm_ext_v6t2_v8m = |
| 224 | ARM_FEATURE_CORE_HIGH (ARM_EXT2_V6T2_V8M); |
| 225 | /* Instructions shared between ARMv8-A and ARMv8-M. */ |
| 226 | static const arm_feature_set arm_ext_atomics = |
| 227 | ARM_FEATURE_CORE_HIGH (ARM_EXT2_ATOMICS); |
| 228 | #ifdef OBJ_ELF |
| 229 | /* DSP instructions Tag_DSP_extension refers to. */ |
| 230 | static const arm_feature_set arm_ext_dsp = |
| 231 | ARM_FEATURE_CORE_LOW (ARM_EXT_V5E | ARM_EXT_V5ExP | ARM_EXT_V6_DSP); |
| 232 | #endif |
| 233 | static const arm_feature_set arm_ext_ras = |
| 234 | ARM_FEATURE_CORE_HIGH (ARM_EXT2_RAS); |
| 235 | /* FP16 instructions. */ |
| 236 | static const arm_feature_set arm_ext_fp16 = |
| 237 | ARM_FEATURE_CORE_HIGH (ARM_EXT2_FP16_INST); |
| 238 | static const arm_feature_set arm_ext_v8_3 = |
| 239 | ARM_FEATURE_CORE_HIGH (ARM_EXT2_V8_3A); |
| 240 | |
| 241 | static const arm_feature_set arm_arch_any = ARM_ANY; |
| 242 | #ifdef OBJ_ELF |
| 243 | static const arm_feature_set fpu_any = FPU_ANY; |
| 244 | #endif |
| 245 | static const arm_feature_set arm_arch_full ATTRIBUTE_UNUSED = ARM_FEATURE (-1, -1, -1); |
| 246 | static const arm_feature_set arm_arch_t2 = ARM_ARCH_THUMB2; |
| 247 | static const arm_feature_set arm_arch_none = ARM_ARCH_NONE; |
| 248 | |
| 249 | static const arm_feature_set arm_cext_iwmmxt2 = |
| 250 | ARM_FEATURE_COPROC (ARM_CEXT_IWMMXT2); |
| 251 | static const arm_feature_set arm_cext_iwmmxt = |
| 252 | ARM_FEATURE_COPROC (ARM_CEXT_IWMMXT); |
| 253 | static const arm_feature_set arm_cext_xscale = |
| 254 | ARM_FEATURE_COPROC (ARM_CEXT_XSCALE); |
| 255 | static const arm_feature_set arm_cext_maverick = |
| 256 | ARM_FEATURE_COPROC (ARM_CEXT_MAVERICK); |
| 257 | static const arm_feature_set fpu_fpa_ext_v1 = |
| 258 | ARM_FEATURE_COPROC (FPU_FPA_EXT_V1); |
| 259 | static const arm_feature_set fpu_fpa_ext_v2 = |
| 260 | ARM_FEATURE_COPROC (FPU_FPA_EXT_V2); |
| 261 | static const arm_feature_set fpu_vfp_ext_v1xd = |
| 262 | ARM_FEATURE_COPROC (FPU_VFP_EXT_V1xD); |
| 263 | static const arm_feature_set fpu_vfp_ext_v1 = |
| 264 | ARM_FEATURE_COPROC (FPU_VFP_EXT_V1); |
| 265 | static const arm_feature_set fpu_vfp_ext_v2 = |
| 266 | ARM_FEATURE_COPROC (FPU_VFP_EXT_V2); |
| 267 | static const arm_feature_set fpu_vfp_ext_v3xd = |
| 268 | ARM_FEATURE_COPROC (FPU_VFP_EXT_V3xD); |
| 269 | static const arm_feature_set fpu_vfp_ext_v3 = |
| 270 | ARM_FEATURE_COPROC (FPU_VFP_EXT_V3); |
| 271 | static const arm_feature_set fpu_vfp_ext_d32 = |
| 272 | ARM_FEATURE_COPROC (FPU_VFP_EXT_D32); |
| 273 | static const arm_feature_set fpu_neon_ext_v1 = |
| 274 | ARM_FEATURE_COPROC (FPU_NEON_EXT_V1); |
| 275 | static const arm_feature_set fpu_vfp_v3_or_neon_ext = |
| 276 | ARM_FEATURE_COPROC (FPU_NEON_EXT_V1 | FPU_VFP_EXT_V3); |
| 277 | #ifdef OBJ_ELF |
| 278 | static const arm_feature_set fpu_vfp_fp16 = |
| 279 | ARM_FEATURE_COPROC (FPU_VFP_EXT_FP16); |
| 280 | static const arm_feature_set fpu_neon_ext_fma = |
| 281 | ARM_FEATURE_COPROC (FPU_NEON_EXT_FMA); |
| 282 | #endif |
| 283 | static const arm_feature_set fpu_vfp_ext_fma = |
| 284 | ARM_FEATURE_COPROC (FPU_VFP_EXT_FMA); |
| 285 | static const arm_feature_set fpu_vfp_ext_armv8 = |
| 286 | ARM_FEATURE_COPROC (FPU_VFP_EXT_ARMV8); |
| 287 | static const arm_feature_set fpu_vfp_ext_armv8xd = |
| 288 | ARM_FEATURE_COPROC (FPU_VFP_EXT_ARMV8xD); |
| 289 | static const arm_feature_set fpu_neon_ext_armv8 = |
| 290 | ARM_FEATURE_COPROC (FPU_NEON_EXT_ARMV8); |
| 291 | static const arm_feature_set fpu_crypto_ext_armv8 = |
| 292 | ARM_FEATURE_COPROC (FPU_CRYPTO_EXT_ARMV8); |
| 293 | static const arm_feature_set crc_ext_armv8 = |
| 294 | ARM_FEATURE_COPROC (CRC_EXT_ARMV8); |
| 295 | static const arm_feature_set fpu_neon_ext_v8_1 = |
| 296 | ARM_FEATURE_COPROC (FPU_NEON_EXT_RDMA); |
| 297 | static const arm_feature_set fpu_neon_ext_dotprod = |
| 298 | ARM_FEATURE_COPROC (FPU_NEON_EXT_DOTPROD); |
| 299 | |
| 300 | static int mfloat_abi_opt = -1; |
| 301 | /* Record user cpu selection for object attributes. */ |
| 302 | static arm_feature_set selected_cpu = ARM_ARCH_NONE; |
| 303 | /* Must be long enough to hold any of the names in arm_cpus. */ |
| 304 | static char selected_cpu_name[20]; |
| 305 | |
| 306 | extern FLONUM_TYPE generic_floating_point_number; |
| 307 | |
| 308 | /* Return if no cpu was selected on command-line. */ |
| 309 | static bfd_boolean |
| 310 | no_cpu_selected (void) |
| 311 | { |
| 312 | return ARM_FEATURE_EQUAL (selected_cpu, arm_arch_none); |
| 313 | } |
| 314 | |
| 315 | #ifdef OBJ_ELF |
| 316 | # ifdef EABI_DEFAULT |
| 317 | static int meabi_flags = EABI_DEFAULT; |
| 318 | # else |
| 319 | static int meabi_flags = EF_ARM_EABI_UNKNOWN; |
| 320 | # endif |
| 321 | |
| 322 | static int attributes_set_explicitly[NUM_KNOWN_OBJ_ATTRIBUTES]; |
| 323 | |
| 324 | bfd_boolean |
| 325 | arm_is_eabi (void) |
| 326 | { |
| 327 | return (EF_ARM_EABI_VERSION (meabi_flags) >= EF_ARM_EABI_VER4); |
| 328 | } |
| 329 | #endif |
| 330 | |
| 331 | #ifdef OBJ_ELF |
| 332 | /* Pre-defined "_GLOBAL_OFFSET_TABLE_" */ |
| 333 | symbolS * GOT_symbol; |
| 334 | #endif |
| 335 | |
| 336 | /* 0: assemble for ARM, |
| 337 | 1: assemble for Thumb, |
| 338 | 2: assemble for Thumb even though target CPU does not support thumb |
| 339 | instructions. */ |
| 340 | static int thumb_mode = 0; |
| 341 | /* A value distinct from the possible values for thumb_mode that we |
| 342 | can use to record whether thumb_mode has been copied into the |
| 343 | tc_frag_data field of a frag. */ |
| 344 | #define MODE_RECORDED (1 << 4) |
| 345 | |
| 346 | /* Specifies the intrinsic IT insn behavior mode. */ |
| 347 | enum implicit_it_mode |
| 348 | { |
| 349 | IMPLICIT_IT_MODE_NEVER = 0x00, |
| 350 | IMPLICIT_IT_MODE_ARM = 0x01, |
| 351 | IMPLICIT_IT_MODE_THUMB = 0x02, |
| 352 | IMPLICIT_IT_MODE_ALWAYS = (IMPLICIT_IT_MODE_ARM | IMPLICIT_IT_MODE_THUMB) |
| 353 | }; |
| 354 | static int implicit_it_mode = IMPLICIT_IT_MODE_ARM; |
| 355 | |
| 356 | /* If unified_syntax is true, we are processing the new unified |
| 357 | ARM/Thumb syntax. Important differences from the old ARM mode: |
| 358 | |
| 359 | - Immediate operands do not require a # prefix. |
| 360 | - Conditional affixes always appear at the end of the |
| 361 | instruction. (For backward compatibility, those instructions |
| 362 | that formerly had them in the middle, continue to accept them |
| 363 | there.) |
| 364 | - The IT instruction may appear, and if it does is validated |
| 365 | against subsequent conditional affixes. It does not generate |
| 366 | machine code. |
| 367 | |
| 368 | Important differences from the old Thumb mode: |
| 369 | |
| 370 | - Immediate operands do not require a # prefix. |
| 371 | - Most of the V6T2 instructions are only available in unified mode. |
| 372 | - The .N and .W suffixes are recognized and honored (it is an error |
| 373 | if they cannot be honored). |
| 374 | - All instructions set the flags if and only if they have an 's' affix. |
| 375 | - Conditional affixes may be used. They are validated against |
| 376 | preceding IT instructions. Unlike ARM mode, you cannot use a |
| 377 | conditional affix except in the scope of an IT instruction. */ |
| 378 | |
| 379 | static bfd_boolean unified_syntax = FALSE; |
| 380 | |
| 381 | /* An immediate operand can start with #, and ld*, st*, pld operands |
| 382 | can contain [ and ]. We need to tell APP not to elide whitespace |
| 383 | before a [, which can appear as the first operand for pld. |
| 384 | Likewise, a { can appear as the first operand for push, pop, vld*, etc. */ |
| 385 | const char arm_symbol_chars[] = "#[]{}"; |
| 386 | |
| 387 | enum neon_el_type |
| 388 | { |
| 389 | NT_invtype, |
| 390 | NT_untyped, |
| 391 | NT_integer, |
| 392 | NT_float, |
| 393 | NT_poly, |
| 394 | NT_signed, |
| 395 | NT_unsigned |
| 396 | }; |
| 397 | |
| 398 | struct neon_type_el |
| 399 | { |
| 400 | enum neon_el_type type; |
| 401 | unsigned size; |
| 402 | }; |
| 403 | |
| 404 | #define NEON_MAX_TYPE_ELS 4 |
| 405 | |
| 406 | struct neon_type |
| 407 | { |
| 408 | struct neon_type_el el[NEON_MAX_TYPE_ELS]; |
| 409 | unsigned elems; |
| 410 | }; |
| 411 | |
| 412 | enum it_instruction_type |
| 413 | { |
| 414 | OUTSIDE_IT_INSN, |
| 415 | INSIDE_IT_INSN, |
| 416 | INSIDE_IT_LAST_INSN, |
| 417 | IF_INSIDE_IT_LAST_INSN, /* Either outside or inside; |
| 418 | if inside, should be the last one. */ |
| 419 | NEUTRAL_IT_INSN, /* This could be either inside or outside, |
| 420 | i.e. BKPT and NOP. */ |
| 421 | IT_INSN /* The IT insn has been parsed. */ |
| 422 | }; |
| 423 | |
| 424 | /* The maximum number of operands we need. */ |
| 425 | #define ARM_IT_MAX_OPERANDS 6 |
| 426 | |
| 427 | struct arm_it |
| 428 | { |
| 429 | const char * error; |
| 430 | unsigned long instruction; |
| 431 | int size; |
| 432 | int size_req; |
| 433 | int cond; |
| 434 | /* "uncond_value" is set to the value in place of the conditional field in |
| 435 | unconditional versions of the instruction, or -1 if nothing is |
| 436 | appropriate. */ |
| 437 | int uncond_value; |
| 438 | struct neon_type vectype; |
| 439 | /* This does not indicate an actual NEON instruction, only that |
| 440 | the mnemonic accepts neon-style type suffixes. */ |
| 441 | int is_neon; |
| 442 | /* Set to the opcode if the instruction needs relaxation. |
| 443 | Zero if the instruction is not relaxed. */ |
| 444 | unsigned long relax; |
| 445 | struct |
| 446 | { |
| 447 | bfd_reloc_code_real_type type; |
| 448 | expressionS exp; |
| 449 | int pc_rel; |
| 450 | } reloc; |
| 451 | |
| 452 | enum it_instruction_type it_insn_type; |
| 453 | |
| 454 | struct |
| 455 | { |
| 456 | unsigned reg; |
| 457 | signed int imm; |
| 458 | struct neon_type_el vectype; |
| 459 | unsigned present : 1; /* Operand present. */ |
| 460 | unsigned isreg : 1; /* Operand was a register. */ |
| 461 | unsigned immisreg : 1; /* .imm field is a second register. */ |
| 462 | unsigned isscalar : 1; /* Operand is a (Neon) scalar. */ |
| 463 | unsigned immisalign : 1; /* Immediate is an alignment specifier. */ |
| 464 | unsigned immisfloat : 1; /* Immediate was parsed as a float. */ |
| 465 | /* Note: we abuse "regisimm" to mean "is Neon register" in VMOV |
| 466 | instructions. This allows us to disambiguate ARM <-> vector insns. */ |
| 467 | unsigned regisimm : 1; /* 64-bit immediate, reg forms high 32 bits. */ |
| 468 | unsigned isvec : 1; /* Is a single, double or quad VFP/Neon reg. */ |
| 469 | unsigned isquad : 1; /* Operand is Neon quad-precision register. */ |
| 470 | unsigned issingle : 1; /* Operand is VFP single-precision register. */ |
| 471 | unsigned hasreloc : 1; /* Operand has relocation suffix. */ |
| 472 | unsigned writeback : 1; /* Operand has trailing ! */ |
| 473 | unsigned preind : 1; /* Preindexed address. */ |
| 474 | unsigned postind : 1; /* Postindexed address. */ |
| 475 | unsigned negative : 1; /* Index register was negated. */ |
| 476 | unsigned shifted : 1; /* Shift applied to operation. */ |
| 477 | unsigned shift_kind : 3; /* Shift operation (enum shift_kind). */ |
| 478 | } operands[ARM_IT_MAX_OPERANDS]; |
| 479 | }; |
| 480 | |
| 481 | static struct arm_it inst; |
| 482 | |
| 483 | #define NUM_FLOAT_VALS 8 |
| 484 | |
| 485 | const char * fp_const[] = |
| 486 | { |
| 487 | "0.0", "1.0", "2.0", "3.0", "4.0", "5.0", "0.5", "10.0", 0 |
| 488 | }; |
| 489 | |
| 490 | /* Number of littlenums required to hold an extended precision number. */ |
| 491 | #define MAX_LITTLENUMS 6 |
| 492 | |
| 493 | LITTLENUM_TYPE fp_values[NUM_FLOAT_VALS][MAX_LITTLENUMS]; |
| 494 | |
| 495 | #define FAIL (-1) |
| 496 | #define SUCCESS (0) |
| 497 | |
| 498 | #define SUFF_S 1 |
| 499 | #define SUFF_D 2 |
| 500 | #define SUFF_E 3 |
| 501 | #define SUFF_P 4 |
| 502 | |
| 503 | #define CP_T_X 0x00008000 |
| 504 | #define CP_T_Y 0x00400000 |
| 505 | |
| 506 | #define CONDS_BIT 0x00100000 |
| 507 | #define LOAD_BIT 0x00100000 |
| 508 | |
| 509 | #define DOUBLE_LOAD_FLAG 0x00000001 |
| 510 | |
| 511 | struct asm_cond |
| 512 | { |
| 513 | const char * template_name; |
| 514 | unsigned long value; |
| 515 | }; |
| 516 | |
| 517 | #define COND_ALWAYS 0xE |
| 518 | |
| 519 | struct asm_psr |
| 520 | { |
| 521 | const char * template_name; |
| 522 | unsigned long field; |
| 523 | }; |
| 524 | |
| 525 | struct asm_barrier_opt |
| 526 | { |
| 527 | const char * template_name; |
| 528 | unsigned long value; |
| 529 | const arm_feature_set arch; |
| 530 | }; |
| 531 | |
| 532 | /* The bit that distinguishes CPSR and SPSR. */ |
| 533 | #define SPSR_BIT (1 << 22) |
| 534 | |
| 535 | /* The individual PSR flag bits. */ |
| 536 | #define PSR_c (1 << 16) |
| 537 | #define PSR_x (1 << 17) |
| 538 | #define PSR_s (1 << 18) |
| 539 | #define PSR_f (1 << 19) |
| 540 | |
| 541 | struct reloc_entry |
| 542 | { |
| 543 | const char * name; |
| 544 | bfd_reloc_code_real_type reloc; |
| 545 | }; |
| 546 | |
| 547 | enum vfp_reg_pos |
| 548 | { |
| 549 | VFP_REG_Sd, VFP_REG_Sm, VFP_REG_Sn, |
| 550 | VFP_REG_Dd, VFP_REG_Dm, VFP_REG_Dn |
| 551 | }; |
| 552 | |
| 553 | enum vfp_ldstm_type |
| 554 | { |
| 555 | VFP_LDSTMIA, VFP_LDSTMDB, VFP_LDSTMIAX, VFP_LDSTMDBX |
| 556 | }; |
| 557 | |
| 558 | /* Bits for DEFINED field in neon_typed_alias. */ |
| 559 | #define NTA_HASTYPE 1 |
| 560 | #define NTA_HASINDEX 2 |
| 561 | |
| 562 | struct neon_typed_alias |
| 563 | { |
| 564 | unsigned char defined; |
| 565 | unsigned char index; |
| 566 | struct neon_type_el eltype; |
| 567 | }; |
| 568 | |
| 569 | /* ARM register categories. This includes coprocessor numbers and various |
| 570 | architecture extensions' registers. */ |
| 571 | enum arm_reg_type |
| 572 | { |
| 573 | REG_TYPE_RN, |
| 574 | REG_TYPE_CP, |
| 575 | REG_TYPE_CN, |
| 576 | REG_TYPE_FN, |
| 577 | REG_TYPE_VFS, |
| 578 | REG_TYPE_VFD, |
| 579 | REG_TYPE_NQ, |
| 580 | REG_TYPE_VFSD, |
| 581 | REG_TYPE_NDQ, |
| 582 | REG_TYPE_NSDQ, |
| 583 | REG_TYPE_VFC, |
| 584 | REG_TYPE_MVF, |
| 585 | REG_TYPE_MVD, |
| 586 | REG_TYPE_MVFX, |
| 587 | REG_TYPE_MVDX, |
| 588 | REG_TYPE_MVAX, |
| 589 | REG_TYPE_DSPSC, |
| 590 | REG_TYPE_MMXWR, |
| 591 | REG_TYPE_MMXWC, |
| 592 | REG_TYPE_MMXWCG, |
| 593 | REG_TYPE_XSCALE, |
| 594 | REG_TYPE_RNB |
| 595 | }; |
| 596 | |
| 597 | /* Structure for a hash table entry for a register. |
| 598 | If TYPE is REG_TYPE_VFD or REG_TYPE_NQ, the NEON field can point to extra |
| 599 | information which states whether a vector type or index is specified (for a |
| 600 | register alias created with .dn or .qn). Otherwise NEON should be NULL. */ |
| 601 | struct reg_entry |
| 602 | { |
| 603 | const char * name; |
| 604 | unsigned int number; |
| 605 | unsigned char type; |
| 606 | unsigned char builtin; |
| 607 | struct neon_typed_alias * neon; |
| 608 | }; |
| 609 | |
| 610 | /* Diagnostics used when we don't get a register of the expected type. */ |
| 611 | const char * const reg_expected_msgs[] = |
| 612 | { |
| 613 | N_("ARM register expected"), |
| 614 | N_("bad or missing co-processor number"), |
| 615 | N_("co-processor register expected"), |
| 616 | N_("FPA register expected"), |
| 617 | N_("VFP single precision register expected"), |
| 618 | N_("VFP/Neon double precision register expected"), |
| 619 | N_("Neon quad precision register expected"), |
| 620 | N_("VFP single or double precision register expected"), |
| 621 | N_("Neon double or quad precision register expected"), |
| 622 | N_("VFP single, double or Neon quad precision register expected"), |
| 623 | N_("VFP system register expected"), |
| 624 | N_("Maverick MVF register expected"), |
| 625 | N_("Maverick MVD register expected"), |
| 626 | N_("Maverick MVFX register expected"), |
| 627 | N_("Maverick MVDX register expected"), |
| 628 | N_("Maverick MVAX register expected"), |
| 629 | N_("Maverick DSPSC register expected"), |
| 630 | N_("iWMMXt data register expected"), |
| 631 | N_("iWMMXt control register expected"), |
| 632 | N_("iWMMXt scalar register expected"), |
| 633 | N_("XScale accumulator register expected"), |
| 634 | }; |
| 635 | |
| 636 | /* Some well known registers that we refer to directly elsewhere. */ |
| 637 | #define REG_R12 12 |
| 638 | #define REG_SP 13 |
| 639 | #define REG_LR 14 |
| 640 | #define REG_PC 15 |
| 641 | |
| 642 | /* ARM instructions take 4bytes in the object file, Thumb instructions |
| 643 | take 2: */ |
| 644 | #define INSN_SIZE 4 |
| 645 | |
| 646 | struct asm_opcode |
| 647 | { |
| 648 | /* Basic string to match. */ |
| 649 | const char * template_name; |
| 650 | |
| 651 | /* Parameters to instruction. */ |
| 652 | unsigned int operands[8]; |
| 653 | |
| 654 | /* Conditional tag - see opcode_lookup. */ |
| 655 | unsigned int tag : 4; |
| 656 | |
| 657 | /* Basic instruction code. */ |
| 658 | unsigned int avalue : 28; |
| 659 | |
| 660 | /* Thumb-format instruction code. */ |
| 661 | unsigned int tvalue; |
| 662 | |
| 663 | /* Which architecture variant provides this instruction. */ |
| 664 | const arm_feature_set * avariant; |
| 665 | const arm_feature_set * tvariant; |
| 666 | |
| 667 | /* Function to call to encode instruction in ARM format. */ |
| 668 | void (* aencode) (void); |
| 669 | |
| 670 | /* Function to call to encode instruction in Thumb format. */ |
| 671 | void (* tencode) (void); |
| 672 | }; |
| 673 | |
| 674 | /* Defines for various bits that we will want to toggle. */ |
| 675 | #define INST_IMMEDIATE 0x02000000 |
| 676 | #define OFFSET_REG 0x02000000 |
| 677 | #define HWOFFSET_IMM 0x00400000 |
| 678 | #define SHIFT_BY_REG 0x00000010 |
| 679 | #define PRE_INDEX 0x01000000 |
| 680 | #define INDEX_UP 0x00800000 |
| 681 | #define WRITE_BACK 0x00200000 |
| 682 | #define LDM_TYPE_2_OR_3 0x00400000 |
| 683 | #define CPSI_MMOD 0x00020000 |
| 684 | |
| 685 | #define LITERAL_MASK 0xf000f000 |
| 686 | #define OPCODE_MASK 0xfe1fffff |
| 687 | #define V4_STR_BIT 0x00000020 |
| 688 | #define VLDR_VMOV_SAME 0x0040f000 |
| 689 | |
| 690 | #define T2_SUBS_PC_LR 0xf3de8f00 |
| 691 | |
| 692 | #define DATA_OP_SHIFT 21 |
| 693 | #define SBIT_SHIFT 20 |
| 694 | |
| 695 | #define T2_OPCODE_MASK 0xfe1fffff |
| 696 | #define T2_DATA_OP_SHIFT 21 |
| 697 | #define T2_SBIT_SHIFT 20 |
| 698 | |
| 699 | #define A_COND_MASK 0xf0000000 |
| 700 | #define A_PUSH_POP_OP_MASK 0x0fff0000 |
| 701 | |
| 702 | /* Opcodes for pushing/poping registers to/from the stack. */ |
| 703 | #define A1_OPCODE_PUSH 0x092d0000 |
| 704 | #define A2_OPCODE_PUSH 0x052d0004 |
| 705 | #define A2_OPCODE_POP 0x049d0004 |
| 706 | |
| 707 | /* Codes to distinguish the arithmetic instructions. */ |
| 708 | #define OPCODE_AND 0 |
| 709 | #define OPCODE_EOR 1 |
| 710 | #define OPCODE_SUB 2 |
| 711 | #define OPCODE_RSB 3 |
| 712 | #define OPCODE_ADD 4 |
| 713 | #define OPCODE_ADC 5 |
| 714 | #define OPCODE_SBC 6 |
| 715 | #define OPCODE_RSC 7 |
| 716 | #define OPCODE_TST 8 |
| 717 | #define OPCODE_TEQ 9 |
| 718 | #define OPCODE_CMP 10 |
| 719 | #define OPCODE_CMN 11 |
| 720 | #define OPCODE_ORR 12 |
| 721 | #define OPCODE_MOV 13 |
| 722 | #define OPCODE_BIC 14 |
| 723 | #define OPCODE_MVN 15 |
| 724 | |
| 725 | #define T2_OPCODE_AND 0 |
| 726 | #define T2_OPCODE_BIC 1 |
| 727 | #define T2_OPCODE_ORR 2 |
| 728 | #define T2_OPCODE_ORN 3 |
| 729 | #define T2_OPCODE_EOR 4 |
| 730 | #define T2_OPCODE_ADD 8 |
| 731 | #define T2_OPCODE_ADC 10 |
| 732 | #define T2_OPCODE_SBC 11 |
| 733 | #define T2_OPCODE_SUB 13 |
| 734 | #define T2_OPCODE_RSB 14 |
| 735 | |
| 736 | #define T_OPCODE_MUL 0x4340 |
| 737 | #define T_OPCODE_TST 0x4200 |
| 738 | #define T_OPCODE_CMN 0x42c0 |
| 739 | #define T_OPCODE_NEG 0x4240 |
| 740 | #define T_OPCODE_MVN 0x43c0 |
| 741 | |
| 742 | #define T_OPCODE_ADD_R3 0x1800 |
| 743 | #define T_OPCODE_SUB_R3 0x1a00 |
| 744 | #define T_OPCODE_ADD_HI 0x4400 |
| 745 | #define T_OPCODE_ADD_ST 0xb000 |
| 746 | #define T_OPCODE_SUB_ST 0xb080 |
| 747 | #define T_OPCODE_ADD_SP 0xa800 |
| 748 | #define T_OPCODE_ADD_PC 0xa000 |
| 749 | #define T_OPCODE_ADD_I8 0x3000 |
| 750 | #define T_OPCODE_SUB_I8 0x3800 |
| 751 | #define T_OPCODE_ADD_I3 0x1c00 |
| 752 | #define T_OPCODE_SUB_I3 0x1e00 |
| 753 | |
| 754 | #define T_OPCODE_ASR_R 0x4100 |
| 755 | #define T_OPCODE_LSL_R 0x4080 |
| 756 | #define T_OPCODE_LSR_R 0x40c0 |
| 757 | #define T_OPCODE_ROR_R 0x41c0 |
| 758 | #define T_OPCODE_ASR_I 0x1000 |
| 759 | #define T_OPCODE_LSL_I 0x0000 |
| 760 | #define T_OPCODE_LSR_I 0x0800 |
| 761 | |
| 762 | #define T_OPCODE_MOV_I8 0x2000 |
| 763 | #define T_OPCODE_CMP_I8 0x2800 |
| 764 | #define T_OPCODE_CMP_LR 0x4280 |
| 765 | #define T_OPCODE_MOV_HR 0x4600 |
| 766 | #define T_OPCODE_CMP_HR 0x4500 |
| 767 | |
| 768 | #define T_OPCODE_LDR_PC 0x4800 |
| 769 | #define T_OPCODE_LDR_SP 0x9800 |
| 770 | #define T_OPCODE_STR_SP 0x9000 |
| 771 | #define T_OPCODE_LDR_IW 0x6800 |
| 772 | #define T_OPCODE_STR_IW 0x6000 |
| 773 | #define T_OPCODE_LDR_IH 0x8800 |
| 774 | #define T_OPCODE_STR_IH 0x8000 |
| 775 | #define T_OPCODE_LDR_IB 0x7800 |
| 776 | #define T_OPCODE_STR_IB 0x7000 |
| 777 | #define T_OPCODE_LDR_RW 0x5800 |
| 778 | #define T_OPCODE_STR_RW 0x5000 |
| 779 | #define T_OPCODE_LDR_RH 0x5a00 |
| 780 | #define T_OPCODE_STR_RH 0x5200 |
| 781 | #define T_OPCODE_LDR_RB 0x5c00 |
| 782 | #define T_OPCODE_STR_RB 0x5400 |
| 783 | |
| 784 | #define T_OPCODE_PUSH 0xb400 |
| 785 | #define T_OPCODE_POP 0xbc00 |
| 786 | |
| 787 | #define T_OPCODE_BRANCH 0xe000 |
| 788 | |
| 789 | #define THUMB_SIZE 2 /* Size of thumb instruction. */ |
| 790 | #define THUMB_PP_PC_LR 0x0100 |
| 791 | #define THUMB_LOAD_BIT 0x0800 |
| 792 | #define THUMB2_LOAD_BIT 0x00100000 |
| 793 | |
| 794 | #define BAD_ARGS _("bad arguments to instruction") |
| 795 | #define BAD_SP _("r13 not allowed here") |
| 796 | #define BAD_PC _("r15 not allowed here") |
| 797 | #define BAD_COND _("instruction cannot be conditional") |
| 798 | #define BAD_OVERLAP _("registers may not be the same") |
| 799 | #define BAD_HIREG _("lo register required") |
| 800 | #define BAD_THUMB32 _("instruction not supported in Thumb16 mode") |
| 801 | #define BAD_ADDR_MODE _("instruction does not accept this addressing mode"); |
| 802 | #define BAD_BRANCH _("branch must be last instruction in IT block") |
| 803 | #define BAD_NOT_IT _("instruction not allowed in IT block") |
| 804 | #define BAD_FPU _("selected FPU does not support instruction") |
| 805 | #define BAD_OUT_IT _("thumb conditional instruction should be in IT block") |
| 806 | #define BAD_IT_COND _("incorrect condition in IT block") |
| 807 | #define BAD_IT_IT _("IT falling in the range of a previous IT block") |
| 808 | #define MISSING_FNSTART _("missing .fnstart before unwinding directive") |
| 809 | #define BAD_PC_ADDRESSING \ |
| 810 | _("cannot use register index with PC-relative addressing") |
| 811 | #define BAD_PC_WRITEBACK \ |
| 812 | _("cannot use writeback with PC-relative addressing") |
| 813 | #define BAD_RANGE _("branch out of range") |
| 814 | #define BAD_FP16 _("selected processor does not support fp16 instruction") |
| 815 | #define UNPRED_REG(R) _("using " R " results in unpredictable behaviour") |
| 816 | #define THUMB1_RELOC_ONLY _("relocation valid in thumb1 code only") |
| 817 | |
| 818 | static struct hash_control * arm_ops_hsh; |
| 819 | static struct hash_control * arm_cond_hsh; |
| 820 | static struct hash_control * arm_shift_hsh; |
| 821 | static struct hash_control * arm_psr_hsh; |
| 822 | static struct hash_control * arm_v7m_psr_hsh; |
| 823 | static struct hash_control * arm_reg_hsh; |
| 824 | static struct hash_control * arm_reloc_hsh; |
| 825 | static struct hash_control * arm_barrier_opt_hsh; |
| 826 | |
| 827 | /* Stuff needed to resolve the label ambiguity |
| 828 | As: |
| 829 | ... |
| 830 | label: <insn> |
| 831 | may differ from: |
| 832 | ... |
| 833 | label: |
| 834 | <insn> */ |
| 835 | |
| 836 | symbolS * last_label_seen; |
| 837 | static int label_is_thumb_function_name = FALSE; |
| 838 | |
| 839 | /* Literal pool structure. Held on a per-section |
| 840 | and per-sub-section basis. */ |
| 841 | |
| 842 | #define MAX_LITERAL_POOL_SIZE 1024 |
| 843 | typedef struct literal_pool |
| 844 | { |
| 845 | expressionS literals [MAX_LITERAL_POOL_SIZE]; |
| 846 | unsigned int next_free_entry; |
| 847 | unsigned int id; |
| 848 | symbolS * symbol; |
| 849 | segT section; |
| 850 | subsegT sub_section; |
| 851 | #ifdef OBJ_ELF |
| 852 | struct dwarf2_line_info locs [MAX_LITERAL_POOL_SIZE]; |
| 853 | #endif |
| 854 | struct literal_pool * next; |
| 855 | unsigned int alignment; |
| 856 | } literal_pool; |
| 857 | |
| 858 | /* Pointer to a linked list of literal pools. */ |
| 859 | literal_pool * list_of_pools = NULL; |
| 860 | |
| 861 | typedef enum asmfunc_states |
| 862 | { |
| 863 | OUTSIDE_ASMFUNC, |
| 864 | WAITING_ASMFUNC_NAME, |
| 865 | WAITING_ENDASMFUNC |
| 866 | } asmfunc_states; |
| 867 | |
| 868 | static asmfunc_states asmfunc_state = OUTSIDE_ASMFUNC; |
| 869 | |
| 870 | #ifdef OBJ_ELF |
| 871 | # define now_it seg_info (now_seg)->tc_segment_info_data.current_it |
| 872 | #else |
| 873 | static struct current_it now_it; |
| 874 | #endif |
| 875 | |
| 876 | static inline int |
| 877 | now_it_compatible (int cond) |
| 878 | { |
| 879 | return (cond & ~1) == (now_it.cc & ~1); |
| 880 | } |
| 881 | |
| 882 | static inline int |
| 883 | conditional_insn (void) |
| 884 | { |
| 885 | return inst.cond != COND_ALWAYS; |
| 886 | } |
| 887 | |
| 888 | static int in_it_block (void); |
| 889 | |
| 890 | static int handle_it_state (void); |
| 891 | |
| 892 | static void force_automatic_it_block_close (void); |
| 893 | |
| 894 | static void it_fsm_post_encode (void); |
| 895 | |
| 896 | #define set_it_insn_type(type) \ |
| 897 | do \ |
| 898 | { \ |
| 899 | inst.it_insn_type = type; \ |
| 900 | if (handle_it_state () == FAIL) \ |
| 901 | return; \ |
| 902 | } \ |
| 903 | while (0) |
| 904 | |
| 905 | #define set_it_insn_type_nonvoid(type, failret) \ |
| 906 | do \ |
| 907 | { \ |
| 908 | inst.it_insn_type = type; \ |
| 909 | if (handle_it_state () == FAIL) \ |
| 910 | return failret; \ |
| 911 | } \ |
| 912 | while(0) |
| 913 | |
| 914 | #define set_it_insn_type_last() \ |
| 915 | do \ |
| 916 | { \ |
| 917 | if (inst.cond == COND_ALWAYS) \ |
| 918 | set_it_insn_type (IF_INSIDE_IT_LAST_INSN); \ |
| 919 | else \ |
| 920 | set_it_insn_type (INSIDE_IT_LAST_INSN); \ |
| 921 | } \ |
| 922 | while (0) |
| 923 | |
| 924 | /* Pure syntax. */ |
| 925 | |
| 926 | /* This array holds the chars that always start a comment. If the |
| 927 | pre-processor is disabled, these aren't very useful. */ |
| 928 | char arm_comment_chars[] = "@"; |
| 929 | |
| 930 | /* This array holds the chars that only start a comment at the beginning of |
| 931 | a line. If the line seems to have the form '# 123 filename' |
| 932 | .line and .file directives will appear in the pre-processed output. */ |
| 933 | /* Note that input_file.c hand checks for '#' at the beginning of the |
| 934 | first line of the input file. This is because the compiler outputs |
| 935 | #NO_APP at the beginning of its output. */ |
| 936 | /* Also note that comments like this one will always work. */ |
| 937 | const char line_comment_chars[] = "#"; |
| 938 | |
| 939 | char arm_line_separator_chars[] = ";"; |
| 940 | |
| 941 | /* Chars that can be used to separate mant |
| 942 | from exp in floating point numbers. */ |
| 943 | const char EXP_CHARS[] = "eE"; |
| 944 | |
| 945 | /* Chars that mean this number is a floating point constant. */ |
| 946 | /* As in 0f12.456 */ |
| 947 | /* or 0d1.2345e12 */ |
| 948 | |
| 949 | const char FLT_CHARS[] = "rRsSfFdDxXeEpP"; |
| 950 | |
| 951 | /* Prefix characters that indicate the start of an immediate |
| 952 | value. */ |
| 953 | #define is_immediate_prefix(C) ((C) == '#' || (C) == '$') |
| 954 | |
| 955 | /* Separator character handling. */ |
| 956 | |
| 957 | #define skip_whitespace(str) do { if (*(str) == ' ') ++(str); } while (0) |
| 958 | |
| 959 | static inline int |
| 960 | skip_past_char (char ** str, char c) |
| 961 | { |
| 962 | /* PR gas/14987: Allow for whitespace before the expected character. */ |
| 963 | skip_whitespace (*str); |
| 964 | |
| 965 | if (**str == c) |
| 966 | { |
| 967 | (*str)++; |
| 968 | return SUCCESS; |
| 969 | } |
| 970 | else |
| 971 | return FAIL; |
| 972 | } |
| 973 | |
| 974 | #define skip_past_comma(str) skip_past_char (str, ',') |
| 975 | |
| 976 | /* Arithmetic expressions (possibly involving symbols). */ |
| 977 | |
| 978 | /* Return TRUE if anything in the expression is a bignum. */ |
| 979 | |
| 980 | static int |
| 981 | walk_no_bignums (symbolS * sp) |
| 982 | { |
| 983 | if (symbol_get_value_expression (sp)->X_op == O_big) |
| 984 | return 1; |
| 985 | |
| 986 | if (symbol_get_value_expression (sp)->X_add_symbol) |
| 987 | { |
| 988 | return (walk_no_bignums (symbol_get_value_expression (sp)->X_add_symbol) |
| 989 | || (symbol_get_value_expression (sp)->X_op_symbol |
| 990 | && walk_no_bignums (symbol_get_value_expression (sp)->X_op_symbol))); |
| 991 | } |
| 992 | |
| 993 | return 0; |
| 994 | } |
| 995 | |
| 996 | static int in_my_get_expression = 0; |
| 997 | |
| 998 | /* Third argument to my_get_expression. */ |
| 999 | #define GE_NO_PREFIX 0 |
| 1000 | #define GE_IMM_PREFIX 1 |
| 1001 | #define GE_OPT_PREFIX 2 |
| 1002 | /* This is a bit of a hack. Use an optional prefix, and also allow big (64-bit) |
| 1003 | immediates, as can be used in Neon VMVN and VMOV immediate instructions. */ |
| 1004 | #define GE_OPT_PREFIX_BIG 3 |
| 1005 | |
| 1006 | static int |
| 1007 | my_get_expression (expressionS * ep, char ** str, int prefix_mode) |
| 1008 | { |
| 1009 | char * save_in; |
| 1010 | segT seg; |
| 1011 | |
| 1012 | /* In unified syntax, all prefixes are optional. */ |
| 1013 | if (unified_syntax) |
| 1014 | prefix_mode = (prefix_mode == GE_OPT_PREFIX_BIG) ? prefix_mode |
| 1015 | : GE_OPT_PREFIX; |
| 1016 | |
| 1017 | switch (prefix_mode) |
| 1018 | { |
| 1019 | case GE_NO_PREFIX: break; |
| 1020 | case GE_IMM_PREFIX: |
| 1021 | if (!is_immediate_prefix (**str)) |
| 1022 | { |
| 1023 | inst.error = _("immediate expression requires a # prefix"); |
| 1024 | return FAIL; |
| 1025 | } |
| 1026 | (*str)++; |
| 1027 | break; |
| 1028 | case GE_OPT_PREFIX: |
| 1029 | case GE_OPT_PREFIX_BIG: |
| 1030 | if (is_immediate_prefix (**str)) |
| 1031 | (*str)++; |
| 1032 | break; |
| 1033 | default: abort (); |
| 1034 | } |
| 1035 | |
| 1036 | memset (ep, 0, sizeof (expressionS)); |
| 1037 | |
| 1038 | save_in = input_line_pointer; |
| 1039 | input_line_pointer = *str; |
| 1040 | in_my_get_expression = 1; |
| 1041 | seg = expression (ep); |
| 1042 | in_my_get_expression = 0; |
| 1043 | |
| 1044 | if (ep->X_op == O_illegal || ep->X_op == O_absent) |
| 1045 | { |
| 1046 | /* We found a bad or missing expression in md_operand(). */ |
| 1047 | *str = input_line_pointer; |
| 1048 | input_line_pointer = save_in; |
| 1049 | if (inst.error == NULL) |
| 1050 | inst.error = (ep->X_op == O_absent |
| 1051 | ? _("missing expression") :_("bad expression")); |
| 1052 | return 1; |
| 1053 | } |
| 1054 | |
| 1055 | #ifdef OBJ_AOUT |
| 1056 | if (seg != absolute_section |
| 1057 | && seg != text_section |
| 1058 | && seg != data_section |
| 1059 | && seg != bss_section |
| 1060 | && seg != undefined_section) |
| 1061 | { |
| 1062 | inst.error = _("bad segment"); |
| 1063 | *str = input_line_pointer; |
| 1064 | input_line_pointer = save_in; |
| 1065 | return 1; |
| 1066 | } |
| 1067 | #else |
| 1068 | (void) seg; |
| 1069 | #endif |
| 1070 | |
| 1071 | /* Get rid of any bignums now, so that we don't generate an error for which |
| 1072 | we can't establish a line number later on. Big numbers are never valid |
| 1073 | in instructions, which is where this routine is always called. */ |
| 1074 | if (prefix_mode != GE_OPT_PREFIX_BIG |
| 1075 | && (ep->X_op == O_big |
| 1076 | || (ep->X_add_symbol |
| 1077 | && (walk_no_bignums (ep->X_add_symbol) |
| 1078 | || (ep->X_op_symbol |
| 1079 | && walk_no_bignums (ep->X_op_symbol)))))) |
| 1080 | { |
| 1081 | inst.error = _("invalid constant"); |
| 1082 | *str = input_line_pointer; |
| 1083 | input_line_pointer = save_in; |
| 1084 | return 1; |
| 1085 | } |
| 1086 | |
| 1087 | *str = input_line_pointer; |
| 1088 | input_line_pointer = save_in; |
| 1089 | return 0; |
| 1090 | } |
| 1091 | |
| 1092 | /* Turn a string in input_line_pointer into a floating point constant |
| 1093 | of type TYPE, and store the appropriate bytes in *LITP. The number |
| 1094 | of LITTLENUMS emitted is stored in *SIZEP. An error message is |
| 1095 | returned, or NULL on OK. |
| 1096 | |
| 1097 | Note that fp constants aren't represent in the normal way on the ARM. |
| 1098 | In big endian mode, things are as expected. However, in little endian |
| 1099 | mode fp constants are big-endian word-wise, and little-endian byte-wise |
| 1100 | within the words. For example, (double) 1.1 in big endian mode is |
| 1101 | the byte sequence 3f f1 99 99 99 99 99 9a, and in little endian mode is |
| 1102 | the byte sequence 99 99 f1 3f 9a 99 99 99. |
| 1103 | |
| 1104 | ??? The format of 12 byte floats is uncertain according to gcc's arm.h. */ |
| 1105 | |
| 1106 | const char * |
| 1107 | md_atof (int type, char * litP, int * sizeP) |
| 1108 | { |
| 1109 | int prec; |
| 1110 | LITTLENUM_TYPE words[MAX_LITTLENUMS]; |
| 1111 | char *t; |
| 1112 | int i; |
| 1113 | |
| 1114 | switch (type) |
| 1115 | { |
| 1116 | case 'f': |
| 1117 | case 'F': |
| 1118 | case 's': |
| 1119 | case 'S': |
| 1120 | prec = 2; |
| 1121 | break; |
| 1122 | |
| 1123 | case 'd': |
| 1124 | case 'D': |
| 1125 | case 'r': |
| 1126 | case 'R': |
| 1127 | prec = 4; |
| 1128 | break; |
| 1129 | |
| 1130 | case 'x': |
| 1131 | case 'X': |
| 1132 | prec = 5; |
| 1133 | break; |
| 1134 | |
| 1135 | case 'p': |
| 1136 | case 'P': |
| 1137 | prec = 5; |
| 1138 | break; |
| 1139 | |
| 1140 | default: |
| 1141 | *sizeP = 0; |
| 1142 | return _("Unrecognized or unsupported floating point constant"); |
| 1143 | } |
| 1144 | |
| 1145 | t = atof_ieee (input_line_pointer, type, words); |
| 1146 | if (t) |
| 1147 | input_line_pointer = t; |
| 1148 | *sizeP = prec * sizeof (LITTLENUM_TYPE); |
| 1149 | |
| 1150 | if (target_big_endian) |
| 1151 | { |
| 1152 | for (i = 0; i < prec; i++) |
| 1153 | { |
| 1154 | md_number_to_chars (litP, (valueT) words[i], sizeof (LITTLENUM_TYPE)); |
| 1155 | litP += sizeof (LITTLENUM_TYPE); |
| 1156 | } |
| 1157 | } |
| 1158 | else |
| 1159 | { |
| 1160 | if (ARM_CPU_HAS_FEATURE (cpu_variant, fpu_endian_pure)) |
| 1161 | for (i = prec - 1; i >= 0; i--) |
| 1162 | { |
| 1163 | md_number_to_chars (litP, (valueT) words[i], sizeof (LITTLENUM_TYPE)); |
| 1164 | litP += sizeof (LITTLENUM_TYPE); |
| 1165 | } |
| 1166 | else |
| 1167 | /* For a 4 byte float the order of elements in `words' is 1 0. |
| 1168 | For an 8 byte float the order is 1 0 3 2. */ |
| 1169 | for (i = 0; i < prec; i += 2) |
| 1170 | { |
| 1171 | md_number_to_chars (litP, (valueT) words[i + 1], |
| 1172 | sizeof (LITTLENUM_TYPE)); |
| 1173 | md_number_to_chars (litP + sizeof (LITTLENUM_TYPE), |
| 1174 | (valueT) words[i], sizeof (LITTLENUM_TYPE)); |
| 1175 | litP += 2 * sizeof (LITTLENUM_TYPE); |
| 1176 | } |
| 1177 | } |
| 1178 | |
| 1179 | return NULL; |
| 1180 | } |
| 1181 | |
| 1182 | /* We handle all bad expressions here, so that we can report the faulty |
| 1183 | instruction in the error message. */ |
| 1184 | void |
| 1185 | md_operand (expressionS * exp) |
| 1186 | { |
| 1187 | if (in_my_get_expression) |
| 1188 | exp->X_op = O_illegal; |
| 1189 | } |
| 1190 | |
| 1191 | /* Immediate values. */ |
| 1192 | |
| 1193 | /* Generic immediate-value read function for use in directives. |
| 1194 | Accepts anything that 'expression' can fold to a constant. |
| 1195 | *val receives the number. */ |
| 1196 | #ifdef OBJ_ELF |
| 1197 | static int |
| 1198 | immediate_for_directive (int *val) |
| 1199 | { |
| 1200 | expressionS exp; |
| 1201 | exp.X_op = O_illegal; |
| 1202 | |
| 1203 | if (is_immediate_prefix (*input_line_pointer)) |
| 1204 | { |
| 1205 | input_line_pointer++; |
| 1206 | expression (&exp); |
| 1207 | } |
| 1208 | |
| 1209 | if (exp.X_op != O_constant) |
| 1210 | { |
| 1211 | as_bad (_("expected #constant")); |
| 1212 | ignore_rest_of_line (); |
| 1213 | return FAIL; |
| 1214 | } |
| 1215 | *val = exp.X_add_number; |
| 1216 | return SUCCESS; |
| 1217 | } |
| 1218 | #endif |
| 1219 | |
| 1220 | /* Register parsing. */ |
| 1221 | |
| 1222 | /* Generic register parser. CCP points to what should be the |
| 1223 | beginning of a register name. If it is indeed a valid register |
| 1224 | name, advance CCP over it and return the reg_entry structure; |
| 1225 | otherwise return NULL. Does not issue diagnostics. */ |
| 1226 | |
| 1227 | static struct reg_entry * |
| 1228 | arm_reg_parse_multi (char **ccp) |
| 1229 | { |
| 1230 | char *start = *ccp; |
| 1231 | char *p; |
| 1232 | struct reg_entry *reg; |
| 1233 | |
| 1234 | skip_whitespace (start); |
| 1235 | |
| 1236 | #ifdef REGISTER_PREFIX |
| 1237 | if (*start != REGISTER_PREFIX) |
| 1238 | return NULL; |
| 1239 | start++; |
| 1240 | #endif |
| 1241 | #ifdef OPTIONAL_REGISTER_PREFIX |
| 1242 | if (*start == OPTIONAL_REGISTER_PREFIX) |
| 1243 | start++; |
| 1244 | #endif |
| 1245 | |
| 1246 | p = start; |
| 1247 | if (!ISALPHA (*p) || !is_name_beginner (*p)) |
| 1248 | return NULL; |
| 1249 | |
| 1250 | do |
| 1251 | p++; |
| 1252 | while (ISALPHA (*p) || ISDIGIT (*p) || *p == '_'); |
| 1253 | |
| 1254 | reg = (struct reg_entry *) hash_find_n (arm_reg_hsh, start, p - start); |
| 1255 | |
| 1256 | if (!reg) |
| 1257 | return NULL; |
| 1258 | |
| 1259 | *ccp = p; |
| 1260 | return reg; |
| 1261 | } |
| 1262 | |
| 1263 | static int |
| 1264 | arm_reg_alt_syntax (char **ccp, char *start, struct reg_entry *reg, |
| 1265 | enum arm_reg_type type) |
| 1266 | { |
| 1267 | /* Alternative syntaxes are accepted for a few register classes. */ |
| 1268 | switch (type) |
| 1269 | { |
| 1270 | case REG_TYPE_MVF: |
| 1271 | case REG_TYPE_MVD: |
| 1272 | case REG_TYPE_MVFX: |
| 1273 | case REG_TYPE_MVDX: |
| 1274 | /* Generic coprocessor register names are allowed for these. */ |
| 1275 | if (reg && reg->type == REG_TYPE_CN) |
| 1276 | return reg->number; |
| 1277 | break; |
| 1278 | |
| 1279 | case REG_TYPE_CP: |
| 1280 | /* For backward compatibility, a bare number is valid here. */ |
| 1281 | { |
| 1282 | unsigned long processor = strtoul (start, ccp, 10); |
| 1283 | if (*ccp != start && processor <= 15) |
| 1284 | return processor; |
| 1285 | } |
| 1286 | /* Fall through. */ |
| 1287 | |
| 1288 | case REG_TYPE_MMXWC: |
| 1289 | /* WC includes WCG. ??? I'm not sure this is true for all |
| 1290 | instructions that take WC registers. */ |
| 1291 | if (reg && reg->type == REG_TYPE_MMXWCG) |
| 1292 | return reg->number; |
| 1293 | break; |
| 1294 | |
| 1295 | default: |
| 1296 | break; |
| 1297 | } |
| 1298 | |
| 1299 | return FAIL; |
| 1300 | } |
| 1301 | |
| 1302 | /* As arm_reg_parse_multi, but the register must be of type TYPE, and the |
| 1303 | return value is the register number or FAIL. */ |
| 1304 | |
| 1305 | static int |
| 1306 | arm_reg_parse (char **ccp, enum arm_reg_type type) |
| 1307 | { |
| 1308 | char *start = *ccp; |
| 1309 | struct reg_entry *reg = arm_reg_parse_multi (ccp); |
| 1310 | int ret; |
| 1311 | |
| 1312 | /* Do not allow a scalar (reg+index) to parse as a register. */ |
| 1313 | if (reg && reg->neon && (reg->neon->defined & NTA_HASINDEX)) |
| 1314 | return FAIL; |
| 1315 | |
| 1316 | if (reg && reg->type == type) |
| 1317 | return reg->number; |
| 1318 | |
| 1319 | if ((ret = arm_reg_alt_syntax (ccp, start, reg, type)) != FAIL) |
| 1320 | return ret; |
| 1321 | |
| 1322 | *ccp = start; |
| 1323 | return FAIL; |
| 1324 | } |
| 1325 | |
| 1326 | /* Parse a Neon type specifier. *STR should point at the leading '.' |
| 1327 | character. Does no verification at this stage that the type fits the opcode |
| 1328 | properly. E.g., |
| 1329 | |
| 1330 | .i32.i32.s16 |
| 1331 | .s32.f32 |
| 1332 | .u16 |
| 1333 | |
| 1334 | Can all be legally parsed by this function. |
| 1335 | |
| 1336 | Fills in neon_type struct pointer with parsed information, and updates STR |
| 1337 | to point after the parsed type specifier. Returns SUCCESS if this was a legal |
| 1338 | type, FAIL if not. */ |
| 1339 | |
| 1340 | static int |
| 1341 | parse_neon_type (struct neon_type *type, char **str) |
| 1342 | { |
| 1343 | char *ptr = *str; |
| 1344 | |
| 1345 | if (type) |
| 1346 | type->elems = 0; |
| 1347 | |
| 1348 | while (type->elems < NEON_MAX_TYPE_ELS) |
| 1349 | { |
| 1350 | enum neon_el_type thistype = NT_untyped; |
| 1351 | unsigned thissize = -1u; |
| 1352 | |
| 1353 | if (*ptr != '.') |
| 1354 | break; |
| 1355 | |
| 1356 | ptr++; |
| 1357 | |
| 1358 | /* Just a size without an explicit type. */ |
| 1359 | if (ISDIGIT (*ptr)) |
| 1360 | goto parsesize; |
| 1361 | |
| 1362 | switch (TOLOWER (*ptr)) |
| 1363 | { |
| 1364 | case 'i': thistype = NT_integer; break; |
| 1365 | case 'f': thistype = NT_float; break; |
| 1366 | case 'p': thistype = NT_poly; break; |
| 1367 | case 's': thistype = NT_signed; break; |
| 1368 | case 'u': thistype = NT_unsigned; break; |
| 1369 | case 'd': |
| 1370 | thistype = NT_float; |
| 1371 | thissize = 64; |
| 1372 | ptr++; |
| 1373 | goto done; |
| 1374 | default: |
| 1375 | as_bad (_("unexpected character `%c' in type specifier"), *ptr); |
| 1376 | return FAIL; |
| 1377 | } |
| 1378 | |
| 1379 | ptr++; |
| 1380 | |
| 1381 | /* .f is an abbreviation for .f32. */ |
| 1382 | if (thistype == NT_float && !ISDIGIT (*ptr)) |
| 1383 | thissize = 32; |
| 1384 | else |
| 1385 | { |
| 1386 | parsesize: |
| 1387 | thissize = strtoul (ptr, &ptr, 10); |
| 1388 | |
| 1389 | if (thissize != 8 && thissize != 16 && thissize != 32 |
| 1390 | && thissize != 64) |
| 1391 | { |
| 1392 | as_bad (_("bad size %d in type specifier"), thissize); |
| 1393 | return FAIL; |
| 1394 | } |
| 1395 | } |
| 1396 | |
| 1397 | done: |
| 1398 | if (type) |
| 1399 | { |
| 1400 | type->el[type->elems].type = thistype; |
| 1401 | type->el[type->elems].size = thissize; |
| 1402 | type->elems++; |
| 1403 | } |
| 1404 | } |
| 1405 | |
| 1406 | /* Empty/missing type is not a successful parse. */ |
| 1407 | if (type->elems == 0) |
| 1408 | return FAIL; |
| 1409 | |
| 1410 | *str = ptr; |
| 1411 | |
| 1412 | return SUCCESS; |
| 1413 | } |
| 1414 | |
| 1415 | /* Errors may be set multiple times during parsing or bit encoding |
| 1416 | (particularly in the Neon bits), but usually the earliest error which is set |
| 1417 | will be the most meaningful. Avoid overwriting it with later (cascading) |
| 1418 | errors by calling this function. */ |
| 1419 | |
| 1420 | static void |
| 1421 | first_error (const char *err) |
| 1422 | { |
| 1423 | if (!inst.error) |
| 1424 | inst.error = err; |
| 1425 | } |
| 1426 | |
| 1427 | /* Parse a single type, e.g. ".s32", leading period included. */ |
| 1428 | static int |
| 1429 | parse_neon_operand_type (struct neon_type_el *vectype, char **ccp) |
| 1430 | { |
| 1431 | char *str = *ccp; |
| 1432 | struct neon_type optype; |
| 1433 | |
| 1434 | if (*str == '.') |
| 1435 | { |
| 1436 | if (parse_neon_type (&optype, &str) == SUCCESS) |
| 1437 | { |
| 1438 | if (optype.elems == 1) |
| 1439 | *vectype = optype.el[0]; |
| 1440 | else |
| 1441 | { |
| 1442 | first_error (_("only one type should be specified for operand")); |
| 1443 | return FAIL; |
| 1444 | } |
| 1445 | } |
| 1446 | else |
| 1447 | { |
| 1448 | first_error (_("vector type expected")); |
| 1449 | return FAIL; |
| 1450 | } |
| 1451 | } |
| 1452 | else |
| 1453 | return FAIL; |
| 1454 | |
| 1455 | *ccp = str; |
| 1456 | |
| 1457 | return SUCCESS; |
| 1458 | } |
| 1459 | |
| 1460 | /* Special meanings for indices (which have a range of 0-7), which will fit into |
| 1461 | a 4-bit integer. */ |
| 1462 | |
| 1463 | #define NEON_ALL_LANES 15 |
| 1464 | #define NEON_INTERLEAVE_LANES 14 |
| 1465 | |
| 1466 | /* Parse either a register or a scalar, with an optional type. Return the |
| 1467 | register number, and optionally fill in the actual type of the register |
| 1468 | when multiple alternatives were given (NEON_TYPE_NDQ) in *RTYPE, and |
| 1469 | type/index information in *TYPEINFO. */ |
| 1470 | |
| 1471 | static int |
| 1472 | parse_typed_reg_or_scalar (char **ccp, enum arm_reg_type type, |
| 1473 | enum arm_reg_type *rtype, |
| 1474 | struct neon_typed_alias *typeinfo) |
| 1475 | { |
| 1476 | char *str = *ccp; |
| 1477 | struct reg_entry *reg = arm_reg_parse_multi (&str); |
| 1478 | struct neon_typed_alias atype; |
| 1479 | struct neon_type_el parsetype; |
| 1480 | |
| 1481 | atype.defined = 0; |
| 1482 | atype.index = -1; |
| 1483 | atype.eltype.type = NT_invtype; |
| 1484 | atype.eltype.size = -1; |
| 1485 | |
| 1486 | /* Try alternate syntax for some types of register. Note these are mutually |
| 1487 | exclusive with the Neon syntax extensions. */ |
| 1488 | if (reg == NULL) |
| 1489 | { |
| 1490 | int altreg = arm_reg_alt_syntax (&str, *ccp, reg, type); |
| 1491 | if (altreg != FAIL) |
| 1492 | *ccp = str; |
| 1493 | if (typeinfo) |
| 1494 | *typeinfo = atype; |
| 1495 | return altreg; |
| 1496 | } |
| 1497 | |
| 1498 | /* Undo polymorphism when a set of register types may be accepted. */ |
| 1499 | if ((type == REG_TYPE_NDQ |
| 1500 | && (reg->type == REG_TYPE_NQ || reg->type == REG_TYPE_VFD)) |
| 1501 | || (type == REG_TYPE_VFSD |
| 1502 | && (reg->type == REG_TYPE_VFS || reg->type == REG_TYPE_VFD)) |
| 1503 | || (type == REG_TYPE_NSDQ |
| 1504 | && (reg->type == REG_TYPE_VFS || reg->type == REG_TYPE_VFD |
| 1505 | || reg->type == REG_TYPE_NQ)) |
| 1506 | || (type == REG_TYPE_MMXWC |
| 1507 | && (reg->type == REG_TYPE_MMXWCG))) |
| 1508 | type = (enum arm_reg_type) reg->type; |
| 1509 | |
| 1510 | if (type != reg->type) |
| 1511 | return FAIL; |
| 1512 | |
| 1513 | if (reg->neon) |
| 1514 | atype = *reg->neon; |
| 1515 | |
| 1516 | if (parse_neon_operand_type (&parsetype, &str) == SUCCESS) |
| 1517 | { |
| 1518 | if ((atype.defined & NTA_HASTYPE) != 0) |
| 1519 | { |
| 1520 | first_error (_("can't redefine type for operand")); |
| 1521 | return FAIL; |
| 1522 | } |
| 1523 | atype.defined |= NTA_HASTYPE; |
| 1524 | atype.eltype = parsetype; |
| 1525 | } |
| 1526 | |
| 1527 | if (skip_past_char (&str, '[') == SUCCESS) |
| 1528 | { |
| 1529 | if (type != REG_TYPE_VFD) |
| 1530 | { |
| 1531 | first_error (_("only D registers may be indexed")); |
| 1532 | return FAIL; |
| 1533 | } |
| 1534 | |
| 1535 | if ((atype.defined & NTA_HASINDEX) != 0) |
| 1536 | { |
| 1537 | first_error (_("can't change index for operand")); |
| 1538 | return FAIL; |
| 1539 | } |
| 1540 | |
| 1541 | atype.defined |= NTA_HASINDEX; |
| 1542 | |
| 1543 | if (skip_past_char (&str, ']') == SUCCESS) |
| 1544 | atype.index = NEON_ALL_LANES; |
| 1545 | else |
| 1546 | { |
| 1547 | expressionS exp; |
| 1548 | |
| 1549 | my_get_expression (&exp, &str, GE_NO_PREFIX); |
| 1550 | |
| 1551 | if (exp.X_op != O_constant) |
| 1552 | { |
| 1553 | first_error (_("constant expression required")); |
| 1554 | return FAIL; |
| 1555 | } |
| 1556 | |
| 1557 | if (skip_past_char (&str, ']') == FAIL) |
| 1558 | return FAIL; |
| 1559 | |
| 1560 | atype.index = exp.X_add_number; |
| 1561 | } |
| 1562 | } |
| 1563 | |
| 1564 | if (typeinfo) |
| 1565 | *typeinfo = atype; |
| 1566 | |
| 1567 | if (rtype) |
| 1568 | *rtype = type; |
| 1569 | |
| 1570 | *ccp = str; |
| 1571 | |
| 1572 | return reg->number; |
| 1573 | } |
| 1574 | |
| 1575 | /* Like arm_reg_parse, but allow allow the following extra features: |
| 1576 | - If RTYPE is non-zero, return the (possibly restricted) type of the |
| 1577 | register (e.g. Neon double or quad reg when either has been requested). |
| 1578 | - If this is a Neon vector type with additional type information, fill |
| 1579 | in the struct pointed to by VECTYPE (if non-NULL). |
| 1580 | This function will fault on encountering a scalar. */ |
| 1581 | |
| 1582 | static int |
| 1583 | arm_typed_reg_parse (char **ccp, enum arm_reg_type type, |
| 1584 | enum arm_reg_type *rtype, struct neon_type_el *vectype) |
| 1585 | { |
| 1586 | struct neon_typed_alias atype; |
| 1587 | char *str = *ccp; |
| 1588 | int reg = parse_typed_reg_or_scalar (&str, type, rtype, &atype); |
| 1589 | |
| 1590 | if (reg == FAIL) |
| 1591 | return FAIL; |
| 1592 | |
| 1593 | /* Do not allow regname(... to parse as a register. */ |
| 1594 | if (*str == '(') |
| 1595 | return FAIL; |
| 1596 | |
| 1597 | /* Do not allow a scalar (reg+index) to parse as a register. */ |
| 1598 | if ((atype.defined & NTA_HASINDEX) != 0) |
| 1599 | { |
| 1600 | first_error (_("register operand expected, but got scalar")); |
| 1601 | return FAIL; |
| 1602 | } |
| 1603 | |
| 1604 | if (vectype) |
| 1605 | *vectype = atype.eltype; |
| 1606 | |
| 1607 | *ccp = str; |
| 1608 | |
| 1609 | return reg; |
| 1610 | } |
| 1611 | |
| 1612 | #define NEON_SCALAR_REG(X) ((X) >> 4) |
| 1613 | #define NEON_SCALAR_INDEX(X) ((X) & 15) |
| 1614 | |
| 1615 | /* Parse a Neon scalar. Most of the time when we're parsing a scalar, we don't |
| 1616 | have enough information to be able to do a good job bounds-checking. So, we |
| 1617 | just do easy checks here, and do further checks later. */ |
| 1618 | |
| 1619 | static int |
| 1620 | parse_scalar (char **ccp, int elsize, struct neon_type_el *type) |
| 1621 | { |
| 1622 | int reg; |
| 1623 | char *str = *ccp; |
| 1624 | struct neon_typed_alias atype; |
| 1625 | |
| 1626 | reg = parse_typed_reg_or_scalar (&str, REG_TYPE_VFD, NULL, &atype); |
| 1627 | |
| 1628 | if (reg == FAIL || (atype.defined & NTA_HASINDEX) == 0) |
| 1629 | return FAIL; |
| 1630 | |
| 1631 | if (atype.index == NEON_ALL_LANES) |
| 1632 | { |
| 1633 | first_error (_("scalar must have an index")); |
| 1634 | return FAIL; |
| 1635 | } |
| 1636 | else if (atype.index >= 64 / elsize) |
| 1637 | { |
| 1638 | first_error (_("scalar index out of range")); |
| 1639 | return FAIL; |
| 1640 | } |
| 1641 | |
| 1642 | if (type) |
| 1643 | *type = atype.eltype; |
| 1644 | |
| 1645 | *ccp = str; |
| 1646 | |
| 1647 | return reg * 16 + atype.index; |
| 1648 | } |
| 1649 | |
| 1650 | /* Parse an ARM register list. Returns the bitmask, or FAIL. */ |
| 1651 | |
| 1652 | static long |
| 1653 | parse_reg_list (char ** strp) |
| 1654 | { |
| 1655 | char * str = * strp; |
| 1656 | long range = 0; |
| 1657 | int another_range; |
| 1658 | |
| 1659 | /* We come back here if we get ranges concatenated by '+' or '|'. */ |
| 1660 | do |
| 1661 | { |
| 1662 | skip_whitespace (str); |
| 1663 | |
| 1664 | another_range = 0; |
| 1665 | |
| 1666 | if (*str == '{') |
| 1667 | { |
| 1668 | int in_range = 0; |
| 1669 | int cur_reg = -1; |
| 1670 | |
| 1671 | str++; |
| 1672 | do |
| 1673 | { |
| 1674 | int reg; |
| 1675 | |
| 1676 | if ((reg = arm_reg_parse (&str, REG_TYPE_RN)) == FAIL) |
| 1677 | { |
| 1678 | first_error (_(reg_expected_msgs[REG_TYPE_RN])); |
| 1679 | return FAIL; |
| 1680 | } |
| 1681 | |
| 1682 | if (in_range) |
| 1683 | { |
| 1684 | int i; |
| 1685 | |
| 1686 | if (reg <= cur_reg) |
| 1687 | { |
| 1688 | first_error (_("bad range in register list")); |
| 1689 | return FAIL; |
| 1690 | } |
| 1691 | |
| 1692 | for (i = cur_reg + 1; i < reg; i++) |
| 1693 | { |
| 1694 | if (range & (1 << i)) |
| 1695 | as_tsktsk |
| 1696 | (_("Warning: duplicated register (r%d) in register list"), |
| 1697 | i); |
| 1698 | else |
| 1699 | range |= 1 << i; |
| 1700 | } |
| 1701 | in_range = 0; |
| 1702 | } |
| 1703 | |
| 1704 | if (range & (1 << reg)) |
| 1705 | as_tsktsk (_("Warning: duplicated register (r%d) in register list"), |
| 1706 | reg); |
| 1707 | else if (reg <= cur_reg) |
| 1708 | as_tsktsk (_("Warning: register range not in ascending order")); |
| 1709 | |
| 1710 | range |= 1 << reg; |
| 1711 | cur_reg = reg; |
| 1712 | } |
| 1713 | while (skip_past_comma (&str) != FAIL |
| 1714 | || (in_range = 1, *str++ == '-')); |
| 1715 | str--; |
| 1716 | |
| 1717 | if (skip_past_char (&str, '}') == FAIL) |
| 1718 | { |
| 1719 | first_error (_("missing `}'")); |
| 1720 | return FAIL; |
| 1721 | } |
| 1722 | } |
| 1723 | else |
| 1724 | { |
| 1725 | expressionS exp; |
| 1726 | |
| 1727 | if (my_get_expression (&exp, &str, GE_NO_PREFIX)) |
| 1728 | return FAIL; |
| 1729 | |
| 1730 | if (exp.X_op == O_constant) |
| 1731 | { |
| 1732 | if (exp.X_add_number |
| 1733 | != (exp.X_add_number & 0x0000ffff)) |
| 1734 | { |
| 1735 | inst.error = _("invalid register mask"); |
| 1736 | return FAIL; |
| 1737 | } |
| 1738 | |
| 1739 | if ((range & exp.X_add_number) != 0) |
| 1740 | { |
| 1741 | int regno = range & exp.X_add_number; |
| 1742 | |
| 1743 | regno &= -regno; |
| 1744 | regno = (1 << regno) - 1; |
| 1745 | as_tsktsk |
| 1746 | (_("Warning: duplicated register (r%d) in register list"), |
| 1747 | regno); |
| 1748 | } |
| 1749 | |
| 1750 | range |= exp.X_add_number; |
| 1751 | } |
| 1752 | else |
| 1753 | { |
| 1754 | if (inst.reloc.type != 0) |
| 1755 | { |
| 1756 | inst.error = _("expression too complex"); |
| 1757 | return FAIL; |
| 1758 | } |
| 1759 | |
| 1760 | memcpy (&inst.reloc.exp, &exp, sizeof (expressionS)); |
| 1761 | inst.reloc.type = BFD_RELOC_ARM_MULTI; |
| 1762 | inst.reloc.pc_rel = 0; |
| 1763 | } |
| 1764 | } |
| 1765 | |
| 1766 | if (*str == '|' || *str == '+') |
| 1767 | { |
| 1768 | str++; |
| 1769 | another_range = 1; |
| 1770 | } |
| 1771 | } |
| 1772 | while (another_range); |
| 1773 | |
| 1774 | *strp = str; |
| 1775 | return range; |
| 1776 | } |
| 1777 | |
| 1778 | /* Types of registers in a list. */ |
| 1779 | |
| 1780 | enum reg_list_els |
| 1781 | { |
| 1782 | REGLIST_VFP_S, |
| 1783 | REGLIST_VFP_D, |
| 1784 | REGLIST_NEON_D |
| 1785 | }; |
| 1786 | |
| 1787 | /* Parse a VFP register list. If the string is invalid return FAIL. |
| 1788 | Otherwise return the number of registers, and set PBASE to the first |
| 1789 | register. Parses registers of type ETYPE. |
| 1790 | If REGLIST_NEON_D is used, several syntax enhancements are enabled: |
| 1791 | - Q registers can be used to specify pairs of D registers |
| 1792 | - { } can be omitted from around a singleton register list |
| 1793 | FIXME: This is not implemented, as it would require backtracking in |
| 1794 | some cases, e.g.: |
| 1795 | vtbl.8 d3,d4,d5 |
| 1796 | This could be done (the meaning isn't really ambiguous), but doesn't |
| 1797 | fit in well with the current parsing framework. |
| 1798 | - 32 D registers may be used (also true for VFPv3). |
| 1799 | FIXME: Types are ignored in these register lists, which is probably a |
| 1800 | bug. */ |
| 1801 | |
| 1802 | static int |
| 1803 | parse_vfp_reg_list (char **ccp, unsigned int *pbase, enum reg_list_els etype) |
| 1804 | { |
| 1805 | char *str = *ccp; |
| 1806 | int base_reg; |
| 1807 | int new_base; |
| 1808 | enum arm_reg_type regtype = (enum arm_reg_type) 0; |
| 1809 | int max_regs = 0; |
| 1810 | int count = 0; |
| 1811 | int warned = 0; |
| 1812 | unsigned long mask = 0; |
| 1813 | int i; |
| 1814 | |
| 1815 | if (skip_past_char (&str, '{') == FAIL) |
| 1816 | { |
| 1817 | inst.error = _("expecting {"); |
| 1818 | return FAIL; |
| 1819 | } |
| 1820 | |
| 1821 | switch (etype) |
| 1822 | { |
| 1823 | case REGLIST_VFP_S: |
| 1824 | regtype = REG_TYPE_VFS; |
| 1825 | max_regs = 32; |
| 1826 | break; |
| 1827 | |
| 1828 | case REGLIST_VFP_D: |
| 1829 | regtype = REG_TYPE_VFD; |
| 1830 | break; |
| 1831 | |
| 1832 | case REGLIST_NEON_D: |
| 1833 | regtype = REG_TYPE_NDQ; |
| 1834 | break; |
| 1835 | } |
| 1836 | |
| 1837 | if (etype != REGLIST_VFP_S) |
| 1838 | { |
| 1839 | /* VFPv3 allows 32 D registers, except for the VFPv3-D16 variant. */ |
| 1840 | if (ARM_CPU_HAS_FEATURE (cpu_variant, fpu_vfp_ext_d32)) |
| 1841 | { |
| 1842 | max_regs = 32; |
| 1843 | if (thumb_mode) |
| 1844 | ARM_MERGE_FEATURE_SETS (thumb_arch_used, thumb_arch_used, |
| 1845 | fpu_vfp_ext_d32); |
| 1846 | else |
| 1847 | ARM_MERGE_FEATURE_SETS (arm_arch_used, arm_arch_used, |
| 1848 | fpu_vfp_ext_d32); |
| 1849 | } |
| 1850 | else |
| 1851 | max_regs = 16; |
| 1852 | } |
| 1853 | |
| 1854 | base_reg = max_regs; |
| 1855 | |
| 1856 | do |
| 1857 | { |
| 1858 | int setmask = 1, addregs = 1; |
| 1859 | |
| 1860 | new_base = arm_typed_reg_parse (&str, regtype, ®type, NULL); |
| 1861 | |
| 1862 | if (new_base == FAIL) |
| 1863 | { |
| 1864 | first_error (_(reg_expected_msgs[regtype])); |
| 1865 | return FAIL; |
| 1866 | } |
| 1867 | |
| 1868 | if (new_base >= max_regs) |
| 1869 | { |
| 1870 | first_error (_("register out of range in list")); |
| 1871 | return FAIL; |
| 1872 | } |
| 1873 | |
| 1874 | /* Note: a value of 2 * n is returned for the register Q<n>. */ |
| 1875 | if (regtype == REG_TYPE_NQ) |
| 1876 | { |
| 1877 | setmask = 3; |
| 1878 | addregs = 2; |
| 1879 | } |
| 1880 | |
| 1881 | if (new_base < base_reg) |
| 1882 | base_reg = new_base; |
| 1883 | |
| 1884 | if (mask & (setmask << new_base)) |
| 1885 | { |
| 1886 | first_error (_("invalid register list")); |
| 1887 | return FAIL; |
| 1888 | } |
| 1889 | |
| 1890 | if ((mask >> new_base) != 0 && ! warned) |
| 1891 | { |
| 1892 | as_tsktsk (_("register list not in ascending order")); |
| 1893 | warned = 1; |
| 1894 | } |
| 1895 | |
| 1896 | mask |= setmask << new_base; |
| 1897 | count += addregs; |
| 1898 | |
| 1899 | if (*str == '-') /* We have the start of a range expression */ |
| 1900 | { |
| 1901 | int high_range; |
| 1902 | |
| 1903 | str++; |
| 1904 | |
| 1905 | if ((high_range = arm_typed_reg_parse (&str, regtype, NULL, NULL)) |
| 1906 | == FAIL) |
| 1907 | { |
| 1908 | inst.error = gettext (reg_expected_msgs[regtype]); |
| 1909 | return FAIL; |
| 1910 | } |
| 1911 | |
| 1912 | if (high_range >= max_regs) |
| 1913 | { |
| 1914 | first_error (_("register out of range in list")); |
| 1915 | return FAIL; |
| 1916 | } |
| 1917 | |
| 1918 | if (regtype == REG_TYPE_NQ) |
| 1919 | high_range = high_range + 1; |
| 1920 | |
| 1921 | if (high_range <= new_base) |
| 1922 | { |
| 1923 | inst.error = _("register range not in ascending order"); |
| 1924 | return FAIL; |
| 1925 | } |
| 1926 | |
| 1927 | for (new_base += addregs; new_base <= high_range; new_base += addregs) |
| 1928 | { |
| 1929 | if (mask & (setmask << new_base)) |
| 1930 | { |
| 1931 | inst.error = _("invalid register list"); |
| 1932 | return FAIL; |
| 1933 | } |
| 1934 | |
| 1935 | mask |= setmask << new_base; |
| 1936 | count += addregs; |
| 1937 | } |
| 1938 | } |
| 1939 | } |
| 1940 | while (skip_past_comma (&str) != FAIL); |
| 1941 | |
| 1942 | str++; |
| 1943 | |
| 1944 | /* Sanity check -- should have raised a parse error above. */ |
| 1945 | if (count == 0 || count > max_regs) |
| 1946 | abort (); |
| 1947 | |
| 1948 | *pbase = base_reg; |
| 1949 | |
| 1950 | /* Final test -- the registers must be consecutive. */ |
| 1951 | mask >>= base_reg; |
| 1952 | for (i = 0; i < count; i++) |
| 1953 | { |
| 1954 | if ((mask & (1u << i)) == 0) |
| 1955 | { |
| 1956 | inst.error = _("non-contiguous register range"); |
| 1957 | return FAIL; |
| 1958 | } |
| 1959 | } |
| 1960 | |
| 1961 | *ccp = str; |
| 1962 | |
| 1963 | return count; |
| 1964 | } |
| 1965 | |
| 1966 | /* True if two alias types are the same. */ |
| 1967 | |
| 1968 | static bfd_boolean |
| 1969 | neon_alias_types_same (struct neon_typed_alias *a, struct neon_typed_alias *b) |
| 1970 | { |
| 1971 | if (!a && !b) |
| 1972 | return TRUE; |
| 1973 | |
| 1974 | if (!a || !b) |
| 1975 | return FALSE; |
| 1976 | |
| 1977 | if (a->defined != b->defined) |
| 1978 | return FALSE; |
| 1979 | |
| 1980 | if ((a->defined & NTA_HASTYPE) != 0 |
| 1981 | && (a->eltype.type != b->eltype.type |
| 1982 | || a->eltype.size != b->eltype.size)) |
| 1983 | return FALSE; |
| 1984 | |
| 1985 | if ((a->defined & NTA_HASINDEX) != 0 |
| 1986 | && (a->index != b->index)) |
| 1987 | return FALSE; |
| 1988 | |
| 1989 | return TRUE; |
| 1990 | } |
| 1991 | |
| 1992 | /* Parse element/structure lists for Neon VLD<n> and VST<n> instructions. |
| 1993 | The base register is put in *PBASE. |
| 1994 | The lane (or one of the NEON_*_LANES constants) is placed in bits [3:0] of |
| 1995 | the return value. |
| 1996 | The register stride (minus one) is put in bit 4 of the return value. |
| 1997 | Bits [6:5] encode the list length (minus one). |
| 1998 | The type of the list elements is put in *ELTYPE, if non-NULL. */ |
| 1999 | |
| 2000 | #define NEON_LANE(X) ((X) & 0xf) |
| 2001 | #define NEON_REG_STRIDE(X) ((((X) >> 4) & 1) + 1) |
| 2002 | #define NEON_REGLIST_LENGTH(X) ((((X) >> 5) & 3) + 1) |
| 2003 | |
| 2004 | static int |
| 2005 | parse_neon_el_struct_list (char **str, unsigned *pbase, |
| 2006 | struct neon_type_el *eltype) |
| 2007 | { |
| 2008 | char *ptr = *str; |
| 2009 | int base_reg = -1; |
| 2010 | int reg_incr = -1; |
| 2011 | int count = 0; |
| 2012 | int lane = -1; |
| 2013 | int leading_brace = 0; |
| 2014 | enum arm_reg_type rtype = REG_TYPE_NDQ; |
| 2015 | const char *const incr_error = _("register stride must be 1 or 2"); |
| 2016 | const char *const type_error = _("mismatched element/structure types in list"); |
| 2017 | struct neon_typed_alias firsttype; |
| 2018 | firsttype.defined = 0; |
| 2019 | firsttype.eltype.type = NT_invtype; |
| 2020 | firsttype.eltype.size = -1; |
| 2021 | firsttype.index = -1; |
| 2022 | |
| 2023 | if (skip_past_char (&ptr, '{') == SUCCESS) |
| 2024 | leading_brace = 1; |
| 2025 | |
| 2026 | do |
| 2027 | { |
| 2028 | struct neon_typed_alias atype; |
| 2029 | int getreg = parse_typed_reg_or_scalar (&ptr, rtype, &rtype, &atype); |
| 2030 | |
| 2031 | if (getreg == FAIL) |
| 2032 | { |
| 2033 | first_error (_(reg_expected_msgs[rtype])); |
| 2034 | return FAIL; |
| 2035 | } |
| 2036 | |
| 2037 | if (base_reg == -1) |
| 2038 | { |
| 2039 | base_reg = getreg; |
| 2040 | if (rtype == REG_TYPE_NQ) |
| 2041 | { |
| 2042 | reg_incr = 1; |
| 2043 | } |
| 2044 | firsttype = atype; |
| 2045 | } |
| 2046 | else if (reg_incr == -1) |
| 2047 | { |
| 2048 | reg_incr = getreg - base_reg; |
| 2049 | if (reg_incr < 1 || reg_incr > 2) |
| 2050 | { |
| 2051 | first_error (_(incr_error)); |
| 2052 | return FAIL; |
| 2053 | } |
| 2054 | } |
| 2055 | else if (getreg != base_reg + reg_incr * count) |
| 2056 | { |
| 2057 | first_error (_(incr_error)); |
| 2058 | return FAIL; |
| 2059 | } |
| 2060 | |
| 2061 | if (! neon_alias_types_same (&atype, &firsttype)) |
| 2062 | { |
| 2063 | first_error (_(type_error)); |
| 2064 | return FAIL; |
| 2065 | } |
| 2066 | |
| 2067 | /* Handle Dn-Dm or Qn-Qm syntax. Can only be used with non-indexed list |
| 2068 | modes. */ |
| 2069 | if (ptr[0] == '-') |
| 2070 | { |
| 2071 | struct neon_typed_alias htype; |
| 2072 | int hireg, dregs = (rtype == REG_TYPE_NQ) ? 2 : 1; |
| 2073 | if (lane == -1) |
| 2074 | lane = NEON_INTERLEAVE_LANES; |
| 2075 | else if (lane != NEON_INTERLEAVE_LANES) |
| 2076 | { |
| 2077 | first_error (_(type_error)); |
| 2078 | return FAIL; |
| 2079 | } |
| 2080 | if (reg_incr == -1) |
| 2081 | reg_incr = 1; |
| 2082 | else if (reg_incr != 1) |
| 2083 | { |
| 2084 | first_error (_("don't use Rn-Rm syntax with non-unit stride")); |
| 2085 | return FAIL; |
| 2086 | } |
| 2087 | ptr++; |
| 2088 | hireg = parse_typed_reg_or_scalar (&ptr, rtype, NULL, &htype); |
| 2089 | if (hireg == FAIL) |
| 2090 | { |
| 2091 | first_error (_(reg_expected_msgs[rtype])); |
| 2092 | return FAIL; |
| 2093 | } |
| 2094 | if (! neon_alias_types_same (&htype, &firsttype)) |
| 2095 | { |
| 2096 | first_error (_(type_error)); |
| 2097 | return FAIL; |
| 2098 | } |
| 2099 | count += hireg + dregs - getreg; |
| 2100 | continue; |
| 2101 | } |
| 2102 | |
| 2103 | /* If we're using Q registers, we can't use [] or [n] syntax. */ |
| 2104 | if (rtype == REG_TYPE_NQ) |
| 2105 | { |
| 2106 | count += 2; |
| 2107 | continue; |
| 2108 | } |
| 2109 | |
| 2110 | if ((atype.defined & NTA_HASINDEX) != 0) |
| 2111 | { |
| 2112 | if (lane == -1) |
| 2113 | lane = atype.index; |
| 2114 | else if (lane != atype.index) |
| 2115 | { |
| 2116 | first_error (_(type_error)); |
| 2117 | return FAIL; |
| 2118 | } |
| 2119 | } |
| 2120 | else if (lane == -1) |
| 2121 | lane = NEON_INTERLEAVE_LANES; |
| 2122 | else if (lane != NEON_INTERLEAVE_LANES) |
| 2123 | { |
| 2124 | first_error (_(type_error)); |
| 2125 | return FAIL; |
| 2126 | } |
| 2127 | count++; |
| 2128 | } |
| 2129 | while ((count != 1 || leading_brace) && skip_past_comma (&ptr) != FAIL); |
| 2130 | |
| 2131 | /* No lane set by [x]. We must be interleaving structures. */ |
| 2132 | if (lane == -1) |
| 2133 | lane = NEON_INTERLEAVE_LANES; |
| 2134 | |
| 2135 | /* Sanity check. */ |
| 2136 | if (lane == -1 || base_reg == -1 || count < 1 || count > 4 |
| 2137 | || (count > 1 && reg_incr == -1)) |
| 2138 | { |
| 2139 | first_error (_("error parsing element/structure list")); |
| 2140 | return FAIL; |
| 2141 | } |
| 2142 | |
| 2143 | if ((count > 1 || leading_brace) && skip_past_char (&ptr, '}') == FAIL) |
| 2144 | { |
| 2145 | first_error (_("expected }")); |
| 2146 | return FAIL; |
| 2147 | } |
| 2148 | |
| 2149 | if (reg_incr == -1) |
| 2150 | reg_incr = 1; |
| 2151 | |
| 2152 | if (eltype) |
| 2153 | *eltype = firsttype.eltype; |
| 2154 | |
| 2155 | *pbase = base_reg; |
| 2156 | *str = ptr; |
| 2157 | |
| 2158 | return lane | ((reg_incr - 1) << 4) | ((count - 1) << 5); |
| 2159 | } |
| 2160 | |
| 2161 | /* Parse an explicit relocation suffix on an expression. This is |
| 2162 | either nothing, or a word in parentheses. Note that if !OBJ_ELF, |
| 2163 | arm_reloc_hsh contains no entries, so this function can only |
| 2164 | succeed if there is no () after the word. Returns -1 on error, |
| 2165 | BFD_RELOC_UNUSED if there wasn't any suffix. */ |
| 2166 | |
| 2167 | static int |
| 2168 | parse_reloc (char **str) |
| 2169 | { |
| 2170 | struct reloc_entry *r; |
| 2171 | char *p, *q; |
| 2172 | |
| 2173 | if (**str != '(') |
| 2174 | return BFD_RELOC_UNUSED; |
| 2175 | |
| 2176 | p = *str + 1; |
| 2177 | q = p; |
| 2178 | |
| 2179 | while (*q && *q != ')' && *q != ',') |
| 2180 | q++; |
| 2181 | if (*q != ')') |
| 2182 | return -1; |
| 2183 | |
| 2184 | if ((r = (struct reloc_entry *) |
| 2185 | hash_find_n (arm_reloc_hsh, p, q - p)) == NULL) |
| 2186 | return -1; |
| 2187 | |
| 2188 | *str = q + 1; |
| 2189 | return r->reloc; |
| 2190 | } |
| 2191 | |
| 2192 | /* Directives: register aliases. */ |
| 2193 | |
| 2194 | static struct reg_entry * |
| 2195 | insert_reg_alias (char *str, unsigned number, int type) |
| 2196 | { |
| 2197 | struct reg_entry *new_reg; |
| 2198 | const char *name; |
| 2199 | |
| 2200 | if ((new_reg = (struct reg_entry *) hash_find (arm_reg_hsh, str)) != 0) |
| 2201 | { |
| 2202 | if (new_reg->builtin) |
| 2203 | as_warn (_("ignoring attempt to redefine built-in register '%s'"), str); |
| 2204 | |
| 2205 | /* Only warn about a redefinition if it's not defined as the |
| 2206 | same register. */ |
| 2207 | else if (new_reg->number != number || new_reg->type != type) |
| 2208 | as_warn (_("ignoring redefinition of register alias '%s'"), str); |
| 2209 | |
| 2210 | return NULL; |
| 2211 | } |
| 2212 | |
| 2213 | name = xstrdup (str); |
| 2214 | new_reg = XNEW (struct reg_entry); |
| 2215 | |
| 2216 | new_reg->name = name; |
| 2217 | new_reg->number = number; |
| 2218 | new_reg->type = type; |
| 2219 | new_reg->builtin = FALSE; |
| 2220 | new_reg->neon = NULL; |
| 2221 | |
| 2222 | if (hash_insert (arm_reg_hsh, name, (void *) new_reg)) |
| 2223 | abort (); |
| 2224 | |
| 2225 | return new_reg; |
| 2226 | } |
| 2227 | |
| 2228 | static void |
| 2229 | insert_neon_reg_alias (char *str, int number, int type, |
| 2230 | struct neon_typed_alias *atype) |
| 2231 | { |
| 2232 | struct reg_entry *reg = insert_reg_alias (str, number, type); |
| 2233 | |
| 2234 | if (!reg) |
| 2235 | { |
| 2236 | first_error (_("attempt to redefine typed alias")); |
| 2237 | return; |
| 2238 | } |
| 2239 | |
| 2240 | if (atype) |
| 2241 | { |
| 2242 | reg->neon = XNEW (struct neon_typed_alias); |
| 2243 | *reg->neon = *atype; |
| 2244 | } |
| 2245 | } |
| 2246 | |
| 2247 | /* Look for the .req directive. This is of the form: |
| 2248 | |
| 2249 | new_register_name .req existing_register_name |
| 2250 | |
| 2251 | If we find one, or if it looks sufficiently like one that we want to |
| 2252 | handle any error here, return TRUE. Otherwise return FALSE. */ |
| 2253 | |
| 2254 | static bfd_boolean |
| 2255 | create_register_alias (char * newname, char *p) |
| 2256 | { |
| 2257 | struct reg_entry *old; |
| 2258 | char *oldname, *nbuf; |
| 2259 | size_t nlen; |
| 2260 | |
| 2261 | /* The input scrubber ensures that whitespace after the mnemonic is |
| 2262 | collapsed to single spaces. */ |
| 2263 | oldname = p; |
| 2264 | if (strncmp (oldname, " .req ", 6) != 0) |
| 2265 | return FALSE; |
| 2266 | |
| 2267 | oldname += 6; |
| 2268 | if (*oldname == '\0') |
| 2269 | return FALSE; |
| 2270 | |
| 2271 | old = (struct reg_entry *) hash_find (arm_reg_hsh, oldname); |
| 2272 | if (!old) |
| 2273 | { |
| 2274 | as_warn (_("unknown register '%s' -- .req ignored"), oldname); |
| 2275 | return TRUE; |
| 2276 | } |
| 2277 | |
| 2278 | /* If TC_CASE_SENSITIVE is defined, then newname already points to |
| 2279 | the desired alias name, and p points to its end. If not, then |
| 2280 | the desired alias name is in the global original_case_string. */ |
| 2281 | #ifdef TC_CASE_SENSITIVE |
| 2282 | nlen = p - newname; |
| 2283 | #else |
| 2284 | newname = original_case_string; |
| 2285 | nlen = strlen (newname); |
| 2286 | #endif |
| 2287 | |
| 2288 | nbuf = xmemdup0 (newname, nlen); |
| 2289 | |
| 2290 | /* Create aliases under the new name as stated; an all-lowercase |
| 2291 | version of the new name; and an all-uppercase version of the new |
| 2292 | name. */ |
| 2293 | if (insert_reg_alias (nbuf, old->number, old->type) != NULL) |
| 2294 | { |
| 2295 | for (p = nbuf; *p; p++) |
| 2296 | *p = TOUPPER (*p); |
| 2297 | |
| 2298 | if (strncmp (nbuf, newname, nlen)) |
| 2299 | { |
| 2300 | /* If this attempt to create an additional alias fails, do not bother |
| 2301 | trying to create the all-lower case alias. We will fail and issue |
| 2302 | a second, duplicate error message. This situation arises when the |
| 2303 | programmer does something like: |
| 2304 | foo .req r0 |
| 2305 | Foo .req r1 |
| 2306 | The second .req creates the "Foo" alias but then fails to create |
| 2307 | the artificial FOO alias because it has already been created by the |
| 2308 | first .req. */ |
| 2309 | if (insert_reg_alias (nbuf, old->number, old->type) == NULL) |
| 2310 | { |
| 2311 | free (nbuf); |
| 2312 | return TRUE; |
| 2313 | } |
| 2314 | } |
| 2315 | |
| 2316 | for (p = nbuf; *p; p++) |
| 2317 | *p = TOLOWER (*p); |
| 2318 | |
| 2319 | if (strncmp (nbuf, newname, nlen)) |
| 2320 | insert_reg_alias (nbuf, old->number, old->type); |
| 2321 | } |
| 2322 | |
| 2323 | free (nbuf); |
| 2324 | return TRUE; |
| 2325 | } |
| 2326 | |
| 2327 | /* Create a Neon typed/indexed register alias using directives, e.g.: |
| 2328 | X .dn d5.s32[1] |
| 2329 | Y .qn 6.s16 |
| 2330 | Z .dn d7 |
| 2331 | T .dn Z[0] |
| 2332 | These typed registers can be used instead of the types specified after the |
| 2333 | Neon mnemonic, so long as all operands given have types. Types can also be |
| 2334 | specified directly, e.g.: |
| 2335 | vadd d0.s32, d1.s32, d2.s32 */ |
| 2336 | |
| 2337 | static bfd_boolean |
| 2338 | create_neon_reg_alias (char *newname, char *p) |
| 2339 | { |
| 2340 | enum arm_reg_type basetype; |
| 2341 | struct reg_entry *basereg; |
| 2342 | struct reg_entry mybasereg; |
| 2343 | struct neon_type ntype; |
| 2344 | struct neon_typed_alias typeinfo; |
| 2345 | char *namebuf, *nameend ATTRIBUTE_UNUSED; |
| 2346 | int namelen; |
| 2347 | |
| 2348 | typeinfo.defined = 0; |
| 2349 | typeinfo.eltype.type = NT_invtype; |
| 2350 | typeinfo.eltype.size = -1; |
| 2351 | typeinfo.index = -1; |
| 2352 | |
| 2353 | nameend = p; |
| 2354 | |
| 2355 | if (strncmp (p, " .dn ", 5) == 0) |
| 2356 | basetype = REG_TYPE_VFD; |
| 2357 | else if (strncmp (p, " .qn ", 5) == 0) |
| 2358 | basetype = REG_TYPE_NQ; |
| 2359 | else |
| 2360 | return FALSE; |
| 2361 | |
| 2362 | p += 5; |
| 2363 | |
| 2364 | if (*p == '\0') |
| 2365 | return FALSE; |
| 2366 | |
| 2367 | basereg = arm_reg_parse_multi (&p); |
| 2368 | |
| 2369 | if (basereg && basereg->type != basetype) |
| 2370 | { |
| 2371 | as_bad (_("bad type for register")); |
| 2372 | return FALSE; |
| 2373 | } |
| 2374 | |
| 2375 | if (basereg == NULL) |
| 2376 | { |
| 2377 | expressionS exp; |
| 2378 | /* Try parsing as an integer. */ |
| 2379 | my_get_expression (&exp, &p, GE_NO_PREFIX); |
| 2380 | if (exp.X_op != O_constant) |
| 2381 | { |
| 2382 | as_bad (_("expression must be constant")); |
| 2383 | return FALSE; |
| 2384 | } |
| 2385 | basereg = &mybasereg; |
| 2386 | basereg->number = (basetype == REG_TYPE_NQ) ? exp.X_add_number * 2 |
| 2387 | : exp.X_add_number; |
| 2388 | basereg->neon = 0; |
| 2389 | } |
| 2390 | |
| 2391 | if (basereg->neon) |
| 2392 | typeinfo = *basereg->neon; |
| 2393 | |
| 2394 | if (parse_neon_type (&ntype, &p) == SUCCESS) |
| 2395 | { |
| 2396 | /* We got a type. */ |
| 2397 | if (typeinfo.defined & NTA_HASTYPE) |
| 2398 | { |
| 2399 | as_bad (_("can't redefine the type of a register alias")); |
| 2400 | return FALSE; |
| 2401 | } |
| 2402 | |
| 2403 | typeinfo.defined |= NTA_HASTYPE; |
| 2404 | if (ntype.elems != 1) |
| 2405 | { |
| 2406 | as_bad (_("you must specify a single type only")); |
| 2407 | return FALSE; |
| 2408 | } |
| 2409 | typeinfo.eltype = ntype.el[0]; |
| 2410 | } |
| 2411 | |
| 2412 | if (skip_past_char (&p, '[') == SUCCESS) |
| 2413 | { |
| 2414 | expressionS exp; |
| 2415 | /* We got a scalar index. */ |
| 2416 | |
| 2417 | if (typeinfo.defined & NTA_HASINDEX) |
| 2418 | { |
| 2419 | as_bad (_("can't redefine the index of a scalar alias")); |
| 2420 | return FALSE; |
| 2421 | } |
| 2422 | |
| 2423 | my_get_expression (&exp, &p, GE_NO_PREFIX); |
| 2424 | |
| 2425 | if (exp.X_op != O_constant) |
| 2426 | { |
| 2427 | as_bad (_("scalar index must be constant")); |
| 2428 | return FALSE; |
| 2429 | } |
| 2430 | |
| 2431 | typeinfo.defined |= NTA_HASINDEX; |
| 2432 | typeinfo.index = exp.X_add_number; |
| 2433 | |
| 2434 | if (skip_past_char (&p, ']') == FAIL) |
| 2435 | { |
| 2436 | as_bad (_("expecting ]")); |
| 2437 | return FALSE; |
| 2438 | } |
| 2439 | } |
| 2440 | |
| 2441 | /* If TC_CASE_SENSITIVE is defined, then newname already points to |
| 2442 | the desired alias name, and p points to its end. If not, then |
| 2443 | the desired alias name is in the global original_case_string. */ |
| 2444 | #ifdef TC_CASE_SENSITIVE |
| 2445 | namelen = nameend - newname; |
| 2446 | #else |
| 2447 | newname = original_case_string; |
| 2448 | namelen = strlen (newname); |
| 2449 | #endif |
| 2450 | |
| 2451 | namebuf = xmemdup0 (newname, namelen); |
| 2452 | |
| 2453 | insert_neon_reg_alias (namebuf, basereg->number, basetype, |
| 2454 | typeinfo.defined != 0 ? &typeinfo : NULL); |
| 2455 | |
| 2456 | /* Insert name in all uppercase. */ |
| 2457 | for (p = namebuf; *p; p++) |
| 2458 | *p = TOUPPER (*p); |
| 2459 | |
| 2460 | if (strncmp (namebuf, newname, namelen)) |
| 2461 | insert_neon_reg_alias (namebuf, basereg->number, basetype, |
| 2462 | typeinfo.defined != 0 ? &typeinfo : NULL); |
| 2463 | |
| 2464 | /* Insert name in all lowercase. */ |
| 2465 | for (p = namebuf; *p; p++) |
| 2466 | *p = TOLOWER (*p); |
| 2467 | |
| 2468 | if (strncmp (namebuf, newname, namelen)) |
| 2469 | insert_neon_reg_alias (namebuf, basereg->number, basetype, |
| 2470 | typeinfo.defined != 0 ? &typeinfo : NULL); |
| 2471 | |
| 2472 | free (namebuf); |
| 2473 | return TRUE; |
| 2474 | } |
| 2475 | |
| 2476 | /* Should never be called, as .req goes between the alias and the |
| 2477 | register name, not at the beginning of the line. */ |
| 2478 | |
| 2479 | static void |
| 2480 | s_req (int a ATTRIBUTE_UNUSED) |
| 2481 | { |
| 2482 | as_bad (_("invalid syntax for .req directive")); |
| 2483 | } |
| 2484 | |
| 2485 | static void |
| 2486 | s_dn (int a ATTRIBUTE_UNUSED) |
| 2487 | { |
| 2488 | as_bad (_("invalid syntax for .dn directive")); |
| 2489 | } |
| 2490 | |
| 2491 | static void |
| 2492 | s_qn (int a ATTRIBUTE_UNUSED) |
| 2493 | { |
| 2494 | as_bad (_("invalid syntax for .qn directive")); |
| 2495 | } |
| 2496 | |
| 2497 | /* The .unreq directive deletes an alias which was previously defined |
| 2498 | by .req. For example: |
| 2499 | |
| 2500 | my_alias .req r11 |
| 2501 | .unreq my_alias */ |
| 2502 | |
| 2503 | static void |
| 2504 | s_unreq (int a ATTRIBUTE_UNUSED) |
| 2505 | { |
| 2506 | char * name; |
| 2507 | char saved_char; |
| 2508 | |
| 2509 | name = input_line_pointer; |
| 2510 | |
| 2511 | while (*input_line_pointer != 0 |
| 2512 | && *input_line_pointer != ' ' |
| 2513 | && *input_line_pointer != '\n') |
| 2514 | ++input_line_pointer; |
| 2515 | |
| 2516 | saved_char = *input_line_pointer; |
| 2517 | *input_line_pointer = 0; |
| 2518 | |
| 2519 | if (!*name) |
| 2520 | as_bad (_("invalid syntax for .unreq directive")); |
| 2521 | else |
| 2522 | { |
| 2523 | struct reg_entry *reg = (struct reg_entry *) hash_find (arm_reg_hsh, |
| 2524 | name); |
| 2525 | |
| 2526 | if (!reg) |
| 2527 | as_bad (_("unknown register alias '%s'"), name); |
| 2528 | else if (reg->builtin) |
| 2529 | as_warn (_("ignoring attempt to use .unreq on fixed register name: '%s'"), |
| 2530 | name); |
| 2531 | else |
| 2532 | { |
| 2533 | char * p; |
| 2534 | char * nbuf; |
| 2535 | |
| 2536 | hash_delete (arm_reg_hsh, name, FALSE); |
| 2537 | free ((char *) reg->name); |
| 2538 | if (reg->neon) |
| 2539 | free (reg->neon); |
| 2540 | free (reg); |
| 2541 | |
| 2542 | /* Also locate the all upper case and all lower case versions. |
| 2543 | Do not complain if we cannot find one or the other as it |
| 2544 | was probably deleted above. */ |
| 2545 | |
| 2546 | nbuf = strdup (name); |
| 2547 | for (p = nbuf; *p; p++) |
| 2548 | *p = TOUPPER (*p); |
| 2549 | reg = (struct reg_entry *) hash_find (arm_reg_hsh, nbuf); |
| 2550 | if (reg) |
| 2551 | { |
| 2552 | hash_delete (arm_reg_hsh, nbuf, FALSE); |
| 2553 | free ((char *) reg->name); |
| 2554 | if (reg->neon) |
| 2555 | free (reg->neon); |
| 2556 | free (reg); |
| 2557 | } |
| 2558 | |
| 2559 | for (p = nbuf; *p; p++) |
| 2560 | *p = TOLOWER (*p); |
| 2561 | reg = (struct reg_entry *) hash_find (arm_reg_hsh, nbuf); |
| 2562 | if (reg) |
| 2563 | { |
| 2564 | hash_delete (arm_reg_hsh, nbuf, FALSE); |
| 2565 | free ((char *) reg->name); |
| 2566 | if (reg->neon) |
| 2567 | free (reg->neon); |
| 2568 | free (reg); |
| 2569 | } |
| 2570 | |
| 2571 | free (nbuf); |
| 2572 | } |
| 2573 | } |
| 2574 | |
| 2575 | *input_line_pointer = saved_char; |
| 2576 | demand_empty_rest_of_line (); |
| 2577 | } |
| 2578 | |
| 2579 | /* Directives: Instruction set selection. */ |
| 2580 | |
| 2581 | #ifdef OBJ_ELF |
| 2582 | /* This code is to handle mapping symbols as defined in the ARM ELF spec. |
| 2583 | (See "Mapping symbols", section 4.5.5, ARM AAELF version 1.0). |
| 2584 | Note that previously, $a and $t has type STT_FUNC (BSF_OBJECT flag), |
| 2585 | and $d has type STT_OBJECT (BSF_OBJECT flag). Now all three are untyped. */ |
| 2586 | |
| 2587 | /* Create a new mapping symbol for the transition to STATE. */ |
| 2588 | |
| 2589 | static void |
| 2590 | make_mapping_symbol (enum mstate state, valueT value, fragS *frag) |
| 2591 | { |
| 2592 | symbolS * symbolP; |
| 2593 | const char * symname; |
| 2594 | int type; |
| 2595 | |
| 2596 | switch (state) |
| 2597 | { |
| 2598 | case MAP_DATA: |
| 2599 | symname = "$d"; |
| 2600 | type = BSF_NO_FLAGS; |
| 2601 | break; |
| 2602 | case MAP_ARM: |
| 2603 | symname = "$a"; |
| 2604 | type = BSF_NO_FLAGS; |
| 2605 | break; |
| 2606 | case MAP_THUMB: |
| 2607 | symname = "$t"; |
| 2608 | type = BSF_NO_FLAGS; |
| 2609 | break; |
| 2610 | default: |
| 2611 | abort (); |
| 2612 | } |
| 2613 | |
| 2614 | symbolP = symbol_new (symname, now_seg, value, frag); |
| 2615 | symbol_get_bfdsym (symbolP)->flags |= type | BSF_LOCAL; |
| 2616 | |
| 2617 | switch (state) |
| 2618 | { |
| 2619 | case MAP_ARM: |
| 2620 | THUMB_SET_FUNC (symbolP, 0); |
| 2621 | ARM_SET_THUMB (symbolP, 0); |
| 2622 | ARM_SET_INTERWORK (symbolP, support_interwork); |
| 2623 | break; |
| 2624 | |
| 2625 | case MAP_THUMB: |
| 2626 | THUMB_SET_FUNC (symbolP, 1); |
| 2627 | ARM_SET_THUMB (symbolP, 1); |
| 2628 | ARM_SET_INTERWORK (symbolP, support_interwork); |
| 2629 | break; |
| 2630 | |
| 2631 | case MAP_DATA: |
| 2632 | default: |
| 2633 | break; |
| 2634 | } |
| 2635 | |
| 2636 | /* Save the mapping symbols for future reference. Also check that |
| 2637 | we do not place two mapping symbols at the same offset within a |
| 2638 | frag. We'll handle overlap between frags in |
| 2639 | check_mapping_symbols. |
| 2640 | |
| 2641 | If .fill or other data filling directive generates zero sized data, |
| 2642 | the mapping symbol for the following code will have the same value |
| 2643 | as the one generated for the data filling directive. In this case, |
| 2644 | we replace the old symbol with the new one at the same address. */ |
| 2645 | if (value == 0) |
| 2646 | { |
| 2647 | if (frag->tc_frag_data.first_map != NULL) |
| 2648 | { |
| 2649 | know (S_GET_VALUE (frag->tc_frag_data.first_map) == 0); |
| 2650 | symbol_remove (frag->tc_frag_data.first_map, &symbol_rootP, &symbol_lastP); |
| 2651 | } |
| 2652 | frag->tc_frag_data.first_map = symbolP; |
| 2653 | } |
| 2654 | if (frag->tc_frag_data.last_map != NULL) |
| 2655 | { |
| 2656 | know (S_GET_VALUE (frag->tc_frag_data.last_map) <= S_GET_VALUE (symbolP)); |
| 2657 | if (S_GET_VALUE (frag->tc_frag_data.last_map) == S_GET_VALUE (symbolP)) |
| 2658 | symbol_remove (frag->tc_frag_data.last_map, &symbol_rootP, &symbol_lastP); |
| 2659 | } |
| 2660 | frag->tc_frag_data.last_map = symbolP; |
| 2661 | } |
| 2662 | |
| 2663 | /* We must sometimes convert a region marked as code to data during |
| 2664 | code alignment, if an odd number of bytes have to be padded. The |
| 2665 | code mapping symbol is pushed to an aligned address. */ |
| 2666 | |
| 2667 | static void |
| 2668 | insert_data_mapping_symbol (enum mstate state, |
| 2669 | valueT value, fragS *frag, offsetT bytes) |
| 2670 | { |
| 2671 | /* If there was already a mapping symbol, remove it. */ |
| 2672 | if (frag->tc_frag_data.last_map != NULL |
| 2673 | && S_GET_VALUE (frag->tc_frag_data.last_map) == frag->fr_address + value) |
| 2674 | { |
| 2675 | symbolS *symp = frag->tc_frag_data.last_map; |
| 2676 | |
| 2677 | if (value == 0) |
| 2678 | { |
| 2679 | know (frag->tc_frag_data.first_map == symp); |
| 2680 | frag->tc_frag_data.first_map = NULL; |
| 2681 | } |
| 2682 | frag->tc_frag_data.last_map = NULL; |
| 2683 | symbol_remove (symp, &symbol_rootP, &symbol_lastP); |
| 2684 | } |
| 2685 | |
| 2686 | make_mapping_symbol (MAP_DATA, value, frag); |
| 2687 | make_mapping_symbol (state, value + bytes, frag); |
| 2688 | } |
| 2689 | |
| 2690 | static void mapping_state_2 (enum mstate state, int max_chars); |
| 2691 | |
| 2692 | /* Set the mapping state to STATE. Only call this when about to |
| 2693 | emit some STATE bytes to the file. */ |
| 2694 | |
| 2695 | #define TRANSITION(from, to) (mapstate == (from) && state == (to)) |
| 2696 | void |
| 2697 | mapping_state (enum mstate state) |
| 2698 | { |
| 2699 | enum mstate mapstate = seg_info (now_seg)->tc_segment_info_data.mapstate; |
| 2700 | |
| 2701 | if (mapstate == state) |
| 2702 | /* The mapping symbol has already been emitted. |
| 2703 | There is nothing else to do. */ |
| 2704 | return; |
| 2705 | |
| 2706 | if (state == MAP_ARM || state == MAP_THUMB) |
| 2707 | /* PR gas/12931 |
| 2708 | All ARM instructions require 4-byte alignment. |
| 2709 | (Almost) all Thumb instructions require 2-byte alignment. |
| 2710 | |
| 2711 | When emitting instructions into any section, mark the section |
| 2712 | appropriately. |
| 2713 | |
| 2714 | Some Thumb instructions are alignment-sensitive modulo 4 bytes, |
| 2715 | but themselves require 2-byte alignment; this applies to some |
| 2716 | PC- relative forms. However, these cases will involve implicit |
| 2717 | literal pool generation or an explicit .align >=2, both of |
| 2718 | which will cause the section to me marked with sufficient |
| 2719 | alignment. Thus, we don't handle those cases here. */ |
| 2720 | record_alignment (now_seg, state == MAP_ARM ? 2 : 1); |
| 2721 | |
| 2722 | if (TRANSITION (MAP_UNDEFINED, MAP_DATA)) |
| 2723 | /* This case will be evaluated later. */ |
| 2724 | return; |
| 2725 | |
| 2726 | mapping_state_2 (state, 0); |
| 2727 | } |
| 2728 | |
| 2729 | /* Same as mapping_state, but MAX_CHARS bytes have already been |
| 2730 | allocated. Put the mapping symbol that far back. */ |
| 2731 | |
| 2732 | static void |
| 2733 | mapping_state_2 (enum mstate state, int max_chars) |
| 2734 | { |
| 2735 | enum mstate mapstate = seg_info (now_seg)->tc_segment_info_data.mapstate; |
| 2736 | |
| 2737 | if (!SEG_NORMAL (now_seg)) |
| 2738 | return; |
| 2739 | |
| 2740 | if (mapstate == state) |
| 2741 | /* The mapping symbol has already been emitted. |
| 2742 | There is nothing else to do. */ |
| 2743 | return; |
| 2744 | |
| 2745 | if (TRANSITION (MAP_UNDEFINED, MAP_ARM) |
| 2746 | || TRANSITION (MAP_UNDEFINED, MAP_THUMB)) |
| 2747 | { |
| 2748 | struct frag * const frag_first = seg_info (now_seg)->frchainP->frch_root; |
| 2749 | const int add_symbol = (frag_now != frag_first) || (frag_now_fix () > 0); |
| 2750 | |
| 2751 | if (add_symbol) |
| 2752 | make_mapping_symbol (MAP_DATA, (valueT) 0, frag_first); |
| 2753 | } |
| 2754 | |
| 2755 | seg_info (now_seg)->tc_segment_info_data.mapstate = state; |
| 2756 | make_mapping_symbol (state, (valueT) frag_now_fix () - max_chars, frag_now); |
| 2757 | } |
| 2758 | #undef TRANSITION |
| 2759 | #else |
| 2760 | #define mapping_state(x) ((void)0) |
| 2761 | #define mapping_state_2(x, y) ((void)0) |
| 2762 | #endif |
| 2763 | |
| 2764 | /* Find the real, Thumb encoded start of a Thumb function. */ |
| 2765 | |
| 2766 | #ifdef OBJ_COFF |
| 2767 | static symbolS * |
| 2768 | find_real_start (symbolS * symbolP) |
| 2769 | { |
| 2770 | char * real_start; |
| 2771 | const char * name = S_GET_NAME (symbolP); |
| 2772 | symbolS * new_target; |
| 2773 | |
| 2774 | /* This definition must agree with the one in gcc/config/arm/thumb.c. */ |
| 2775 | #define STUB_NAME ".real_start_of" |
| 2776 | |
| 2777 | if (name == NULL) |
| 2778 | abort (); |
| 2779 | |
| 2780 | /* The compiler may generate BL instructions to local labels because |
| 2781 | it needs to perform a branch to a far away location. These labels |
| 2782 | do not have a corresponding ".real_start_of" label. We check |
| 2783 | both for S_IS_LOCAL and for a leading dot, to give a way to bypass |
| 2784 | the ".real_start_of" convention for nonlocal branches. */ |
| 2785 | if (S_IS_LOCAL (symbolP) || name[0] == '.') |
| 2786 | return symbolP; |
| 2787 | |
| 2788 | real_start = concat (STUB_NAME, name, NULL); |
| 2789 | new_target = symbol_find (real_start); |
| 2790 | free (real_start); |
| 2791 | |
| 2792 | if (new_target == NULL) |
| 2793 | { |
| 2794 | as_warn (_("Failed to find real start of function: %s\n"), name); |
| 2795 | new_target = symbolP; |
| 2796 | } |
| 2797 | |
| 2798 | return new_target; |
| 2799 | } |
| 2800 | #endif |
| 2801 | |
| 2802 | static void |
| 2803 | opcode_select (int width) |
| 2804 | { |
| 2805 | switch (width) |
| 2806 | { |
| 2807 | case 16: |
| 2808 | if (! thumb_mode) |
| 2809 | { |
| 2810 | if (!ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v4t)) |
| 2811 | as_bad (_("selected processor does not support THUMB opcodes")); |
| 2812 | |
| 2813 | thumb_mode = 1; |
| 2814 | /* No need to force the alignment, since we will have been |
| 2815 | coming from ARM mode, which is word-aligned. */ |
| 2816 | record_alignment (now_seg, 1); |
| 2817 | } |
| 2818 | break; |
| 2819 | |
| 2820 | case 32: |
| 2821 | if (thumb_mode) |
| 2822 | { |
| 2823 | if (!ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v1)) |
| 2824 | as_bad (_("selected processor does not support ARM opcodes")); |
| 2825 | |
| 2826 | thumb_mode = 0; |
| 2827 | |
| 2828 | if (!need_pass_2) |
| 2829 | frag_align (2, 0, 0); |
| 2830 | |
| 2831 | record_alignment (now_seg, 1); |
| 2832 | } |
| 2833 | break; |
| 2834 | |
| 2835 | default: |
| 2836 | as_bad (_("invalid instruction size selected (%d)"), width); |
| 2837 | } |
| 2838 | } |
| 2839 | |
| 2840 | static void |
| 2841 | s_arm (int ignore ATTRIBUTE_UNUSED) |
| 2842 | { |
| 2843 | opcode_select (32); |
| 2844 | demand_empty_rest_of_line (); |
| 2845 | } |
| 2846 | |
| 2847 | static void |
| 2848 | s_thumb (int ignore ATTRIBUTE_UNUSED) |
| 2849 | { |
| 2850 | opcode_select (16); |
| 2851 | demand_empty_rest_of_line (); |
| 2852 | } |
| 2853 | |
| 2854 | static void |
| 2855 | s_code (int unused ATTRIBUTE_UNUSED) |
| 2856 | { |
| 2857 | int temp; |
| 2858 | |
| 2859 | temp = get_absolute_expression (); |
| 2860 | switch (temp) |
| 2861 | { |
| 2862 | case 16: |
| 2863 | case 32: |
| 2864 | opcode_select (temp); |
| 2865 | break; |
| 2866 | |
| 2867 | default: |
| 2868 | as_bad (_("invalid operand to .code directive (%d) (expecting 16 or 32)"), temp); |
| 2869 | } |
| 2870 | } |
| 2871 | |
| 2872 | static void |
| 2873 | s_force_thumb (int ignore ATTRIBUTE_UNUSED) |
| 2874 | { |
| 2875 | /* If we are not already in thumb mode go into it, EVEN if |
| 2876 | the target processor does not support thumb instructions. |
| 2877 | This is used by gcc/config/arm/lib1funcs.asm for example |
| 2878 | to compile interworking support functions even if the |
| 2879 | target processor should not support interworking. */ |
| 2880 | if (! thumb_mode) |
| 2881 | { |
| 2882 | thumb_mode = 2; |
| 2883 | record_alignment (now_seg, 1); |
| 2884 | } |
| 2885 | |
| 2886 | demand_empty_rest_of_line (); |
| 2887 | } |
| 2888 | |
| 2889 | static void |
| 2890 | s_thumb_func (int ignore ATTRIBUTE_UNUSED) |
| 2891 | { |
| 2892 | s_thumb (0); |
| 2893 | |
| 2894 | /* The following label is the name/address of the start of a Thumb function. |
| 2895 | We need to know this for the interworking support. */ |
| 2896 | label_is_thumb_function_name = TRUE; |
| 2897 | } |
| 2898 | |
| 2899 | /* Perform a .set directive, but also mark the alias as |
| 2900 | being a thumb function. */ |
| 2901 | |
| 2902 | static void |
| 2903 | s_thumb_set (int equiv) |
| 2904 | { |
| 2905 | /* XXX the following is a duplicate of the code for s_set() in read.c |
| 2906 | We cannot just call that code as we need to get at the symbol that |
| 2907 | is created. */ |
| 2908 | char * name; |
| 2909 | char delim; |
| 2910 | char * end_name; |
| 2911 | symbolS * symbolP; |
| 2912 | |
| 2913 | /* Especial apologies for the random logic: |
| 2914 | This just grew, and could be parsed much more simply! |
| 2915 | Dean - in haste. */ |
| 2916 | delim = get_symbol_name (& name); |
| 2917 | end_name = input_line_pointer; |
| 2918 | (void) restore_line_pointer (delim); |
| 2919 | |
| 2920 | if (*input_line_pointer != ',') |
| 2921 | { |
| 2922 | *end_name = 0; |
| 2923 | as_bad (_("expected comma after name \"%s\""), name); |
| 2924 | *end_name = delim; |
| 2925 | ignore_rest_of_line (); |
| 2926 | return; |
| 2927 | } |
| 2928 | |
| 2929 | input_line_pointer++; |
| 2930 | *end_name = 0; |
| 2931 | |
| 2932 | if (name[0] == '.' && name[1] == '\0') |
| 2933 | { |
| 2934 | /* XXX - this should not happen to .thumb_set. */ |
| 2935 | abort (); |
| 2936 | } |
| 2937 | |
| 2938 | if ((symbolP = symbol_find (name)) == NULL |
| 2939 | && (symbolP = md_undefined_symbol (name)) == NULL) |
| 2940 | { |
| 2941 | #ifndef NO_LISTING |
| 2942 | /* When doing symbol listings, play games with dummy fragments living |
| 2943 | outside the normal fragment chain to record the file and line info |
| 2944 | for this symbol. */ |
| 2945 | if (listing & LISTING_SYMBOLS) |
| 2946 | { |
| 2947 | extern struct list_info_struct * listing_tail; |
| 2948 | fragS * dummy_frag = (fragS * ) xmalloc (sizeof (fragS)); |
| 2949 | |
| 2950 | memset (dummy_frag, 0, sizeof (fragS)); |
| 2951 | dummy_frag->fr_type = rs_fill; |
| 2952 | dummy_frag->line = listing_tail; |
| 2953 | symbolP = symbol_new (name, undefined_section, 0, dummy_frag); |
| 2954 | dummy_frag->fr_symbol = symbolP; |
| 2955 | } |
| 2956 | else |
| 2957 | #endif |
| 2958 | symbolP = symbol_new (name, undefined_section, 0, &zero_address_frag); |
| 2959 | |
| 2960 | #ifdef OBJ_COFF |
| 2961 | /* "set" symbols are local unless otherwise specified. */ |
| 2962 | SF_SET_LOCAL (symbolP); |
| 2963 | #endif /* OBJ_COFF */ |
| 2964 | } /* Make a new symbol. */ |
| 2965 | |
| 2966 | symbol_table_insert (symbolP); |
| 2967 | |
| 2968 | * end_name = delim; |
| 2969 | |
| 2970 | if (equiv |
| 2971 | && S_IS_DEFINED (symbolP) |
| 2972 | && S_GET_SEGMENT (symbolP) != reg_section) |
| 2973 | as_bad (_("symbol `%s' already defined"), S_GET_NAME (symbolP)); |
| 2974 | |
| 2975 | pseudo_set (symbolP); |
| 2976 | |
| 2977 | demand_empty_rest_of_line (); |
| 2978 | |
| 2979 | /* XXX Now we come to the Thumb specific bit of code. */ |
| 2980 | |
| 2981 | THUMB_SET_FUNC (symbolP, 1); |
| 2982 | ARM_SET_THUMB (symbolP, 1); |
| 2983 | #if defined OBJ_ELF || defined OBJ_COFF |
| 2984 | ARM_SET_INTERWORK (symbolP, support_interwork); |
| 2985 | #endif |
| 2986 | } |
| 2987 | |
| 2988 | /* Directives: Mode selection. */ |
| 2989 | |
| 2990 | /* .syntax [unified|divided] - choose the new unified syntax |
| 2991 | (same for Arm and Thumb encoding, modulo slight differences in what |
| 2992 | can be represented) or the old divergent syntax for each mode. */ |
| 2993 | static void |
| 2994 | s_syntax (int unused ATTRIBUTE_UNUSED) |
| 2995 | { |
| 2996 | char *name, delim; |
| 2997 | |
| 2998 | delim = get_symbol_name (& name); |
| 2999 | |
| 3000 | if (!strcasecmp (name, "unified")) |
| 3001 | unified_syntax = TRUE; |
| 3002 | else if (!strcasecmp (name, "divided")) |
| 3003 | unified_syntax = FALSE; |
| 3004 | else |
| 3005 | { |
| 3006 | as_bad (_("unrecognized syntax mode \"%s\""), name); |
| 3007 | return; |
| 3008 | } |
| 3009 | (void) restore_line_pointer (delim); |
| 3010 | demand_empty_rest_of_line (); |
| 3011 | } |
| 3012 | |
| 3013 | /* Directives: sectioning and alignment. */ |
| 3014 | |
| 3015 | static void |
| 3016 | s_bss (int ignore ATTRIBUTE_UNUSED) |
| 3017 | { |
| 3018 | /* We don't support putting frags in the BSS segment, we fake it by |
| 3019 | marking in_bss, then looking at s_skip for clues. */ |
| 3020 | subseg_set (bss_section, 0); |
| 3021 | demand_empty_rest_of_line (); |
| 3022 | |
| 3023 | #ifdef md_elf_section_change_hook |
| 3024 | md_elf_section_change_hook (); |
| 3025 | #endif |
| 3026 | } |
| 3027 | |
| 3028 | static void |
| 3029 | s_even (int ignore ATTRIBUTE_UNUSED) |
| 3030 | { |
| 3031 | /* Never make frag if expect extra pass. */ |
| 3032 | if (!need_pass_2) |
| 3033 | frag_align (1, 0, 0); |
| 3034 | |
| 3035 | record_alignment (now_seg, 1); |
| 3036 | |
| 3037 | demand_empty_rest_of_line (); |
| 3038 | } |
| 3039 | |
| 3040 | /* Directives: CodeComposer Studio. */ |
| 3041 | |
| 3042 | /* .ref (for CodeComposer Studio syntax only). */ |
| 3043 | static void |
| 3044 | s_ccs_ref (int unused ATTRIBUTE_UNUSED) |
| 3045 | { |
| 3046 | if (codecomposer_syntax) |
| 3047 | ignore_rest_of_line (); |
| 3048 | else |
| 3049 | as_bad (_(".ref pseudo-op only available with -mccs flag.")); |
| 3050 | } |
| 3051 | |
| 3052 | /* If name is not NULL, then it is used for marking the beginning of a |
| 3053 | function, whereas if it is NULL then it means the function end. */ |
| 3054 | static void |
| 3055 | asmfunc_debug (const char * name) |
| 3056 | { |
| 3057 | static const char * last_name = NULL; |
| 3058 | |
| 3059 | if (name != NULL) |
| 3060 | { |
| 3061 | gas_assert (last_name == NULL); |
| 3062 | last_name = name; |
| 3063 | |
| 3064 | if (debug_type == DEBUG_STABS) |
| 3065 | stabs_generate_asm_func (name, name); |
| 3066 | } |
| 3067 | else |
| 3068 | { |
| 3069 | gas_assert (last_name != NULL); |
| 3070 | |
| 3071 | if (debug_type == DEBUG_STABS) |
| 3072 | stabs_generate_asm_endfunc (last_name, last_name); |
| 3073 | |
| 3074 | last_name = NULL; |
| 3075 | } |
| 3076 | } |
| 3077 | |
| 3078 | static void |
| 3079 | s_ccs_asmfunc (int unused ATTRIBUTE_UNUSED) |
| 3080 | { |
| 3081 | if (codecomposer_syntax) |
| 3082 | { |
| 3083 | switch (asmfunc_state) |
| 3084 | { |
| 3085 | case OUTSIDE_ASMFUNC: |
| 3086 | asmfunc_state = WAITING_ASMFUNC_NAME; |
| 3087 | break; |
| 3088 | |
| 3089 | case WAITING_ASMFUNC_NAME: |
| 3090 | as_bad (_(".asmfunc repeated.")); |
| 3091 | break; |
| 3092 | |
| 3093 | case WAITING_ENDASMFUNC: |
| 3094 | as_bad (_(".asmfunc without function.")); |
| 3095 | break; |
| 3096 | } |
| 3097 | demand_empty_rest_of_line (); |
| 3098 | } |
| 3099 | else |
| 3100 | as_bad (_(".asmfunc pseudo-op only available with -mccs flag.")); |
| 3101 | } |
| 3102 | |
| 3103 | static void |
| 3104 | s_ccs_endasmfunc (int unused ATTRIBUTE_UNUSED) |
| 3105 | { |
| 3106 | if (codecomposer_syntax) |
| 3107 | { |
| 3108 | switch (asmfunc_state) |
| 3109 | { |
| 3110 | case OUTSIDE_ASMFUNC: |
| 3111 | as_bad (_(".endasmfunc without a .asmfunc.")); |
| 3112 | break; |
| 3113 | |
| 3114 | case WAITING_ASMFUNC_NAME: |
| 3115 | as_bad (_(".endasmfunc without function.")); |
| 3116 | break; |
| 3117 | |
| 3118 | case WAITING_ENDASMFUNC: |
| 3119 | asmfunc_state = OUTSIDE_ASMFUNC; |
| 3120 | asmfunc_debug (NULL); |
| 3121 | break; |
| 3122 | } |
| 3123 | demand_empty_rest_of_line (); |
| 3124 | } |
| 3125 | else |
| 3126 | as_bad (_(".endasmfunc pseudo-op only available with -mccs flag.")); |
| 3127 | } |
| 3128 | |
| 3129 | static void |
| 3130 | s_ccs_def (int name) |
| 3131 | { |
| 3132 | if (codecomposer_syntax) |
| 3133 | s_globl (name); |
| 3134 | else |
| 3135 | as_bad (_(".def pseudo-op only available with -mccs flag.")); |
| 3136 | } |
| 3137 | |
| 3138 | /* Directives: Literal pools. */ |
| 3139 | |
| 3140 | static literal_pool * |
| 3141 | find_literal_pool (void) |
| 3142 | { |
| 3143 | literal_pool * pool; |
| 3144 | |
| 3145 | for (pool = list_of_pools; pool != NULL; pool = pool->next) |
| 3146 | { |
| 3147 | if (pool->section == now_seg |
| 3148 | && pool->sub_section == now_subseg) |
| 3149 | break; |
| 3150 | } |
| 3151 | |
| 3152 | return pool; |
| 3153 | } |
| 3154 | |
| 3155 | static literal_pool * |
| 3156 | find_or_make_literal_pool (void) |
| 3157 | { |
| 3158 | /* Next literal pool ID number. */ |
| 3159 | static unsigned int latest_pool_num = 1; |
| 3160 | literal_pool * pool; |
| 3161 | |
| 3162 | pool = find_literal_pool (); |
| 3163 | |
| 3164 | if (pool == NULL) |
| 3165 | { |
| 3166 | /* Create a new pool. */ |
| 3167 | pool = XNEW (literal_pool); |
| 3168 | if (! pool) |
| 3169 | return NULL; |
| 3170 | |
| 3171 | pool->next_free_entry = 0; |
| 3172 | pool->section = now_seg; |
| 3173 | pool->sub_section = now_subseg; |
| 3174 | pool->next = list_of_pools; |
| 3175 | pool->symbol = NULL; |
| 3176 | pool->alignment = 2; |
| 3177 | |
| 3178 | /* Add it to the list. */ |
| 3179 | list_of_pools = pool; |
| 3180 | } |
| 3181 | |
| 3182 | /* New pools, and emptied pools, will have a NULL symbol. */ |
| 3183 | if (pool->symbol == NULL) |
| 3184 | { |
| 3185 | pool->symbol = symbol_create (FAKE_LABEL_NAME, undefined_section, |
| 3186 | (valueT) 0, &zero_address_frag); |
| 3187 | pool->id = latest_pool_num ++; |
| 3188 | } |
| 3189 | |
| 3190 | /* Done. */ |
| 3191 | return pool; |
| 3192 | } |
| 3193 | |
| 3194 | /* Add the literal in the global 'inst' |
| 3195 | structure to the relevant literal pool. */ |
| 3196 | |
| 3197 | static int |
| 3198 | add_to_lit_pool (unsigned int nbytes) |
| 3199 | { |
| 3200 | #define PADDING_SLOT 0x1 |
| 3201 | #define LIT_ENTRY_SIZE_MASK 0xFF |
| 3202 | literal_pool * pool; |
| 3203 | unsigned int entry, pool_size = 0; |
| 3204 | bfd_boolean padding_slot_p = FALSE; |
| 3205 | unsigned imm1 = 0; |
| 3206 | unsigned imm2 = 0; |
| 3207 | |
| 3208 | if (nbytes == 8) |
| 3209 | { |
| 3210 | imm1 = inst.operands[1].imm; |
| 3211 | imm2 = (inst.operands[1].regisimm ? inst.operands[1].reg |
| 3212 | : inst.reloc.exp.X_unsigned ? 0 |
| 3213 | : ((bfd_int64_t) inst.operands[1].imm) >> 32); |
| 3214 | if (target_big_endian) |
| 3215 | { |
| 3216 | imm1 = imm2; |
| 3217 | imm2 = inst.operands[1].imm; |
| 3218 | } |
| 3219 | } |
| 3220 | |
| 3221 | pool = find_or_make_literal_pool (); |
| 3222 | |
| 3223 | /* Check if this literal value is already in the pool. */ |
| 3224 | for (entry = 0; entry < pool->next_free_entry; entry ++) |
| 3225 | { |
| 3226 | if (nbytes == 4) |
| 3227 | { |
| 3228 | if ((pool->literals[entry].X_op == inst.reloc.exp.X_op) |
| 3229 | && (inst.reloc.exp.X_op == O_constant) |
| 3230 | && (pool->literals[entry].X_add_number |
| 3231 | == inst.reloc.exp.X_add_number) |
| 3232 | && (pool->literals[entry].X_md == nbytes) |
| 3233 | && (pool->literals[entry].X_unsigned |
| 3234 | == inst.reloc.exp.X_unsigned)) |
| 3235 | break; |
| 3236 | |
| 3237 | if ((pool->literals[entry].X_op == inst.reloc.exp.X_op) |
| 3238 | && (inst.reloc.exp.X_op == O_symbol) |
| 3239 | && (pool->literals[entry].X_add_number |
| 3240 | == inst.reloc.exp.X_add_number) |
| 3241 | && (pool->literals[entry].X_add_symbol |
| 3242 | == inst.reloc.exp.X_add_symbol) |
| 3243 | && (pool->literals[entry].X_op_symbol |
| 3244 | == inst.reloc.exp.X_op_symbol) |
| 3245 | && (pool->literals[entry].X_md == nbytes)) |
| 3246 | break; |
| 3247 | } |
| 3248 | else if ((nbytes == 8) |
| 3249 | && !(pool_size & 0x7) |
| 3250 | && ((entry + 1) != pool->next_free_entry) |
| 3251 | && (pool->literals[entry].X_op == O_constant) |
| 3252 | && (pool->literals[entry].X_add_number == (offsetT) imm1) |
| 3253 | && (pool->literals[entry].X_unsigned |
| 3254 | == inst.reloc.exp.X_unsigned) |
| 3255 | && (pool->literals[entry + 1].X_op == O_constant) |
| 3256 | && (pool->literals[entry + 1].X_add_number == (offsetT) imm2) |
| 3257 | && (pool->literals[entry + 1].X_unsigned |
| 3258 | == inst.reloc.exp.X_unsigned)) |
| 3259 | break; |
| 3260 | |
| 3261 | padding_slot_p = ((pool->literals[entry].X_md >> 8) == PADDING_SLOT); |
| 3262 | if (padding_slot_p && (nbytes == 4)) |
| 3263 | break; |
| 3264 | |
| 3265 | pool_size += 4; |
| 3266 | } |
| 3267 | |
| 3268 | /* Do we need to create a new entry? */ |
| 3269 | if (entry == pool->next_free_entry) |
| 3270 | { |
| 3271 | if (entry >= MAX_LITERAL_POOL_SIZE) |
| 3272 | { |
| 3273 | inst.error = _("literal pool overflow"); |
| 3274 | return FAIL; |
| 3275 | } |
| 3276 | |
| 3277 | if (nbytes == 8) |
| 3278 | { |
| 3279 | /* For 8-byte entries, we align to an 8-byte boundary, |
| 3280 | and split it into two 4-byte entries, because on 32-bit |
| 3281 | host, 8-byte constants are treated as big num, thus |
| 3282 | saved in "generic_bignum" which will be overwritten |
| 3283 | by later assignments. |
| 3284 | |
| 3285 | We also need to make sure there is enough space for |
| 3286 | the split. |
| 3287 | |
| 3288 | We also check to make sure the literal operand is a |
| 3289 | constant number. */ |
| 3290 | if (!(inst.reloc.exp.X_op == O_constant |
| 3291 | || inst.reloc.exp.X_op == O_big)) |
| 3292 | { |
| 3293 | inst.error = _("invalid type for literal pool"); |
| 3294 | return FAIL; |
| 3295 | } |
| 3296 | else if (pool_size & 0x7) |
| 3297 | { |
| 3298 | if ((entry + 2) >= MAX_LITERAL_POOL_SIZE) |
| 3299 | { |
| 3300 | inst.error = _("literal pool overflow"); |
| 3301 | return FAIL; |
| 3302 | } |
| 3303 | |
| 3304 | pool->literals[entry] = inst.reloc.exp; |
| 3305 | pool->literals[entry].X_op = O_constant; |
| 3306 | pool->literals[entry].X_add_number = 0; |
| 3307 | pool->literals[entry++].X_md = (PADDING_SLOT << 8) | 4; |
| 3308 | pool->next_free_entry += 1; |
| 3309 | pool_size += 4; |
| 3310 | } |
| 3311 | else if ((entry + 1) >= MAX_LITERAL_POOL_SIZE) |
| 3312 | { |
| 3313 | inst.error = _("literal pool overflow"); |
| 3314 | return FAIL; |
| 3315 | } |
| 3316 | |
| 3317 | pool->literals[entry] = inst.reloc.exp; |
| 3318 | pool->literals[entry].X_op = O_constant; |
| 3319 | pool->literals[entry].X_add_number = imm1; |
| 3320 | pool->literals[entry].X_unsigned = inst.reloc.exp.X_unsigned; |
| 3321 | pool->literals[entry++].X_md = 4; |
| 3322 | pool->literals[entry] = inst.reloc.exp; |
| 3323 | pool->literals[entry].X_op = O_constant; |
| 3324 | pool->literals[entry].X_add_number = imm2; |
| 3325 | pool->literals[entry].X_unsigned = inst.reloc.exp.X_unsigned; |
| 3326 | pool->literals[entry].X_md = 4; |
| 3327 | pool->alignment = 3; |
| 3328 | pool->next_free_entry += 1; |
| 3329 | } |
| 3330 | else |
| 3331 | { |
| 3332 | pool->literals[entry] = inst.reloc.exp; |
| 3333 | pool->literals[entry].X_md = 4; |
| 3334 | } |
| 3335 | |
| 3336 | #ifdef OBJ_ELF |
| 3337 | /* PR ld/12974: Record the location of the first source line to reference |
| 3338 | this entry in the literal pool. If it turns out during linking that the |
| 3339 | symbol does not exist we will be able to give an accurate line number for |
| 3340 | the (first use of the) missing reference. */ |
| 3341 | if (debug_type == DEBUG_DWARF2) |
| 3342 | dwarf2_where (pool->locs + entry); |
| 3343 | #endif |
| 3344 | pool->next_free_entry += 1; |
| 3345 | } |
| 3346 | else if (padding_slot_p) |
| 3347 | { |
| 3348 | pool->literals[entry] = inst.reloc.exp; |
| 3349 | pool->literals[entry].X_md = nbytes; |
| 3350 | } |
| 3351 | |
| 3352 | inst.reloc.exp.X_op = O_symbol; |
| 3353 | inst.reloc.exp.X_add_number = pool_size; |
| 3354 | inst.reloc.exp.X_add_symbol = pool->symbol; |
| 3355 | |
| 3356 | return SUCCESS; |
| 3357 | } |
| 3358 | |
| 3359 | bfd_boolean |
| 3360 | tc_start_label_without_colon (void) |
| 3361 | { |
| 3362 | bfd_boolean ret = TRUE; |
| 3363 | |
| 3364 | if (codecomposer_syntax && asmfunc_state == WAITING_ASMFUNC_NAME) |
| 3365 | { |
| 3366 | const char *label = input_line_pointer; |
| 3367 | |
| 3368 | while (!is_end_of_line[(int) label[-1]]) |
| 3369 | --label; |
| 3370 | |
| 3371 | if (*label == '.') |
| 3372 | { |
| 3373 | as_bad (_("Invalid label '%s'"), label); |
| 3374 | ret = FALSE; |
| 3375 | } |
| 3376 | |
| 3377 | asmfunc_debug (label); |
| 3378 | |
| 3379 | asmfunc_state = WAITING_ENDASMFUNC; |
| 3380 | } |
| 3381 | |
| 3382 | return ret; |
| 3383 | } |
| 3384 | |
| 3385 | /* Can't use symbol_new here, so have to create a symbol and then at |
| 3386 | a later date assign it a value. That's what these functions do. */ |
| 3387 | |
| 3388 | static void |
| 3389 | symbol_locate (symbolS * symbolP, |
| 3390 | const char * name, /* It is copied, the caller can modify. */ |
| 3391 | segT segment, /* Segment identifier (SEG_<something>). */ |
| 3392 | valueT valu, /* Symbol value. */ |
| 3393 | fragS * frag) /* Associated fragment. */ |
| 3394 | { |
| 3395 | size_t name_length; |
| 3396 | char * preserved_copy_of_name; |
| 3397 | |
| 3398 | name_length = strlen (name) + 1; /* +1 for \0. */ |
| 3399 | obstack_grow (¬es, name, name_length); |
| 3400 | preserved_copy_of_name = (char *) obstack_finish (¬es); |
| 3401 | |
| 3402 | #ifdef tc_canonicalize_symbol_name |
| 3403 | preserved_copy_of_name = |
| 3404 | tc_canonicalize_symbol_name (preserved_copy_of_name); |
| 3405 | #endif |
| 3406 | |
| 3407 | S_SET_NAME (symbolP, preserved_copy_of_name); |
| 3408 | |
| 3409 | S_SET_SEGMENT (symbolP, segment); |
| 3410 | S_SET_VALUE (symbolP, valu); |
| 3411 | symbol_clear_list_pointers (symbolP); |
| 3412 | |
| 3413 | symbol_set_frag (symbolP, frag); |
| 3414 | |
| 3415 | /* Link to end of symbol chain. */ |
| 3416 | { |
| 3417 | extern int symbol_table_frozen; |
| 3418 | |
| 3419 | if (symbol_table_frozen) |
| 3420 | abort (); |
| 3421 | } |
| 3422 | |
| 3423 | symbol_append (symbolP, symbol_lastP, & symbol_rootP, & symbol_lastP); |
| 3424 | |
| 3425 | obj_symbol_new_hook (symbolP); |
| 3426 | |
| 3427 | #ifdef tc_symbol_new_hook |
| 3428 | tc_symbol_new_hook (symbolP); |
| 3429 | #endif |
| 3430 | |
| 3431 | #ifdef DEBUG_SYMS |
| 3432 | verify_symbol_chain (symbol_rootP, symbol_lastP); |
| 3433 | #endif /* DEBUG_SYMS */ |
| 3434 | } |
| 3435 | |
| 3436 | static void |
| 3437 | s_ltorg (int ignored ATTRIBUTE_UNUSED) |
| 3438 | { |
| 3439 | unsigned int entry; |
| 3440 | literal_pool * pool; |
| 3441 | char sym_name[20]; |
| 3442 | |
| 3443 | pool = find_literal_pool (); |
| 3444 | if (pool == NULL |
| 3445 | || pool->symbol == NULL |
| 3446 | || pool->next_free_entry == 0) |
| 3447 | return; |
| 3448 | |
| 3449 | /* Align pool as you have word accesses. |
| 3450 | Only make a frag if we have to. */ |
| 3451 | if (!need_pass_2) |
| 3452 | frag_align (pool->alignment, 0, 0); |
| 3453 | |
| 3454 | record_alignment (now_seg, 2); |
| 3455 | |
| 3456 | #ifdef OBJ_ELF |
| 3457 | seg_info (now_seg)->tc_segment_info_data.mapstate = MAP_DATA; |
| 3458 | make_mapping_symbol (MAP_DATA, (valueT) frag_now_fix (), frag_now); |
| 3459 | #endif |
| 3460 | sprintf (sym_name, "$$lit_\002%x", pool->id); |
| 3461 | |
| 3462 | symbol_locate (pool->symbol, sym_name, now_seg, |
| 3463 | (valueT) frag_now_fix (), frag_now); |
| 3464 | symbol_table_insert (pool->symbol); |
| 3465 | |
| 3466 | ARM_SET_THUMB (pool->symbol, thumb_mode); |
| 3467 | |
| 3468 | #if defined OBJ_COFF || defined OBJ_ELF |
| 3469 | ARM_SET_INTERWORK (pool->symbol, support_interwork); |
| 3470 | #endif |
| 3471 | |
| 3472 | for (entry = 0; entry < pool->next_free_entry; entry ++) |
| 3473 | { |
| 3474 | #ifdef OBJ_ELF |
| 3475 | if (debug_type == DEBUG_DWARF2) |
| 3476 | dwarf2_gen_line_info (frag_now_fix (), pool->locs + entry); |
| 3477 | #endif |
| 3478 | /* First output the expression in the instruction to the pool. */ |
| 3479 | emit_expr (&(pool->literals[entry]), |
| 3480 | pool->literals[entry].X_md & LIT_ENTRY_SIZE_MASK); |
| 3481 | } |
| 3482 | |
| 3483 | /* Mark the pool as empty. */ |
| 3484 | pool->next_free_entry = 0; |
| 3485 | pool->symbol = NULL; |
| 3486 | } |
| 3487 | |
| 3488 | #ifdef OBJ_ELF |
| 3489 | /* Forward declarations for functions below, in the MD interface |
| 3490 | section. */ |
| 3491 | static void fix_new_arm (fragS *, int, short, expressionS *, int, int); |
| 3492 | static valueT create_unwind_entry (int); |
| 3493 | static void start_unwind_section (const segT, int); |
| 3494 | static void add_unwind_opcode (valueT, int); |
| 3495 | static void flush_pending_unwind (void); |
| 3496 | |
| 3497 | /* Directives: Data. */ |
| 3498 | |
| 3499 | static void |
| 3500 | s_arm_elf_cons (int nbytes) |
| 3501 | { |
| 3502 | expressionS exp; |
| 3503 | |
| 3504 | #ifdef md_flush_pending_output |
| 3505 | md_flush_pending_output (); |
| 3506 | #endif |
| 3507 | |
| 3508 | if (is_it_end_of_statement ()) |
| 3509 | { |
| 3510 | demand_empty_rest_of_line (); |
| 3511 | return; |
| 3512 | } |
| 3513 | |
| 3514 | #ifdef md_cons_align |
| 3515 | md_cons_align (nbytes); |
| 3516 | #endif |
| 3517 | |
| 3518 | mapping_state (MAP_DATA); |
| 3519 | do |
| 3520 | { |
| 3521 | int reloc; |
| 3522 | char *base = input_line_pointer; |
| 3523 | |
| 3524 | expression (& exp); |
| 3525 | |
| 3526 | if (exp.X_op != O_symbol) |
| 3527 | emit_expr (&exp, (unsigned int) nbytes); |
| 3528 | else |
| 3529 | { |
| 3530 | char *before_reloc = input_line_pointer; |
| 3531 | reloc = parse_reloc (&input_line_pointer); |
| 3532 | if (reloc == -1) |
| 3533 | { |
| 3534 | as_bad (_("unrecognized relocation suffix")); |
| 3535 | ignore_rest_of_line (); |
| 3536 | return; |
| 3537 | } |
| 3538 | else if (reloc == BFD_RELOC_UNUSED) |
| 3539 | emit_expr (&exp, (unsigned int) nbytes); |
| 3540 | else |
| 3541 | { |
| 3542 | reloc_howto_type *howto = (reloc_howto_type *) |
| 3543 | bfd_reloc_type_lookup (stdoutput, |
| 3544 | (bfd_reloc_code_real_type) reloc); |
| 3545 | int size = bfd_get_reloc_size (howto); |
| 3546 | |
| 3547 | if (reloc == BFD_RELOC_ARM_PLT32) |
| 3548 | { |
| 3549 | as_bad (_("(plt) is only valid on branch targets")); |
| 3550 | reloc = BFD_RELOC_UNUSED; |
| 3551 | size = 0; |
| 3552 | } |
| 3553 | |
| 3554 | if (size > nbytes) |
| 3555 | as_bad (_("%s relocations do not fit in %d bytes"), |
| 3556 | howto->name, nbytes); |
| 3557 | else |
| 3558 | { |
| 3559 | /* We've parsed an expression stopping at O_symbol. |
| 3560 | But there may be more expression left now that we |
| 3561 | have parsed the relocation marker. Parse it again. |
| 3562 | XXX Surely there is a cleaner way to do this. */ |
| 3563 | char *p = input_line_pointer; |
| 3564 | int offset; |
| 3565 | char *save_buf = XNEWVEC (char, input_line_pointer - base); |
| 3566 | |
| 3567 | memcpy (save_buf, base, input_line_pointer - base); |
| 3568 | memmove (base + (input_line_pointer - before_reloc), |
| 3569 | base, before_reloc - base); |
| 3570 | |
| 3571 | input_line_pointer = base + (input_line_pointer-before_reloc); |
| 3572 | expression (&exp); |
| 3573 | memcpy (base, save_buf, p - base); |
| 3574 | |
| 3575 | offset = nbytes - size; |
| 3576 | p = frag_more (nbytes); |
| 3577 | memset (p, 0, nbytes); |
| 3578 | fix_new_exp (frag_now, p - frag_now->fr_literal + offset, |
| 3579 | size, &exp, 0, (enum bfd_reloc_code_real) reloc); |
| 3580 | free (save_buf); |
| 3581 | } |
| 3582 | } |
| 3583 | } |
| 3584 | } |
| 3585 | while (*input_line_pointer++ == ','); |
| 3586 | |
| 3587 | /* Put terminator back into stream. */ |
| 3588 | input_line_pointer --; |
| 3589 | demand_empty_rest_of_line (); |
| 3590 | } |
| 3591 | |
| 3592 | /* Emit an expression containing a 32-bit thumb instruction. |
| 3593 | Implementation based on put_thumb32_insn. */ |
| 3594 | |
| 3595 | static void |
| 3596 | emit_thumb32_expr (expressionS * exp) |
| 3597 | { |
| 3598 | expressionS exp_high = *exp; |
| 3599 | |
| 3600 | exp_high.X_add_number = (unsigned long)exp_high.X_add_number >> 16; |
| 3601 | emit_expr (& exp_high, (unsigned int) THUMB_SIZE); |
| 3602 | exp->X_add_number &= 0xffff; |
| 3603 | emit_expr (exp, (unsigned int) THUMB_SIZE); |
| 3604 | } |
| 3605 | |
| 3606 | /* Guess the instruction size based on the opcode. */ |
| 3607 | |
| 3608 | static int |
| 3609 | thumb_insn_size (int opcode) |
| 3610 | { |
| 3611 | if ((unsigned int) opcode < 0xe800u) |
| 3612 | return 2; |
| 3613 | else if ((unsigned int) opcode >= 0xe8000000u) |
| 3614 | return 4; |
| 3615 | else |
| 3616 | return 0; |
| 3617 | } |
| 3618 | |
| 3619 | static bfd_boolean |
| 3620 | emit_insn (expressionS *exp, int nbytes) |
| 3621 | { |
| 3622 | int size = 0; |
| 3623 | |
| 3624 | if (exp->X_op == O_constant) |
| 3625 | { |
| 3626 | size = nbytes; |
| 3627 | |
| 3628 | if (size == 0) |
| 3629 | size = thumb_insn_size (exp->X_add_number); |
| 3630 | |
| 3631 | if (size != 0) |
| 3632 | { |
| 3633 | if (size == 2 && (unsigned int)exp->X_add_number > 0xffffu) |
| 3634 | { |
| 3635 | as_bad (_(".inst.n operand too big. "\ |
| 3636 | "Use .inst.w instead")); |
| 3637 | size = 0; |
| 3638 | } |
| 3639 | else |
| 3640 | { |
| 3641 | if (now_it.state == AUTOMATIC_IT_BLOCK) |
| 3642 | set_it_insn_type_nonvoid (OUTSIDE_IT_INSN, 0); |
| 3643 | else |
| 3644 | set_it_insn_type_nonvoid (NEUTRAL_IT_INSN, 0); |
| 3645 | |
| 3646 | if (thumb_mode && (size > THUMB_SIZE) && !target_big_endian) |
| 3647 | emit_thumb32_expr (exp); |
| 3648 | else |
| 3649 | emit_expr (exp, (unsigned int) size); |
| 3650 | |
| 3651 | it_fsm_post_encode (); |
| 3652 | } |
| 3653 | } |
| 3654 | else |
| 3655 | as_bad (_("cannot determine Thumb instruction size. " \ |
| 3656 | "Use .inst.n/.inst.w instead")); |
| 3657 | } |
| 3658 | else |
| 3659 | as_bad (_("constant expression required")); |
| 3660 | |
| 3661 | return (size != 0); |
| 3662 | } |
| 3663 | |
| 3664 | /* Like s_arm_elf_cons but do not use md_cons_align and |
| 3665 | set the mapping state to MAP_ARM/MAP_THUMB. */ |
| 3666 | |
| 3667 | static void |
| 3668 | s_arm_elf_inst (int nbytes) |
| 3669 | { |
| 3670 | if (is_it_end_of_statement ()) |
| 3671 | { |
| 3672 | demand_empty_rest_of_line (); |
| 3673 | return; |
| 3674 | } |
| 3675 | |
| 3676 | /* Calling mapping_state () here will not change ARM/THUMB, |
| 3677 | but will ensure not to be in DATA state. */ |
| 3678 | |
| 3679 | if (thumb_mode) |
| 3680 | mapping_state (MAP_THUMB); |
| 3681 | else |
| 3682 | { |
| 3683 | if (nbytes != 0) |
| 3684 | { |
| 3685 | as_bad (_("width suffixes are invalid in ARM mode")); |
| 3686 | ignore_rest_of_line (); |
| 3687 | return; |
| 3688 | } |
| 3689 | |
| 3690 | nbytes = 4; |
| 3691 | |
| 3692 | mapping_state (MAP_ARM); |
| 3693 | } |
| 3694 | |
| 3695 | do |
| 3696 | { |
| 3697 | expressionS exp; |
| 3698 | |
| 3699 | expression (& exp); |
| 3700 | |
| 3701 | if (! emit_insn (& exp, nbytes)) |
| 3702 | { |
| 3703 | ignore_rest_of_line (); |
| 3704 | return; |
| 3705 | } |
| 3706 | } |
| 3707 | while (*input_line_pointer++ == ','); |
| 3708 | |
| 3709 | /* Put terminator back into stream. */ |
| 3710 | input_line_pointer --; |
| 3711 | demand_empty_rest_of_line (); |
| 3712 | } |
| 3713 | |
| 3714 | /* Parse a .rel31 directive. */ |
| 3715 | |
| 3716 | static void |
| 3717 | s_arm_rel31 (int ignored ATTRIBUTE_UNUSED) |
| 3718 | { |
| 3719 | expressionS exp; |
| 3720 | char *p; |
| 3721 | valueT highbit; |
| 3722 | |
| 3723 | highbit = 0; |
| 3724 | if (*input_line_pointer == '1') |
| 3725 | highbit = 0x80000000; |
| 3726 | else if (*input_line_pointer != '0') |
| 3727 | as_bad (_("expected 0 or 1")); |
| 3728 | |
| 3729 | input_line_pointer++; |
| 3730 | if (*input_line_pointer != ',') |
| 3731 | as_bad (_("missing comma")); |
| 3732 | input_line_pointer++; |
| 3733 | |
| 3734 | #ifdef md_flush_pending_output |
| 3735 | md_flush_pending_output (); |
| 3736 | #endif |
| 3737 | |
| 3738 | #ifdef md_cons_align |
| 3739 | md_cons_align (4); |
| 3740 | #endif |
| 3741 | |
| 3742 | mapping_state (MAP_DATA); |
| 3743 | |
| 3744 | expression (&exp); |
| 3745 | |
| 3746 | p = frag_more (4); |
| 3747 | md_number_to_chars (p, highbit, 4); |
| 3748 | fix_new_arm (frag_now, p - frag_now->fr_literal, 4, &exp, 1, |
| 3749 | BFD_RELOC_ARM_PREL31); |
| 3750 | |
| 3751 | demand_empty_rest_of_line (); |
| 3752 | } |
| 3753 | |
| 3754 | /* Directives: AEABI stack-unwind tables. */ |
| 3755 | |
| 3756 | /* Parse an unwind_fnstart directive. Simply records the current location. */ |
| 3757 | |
| 3758 | static void |
| 3759 | s_arm_unwind_fnstart (int ignored ATTRIBUTE_UNUSED) |
| 3760 | { |
| 3761 | demand_empty_rest_of_line (); |
| 3762 | if (unwind.proc_start) |
| 3763 | { |
| 3764 | as_bad (_("duplicate .fnstart directive")); |
| 3765 | return; |
| 3766 | } |
| 3767 | |
| 3768 | /* Mark the start of the function. */ |
| 3769 | unwind.proc_start = expr_build_dot (); |
| 3770 | |
| 3771 | /* Reset the rest of the unwind info. */ |
| 3772 | unwind.opcode_count = 0; |
| 3773 | unwind.table_entry = NULL; |
| 3774 | unwind.personality_routine = NULL; |
| 3775 | unwind.personality_index = -1; |
| 3776 | unwind.frame_size = 0; |
| 3777 | unwind.fp_offset = 0; |
| 3778 | unwind.fp_reg = REG_SP; |
| 3779 | unwind.fp_used = 0; |
| 3780 | unwind.sp_restored = 0; |
| 3781 | } |
| 3782 | |
| 3783 | |
| 3784 | /* Parse a handlerdata directive. Creates the exception handling table entry |
| 3785 | for the function. */ |
| 3786 | |
| 3787 | static void |
| 3788 | s_arm_unwind_handlerdata (int ignored ATTRIBUTE_UNUSED) |
| 3789 | { |
| 3790 | demand_empty_rest_of_line (); |
| 3791 | if (!unwind.proc_start) |
| 3792 | as_bad (MISSING_FNSTART); |
| 3793 | |
| 3794 | if (unwind.table_entry) |
| 3795 | as_bad (_("duplicate .handlerdata directive")); |
| 3796 | |
| 3797 | create_unwind_entry (1); |
| 3798 | } |
| 3799 | |
| 3800 | /* Parse an unwind_fnend directive. Generates the index table entry. */ |
| 3801 | |
| 3802 | static void |
| 3803 | s_arm_unwind_fnend (int ignored ATTRIBUTE_UNUSED) |
| 3804 | { |
| 3805 | long where; |
| 3806 | char *ptr; |
| 3807 | valueT val; |
| 3808 | unsigned int marked_pr_dependency; |
| 3809 | |
| 3810 | demand_empty_rest_of_line (); |
| 3811 | |
| 3812 | if (!unwind.proc_start) |
| 3813 | { |
| 3814 | as_bad (_(".fnend directive without .fnstart")); |
| 3815 | return; |
| 3816 | } |
| 3817 | |
| 3818 | /* Add eh table entry. */ |
| 3819 | if (unwind.table_entry == NULL) |
| 3820 | val = create_unwind_entry (0); |
| 3821 | else |
| 3822 | val = 0; |
| 3823 | |
| 3824 | /* Add index table entry. This is two words. */ |
| 3825 | start_unwind_section (unwind.saved_seg, 1); |
| 3826 | frag_align (2, 0, 0); |
| 3827 | record_alignment (now_seg, 2); |
| 3828 | |
| 3829 | ptr = frag_more (8); |
| 3830 | memset (ptr, 0, 8); |
| 3831 | where = frag_now_fix () - 8; |
| 3832 | |
| 3833 | /* Self relative offset of the function start. */ |
| 3834 | fix_new (frag_now, where, 4, unwind.proc_start, 0, 1, |
| 3835 | BFD_RELOC_ARM_PREL31); |
| 3836 | |
| 3837 | /* Indicate dependency on EHABI-defined personality routines to the |
| 3838 | linker, if it hasn't been done already. */ |
| 3839 | marked_pr_dependency |
| 3840 | = seg_info (now_seg)->tc_segment_info_data.marked_pr_dependency; |
| 3841 | if (unwind.personality_index >= 0 && unwind.personality_index < 3 |
| 3842 | && !(marked_pr_dependency & (1 << unwind.personality_index))) |
| 3843 | { |
| 3844 | static const char *const name[] = |
| 3845 | { |
| 3846 | "__aeabi_unwind_cpp_pr0", |
| 3847 | "__aeabi_unwind_cpp_pr1", |
| 3848 | "__aeabi_unwind_cpp_pr2" |
| 3849 | }; |
| 3850 | symbolS *pr = symbol_find_or_make (name[unwind.personality_index]); |
| 3851 | fix_new (frag_now, where, 0, pr, 0, 1, BFD_RELOC_NONE); |
| 3852 | seg_info (now_seg)->tc_segment_info_data.marked_pr_dependency |
| 3853 | |= 1 << unwind.personality_index; |
| 3854 | } |
| 3855 | |
| 3856 | if (val) |
| 3857 | /* Inline exception table entry. */ |
| 3858 | md_number_to_chars (ptr + 4, val, 4); |
| 3859 | else |
| 3860 | /* Self relative offset of the table entry. */ |
| 3861 | fix_new (frag_now, where + 4, 4, unwind.table_entry, 0, 1, |
| 3862 | BFD_RELOC_ARM_PREL31); |
| 3863 | |
| 3864 | /* Restore the original section. */ |
| 3865 | subseg_set (unwind.saved_seg, unwind.saved_subseg); |
| 3866 | |
| 3867 | unwind.proc_start = NULL; |
| 3868 | } |
| 3869 | |
| 3870 | |
| 3871 | /* Parse an unwind_cantunwind directive. */ |
| 3872 | |
| 3873 | static void |
| 3874 | s_arm_unwind_cantunwind (int ignored ATTRIBUTE_UNUSED) |
| 3875 | { |
| 3876 | demand_empty_rest_of_line (); |
| 3877 | if (!unwind.proc_start) |
| 3878 | as_bad (MISSING_FNSTART); |
| 3879 | |
| 3880 | if (unwind.personality_routine || unwind.personality_index != -1) |
| 3881 | as_bad (_("personality routine specified for cantunwind frame")); |
| 3882 | |
| 3883 | unwind.personality_index = -2; |
| 3884 | } |
| 3885 | |
| 3886 | |
| 3887 | /* Parse a personalityindex directive. */ |
| 3888 | |
| 3889 | static void |
| 3890 | s_arm_unwind_personalityindex (int ignored ATTRIBUTE_UNUSED) |
| 3891 | { |
| 3892 | expressionS exp; |
| 3893 | |
| 3894 | if (!unwind.proc_start) |
| 3895 | as_bad (MISSING_FNSTART); |
| 3896 | |
| 3897 | if (unwind.personality_routine || unwind.personality_index != -1) |
| 3898 | as_bad (_("duplicate .personalityindex directive")); |
| 3899 | |
| 3900 | expression (&exp); |
| 3901 | |
| 3902 | if (exp.X_op != O_constant |
| 3903 | || exp.X_add_number < 0 || exp.X_add_number > 15) |
| 3904 | { |
| 3905 | as_bad (_("bad personality routine number")); |
| 3906 | ignore_rest_of_line (); |
| 3907 | return; |
| 3908 | } |
| 3909 | |
| 3910 | unwind.personality_index = exp.X_add_number; |
| 3911 | |
| 3912 | demand_empty_rest_of_line (); |
| 3913 | } |
| 3914 | |
| 3915 | |
| 3916 | /* Parse a personality directive. */ |
| 3917 | |
| 3918 | static void |
| 3919 | s_arm_unwind_personality (int ignored ATTRIBUTE_UNUSED) |
| 3920 | { |
| 3921 | char *name, *p, c; |
| 3922 | |
| 3923 | if (!unwind.proc_start) |
| 3924 | as_bad (MISSING_FNSTART); |
| 3925 | |
| 3926 | if (unwind.personality_routine || unwind.personality_index != -1) |
| 3927 | as_bad (_("duplicate .personality directive")); |
| 3928 | |
| 3929 | c = get_symbol_name (& name); |
| 3930 | p = input_line_pointer; |
| 3931 | if (c == '"') |
| 3932 | ++ input_line_pointer; |
| 3933 | unwind.personality_routine = symbol_find_or_make (name); |
| 3934 | *p = c; |
| 3935 | demand_empty_rest_of_line (); |
| 3936 | } |
| 3937 | |
| 3938 | |
| 3939 | /* Parse a directive saving core registers. */ |
| 3940 | |
| 3941 | static void |
| 3942 | s_arm_unwind_save_core (void) |
| 3943 | { |
| 3944 | valueT op; |
| 3945 | long range; |
| 3946 | int n; |
| 3947 | |
| 3948 | range = parse_reg_list (&input_line_pointer); |
| 3949 | if (range == FAIL) |
| 3950 | { |
| 3951 | as_bad (_("expected register list")); |
| 3952 | ignore_rest_of_line (); |
| 3953 | return; |
| 3954 | } |
| 3955 | |
| 3956 | demand_empty_rest_of_line (); |
| 3957 | |
| 3958 | /* Turn .unwind_movsp ip followed by .unwind_save {..., ip, ...} |
| 3959 | into .unwind_save {..., sp...}. We aren't bothered about the value of |
| 3960 | ip because it is clobbered by calls. */ |
| 3961 | if (unwind.sp_restored && unwind.fp_reg == 12 |
| 3962 | && (range & 0x3000) == 0x1000) |
| 3963 | { |
| 3964 | unwind.opcode_count--; |
| 3965 | unwind.sp_restored = 0; |
| 3966 | range = (range | 0x2000) & ~0x1000; |
| 3967 | unwind.pending_offset = 0; |
| 3968 | } |
| 3969 | |
| 3970 | /* Pop r4-r15. */ |
| 3971 | if (range & 0xfff0) |
| 3972 | { |
| 3973 | /* See if we can use the short opcodes. These pop a block of up to 8 |
| 3974 | registers starting with r4, plus maybe r14. */ |
| 3975 | for (n = 0; n < 8; n++) |
| 3976 | { |
| 3977 | /* Break at the first non-saved register. */ |
| 3978 | if ((range & (1 << (n + 4))) == 0) |
| 3979 | break; |
| 3980 | } |
| 3981 | /* See if there are any other bits set. */ |
| 3982 | if (n == 0 || (range & (0xfff0 << n) & 0xbff0) != 0) |
| 3983 | { |
| 3984 | /* Use the long form. */ |
| 3985 | op = 0x8000 | ((range >> 4) & 0xfff); |
| 3986 | add_unwind_opcode (op, 2); |
| 3987 | } |
| 3988 | else |
| 3989 | { |
| 3990 | /* Use the short form. */ |
| 3991 | if (range & 0x4000) |
| 3992 | op = 0xa8; /* Pop r14. */ |
| 3993 | else |
| 3994 | op = 0xa0; /* Do not pop r14. */ |
| 3995 | op |= (n - 1); |
| 3996 | add_unwind_opcode (op, 1); |
| 3997 | } |
| 3998 | } |
| 3999 | |
| 4000 | /* Pop r0-r3. */ |
| 4001 | if (range & 0xf) |
| 4002 | { |
| 4003 | op = 0xb100 | (range & 0xf); |
| 4004 | add_unwind_opcode (op, 2); |
| 4005 | } |
| 4006 | |
| 4007 | /* Record the number of bytes pushed. */ |
| 4008 | for (n = 0; n < 16; n++) |
| 4009 | { |
| 4010 | if (range & (1 << n)) |
| 4011 | unwind.frame_size += 4; |
| 4012 | } |
| 4013 | } |
| 4014 | |
| 4015 | |
| 4016 | /* Parse a directive saving FPA registers. */ |
| 4017 | |
| 4018 | static void |
| 4019 | s_arm_unwind_save_fpa (int reg) |
| 4020 | { |
| 4021 | expressionS exp; |
| 4022 | int num_regs; |
| 4023 | valueT op; |
| 4024 | |
| 4025 | /* Get Number of registers to transfer. */ |
| 4026 | if (skip_past_comma (&input_line_pointer) != FAIL) |
| 4027 | expression (&exp); |
| 4028 | else |
| 4029 | exp.X_op = O_illegal; |
| 4030 | |
| 4031 | if (exp.X_op != O_constant) |
| 4032 | { |
| 4033 | as_bad (_("expected , <constant>")); |
| 4034 | ignore_rest_of_line (); |
| 4035 | return; |
| 4036 | } |
| 4037 | |
| 4038 | num_regs = exp.X_add_number; |
| 4039 | |
| 4040 | if (num_regs < 1 || num_regs > 4) |
| 4041 | { |
| 4042 | as_bad (_("number of registers must be in the range [1:4]")); |
| 4043 | ignore_rest_of_line (); |
| 4044 | return; |
| 4045 | } |
| 4046 | |
| 4047 | demand_empty_rest_of_line (); |
| 4048 | |
| 4049 | if (reg == 4) |
| 4050 | { |
| 4051 | /* Short form. */ |
| 4052 | op = 0xb4 | (num_regs - 1); |
| 4053 | add_unwind_opcode (op, 1); |
| 4054 | } |
| 4055 | else |
| 4056 | { |
| 4057 | /* Long form. */ |
| 4058 | op = 0xc800 | (reg << 4) | (num_regs - 1); |
| 4059 | add_unwind_opcode (op, 2); |
| 4060 | } |
| 4061 | unwind.frame_size += num_regs * 12; |
| 4062 | } |
| 4063 | |
| 4064 | |
| 4065 | /* Parse a directive saving VFP registers for ARMv6 and above. */ |
| 4066 | |
| 4067 | static void |
| 4068 | s_arm_unwind_save_vfp_armv6 (void) |
| 4069 | { |
| 4070 | int count; |
| 4071 | unsigned int start; |
| 4072 | valueT op; |
| 4073 | int num_vfpv3_regs = 0; |
| 4074 | int num_regs_below_16; |
| 4075 | |
| 4076 | count = parse_vfp_reg_list (&input_line_pointer, &start, REGLIST_VFP_D); |
| 4077 | if (count == FAIL) |
| 4078 | { |
| 4079 | as_bad (_("expected register list")); |
| 4080 | ignore_rest_of_line (); |
| 4081 | return; |
| 4082 | } |
| 4083 | |
| 4084 | demand_empty_rest_of_line (); |
| 4085 | |
| 4086 | /* We always generate FSTMD/FLDMD-style unwinding opcodes (rather |
| 4087 | than FSTMX/FLDMX-style ones). */ |
| 4088 | |
| 4089 | /* Generate opcode for (VFPv3) registers numbered in the range 16 .. 31. */ |
| 4090 | if (start >= 16) |
| 4091 | num_vfpv3_regs = count; |
| 4092 | else if (start + count > 16) |
| 4093 | num_vfpv3_regs = start + count - 16; |
| 4094 | |
| 4095 | if (num_vfpv3_regs > 0) |
| 4096 | { |
| 4097 | int start_offset = start > 16 ? start - 16 : 0; |
| 4098 | op = 0xc800 | (start_offset << 4) | (num_vfpv3_regs - 1); |
| 4099 | add_unwind_opcode (op, 2); |
| 4100 | } |
| 4101 | |
| 4102 | /* Generate opcode for registers numbered in the range 0 .. 15. */ |
| 4103 | num_regs_below_16 = num_vfpv3_regs > 0 ? 16 - (int) start : count; |
| 4104 | gas_assert (num_regs_below_16 + num_vfpv3_regs == count); |
| 4105 | if (num_regs_below_16 > 0) |
| 4106 | { |
| 4107 | op = 0xc900 | (start << 4) | (num_regs_below_16 - 1); |
| 4108 | add_unwind_opcode (op, 2); |
| 4109 | } |
| 4110 | |
| 4111 | unwind.frame_size += count * 8; |
| 4112 | } |
| 4113 | |
| 4114 | |
| 4115 | /* Parse a directive saving VFP registers for pre-ARMv6. */ |
| 4116 | |
| 4117 | static void |
| 4118 | s_arm_unwind_save_vfp (void) |
| 4119 | { |
| 4120 | int count; |
| 4121 | unsigned int reg; |
| 4122 | valueT op; |
| 4123 | |
| 4124 | count = parse_vfp_reg_list (&input_line_pointer, ®, REGLIST_VFP_D); |
| 4125 | if (count == FAIL) |
| 4126 | { |
| 4127 | as_bad (_("expected register list")); |
| 4128 | ignore_rest_of_line (); |
| 4129 | return; |
| 4130 | } |
| 4131 | |
| 4132 | demand_empty_rest_of_line (); |
| 4133 | |
| 4134 | if (reg == 8) |
| 4135 | { |
| 4136 | /* Short form. */ |
| 4137 | op = 0xb8 | (count - 1); |
| 4138 | add_unwind_opcode (op, 1); |
| 4139 | } |
| 4140 | else |
| 4141 | { |
| 4142 | /* Long form. */ |
| 4143 | op = 0xb300 | (reg << 4) | (count - 1); |
| 4144 | add_unwind_opcode (op, 2); |
| 4145 | } |
| 4146 | unwind.frame_size += count * 8 + 4; |
| 4147 | } |
| 4148 | |
| 4149 | |
| 4150 | /* Parse a directive saving iWMMXt data registers. */ |
| 4151 | |
| 4152 | static void |
| 4153 | s_arm_unwind_save_mmxwr (void) |
| 4154 | { |
| 4155 | int reg; |
| 4156 | int hi_reg; |
| 4157 | int i; |
| 4158 | unsigned mask = 0; |
| 4159 | valueT op; |
| 4160 | |
| 4161 | if (*input_line_pointer == '{') |
| 4162 | input_line_pointer++; |
| 4163 | |
| 4164 | do |
| 4165 | { |
| 4166 | reg = arm_reg_parse (&input_line_pointer, REG_TYPE_MMXWR); |
| 4167 | |
| 4168 | if (reg == FAIL) |
| 4169 | { |
| 4170 | as_bad ("%s", _(reg_expected_msgs[REG_TYPE_MMXWR])); |
| 4171 | goto error; |
| 4172 | } |
| 4173 | |
| 4174 | if (mask >> reg) |
| 4175 | as_tsktsk (_("register list not in ascending order")); |
| 4176 | mask |= 1 << reg; |
| 4177 | |
| 4178 | if (*input_line_pointer == '-') |
| 4179 | { |
| 4180 | input_line_pointer++; |
| 4181 | hi_reg = arm_reg_parse (&input_line_pointer, REG_TYPE_MMXWR); |
| 4182 | if (hi_reg == FAIL) |
| 4183 | { |
| 4184 | as_bad ("%s", _(reg_expected_msgs[REG_TYPE_MMXWR])); |
| 4185 | goto error; |
| 4186 | } |
| 4187 | else if (reg >= hi_reg) |
| 4188 | { |
| 4189 | as_bad (_("bad register range")); |
| 4190 | goto error; |
| 4191 | } |
| 4192 | for (; reg < hi_reg; reg++) |
| 4193 | mask |= 1 << reg; |
| 4194 | } |
| 4195 | } |
| 4196 | while (skip_past_comma (&input_line_pointer) != FAIL); |
| 4197 | |
| 4198 | skip_past_char (&input_line_pointer, '}'); |
| 4199 | |
| 4200 | demand_empty_rest_of_line (); |
| 4201 | |
| 4202 | /* Generate any deferred opcodes because we're going to be looking at |
| 4203 | the list. */ |
| 4204 | flush_pending_unwind (); |
| 4205 | |
| 4206 | for (i = 0; i < 16; i++) |
| 4207 | { |
| 4208 | if (mask & (1 << i)) |
| 4209 | unwind.frame_size += 8; |
| 4210 | } |
| 4211 | |
| 4212 | /* Attempt to combine with a previous opcode. We do this because gcc |
| 4213 | likes to output separate unwind directives for a single block of |
| 4214 | registers. */ |
| 4215 | if (unwind.opcode_count > 0) |
| 4216 | { |
| 4217 | i = unwind.opcodes[unwind.opcode_count - 1]; |
| 4218 | if ((i & 0xf8) == 0xc0) |
| 4219 | { |
| 4220 | i &= 7; |
| 4221 | /* Only merge if the blocks are contiguous. */ |
| 4222 | if (i < 6) |
| 4223 | { |
| 4224 | if ((mask & 0xfe00) == (1 << 9)) |
| 4225 | { |
| 4226 | mask |= ((1 << (i + 11)) - 1) & 0xfc00; |
| 4227 | unwind.opcode_count--; |
| 4228 | } |
| 4229 | } |
| 4230 | else if (i == 6 && unwind.opcode_count >= 2) |
| 4231 | { |
| 4232 | i = unwind.opcodes[unwind.opcode_count - 2]; |
| 4233 | reg = i >> 4; |
| 4234 | i &= 0xf; |
| 4235 | |
| 4236 | op = 0xffff << (reg - 1); |
| 4237 | if (reg > 0 |
| 4238 | && ((mask & op) == (1u << (reg - 1)))) |
| 4239 | { |
| 4240 | op = (1 << (reg + i + 1)) - 1; |
| 4241 | op &= ~((1 << reg) - 1); |
| 4242 | mask |= op; |
| 4243 | unwind.opcode_count -= 2; |
| 4244 | } |
| 4245 | } |
| 4246 | } |
| 4247 | } |
| 4248 | |
| 4249 | hi_reg = 15; |
| 4250 | /* We want to generate opcodes in the order the registers have been |
| 4251 | saved, ie. descending order. */ |
| 4252 | for (reg = 15; reg >= -1; reg--) |
| 4253 | { |
| 4254 | /* Save registers in blocks. */ |
| 4255 | if (reg < 0 |
| 4256 | || !(mask & (1 << reg))) |
| 4257 | { |
| 4258 | /* We found an unsaved reg. Generate opcodes to save the |
| 4259 | preceding block. */ |
| 4260 | if (reg != hi_reg) |
| 4261 | { |
| 4262 | if (reg == 9) |
| 4263 | { |
| 4264 | /* Short form. */ |
| 4265 | op = 0xc0 | (hi_reg - 10); |
| 4266 | add_unwind_opcode (op, 1); |
| 4267 | } |
| 4268 | else |
| 4269 | { |
| 4270 | /* Long form. */ |
| 4271 | op = 0xc600 | ((reg + 1) << 4) | ((hi_reg - reg) - 1); |
| 4272 | add_unwind_opcode (op, 2); |
| 4273 | } |
| 4274 | } |
| 4275 | hi_reg = reg - 1; |
| 4276 | } |
| 4277 | } |
| 4278 | |
| 4279 | return; |
| 4280 | error: |
| 4281 | ignore_rest_of_line (); |
| 4282 | } |
| 4283 | |
| 4284 | static void |
| 4285 | s_arm_unwind_save_mmxwcg (void) |
| 4286 | { |
| 4287 | int reg; |
| 4288 | int hi_reg; |
| 4289 | unsigned mask = 0; |
| 4290 | valueT op; |
| 4291 | |
| 4292 | if (*input_line_pointer == '{') |
| 4293 | input_line_pointer++; |
| 4294 | |
| 4295 | skip_whitespace (input_line_pointer); |
| 4296 | |
| 4297 | do |
| 4298 | { |
| 4299 | reg = arm_reg_parse (&input_line_pointer, REG_TYPE_MMXWCG); |
| 4300 | |
| 4301 | if (reg == FAIL) |
| 4302 | { |
| 4303 | as_bad ("%s", _(reg_expected_msgs[REG_TYPE_MMXWCG])); |
| 4304 | goto error; |
| 4305 | } |
| 4306 | |
| 4307 | reg -= 8; |
| 4308 | if (mask >> reg) |
| 4309 | as_tsktsk (_("register list not in ascending order")); |
| 4310 | mask |= 1 << reg; |
| 4311 | |
| 4312 | if (*input_line_pointer == '-') |
| 4313 | { |
| 4314 | input_line_pointer++; |
| 4315 | hi_reg = arm_reg_parse (&input_line_pointer, REG_TYPE_MMXWCG); |
| 4316 | if (hi_reg == FAIL) |
| 4317 | { |
| 4318 | as_bad ("%s", _(reg_expected_msgs[REG_TYPE_MMXWCG])); |
| 4319 | goto error; |
| 4320 | } |
| 4321 | else if (reg >= hi_reg) |
| 4322 | { |
| 4323 | as_bad (_("bad register range")); |
| 4324 | goto error; |
| 4325 | } |
| 4326 | for (; reg < hi_reg; reg++) |
| 4327 | mask |= 1 << reg; |
| 4328 | } |
| 4329 | } |
| 4330 | while (skip_past_comma (&input_line_pointer) != FAIL); |
| 4331 | |
| 4332 | skip_past_char (&input_line_pointer, '}'); |
| 4333 | |
| 4334 | demand_empty_rest_of_line (); |
| 4335 | |
| 4336 | /* Generate any deferred opcodes because we're going to be looking at |
| 4337 | the list. */ |
| 4338 | flush_pending_unwind (); |
| 4339 | |
| 4340 | for (reg = 0; reg < 16; reg++) |
| 4341 | { |
| 4342 | if (mask & (1 << reg)) |
| 4343 | unwind.frame_size += 4; |
| 4344 | } |
| 4345 | op = 0xc700 | mask; |
| 4346 | add_unwind_opcode (op, 2); |
| 4347 | return; |
| 4348 | error: |
| 4349 | ignore_rest_of_line (); |
| 4350 | } |
| 4351 | |
| 4352 | |
| 4353 | /* Parse an unwind_save directive. |
| 4354 | If the argument is non-zero, this is a .vsave directive. */ |
| 4355 | |
| 4356 | static void |
| 4357 | s_arm_unwind_save (int arch_v6) |
| 4358 | { |
| 4359 | char *peek; |
| 4360 | struct reg_entry *reg; |
| 4361 | bfd_boolean had_brace = FALSE; |
| 4362 | |
| 4363 | if (!unwind.proc_start) |
| 4364 | as_bad (MISSING_FNSTART); |
| 4365 | |
| 4366 | /* Figure out what sort of save we have. */ |
| 4367 | peek = input_line_pointer; |
| 4368 | |
| 4369 | if (*peek == '{') |
| 4370 | { |
| 4371 | had_brace = TRUE; |
| 4372 | peek++; |
| 4373 | } |
| 4374 | |
| 4375 | reg = arm_reg_parse_multi (&peek); |
| 4376 | |
| 4377 | if (!reg) |
| 4378 | { |
| 4379 | as_bad (_("register expected")); |
| 4380 | ignore_rest_of_line (); |
| 4381 | return; |
| 4382 | } |
| 4383 | |
| 4384 | switch (reg->type) |
| 4385 | { |
| 4386 | case REG_TYPE_FN: |
| 4387 | if (had_brace) |
| 4388 | { |
| 4389 | as_bad (_("FPA .unwind_save does not take a register list")); |
| 4390 | ignore_rest_of_line (); |
| 4391 | return; |
| 4392 | } |
| 4393 | input_line_pointer = peek; |
| 4394 | s_arm_unwind_save_fpa (reg->number); |
| 4395 | return; |
| 4396 | |
| 4397 | case REG_TYPE_RN: |
| 4398 | s_arm_unwind_save_core (); |
| 4399 | return; |
| 4400 | |
| 4401 | case REG_TYPE_VFD: |
| 4402 | if (arch_v6) |
| 4403 | s_arm_unwind_save_vfp_armv6 (); |
| 4404 | else |
| 4405 | s_arm_unwind_save_vfp (); |
| 4406 | return; |
| 4407 | |
| 4408 | case REG_TYPE_MMXWR: |
| 4409 | s_arm_unwind_save_mmxwr (); |
| 4410 | return; |
| 4411 | |
| 4412 | case REG_TYPE_MMXWCG: |
| 4413 | s_arm_unwind_save_mmxwcg (); |
| 4414 | return; |
| 4415 | |
| 4416 | default: |
| 4417 | as_bad (_(".unwind_save does not support this kind of register")); |
| 4418 | ignore_rest_of_line (); |
| 4419 | } |
| 4420 | } |
| 4421 | |
| 4422 | |
| 4423 | /* Parse an unwind_movsp directive. */ |
| 4424 | |
| 4425 | static void |
| 4426 | s_arm_unwind_movsp (int ignored ATTRIBUTE_UNUSED) |
| 4427 | { |
| 4428 | int reg; |
| 4429 | valueT op; |
| 4430 | int offset; |
| 4431 | |
| 4432 | if (!unwind.proc_start) |
| 4433 | as_bad (MISSING_FNSTART); |
| 4434 | |
| 4435 | reg = arm_reg_parse (&input_line_pointer, REG_TYPE_RN); |
| 4436 | if (reg == FAIL) |
| 4437 | { |
| 4438 | as_bad ("%s", _(reg_expected_msgs[REG_TYPE_RN])); |
| 4439 | ignore_rest_of_line (); |
| 4440 | return; |
| 4441 | } |
| 4442 | |
| 4443 | /* Optional constant. */ |
| 4444 | if (skip_past_comma (&input_line_pointer) != FAIL) |
| 4445 | { |
| 4446 | if (immediate_for_directive (&offset) == FAIL) |
| 4447 | return; |
| 4448 | } |
| 4449 | else |
| 4450 | offset = 0; |
| 4451 | |
| 4452 | demand_empty_rest_of_line (); |
| 4453 | |
| 4454 | if (reg == REG_SP || reg == REG_PC) |
| 4455 | { |
| 4456 | as_bad (_("SP and PC not permitted in .unwind_movsp directive")); |
| 4457 | return; |
| 4458 | } |
| 4459 | |
| 4460 | if (unwind.fp_reg != REG_SP) |
| 4461 | as_bad (_("unexpected .unwind_movsp directive")); |
| 4462 | |
| 4463 | /* Generate opcode to restore the value. */ |
| 4464 | op = 0x90 | reg; |
| 4465 | add_unwind_opcode (op, 1); |
| 4466 | |
| 4467 | /* Record the information for later. */ |
| 4468 | unwind.fp_reg = reg; |
| 4469 | unwind.fp_offset = unwind.frame_size - offset; |
| 4470 | unwind.sp_restored = 1; |
| 4471 | } |
| 4472 | |
| 4473 | /* Parse an unwind_pad directive. */ |
| 4474 | |
| 4475 | static void |
| 4476 | s_arm_unwind_pad (int ignored ATTRIBUTE_UNUSED) |
| 4477 | { |
| 4478 | int offset; |
| 4479 | |
| 4480 | if (!unwind.proc_start) |
| 4481 | as_bad (MISSING_FNSTART); |
| 4482 | |
| 4483 | if (immediate_for_directive (&offset) == FAIL) |
| 4484 | return; |
| 4485 | |
| 4486 | if (offset & 3) |
| 4487 | { |
| 4488 | as_bad (_("stack increment must be multiple of 4")); |
| 4489 | ignore_rest_of_line (); |
| 4490 | return; |
| 4491 | } |
| 4492 | |
| 4493 | /* Don't generate any opcodes, just record the details for later. */ |
| 4494 | unwind.frame_size += offset; |
| 4495 | unwind.pending_offset += offset; |
| 4496 | |
| 4497 | demand_empty_rest_of_line (); |
| 4498 | } |
| 4499 | |
| 4500 | /* Parse an unwind_setfp directive. */ |
| 4501 | |
| 4502 | static void |
| 4503 | s_arm_unwind_setfp (int ignored ATTRIBUTE_UNUSED) |
| 4504 | { |
| 4505 | int sp_reg; |
| 4506 | int fp_reg; |
| 4507 | int offset; |
| 4508 | |
| 4509 | if (!unwind.proc_start) |
| 4510 | as_bad (MISSING_FNSTART); |
| 4511 | |
| 4512 | fp_reg = arm_reg_parse (&input_line_pointer, REG_TYPE_RN); |
| 4513 | if (skip_past_comma (&input_line_pointer) == FAIL) |
| 4514 | sp_reg = FAIL; |
| 4515 | else |
| 4516 | sp_reg = arm_reg_parse (&input_line_pointer, REG_TYPE_RN); |
| 4517 | |
| 4518 | if (fp_reg == FAIL || sp_reg == FAIL) |
| 4519 | { |
| 4520 | as_bad (_("expected <reg>, <reg>")); |
| 4521 | ignore_rest_of_line (); |
| 4522 | return; |
| 4523 | } |
| 4524 | |
| 4525 | /* Optional constant. */ |
| 4526 | if (skip_past_comma (&input_line_pointer) != FAIL) |
| 4527 | { |
| 4528 | if (immediate_for_directive (&offset) == FAIL) |
| 4529 | return; |
| 4530 | } |
| 4531 | else |
| 4532 | offset = 0; |
| 4533 | |
| 4534 | demand_empty_rest_of_line (); |
| 4535 | |
| 4536 | if (sp_reg != REG_SP && sp_reg != unwind.fp_reg) |
| 4537 | { |
| 4538 | as_bad (_("register must be either sp or set by a previous" |
| 4539 | "unwind_movsp directive")); |
| 4540 | return; |
| 4541 | } |
| 4542 | |
| 4543 | /* Don't generate any opcodes, just record the information for later. */ |
| 4544 | unwind.fp_reg = fp_reg; |
| 4545 | unwind.fp_used = 1; |
| 4546 | if (sp_reg == REG_SP) |
| 4547 | unwind.fp_offset = unwind.frame_size - offset; |
| 4548 | else |
| 4549 | unwind.fp_offset -= offset; |
| 4550 | } |
| 4551 | |
| 4552 | /* Parse an unwind_raw directive. */ |
| 4553 | |
| 4554 | static void |
| 4555 | s_arm_unwind_raw (int ignored ATTRIBUTE_UNUSED) |
| 4556 | { |
| 4557 | expressionS exp; |
| 4558 | /* This is an arbitrary limit. */ |
| 4559 | unsigned char op[16]; |
| 4560 | int count; |
| 4561 | |
| 4562 | if (!unwind.proc_start) |
| 4563 | as_bad (MISSING_FNSTART); |
| 4564 | |
| 4565 | expression (&exp); |
| 4566 | if (exp.X_op == O_constant |
| 4567 | && skip_past_comma (&input_line_pointer) != FAIL) |
| 4568 | { |
| 4569 | unwind.frame_size += exp.X_add_number; |
| 4570 | expression (&exp); |
| 4571 | } |
| 4572 | else |
| 4573 | exp.X_op = O_illegal; |
| 4574 | |
| 4575 | if (exp.X_op != O_constant) |
| 4576 | { |
| 4577 | as_bad (_("expected <offset>, <opcode>")); |
| 4578 | ignore_rest_of_line (); |
| 4579 | return; |
| 4580 | } |
| 4581 | |
| 4582 | count = 0; |
| 4583 | |
| 4584 | /* Parse the opcode. */ |
| 4585 | for (;;) |
| 4586 | { |
| 4587 | if (count >= 16) |
| 4588 | { |
| 4589 | as_bad (_("unwind opcode too long")); |
| 4590 | ignore_rest_of_line (); |
| 4591 | } |
| 4592 | if (exp.X_op != O_constant || exp.X_add_number & ~0xff) |
| 4593 | { |
| 4594 | as_bad (_("invalid unwind opcode")); |
| 4595 | ignore_rest_of_line (); |
| 4596 | return; |
| 4597 | } |
| 4598 | op[count++] = exp.X_add_number; |
| 4599 | |
| 4600 | /* Parse the next byte. */ |
| 4601 | if (skip_past_comma (&input_line_pointer) == FAIL) |
| 4602 | break; |
| 4603 | |
| 4604 | expression (&exp); |
| 4605 | } |
| 4606 | |
| 4607 | /* Add the opcode bytes in reverse order. */ |
| 4608 | while (count--) |
| 4609 | add_unwind_opcode (op[count], 1); |
| 4610 | |
| 4611 | demand_empty_rest_of_line (); |
| 4612 | } |
| 4613 | |
| 4614 | |
| 4615 | /* Parse a .eabi_attribute directive. */ |
| 4616 | |
| 4617 | static void |
| 4618 | s_arm_eabi_attribute (int ignored ATTRIBUTE_UNUSED) |
| 4619 | { |
| 4620 | int tag = obj_elf_vendor_attribute (OBJ_ATTR_PROC); |
| 4621 | |
| 4622 | if (tag < NUM_KNOWN_OBJ_ATTRIBUTES) |
| 4623 | attributes_set_explicitly[tag] = 1; |
| 4624 | } |
| 4625 | |
| 4626 | /* Emit a tls fix for the symbol. */ |
| 4627 | |
| 4628 | static void |
| 4629 | s_arm_tls_descseq (int ignored ATTRIBUTE_UNUSED) |
| 4630 | { |
| 4631 | char *p; |
| 4632 | expressionS exp; |
| 4633 | #ifdef md_flush_pending_output |
| 4634 | md_flush_pending_output (); |
| 4635 | #endif |
| 4636 | |
| 4637 | #ifdef md_cons_align |
| 4638 | md_cons_align (4); |
| 4639 | #endif |
| 4640 | |
| 4641 | /* Since we're just labelling the code, there's no need to define a |
| 4642 | mapping symbol. */ |
| 4643 | expression (&exp); |
| 4644 | p = obstack_next_free (&frchain_now->frch_obstack); |
| 4645 | fix_new_arm (frag_now, p - frag_now->fr_literal, 4, &exp, 0, |
| 4646 | thumb_mode ? BFD_RELOC_ARM_THM_TLS_DESCSEQ |
| 4647 | : BFD_RELOC_ARM_TLS_DESCSEQ); |
| 4648 | } |
| 4649 | #endif /* OBJ_ELF */ |
| 4650 | |
| 4651 | static void s_arm_arch (int); |
| 4652 | static void s_arm_object_arch (int); |
| 4653 | static void s_arm_cpu (int); |
| 4654 | static void s_arm_fpu (int); |
| 4655 | static void s_arm_arch_extension (int); |
| 4656 | |
| 4657 | #ifdef TE_PE |
| 4658 | |
| 4659 | static void |
| 4660 | pe_directive_secrel (int dummy ATTRIBUTE_UNUSED) |
| 4661 | { |
| 4662 | expressionS exp; |
| 4663 | |
| 4664 | do |
| 4665 | { |
| 4666 | expression (&exp); |
| 4667 | if (exp.X_op == O_symbol) |
| 4668 | exp.X_op = O_secrel; |
| 4669 | |
| 4670 | emit_expr (&exp, 4); |
| 4671 | } |
| 4672 | while (*input_line_pointer++ == ','); |
| 4673 | |
| 4674 | input_line_pointer--; |
| 4675 | demand_empty_rest_of_line (); |
| 4676 | } |
| 4677 | #endif /* TE_PE */ |
| 4678 | |
| 4679 | /* This table describes all the machine specific pseudo-ops the assembler |
| 4680 | has to support. The fields are: |
| 4681 | pseudo-op name without dot |
| 4682 | function to call to execute this pseudo-op |
| 4683 | Integer arg to pass to the function. */ |
| 4684 | |
| 4685 | const pseudo_typeS md_pseudo_table[] = |
| 4686 | { |
| 4687 | /* Never called because '.req' does not start a line. */ |
| 4688 | { "req", s_req, 0 }, |
| 4689 | /* Following two are likewise never called. */ |
| 4690 | { "dn", s_dn, 0 }, |
| 4691 | { "qn", s_qn, 0 }, |
| 4692 | { "unreq", s_unreq, 0 }, |
| 4693 | { "bss", s_bss, 0 }, |
| 4694 | { "align", s_align_ptwo, 2 }, |
| 4695 | { "arm", s_arm, 0 }, |
| 4696 | { "thumb", s_thumb, 0 }, |
| 4697 | { "code", s_code, 0 }, |
| 4698 | { "force_thumb", s_force_thumb, 0 }, |
| 4699 | { "thumb_func", s_thumb_func, 0 }, |
| 4700 | { "thumb_set", s_thumb_set, 0 }, |
| 4701 | { "even", s_even, 0 }, |
| 4702 | { "ltorg", s_ltorg, 0 }, |
| 4703 | { "pool", s_ltorg, 0 }, |
| 4704 | { "syntax", s_syntax, 0 }, |
| 4705 | { "cpu", s_arm_cpu, 0 }, |
| 4706 | { "arch", s_arm_arch, 0 }, |
| 4707 | { "object_arch", s_arm_object_arch, 0 }, |
| 4708 | { "fpu", s_arm_fpu, 0 }, |
| 4709 | { "arch_extension", s_arm_arch_extension, 0 }, |
| 4710 | #ifdef OBJ_ELF |
| 4711 | { "word", s_arm_elf_cons, 4 }, |
| 4712 | { "long", s_arm_elf_cons, 4 }, |
| 4713 | { "inst.n", s_arm_elf_inst, 2 }, |
| 4714 | { "inst.w", s_arm_elf_inst, 4 }, |
| 4715 | { "inst", s_arm_elf_inst, 0 }, |
| 4716 | { "rel31", s_arm_rel31, 0 }, |
| 4717 | { "fnstart", s_arm_unwind_fnstart, 0 }, |
| 4718 | { "fnend", s_arm_unwind_fnend, 0 }, |
| 4719 | { "cantunwind", s_arm_unwind_cantunwind, 0 }, |
| 4720 | { "personality", s_arm_unwind_personality, 0 }, |
| 4721 | { "personalityindex", s_arm_unwind_personalityindex, 0 }, |
| 4722 | { "handlerdata", s_arm_unwind_handlerdata, 0 }, |
| 4723 | { "save", s_arm_unwind_save, 0 }, |
| 4724 | { "vsave", s_arm_unwind_save, 1 }, |
| 4725 | { "movsp", s_arm_unwind_movsp, 0 }, |
| 4726 | { "pad", s_arm_unwind_pad, 0 }, |
| 4727 | { "setfp", s_arm_unwind_setfp, 0 }, |
| 4728 | { "unwind_raw", s_arm_unwind_raw, 0 }, |
| 4729 | { "eabi_attribute", s_arm_eabi_attribute, 0 }, |
| 4730 | { "tlsdescseq", s_arm_tls_descseq, 0 }, |
| 4731 | #else |
| 4732 | { "word", cons, 4}, |
| 4733 | |
| 4734 | /* These are used for dwarf. */ |
| 4735 | {"2byte", cons, 2}, |
| 4736 | {"4byte", cons, 4}, |
| 4737 | {"8byte", cons, 8}, |
| 4738 | /* These are used for dwarf2. */ |
| 4739 | { "file", (void (*) (int)) dwarf2_directive_file, 0 }, |
| 4740 | { "loc", dwarf2_directive_loc, 0 }, |
| 4741 | { "loc_mark_labels", dwarf2_directive_loc_mark_labels, 0 }, |
| 4742 | #endif |
| 4743 | { "extend", float_cons, 'x' }, |
| 4744 | { "ldouble", float_cons, 'x' }, |
| 4745 | { "packed", float_cons, 'p' }, |
| 4746 | #ifdef TE_PE |
| 4747 | {"secrel32", pe_directive_secrel, 0}, |
| 4748 | #endif |
| 4749 | |
| 4750 | /* These are for compatibility with CodeComposer Studio. */ |
| 4751 | {"ref", s_ccs_ref, 0}, |
| 4752 | {"def", s_ccs_def, 0}, |
| 4753 | {"asmfunc", s_ccs_asmfunc, 0}, |
| 4754 | {"endasmfunc", s_ccs_endasmfunc, 0}, |
| 4755 | |
| 4756 | { 0, 0, 0 } |
| 4757 | }; |
| 4758 | \f |
| 4759 | /* Parser functions used exclusively in instruction operands. */ |
| 4760 | |
| 4761 | /* Generic immediate-value read function for use in insn parsing. |
| 4762 | STR points to the beginning of the immediate (the leading #); |
| 4763 | VAL receives the value; if the value is outside [MIN, MAX] |
| 4764 | issue an error. PREFIX_OPT is true if the immediate prefix is |
| 4765 | optional. */ |
| 4766 | |
| 4767 | static int |
| 4768 | parse_immediate (char **str, int *val, int min, int max, |
| 4769 | bfd_boolean prefix_opt) |
| 4770 | { |
| 4771 | expressionS exp; |
| 4772 | my_get_expression (&exp, str, prefix_opt ? GE_OPT_PREFIX : GE_IMM_PREFIX); |
| 4773 | if (exp.X_op != O_constant) |
| 4774 | { |
| 4775 | inst.error = _("constant expression required"); |
| 4776 | return FAIL; |
| 4777 | } |
| 4778 | |
| 4779 | if (exp.X_add_number < min || exp.X_add_number > max) |
| 4780 | { |
| 4781 | inst.error = _("immediate value out of range"); |
| 4782 | return FAIL; |
| 4783 | } |
| 4784 | |
| 4785 | *val = exp.X_add_number; |
| 4786 | return SUCCESS; |
| 4787 | } |
| 4788 | |
| 4789 | /* Less-generic immediate-value read function with the possibility of loading a |
| 4790 | big (64-bit) immediate, as required by Neon VMOV, VMVN and logic immediate |
| 4791 | instructions. Puts the result directly in inst.operands[i]. */ |
| 4792 | |
| 4793 | static int |
| 4794 | parse_big_immediate (char **str, int i, expressionS *in_exp, |
| 4795 | bfd_boolean allow_symbol_p) |
| 4796 | { |
| 4797 | expressionS exp; |
| 4798 | expressionS *exp_p = in_exp ? in_exp : &exp; |
| 4799 | char *ptr = *str; |
| 4800 | |
| 4801 | my_get_expression (exp_p, &ptr, GE_OPT_PREFIX_BIG); |
| 4802 | |
| 4803 | if (exp_p->X_op == O_constant) |
| 4804 | { |
| 4805 | inst.operands[i].imm = exp_p->X_add_number & 0xffffffff; |
| 4806 | /* If we're on a 64-bit host, then a 64-bit number can be returned using |
| 4807 | O_constant. We have to be careful not to break compilation for |
| 4808 | 32-bit X_add_number, though. */ |
| 4809 | if ((exp_p->X_add_number & ~(offsetT)(0xffffffffU)) != 0) |
| 4810 | { |
| 4811 | /* X >> 32 is illegal if sizeof (exp_p->X_add_number) == 4. */ |
| 4812 | inst.operands[i].reg = (((exp_p->X_add_number >> 16) >> 16) |
| 4813 | & 0xffffffff); |
| 4814 | inst.operands[i].regisimm = 1; |
| 4815 | } |
| 4816 | } |
| 4817 | else if (exp_p->X_op == O_big |
| 4818 | && LITTLENUM_NUMBER_OF_BITS * exp_p->X_add_number > 32) |
| 4819 | { |
| 4820 | unsigned parts = 32 / LITTLENUM_NUMBER_OF_BITS, j, idx = 0; |
| 4821 | |
| 4822 | /* Bignums have their least significant bits in |
| 4823 | generic_bignum[0]. Make sure we put 32 bits in imm and |
| 4824 | 32 bits in reg, in a (hopefully) portable way. */ |
| 4825 | gas_assert (parts != 0); |
| 4826 | |
| 4827 | /* Make sure that the number is not too big. |
| 4828 | PR 11972: Bignums can now be sign-extended to the |
| 4829 | size of a .octa so check that the out of range bits |
| 4830 | are all zero or all one. */ |
| 4831 | if (LITTLENUM_NUMBER_OF_BITS * exp_p->X_add_number > 64) |
| 4832 | { |
| 4833 | LITTLENUM_TYPE m = -1; |
| 4834 | |
| 4835 | if (generic_bignum[parts * 2] != 0 |
| 4836 | && generic_bignum[parts * 2] != m) |
| 4837 | return FAIL; |
| 4838 | |
| 4839 | for (j = parts * 2 + 1; j < (unsigned) exp_p->X_add_number; j++) |
| 4840 | if (generic_bignum[j] != generic_bignum[j-1]) |
| 4841 | return FAIL; |
| 4842 | } |
| 4843 | |
| 4844 | inst.operands[i].imm = 0; |
| 4845 | for (j = 0; j < parts; j++, idx++) |
| 4846 | inst.operands[i].imm |= generic_bignum[idx] |
| 4847 | << (LITTLENUM_NUMBER_OF_BITS * j); |
| 4848 | inst.operands[i].reg = 0; |
| 4849 | for (j = 0; j < parts; j++, idx++) |
| 4850 | inst.operands[i].reg |= generic_bignum[idx] |
| 4851 | << (LITTLENUM_NUMBER_OF_BITS * j); |
| 4852 | inst.operands[i].regisimm = 1; |
| 4853 | } |
| 4854 | else if (!(exp_p->X_op == O_symbol && allow_symbol_p)) |
| 4855 | return FAIL; |
| 4856 | |
| 4857 | *str = ptr; |
| 4858 | |
| 4859 | return SUCCESS; |
| 4860 | } |
| 4861 | |
| 4862 | /* Returns the pseudo-register number of an FPA immediate constant, |
| 4863 | or FAIL if there isn't a valid constant here. */ |
| 4864 | |
| 4865 | static int |
| 4866 | parse_fpa_immediate (char ** str) |
| 4867 | { |
| 4868 | LITTLENUM_TYPE words[MAX_LITTLENUMS]; |
| 4869 | char * save_in; |
| 4870 | expressionS exp; |
| 4871 | int i; |
| 4872 | int j; |
| 4873 | |
| 4874 | /* First try and match exact strings, this is to guarantee |
| 4875 | that some formats will work even for cross assembly. */ |
| 4876 | |
| 4877 | for (i = 0; fp_const[i]; i++) |
| 4878 | { |
| 4879 | if (strncmp (*str, fp_const[i], strlen (fp_const[i])) == 0) |
| 4880 | { |
| 4881 | char *start = *str; |
| 4882 | |
| 4883 | *str += strlen (fp_const[i]); |
| 4884 | if (is_end_of_line[(unsigned char) **str]) |
| 4885 | return i + 8; |
| 4886 | *str = start; |
| 4887 | } |
| 4888 | } |
| 4889 | |
| 4890 | /* Just because we didn't get a match doesn't mean that the constant |
| 4891 | isn't valid, just that it is in a format that we don't |
| 4892 | automatically recognize. Try parsing it with the standard |
| 4893 | expression routines. */ |
| 4894 | |
| 4895 | memset (words, 0, MAX_LITTLENUMS * sizeof (LITTLENUM_TYPE)); |
| 4896 | |
| 4897 | /* Look for a raw floating point number. */ |
| 4898 | if ((save_in = atof_ieee (*str, 'x', words)) != NULL |
| 4899 | && is_end_of_line[(unsigned char) *save_in]) |
| 4900 | { |
| 4901 | for (i = 0; i < NUM_FLOAT_VALS; i++) |
| 4902 | { |
| 4903 | for (j = 0; j < MAX_LITTLENUMS; j++) |
| 4904 | { |
| 4905 | if (words[j] != fp_values[i][j]) |
| 4906 | break; |
| 4907 | } |
| 4908 | |
| 4909 | if (j == MAX_LITTLENUMS) |
| 4910 | { |
| 4911 | *str = save_in; |
| 4912 | return i + 8; |
| 4913 | } |
| 4914 | } |
| 4915 | } |
| 4916 | |
| 4917 | /* Try and parse a more complex expression, this will probably fail |
| 4918 | unless the code uses a floating point prefix (eg "0f"). */ |
| 4919 | save_in = input_line_pointer; |
| 4920 | input_line_pointer = *str; |
| 4921 | if (expression (&exp) == absolute_section |
| 4922 | && exp.X_op == O_big |
| 4923 | && exp.X_add_number < 0) |
| 4924 | { |
| 4925 | /* FIXME: 5 = X_PRECISION, should be #define'd where we can use it. |
| 4926 | Ditto for 15. */ |
| 4927 | #define X_PRECISION 5 |
| 4928 | #define E_PRECISION 15L |
| 4929 | if (gen_to_words (words, X_PRECISION, E_PRECISION) == 0) |
| 4930 | { |
| 4931 | for (i = 0; i < NUM_FLOAT_VALS; i++) |
| 4932 | { |
| 4933 | for (j = 0; j < MAX_LITTLENUMS; j++) |
| 4934 | { |
| 4935 | if (words[j] != fp_values[i][j]) |
| 4936 | break; |
| 4937 | } |
| 4938 | |
| 4939 | if (j == MAX_LITTLENUMS) |
| 4940 | { |
| 4941 | *str = input_line_pointer; |
| 4942 | input_line_pointer = save_in; |
| 4943 | return i + 8; |
| 4944 | } |
| 4945 | } |
| 4946 | } |
| 4947 | } |
| 4948 | |
| 4949 | *str = input_line_pointer; |
| 4950 | input_line_pointer = save_in; |
| 4951 | inst.error = _("invalid FPA immediate expression"); |
| 4952 | return FAIL; |
| 4953 | } |
| 4954 | |
| 4955 | /* Returns 1 if a number has "quarter-precision" float format |
| 4956 | 0baBbbbbbc defgh000 00000000 00000000. */ |
| 4957 | |
| 4958 | static int |
| 4959 | is_quarter_float (unsigned imm) |
| 4960 | { |
| 4961 | int bs = (imm & 0x20000000) ? 0x3e000000 : 0x40000000; |
| 4962 | return (imm & 0x7ffff) == 0 && ((imm & 0x7e000000) ^ bs) == 0; |
| 4963 | } |
| 4964 | |
| 4965 | |
| 4966 | /* Detect the presence of a floating point or integer zero constant, |
| 4967 | i.e. #0.0 or #0. */ |
| 4968 | |
| 4969 | static bfd_boolean |
| 4970 | parse_ifimm_zero (char **in) |
| 4971 | { |
| 4972 | int error_code; |
| 4973 | |
| 4974 | if (!is_immediate_prefix (**in)) |
| 4975 | { |
| 4976 | /* In unified syntax, all prefixes are optional. */ |
| 4977 | if (!unified_syntax) |
| 4978 | return FALSE; |
| 4979 | } |
| 4980 | else |
| 4981 | ++*in; |
| 4982 | |
| 4983 | /* Accept #0x0 as a synonym for #0. */ |
| 4984 | if (strncmp (*in, "0x", 2) == 0) |
| 4985 | { |
| 4986 | int val; |
| 4987 | if (parse_immediate (in, &val, 0, 0, TRUE) == FAIL) |
| 4988 | return FALSE; |
| 4989 | return TRUE; |
| 4990 | } |
| 4991 | |
| 4992 | error_code = atof_generic (in, ".", EXP_CHARS, |
| 4993 | &generic_floating_point_number); |
| 4994 | |
| 4995 | if (!error_code |
| 4996 | && generic_floating_point_number.sign == '+' |
| 4997 | && (generic_floating_point_number.low |
| 4998 | > generic_floating_point_number.leader)) |
| 4999 | return TRUE; |
| 5000 | |
| 5001 | return FALSE; |
| 5002 | } |
| 5003 | |
| 5004 | /* Parse an 8-bit "quarter-precision" floating point number of the form: |
| 5005 | 0baBbbbbbc defgh000 00000000 00000000. |
| 5006 | The zero and minus-zero cases need special handling, since they can't be |
| 5007 | encoded in the "quarter-precision" float format, but can nonetheless be |
| 5008 | loaded as integer constants. */ |
| 5009 | |
| 5010 | static unsigned |
| 5011 | parse_qfloat_immediate (char **ccp, int *immed) |
| 5012 | { |
| 5013 | char *str = *ccp; |
| 5014 | char *fpnum; |
| 5015 | LITTLENUM_TYPE words[MAX_LITTLENUMS]; |
| 5016 | int found_fpchar = 0; |
| 5017 | |
| 5018 | skip_past_char (&str, '#'); |
| 5019 | |
| 5020 | /* We must not accidentally parse an integer as a floating-point number. Make |
| 5021 | sure that the value we parse is not an integer by checking for special |
| 5022 | characters '.' or 'e'. |
| 5023 | FIXME: This is a horrible hack, but doing better is tricky because type |
| 5024 | information isn't in a very usable state at parse time. */ |
| 5025 | fpnum = str; |
| 5026 | skip_whitespace (fpnum); |
| 5027 | |
| 5028 | if (strncmp (fpnum, "0x", 2) == 0) |
| 5029 | return FAIL; |
| 5030 | else |
| 5031 | { |
| 5032 | for (; *fpnum != '\0' && *fpnum != ' ' && *fpnum != '\n'; fpnum++) |
| 5033 | if (*fpnum == '.' || *fpnum == 'e' || *fpnum == 'E') |
| 5034 | { |
| 5035 | found_fpchar = 1; |
| 5036 | break; |
| 5037 | } |
| 5038 | |
| 5039 | if (!found_fpchar) |
| 5040 | return FAIL; |
| 5041 | } |
| 5042 | |
| 5043 | if ((str = atof_ieee (str, 's', words)) != NULL) |
| 5044 | { |
| 5045 | unsigned fpword = 0; |
| 5046 | int i; |
| 5047 | |
| 5048 | /* Our FP word must be 32 bits (single-precision FP). */ |
| 5049 | for (i = 0; i < 32 / LITTLENUM_NUMBER_OF_BITS; i++) |
| 5050 | { |
| 5051 | fpword <<= LITTLENUM_NUMBER_OF_BITS; |
| 5052 | fpword |= words[i]; |
| 5053 | } |
| 5054 | |
| 5055 | if (is_quarter_float (fpword) || (fpword & 0x7fffffff) == 0) |
| 5056 | *immed = fpword; |
| 5057 | else |
| 5058 | return FAIL; |
| 5059 | |
| 5060 | *ccp = str; |
| 5061 | |
| 5062 | return SUCCESS; |
| 5063 | } |
| 5064 | |
| 5065 | return FAIL; |
| 5066 | } |
| 5067 | |
| 5068 | /* Shift operands. */ |
| 5069 | enum shift_kind |
| 5070 | { |
| 5071 | SHIFT_LSL, SHIFT_LSR, SHIFT_ASR, SHIFT_ROR, SHIFT_RRX |
| 5072 | }; |
| 5073 | |
| 5074 | struct asm_shift_name |
| 5075 | { |
| 5076 | const char *name; |
| 5077 | enum shift_kind kind; |
| 5078 | }; |
| 5079 | |
| 5080 | /* Third argument to parse_shift. */ |
| 5081 | enum parse_shift_mode |
| 5082 | { |
| 5083 | NO_SHIFT_RESTRICT, /* Any kind of shift is accepted. */ |
| 5084 | SHIFT_IMMEDIATE, /* Shift operand must be an immediate. */ |
| 5085 | SHIFT_LSL_OR_ASR_IMMEDIATE, /* Shift must be LSL or ASR immediate. */ |
| 5086 | SHIFT_ASR_IMMEDIATE, /* Shift must be ASR immediate. */ |
| 5087 | SHIFT_LSL_IMMEDIATE, /* Shift must be LSL immediate. */ |
| 5088 | }; |
| 5089 | |
| 5090 | /* Parse a <shift> specifier on an ARM data processing instruction. |
| 5091 | This has three forms: |
| 5092 | |
| 5093 | (LSL|LSR|ASL|ASR|ROR) Rs |
| 5094 | (LSL|LSR|ASL|ASR|ROR) #imm |
| 5095 | RRX |
| 5096 | |
| 5097 | Note that ASL is assimilated to LSL in the instruction encoding, and |
| 5098 | RRX to ROR #0 (which cannot be written as such). */ |
| 5099 | |
| 5100 | static int |
| 5101 | parse_shift (char **str, int i, enum parse_shift_mode mode) |
| 5102 | { |
| 5103 | const struct asm_shift_name *shift_name; |
| 5104 | enum shift_kind shift; |
| 5105 | char *s = *str; |
| 5106 | char *p = s; |
| 5107 | int reg; |
| 5108 | |
| 5109 | for (p = *str; ISALPHA (*p); p++) |
| 5110 | ; |
| 5111 | |
| 5112 | if (p == *str) |
| 5113 | { |
| 5114 | inst.error = _("shift expression expected"); |
| 5115 | return FAIL; |
| 5116 | } |
| 5117 | |
| 5118 | shift_name = (const struct asm_shift_name *) hash_find_n (arm_shift_hsh, *str, |
| 5119 | p - *str); |
| 5120 | |
| 5121 | if (shift_name == NULL) |
| 5122 | { |
| 5123 | inst.error = _("shift expression expected"); |
| 5124 | return FAIL; |
| 5125 | } |
| 5126 | |
| 5127 | shift = shift_name->kind; |
| 5128 | |
| 5129 | switch (mode) |
| 5130 | { |
| 5131 | case NO_SHIFT_RESTRICT: |
| 5132 | case SHIFT_IMMEDIATE: break; |
| 5133 | |
| 5134 | case SHIFT_LSL_OR_ASR_IMMEDIATE: |
| 5135 | if (shift != SHIFT_LSL && shift != SHIFT_ASR) |
| 5136 | { |
| 5137 | inst.error = _("'LSL' or 'ASR' required"); |
| 5138 | return FAIL; |
| 5139 | } |
| 5140 | break; |
| 5141 | |
| 5142 | case SHIFT_LSL_IMMEDIATE: |
| 5143 | if (shift != SHIFT_LSL) |
| 5144 | { |
| 5145 | inst.error = _("'LSL' required"); |
| 5146 | return FAIL; |
| 5147 | } |
| 5148 | break; |
| 5149 | |
| 5150 | case SHIFT_ASR_IMMEDIATE: |
| 5151 | if (shift != SHIFT_ASR) |
| 5152 | { |
| 5153 | inst.error = _("'ASR' required"); |
| 5154 | return FAIL; |
| 5155 | } |
| 5156 | break; |
| 5157 | |
| 5158 | default: abort (); |
| 5159 | } |
| 5160 | |
| 5161 | if (shift != SHIFT_RRX) |
| 5162 | { |
| 5163 | /* Whitespace can appear here if the next thing is a bare digit. */ |
| 5164 | skip_whitespace (p); |
| 5165 | |
| 5166 | if (mode == NO_SHIFT_RESTRICT |
| 5167 | && (reg = arm_reg_parse (&p, REG_TYPE_RN)) != FAIL) |
| 5168 | { |
| 5169 | inst.operands[i].imm = reg; |
| 5170 | inst.operands[i].immisreg = 1; |
| 5171 | } |
| 5172 | else if (my_get_expression (&inst.reloc.exp, &p, GE_IMM_PREFIX)) |
| 5173 | return FAIL; |
| 5174 | } |
| 5175 | inst.operands[i].shift_kind = shift; |
| 5176 | inst.operands[i].shifted = 1; |
| 5177 | *str = p; |
| 5178 | return SUCCESS; |
| 5179 | } |
| 5180 | |
| 5181 | /* Parse a <shifter_operand> for an ARM data processing instruction: |
| 5182 | |
| 5183 | #<immediate> |
| 5184 | #<immediate>, <rotate> |
| 5185 | <Rm> |
| 5186 | <Rm>, <shift> |
| 5187 | |
| 5188 | where <shift> is defined by parse_shift above, and <rotate> is a |
| 5189 | multiple of 2 between 0 and 30. Validation of immediate operands |
| 5190 | is deferred to md_apply_fix. */ |
| 5191 | |
| 5192 | static int |
| 5193 | parse_shifter_operand (char **str, int i) |
| 5194 | { |
| 5195 | int value; |
| 5196 | expressionS exp; |
| 5197 | |
| 5198 | if ((value = arm_reg_parse (str, REG_TYPE_RN)) != FAIL) |
| 5199 | { |
| 5200 | inst.operands[i].reg = value; |
| 5201 | inst.operands[i].isreg = 1; |
| 5202 | |
| 5203 | /* parse_shift will override this if appropriate */ |
| 5204 | inst.reloc.exp.X_op = O_constant; |
| 5205 | inst.reloc.exp.X_add_number = 0; |
| 5206 | |
| 5207 | if (skip_past_comma (str) == FAIL) |
| 5208 | return SUCCESS; |
| 5209 | |
| 5210 | /* Shift operation on register. */ |
| 5211 | return parse_shift (str, i, NO_SHIFT_RESTRICT); |
| 5212 | } |
| 5213 | |
| 5214 | if (my_get_expression (&inst.reloc.exp, str, GE_IMM_PREFIX)) |
| 5215 | return FAIL; |
| 5216 | |
| 5217 | if (skip_past_comma (str) == SUCCESS) |
| 5218 | { |
| 5219 | /* #x, y -- ie explicit rotation by Y. */ |
| 5220 | if (my_get_expression (&exp, str, GE_NO_PREFIX)) |
| 5221 | return FAIL; |
| 5222 | |
| 5223 | if (exp.X_op != O_constant || inst.reloc.exp.X_op != O_constant) |
| 5224 | { |
| 5225 | inst.error = _("constant expression expected"); |
| 5226 | return FAIL; |
| 5227 | } |
| 5228 | |
| 5229 | value = exp.X_add_number; |
| 5230 | if (value < 0 || value > 30 || value % 2 != 0) |
| 5231 | { |
| 5232 | inst.error = _("invalid rotation"); |
| 5233 | return FAIL; |
| 5234 | } |
| 5235 | if (inst.reloc.exp.X_add_number < 0 || inst.reloc.exp.X_add_number > 255) |
| 5236 | { |
| 5237 | inst.error = _("invalid constant"); |
| 5238 | return FAIL; |
| 5239 | } |
| 5240 | |
| 5241 | /* Encode as specified. */ |
| 5242 | inst.operands[i].imm = inst.reloc.exp.X_add_number | value << 7; |
| 5243 | return SUCCESS; |
| 5244 | } |
| 5245 | |
| 5246 | inst.reloc.type = BFD_RELOC_ARM_IMMEDIATE; |
| 5247 | inst.reloc.pc_rel = 0; |
| 5248 | return SUCCESS; |
| 5249 | } |
| 5250 | |
| 5251 | /* Group relocation information. Each entry in the table contains the |
| 5252 | textual name of the relocation as may appear in assembler source |
| 5253 | and must end with a colon. |
| 5254 | Along with this textual name are the relocation codes to be used if |
| 5255 | the corresponding instruction is an ALU instruction (ADD or SUB only), |
| 5256 | an LDR, an LDRS, or an LDC. */ |
| 5257 | |
| 5258 | struct group_reloc_table_entry |
| 5259 | { |
| 5260 | const char *name; |
| 5261 | int alu_code; |
| 5262 | int ldr_code; |
| 5263 | int ldrs_code; |
| 5264 | int ldc_code; |
| 5265 | }; |
| 5266 | |
| 5267 | typedef enum |
| 5268 | { |
| 5269 | /* Varieties of non-ALU group relocation. */ |
| 5270 | |
| 5271 | GROUP_LDR, |
| 5272 | GROUP_LDRS, |
| 5273 | GROUP_LDC |
| 5274 | } group_reloc_type; |
| 5275 | |
| 5276 | static struct group_reloc_table_entry group_reloc_table[] = |
| 5277 | { /* Program counter relative: */ |
| 5278 | { "pc_g0_nc", |
| 5279 | BFD_RELOC_ARM_ALU_PC_G0_NC, /* ALU */ |
| 5280 | 0, /* LDR */ |
| 5281 | 0, /* LDRS */ |
| 5282 | 0 }, /* LDC */ |
| 5283 | { "pc_g0", |
| 5284 | BFD_RELOC_ARM_ALU_PC_G0, /* ALU */ |
| 5285 | BFD_RELOC_ARM_LDR_PC_G0, /* LDR */ |
| 5286 | BFD_RELOC_ARM_LDRS_PC_G0, /* LDRS */ |
| 5287 | BFD_RELOC_ARM_LDC_PC_G0 }, /* LDC */ |
| 5288 | { "pc_g1_nc", |
| 5289 | BFD_RELOC_ARM_ALU_PC_G1_NC, /* ALU */ |
| 5290 | 0, /* LDR */ |
| 5291 | 0, /* LDRS */ |
| 5292 | 0 }, /* LDC */ |
| 5293 | { "pc_g1", |
| 5294 | BFD_RELOC_ARM_ALU_PC_G1, /* ALU */ |
| 5295 | BFD_RELOC_ARM_LDR_PC_G1, /* LDR */ |
| 5296 | BFD_RELOC_ARM_LDRS_PC_G1, /* LDRS */ |
| 5297 | BFD_RELOC_ARM_LDC_PC_G1 }, /* LDC */ |
| 5298 | { "pc_g2", |
| 5299 | BFD_RELOC_ARM_ALU_PC_G2, /* ALU */ |
| 5300 | BFD_RELOC_ARM_LDR_PC_G2, /* LDR */ |
| 5301 | BFD_RELOC_ARM_LDRS_PC_G2, /* LDRS */ |
| 5302 | BFD_RELOC_ARM_LDC_PC_G2 }, /* LDC */ |
| 5303 | /* Section base relative */ |
| 5304 | { "sb_g0_nc", |
| 5305 | BFD_RELOC_ARM_ALU_SB_G0_NC, /* ALU */ |
| 5306 | 0, /* LDR */ |
| 5307 | 0, /* LDRS */ |
| 5308 | 0 }, /* LDC */ |
| 5309 | { "sb_g0", |
| 5310 | BFD_RELOC_ARM_ALU_SB_G0, /* ALU */ |
| 5311 | BFD_RELOC_ARM_LDR_SB_G0, /* LDR */ |
| 5312 | BFD_RELOC_ARM_LDRS_SB_G0, /* LDRS */ |
| 5313 | BFD_RELOC_ARM_LDC_SB_G0 }, /* LDC */ |
| 5314 | { "sb_g1_nc", |
| 5315 | BFD_RELOC_ARM_ALU_SB_G1_NC, /* ALU */ |
| 5316 | 0, /* LDR */ |
| 5317 | 0, /* LDRS */ |
| 5318 | 0 }, /* LDC */ |
| 5319 | { "sb_g1", |
| 5320 | BFD_RELOC_ARM_ALU_SB_G1, /* ALU */ |
| 5321 | BFD_RELOC_ARM_LDR_SB_G1, /* LDR */ |
| 5322 | BFD_RELOC_ARM_LDRS_SB_G1, /* LDRS */ |
| 5323 | BFD_RELOC_ARM_LDC_SB_G1 }, /* LDC */ |
| 5324 | { "sb_g2", |
| 5325 | BFD_RELOC_ARM_ALU_SB_G2, /* ALU */ |
| 5326 | BFD_RELOC_ARM_LDR_SB_G2, /* LDR */ |
| 5327 | BFD_RELOC_ARM_LDRS_SB_G2, /* LDRS */ |
| 5328 | BFD_RELOC_ARM_LDC_SB_G2 }, /* LDC */ |
| 5329 | /* Absolute thumb alu relocations. */ |
| 5330 | { "lower0_7", |
| 5331 | BFD_RELOC_ARM_THUMB_ALU_ABS_G0_NC,/* ALU. */ |
| 5332 | 0, /* LDR. */ |
| 5333 | 0, /* LDRS. */ |
| 5334 | 0 }, /* LDC. */ |
| 5335 | { "lower8_15", |
| 5336 | BFD_RELOC_ARM_THUMB_ALU_ABS_G1_NC,/* ALU. */ |
| 5337 | 0, /* LDR. */ |
| 5338 | 0, /* LDRS. */ |
| 5339 | 0 }, /* LDC. */ |
| 5340 | { "upper0_7", |
| 5341 | BFD_RELOC_ARM_THUMB_ALU_ABS_G2_NC,/* ALU. */ |
| 5342 | 0, /* LDR. */ |
| 5343 | 0, /* LDRS. */ |
| 5344 | 0 }, /* LDC. */ |
| 5345 | { "upper8_15", |
| 5346 | BFD_RELOC_ARM_THUMB_ALU_ABS_G3_NC,/* ALU. */ |
| 5347 | 0, /* LDR. */ |
| 5348 | 0, /* LDRS. */ |
| 5349 | 0 } }; /* LDC. */ |
| 5350 | |
| 5351 | /* Given the address of a pointer pointing to the textual name of a group |
| 5352 | relocation as may appear in assembler source, attempt to find its details |
| 5353 | in group_reloc_table. The pointer will be updated to the character after |
| 5354 | the trailing colon. On failure, FAIL will be returned; SUCCESS |
| 5355 | otherwise. On success, *entry will be updated to point at the relevant |
| 5356 | group_reloc_table entry. */ |
| 5357 | |
| 5358 | static int |
| 5359 | find_group_reloc_table_entry (char **str, struct group_reloc_table_entry **out) |
| 5360 | { |
| 5361 | unsigned int i; |
| 5362 | for (i = 0; i < ARRAY_SIZE (group_reloc_table); i++) |
| 5363 | { |
| 5364 | int length = strlen (group_reloc_table[i].name); |
| 5365 | |
| 5366 | if (strncasecmp (group_reloc_table[i].name, *str, length) == 0 |
| 5367 | && (*str)[length] == ':') |
| 5368 | { |
| 5369 | *out = &group_reloc_table[i]; |
| 5370 | *str += (length + 1); |
| 5371 | return SUCCESS; |
| 5372 | } |
| 5373 | } |
| 5374 | |
| 5375 | return FAIL; |
| 5376 | } |
| 5377 | |
| 5378 | /* Parse a <shifter_operand> for an ARM data processing instruction |
| 5379 | (as for parse_shifter_operand) where group relocations are allowed: |
| 5380 | |
| 5381 | #<immediate> |
| 5382 | #<immediate>, <rotate> |
| 5383 | #:<group_reloc>:<expression> |
| 5384 | <Rm> |
| 5385 | <Rm>, <shift> |
| 5386 | |
| 5387 | where <group_reloc> is one of the strings defined in group_reloc_table. |
| 5388 | The hashes are optional. |
| 5389 | |
| 5390 | Everything else is as for parse_shifter_operand. */ |
| 5391 | |
| 5392 | static parse_operand_result |
| 5393 | parse_shifter_operand_group_reloc (char **str, int i) |
| 5394 | { |
| 5395 | /* Determine if we have the sequence of characters #: or just : |
| 5396 | coming next. If we do, then we check for a group relocation. |
| 5397 | If we don't, punt the whole lot to parse_shifter_operand. */ |
| 5398 | |
| 5399 | if (((*str)[0] == '#' && (*str)[1] == ':') |
| 5400 | || (*str)[0] == ':') |
| 5401 | { |
| 5402 | struct group_reloc_table_entry *entry; |
| 5403 | |
| 5404 | if ((*str)[0] == '#') |
| 5405 | (*str) += 2; |
| 5406 | else |
| 5407 | (*str)++; |
| 5408 | |
| 5409 | /* Try to parse a group relocation. Anything else is an error. */ |
| 5410 | if (find_group_reloc_table_entry (str, &entry) == FAIL) |
| 5411 | { |
| 5412 | inst.error = _("unknown group relocation"); |
| 5413 | return PARSE_OPERAND_FAIL_NO_BACKTRACK; |
| 5414 | } |
| 5415 | |
| 5416 | /* We now have the group relocation table entry corresponding to |
| 5417 | the name in the assembler source. Next, we parse the expression. */ |
| 5418 | if (my_get_expression (&inst.reloc.exp, str, GE_NO_PREFIX)) |
| 5419 | return PARSE_OPERAND_FAIL_NO_BACKTRACK; |
| 5420 | |
| 5421 | /* Record the relocation type (always the ALU variant here). */ |
| 5422 | inst.reloc.type = (bfd_reloc_code_real_type) entry->alu_code; |
| 5423 | gas_assert (inst.reloc.type != 0); |
| 5424 | |
| 5425 | return PARSE_OPERAND_SUCCESS; |
| 5426 | } |
| 5427 | else |
| 5428 | return parse_shifter_operand (str, i) == SUCCESS |
| 5429 | ? PARSE_OPERAND_SUCCESS : PARSE_OPERAND_FAIL; |
| 5430 | |
| 5431 | /* Never reached. */ |
| 5432 | } |
| 5433 | |
| 5434 | /* Parse a Neon alignment expression. Information is written to |
| 5435 | inst.operands[i]. We assume the initial ':' has been skipped. |
| 5436 | |
| 5437 | align .imm = align << 8, .immisalign=1, .preind=0 */ |
| 5438 | static parse_operand_result |
| 5439 | parse_neon_alignment (char **str, int i) |
| 5440 | { |
| 5441 | char *p = *str; |
| 5442 | expressionS exp; |
| 5443 | |
| 5444 | my_get_expression (&exp, &p, GE_NO_PREFIX); |
| 5445 | |
| 5446 | if (exp.X_op != O_constant) |
| 5447 | { |
| 5448 | inst.error = _("alignment must be constant"); |
| 5449 | return PARSE_OPERAND_FAIL; |
| 5450 | } |
| 5451 | |
| 5452 | inst.operands[i].imm = exp.X_add_number << 8; |
| 5453 | inst.operands[i].immisalign = 1; |
| 5454 | /* Alignments are not pre-indexes. */ |
| 5455 | inst.operands[i].preind = 0; |
| 5456 | |
| 5457 | *str = p; |
| 5458 | return PARSE_OPERAND_SUCCESS; |
| 5459 | } |
| 5460 | |
| 5461 | /* Parse all forms of an ARM address expression. Information is written |
| 5462 | to inst.operands[i] and/or inst.reloc. |
| 5463 | |
| 5464 | Preindexed addressing (.preind=1): |
| 5465 | |
| 5466 | [Rn, #offset] .reg=Rn .reloc.exp=offset |
| 5467 | [Rn, +/-Rm] .reg=Rn .imm=Rm .immisreg=1 .negative=0/1 |
| 5468 | [Rn, +/-Rm, shift] .reg=Rn .imm=Rm .immisreg=1 .negative=0/1 |
| 5469 | .shift_kind=shift .reloc.exp=shift_imm |
| 5470 | |
| 5471 | These three may have a trailing ! which causes .writeback to be set also. |
| 5472 | |
| 5473 | Postindexed addressing (.postind=1, .writeback=1): |
| 5474 | |
| 5475 | [Rn], #offset .reg=Rn .reloc.exp=offset |
| 5476 | [Rn], +/-Rm .reg=Rn .imm=Rm .immisreg=1 .negative=0/1 |
| 5477 | [Rn], +/-Rm, shift .reg=Rn .imm=Rm .immisreg=1 .negative=0/1 |
| 5478 | .shift_kind=shift .reloc.exp=shift_imm |
| 5479 | |
| 5480 | Unindexed addressing (.preind=0, .postind=0): |
| 5481 | |
| 5482 | [Rn], {option} .reg=Rn .imm=option .immisreg=0 |
| 5483 | |
| 5484 | Other: |
| 5485 | |
| 5486 | [Rn]{!} shorthand for [Rn,#0]{!} |
| 5487 | =immediate .isreg=0 .reloc.exp=immediate |
| 5488 | label .reg=PC .reloc.pc_rel=1 .reloc.exp=label |
| 5489 | |
| 5490 | It is the caller's responsibility to check for addressing modes not |
| 5491 | supported by the instruction, and to set inst.reloc.type. */ |
| 5492 | |
| 5493 | static parse_operand_result |
| 5494 | parse_address_main (char **str, int i, int group_relocations, |
| 5495 | group_reloc_type group_type) |
| 5496 | { |
| 5497 | char *p = *str; |
| 5498 | int reg; |
| 5499 | |
| 5500 | if (skip_past_char (&p, '[') == FAIL) |
| 5501 | { |
| 5502 | if (skip_past_char (&p, '=') == FAIL) |
| 5503 | { |
| 5504 | /* Bare address - translate to PC-relative offset. */ |
| 5505 | inst.reloc.pc_rel = 1; |
| 5506 | inst.operands[i].reg = REG_PC; |
| 5507 | inst.operands[i].isreg = 1; |
| 5508 | inst.operands[i].preind = 1; |
| 5509 | |
| 5510 | if (my_get_expression (&inst.reloc.exp, &p, GE_OPT_PREFIX_BIG)) |
| 5511 | return PARSE_OPERAND_FAIL; |
| 5512 | } |
| 5513 | else if (parse_big_immediate (&p, i, &inst.reloc.exp, |
| 5514 | /*allow_symbol_p=*/TRUE)) |
| 5515 | return PARSE_OPERAND_FAIL; |
| 5516 | |
| 5517 | *str = p; |
| 5518 | return PARSE_OPERAND_SUCCESS; |
| 5519 | } |
| 5520 | |
| 5521 | /* PR gas/14887: Allow for whitespace after the opening bracket. */ |
| 5522 | skip_whitespace (p); |
| 5523 | |
| 5524 | if ((reg = arm_reg_parse (&p, REG_TYPE_RN)) == FAIL) |
| 5525 | { |
| 5526 | inst.error = _(reg_expected_msgs[REG_TYPE_RN]); |
| 5527 | return PARSE_OPERAND_FAIL; |
| 5528 | } |
| 5529 | inst.operands[i].reg = reg; |
| 5530 | inst.operands[i].isreg = 1; |
| 5531 | |
| 5532 | if (skip_past_comma (&p) == SUCCESS) |
| 5533 | { |
| 5534 | inst.operands[i].preind = 1; |
| 5535 | |
| 5536 | if (*p == '+') p++; |
| 5537 | else if (*p == '-') p++, inst.operands[i].negative = 1; |
| 5538 | |
| 5539 | if ((reg = arm_reg_parse (&p, REG_TYPE_RN)) != FAIL) |
| 5540 | { |
| 5541 | inst.operands[i].imm = reg; |
| 5542 | inst.operands[i].immisreg = 1; |
| 5543 | |
| 5544 | if (skip_past_comma (&p) == SUCCESS) |
| 5545 | if (parse_shift (&p, i, SHIFT_IMMEDIATE) == FAIL) |
| 5546 | return PARSE_OPERAND_FAIL; |
| 5547 | } |
| 5548 | else if (skip_past_char (&p, ':') == SUCCESS) |
| 5549 | { |
| 5550 | /* FIXME: '@' should be used here, but it's filtered out by generic |
| 5551 | code before we get to see it here. This may be subject to |
| 5552 | change. */ |
| 5553 | parse_operand_result result = parse_neon_alignment (&p, i); |
| 5554 | |
| 5555 | if (result != PARSE_OPERAND_SUCCESS) |
| 5556 | return result; |
| 5557 | } |
| 5558 | else |
| 5559 | { |
| 5560 | if (inst.operands[i].negative) |
| 5561 | { |
| 5562 | inst.operands[i].negative = 0; |
| 5563 | p--; |
| 5564 | } |
| 5565 | |
| 5566 | if (group_relocations |
| 5567 | && ((*p == '#' && *(p + 1) == ':') || *p == ':')) |
| 5568 | { |
| 5569 | struct group_reloc_table_entry *entry; |
| 5570 | |
| 5571 | /* Skip over the #: or : sequence. */ |
| 5572 | if (*p == '#') |
| 5573 | p += 2; |
| 5574 | else |
| 5575 | p++; |
| 5576 | |
| 5577 | /* Try to parse a group relocation. Anything else is an |
| 5578 | error. */ |
| 5579 | if (find_group_reloc_table_entry (&p, &entry) == FAIL) |
| 5580 | { |
| 5581 | inst.error = _("unknown group relocation"); |
| 5582 | return PARSE_OPERAND_FAIL_NO_BACKTRACK; |
| 5583 | } |
| 5584 | |
| 5585 | /* We now have the group relocation table entry corresponding to |
| 5586 | the name in the assembler source. Next, we parse the |
| 5587 | expression. */ |
| 5588 | if (my_get_expression (&inst.reloc.exp, &p, GE_NO_PREFIX)) |
| 5589 | return PARSE_OPERAND_FAIL_NO_BACKTRACK; |
| 5590 | |
| 5591 | /* Record the relocation type. */ |
| 5592 | switch (group_type) |
| 5593 | { |
| 5594 | case GROUP_LDR: |
| 5595 | inst.reloc.type = (bfd_reloc_code_real_type) entry->ldr_code; |
| 5596 | break; |
| 5597 | |
| 5598 | case GROUP_LDRS: |
| 5599 | inst.reloc.type = (bfd_reloc_code_real_type) entry->ldrs_code; |
| 5600 | break; |
| 5601 | |
| 5602 | case GROUP_LDC: |
| 5603 | inst.reloc.type = (bfd_reloc_code_real_type) entry->ldc_code; |
| 5604 | break; |
| 5605 | |
| 5606 | default: |
| 5607 | gas_assert (0); |
| 5608 | } |
| 5609 | |
| 5610 | if (inst.reloc.type == 0) |
| 5611 | { |
| 5612 | inst.error = _("this group relocation is not allowed on this instruction"); |
| 5613 | return PARSE_OPERAND_FAIL_NO_BACKTRACK; |
| 5614 | } |
| 5615 | } |
| 5616 | else |
| 5617 | { |
| 5618 | char *q = p; |
| 5619 | if (my_get_expression (&inst.reloc.exp, &p, GE_IMM_PREFIX)) |
| 5620 | return PARSE_OPERAND_FAIL; |
| 5621 | /* If the offset is 0, find out if it's a +0 or -0. */ |
| 5622 | if (inst.reloc.exp.X_op == O_constant |
| 5623 | && inst.reloc.exp.X_add_number == 0) |
| 5624 | { |
| 5625 | skip_whitespace (q); |
| 5626 | if (*q == '#') |
| 5627 | { |
| 5628 | q++; |
| 5629 | skip_whitespace (q); |
| 5630 | } |
| 5631 | if (*q == '-') |
| 5632 | inst.operands[i].negative = 1; |
| 5633 | } |
| 5634 | } |
| 5635 | } |
| 5636 | } |
| 5637 | else if (skip_past_char (&p, ':') == SUCCESS) |
| 5638 | { |
| 5639 | /* FIXME: '@' should be used here, but it's filtered out by generic code |
| 5640 | before we get to see it here. This may be subject to change. */ |
| 5641 | parse_operand_result result = parse_neon_alignment (&p, i); |
| 5642 | |
| 5643 | if (result != PARSE_OPERAND_SUCCESS) |
| 5644 | return result; |
| 5645 | } |
| 5646 | |
| 5647 | if (skip_past_char (&p, ']') == FAIL) |
| 5648 | { |
| 5649 | inst.error = _("']' expected"); |
| 5650 | return PARSE_OPERAND_FAIL; |
| 5651 | } |
| 5652 | |
| 5653 | if (skip_past_char (&p, '!') == SUCCESS) |
| 5654 | inst.operands[i].writeback = 1; |
| 5655 | |
| 5656 | else if (skip_past_comma (&p) == SUCCESS) |
| 5657 | { |
| 5658 | if (skip_past_char (&p, '{') == SUCCESS) |
| 5659 | { |
| 5660 | /* [Rn], {expr} - unindexed, with option */ |
| 5661 | if (parse_immediate (&p, &inst.operands[i].imm, |
| 5662 | 0, 255, TRUE) == FAIL) |
| 5663 | return PARSE_OPERAND_FAIL; |
| 5664 | |
| 5665 | if (skip_past_char (&p, '}') == FAIL) |
| 5666 | { |
| 5667 | inst.error = _("'}' expected at end of 'option' field"); |
| 5668 | return PARSE_OPERAND_FAIL; |
| 5669 | } |
| 5670 | if (inst.operands[i].preind) |
| 5671 | { |
| 5672 | inst.error = _("cannot combine index with option"); |
| 5673 | return PARSE_OPERAND_FAIL; |
| 5674 | } |
| 5675 | *str = p; |
| 5676 | return PARSE_OPERAND_SUCCESS; |
| 5677 | } |
| 5678 | else |
| 5679 | { |
| 5680 | inst.operands[i].postind = 1; |
| 5681 | inst.operands[i].writeback = 1; |
| 5682 | |
| 5683 | if (inst.operands[i].preind) |
| 5684 | { |
| 5685 | inst.error = _("cannot combine pre- and post-indexing"); |
| 5686 | return PARSE_OPERAND_FAIL; |
| 5687 | } |
| 5688 | |
| 5689 | if (*p == '+') p++; |
| 5690 | else if (*p == '-') p++, inst.operands[i].negative = 1; |
| 5691 | |
| 5692 | if ((reg = arm_reg_parse (&p, REG_TYPE_RN)) != FAIL) |
| 5693 | { |
| 5694 | /* We might be using the immediate for alignment already. If we |
| 5695 | are, OR the register number into the low-order bits. */ |
| 5696 | if (inst.operands[i].immisalign) |
| 5697 | inst.operands[i].imm |= reg; |
| 5698 | else |
| 5699 | inst.operands[i].imm = reg; |
| 5700 | inst.operands[i].immisreg = 1; |
| 5701 | |
| 5702 | if (skip_past_comma (&p) == SUCCESS) |
| 5703 | if (parse_shift (&p, i, SHIFT_IMMEDIATE) == FAIL) |
| 5704 | return PARSE_OPERAND_FAIL; |
| 5705 | } |
| 5706 | else |
| 5707 | { |
| 5708 | char *q = p; |
| 5709 | if (inst.operands[i].negative) |
| 5710 | { |
| 5711 | inst.operands[i].negative = 0; |
| 5712 | p--; |
| 5713 | } |
| 5714 | if (my_get_expression (&inst.reloc.exp, &p, GE_IMM_PREFIX)) |
| 5715 | return PARSE_OPERAND_FAIL; |
| 5716 | /* If the offset is 0, find out if it's a +0 or -0. */ |
| 5717 | if (inst.reloc.exp.X_op == O_constant |
| 5718 | && inst.reloc.exp.X_add_number == 0) |
| 5719 | { |
| 5720 | skip_whitespace (q); |
| 5721 | if (*q == '#') |
| 5722 | { |
| 5723 | q++; |
| 5724 | skip_whitespace (q); |
| 5725 | } |
| 5726 | if (*q == '-') |
| 5727 | inst.operands[i].negative = 1; |
| 5728 | } |
| 5729 | } |
| 5730 | } |
| 5731 | } |
| 5732 | |
| 5733 | /* If at this point neither .preind nor .postind is set, we have a |
| 5734 | bare [Rn]{!}, which is shorthand for [Rn,#0]{!}. */ |
| 5735 | if (inst.operands[i].preind == 0 && inst.operands[i].postind == 0) |
| 5736 | { |
| 5737 | inst.operands[i].preind = 1; |
| 5738 | inst.reloc.exp.X_op = O_constant; |
| 5739 | inst.reloc.exp.X_add_number = 0; |
| 5740 | } |
| 5741 | *str = p; |
| 5742 | return PARSE_OPERAND_SUCCESS; |
| 5743 | } |
| 5744 | |
| 5745 | static int |
| 5746 | parse_address (char **str, int i) |
| 5747 | { |
| 5748 | return parse_address_main (str, i, 0, GROUP_LDR) == PARSE_OPERAND_SUCCESS |
| 5749 | ? SUCCESS : FAIL; |
| 5750 | } |
| 5751 | |
| 5752 | static parse_operand_result |
| 5753 | parse_address_group_reloc (char **str, int i, group_reloc_type type) |
| 5754 | { |
| 5755 | return parse_address_main (str, i, 1, type); |
| 5756 | } |
| 5757 | |
| 5758 | /* Parse an operand for a MOVW or MOVT instruction. */ |
| 5759 | static int |
| 5760 | parse_half (char **str) |
| 5761 | { |
| 5762 | char * p; |
| 5763 | |
| 5764 | p = *str; |
| 5765 | skip_past_char (&p, '#'); |
| 5766 | if (strncasecmp (p, ":lower16:", 9) == 0) |
| 5767 | inst.reloc.type = BFD_RELOC_ARM_MOVW; |
| 5768 | else if (strncasecmp (p, ":upper16:", 9) == 0) |
| 5769 | inst.reloc.type = BFD_RELOC_ARM_MOVT; |
| 5770 | |
| 5771 | if (inst.reloc.type != BFD_RELOC_UNUSED) |
| 5772 | { |
| 5773 | p += 9; |
| 5774 | skip_whitespace (p); |
| 5775 | } |
| 5776 | |
| 5777 | if (my_get_expression (&inst.reloc.exp, &p, GE_NO_PREFIX)) |
| 5778 | return FAIL; |
| 5779 | |
| 5780 | if (inst.reloc.type == BFD_RELOC_UNUSED) |
| 5781 | { |
| 5782 | if (inst.reloc.exp.X_op != O_constant) |
| 5783 | { |
| 5784 | inst.error = _("constant expression expected"); |
| 5785 | return FAIL; |
| 5786 | } |
| 5787 | if (inst.reloc.exp.X_add_number < 0 |
| 5788 | || inst.reloc.exp.X_add_number > 0xffff) |
| 5789 | { |
| 5790 | inst.error = _("immediate value out of range"); |
| 5791 | return FAIL; |
| 5792 | } |
| 5793 | } |
| 5794 | *str = p; |
| 5795 | return SUCCESS; |
| 5796 | } |
| 5797 | |
| 5798 | /* Miscellaneous. */ |
| 5799 | |
| 5800 | /* Parse a PSR flag operand. The value returned is FAIL on syntax error, |
| 5801 | or a bitmask suitable to be or-ed into the ARM msr instruction. */ |
| 5802 | static int |
| 5803 | parse_psr (char **str, bfd_boolean lhs) |
| 5804 | { |
| 5805 | char *p; |
| 5806 | unsigned long psr_field; |
| 5807 | const struct asm_psr *psr; |
| 5808 | char *start; |
| 5809 | bfd_boolean is_apsr = FALSE; |
| 5810 | bfd_boolean m_profile = ARM_CPU_HAS_FEATURE (selected_cpu, arm_ext_m); |
| 5811 | |
| 5812 | /* PR gas/12698: If the user has specified -march=all then m_profile will |
| 5813 | be TRUE, but we want to ignore it in this case as we are building for any |
| 5814 | CPU type, including non-m variants. */ |
| 5815 | if (ARM_FEATURE_CORE_EQUAL (selected_cpu, arm_arch_any)) |
| 5816 | m_profile = FALSE; |
| 5817 | |
| 5818 | /* CPSR's and SPSR's can now be lowercase. This is just a convenience |
| 5819 | feature for ease of use and backwards compatibility. */ |
| 5820 | p = *str; |
| 5821 | if (strncasecmp (p, "SPSR", 4) == 0) |
| 5822 | { |
| 5823 | if (m_profile) |
| 5824 | goto unsupported_psr; |
| 5825 | |
| 5826 | psr_field = SPSR_BIT; |
| 5827 | } |
| 5828 | else if (strncasecmp (p, "CPSR", 4) == 0) |
| 5829 | { |
| 5830 | if (m_profile) |
| 5831 | goto unsupported_psr; |
| 5832 | |
| 5833 | psr_field = 0; |
| 5834 | } |
| 5835 | else if (strncasecmp (p, "APSR", 4) == 0) |
| 5836 | { |
| 5837 | /* APSR[_<bits>] can be used as a synonym for CPSR[_<flags>] on ARMv7-A |
| 5838 | and ARMv7-R architecture CPUs. */ |
| 5839 | is_apsr = TRUE; |
| 5840 | psr_field = 0; |
| 5841 | } |
| 5842 | else if (m_profile) |
| 5843 | { |
| 5844 | start = p; |
| 5845 | do |
| 5846 | p++; |
| 5847 | while (ISALNUM (*p) || *p == '_'); |
| 5848 | |
| 5849 | if (strncasecmp (start, "iapsr", 5) == 0 |
| 5850 | || strncasecmp (start, "eapsr", 5) == 0 |
| 5851 | || strncasecmp (start, "xpsr", 4) == 0 |
| 5852 | || strncasecmp (start, "psr", 3) == 0) |
| 5853 | p = start + strcspn (start, "rR") + 1; |
| 5854 | |
| 5855 | psr = (const struct asm_psr *) hash_find_n (arm_v7m_psr_hsh, start, |
| 5856 | p - start); |
| 5857 | |
| 5858 | if (!psr) |
| 5859 | return FAIL; |
| 5860 | |
| 5861 | /* If APSR is being written, a bitfield may be specified. Note that |
| 5862 | APSR itself is handled above. */ |
| 5863 | if (psr->field <= 3) |
| 5864 | { |
| 5865 | psr_field = psr->field; |
| 5866 | is_apsr = TRUE; |
| 5867 | goto check_suffix; |
| 5868 | } |
| 5869 | |
| 5870 | *str = p; |
| 5871 | /* M-profile MSR instructions have the mask field set to "10", except |
| 5872 | *PSR variants which modify APSR, which may use a different mask (and |
| 5873 | have been handled already). Do that by setting the PSR_f field |
| 5874 | here. */ |
| 5875 | return psr->field | (lhs ? PSR_f : 0); |
| 5876 | } |
| 5877 | else |
| 5878 | goto unsupported_psr; |
| 5879 | |
| 5880 | p += 4; |
| 5881 | check_suffix: |
| 5882 | if (*p == '_') |
| 5883 | { |
| 5884 | /* A suffix follows. */ |
| 5885 | p++; |
| 5886 | start = p; |
| 5887 | |
| 5888 | do |
| 5889 | p++; |
| 5890 | while (ISALNUM (*p) || *p == '_'); |
| 5891 | |
| 5892 | if (is_apsr) |
| 5893 | { |
| 5894 | /* APSR uses a notation for bits, rather than fields. */ |
| 5895 | unsigned int nzcvq_bits = 0; |
| 5896 | unsigned int g_bit = 0; |
| 5897 | char *bit; |
| 5898 | |
| 5899 | for (bit = start; bit != p; bit++) |
| 5900 | { |
| 5901 | switch (TOLOWER (*bit)) |
| 5902 | { |
| 5903 | case 'n': |
| 5904 | nzcvq_bits |= (nzcvq_bits & 0x01) ? 0x20 : 0x01; |
| 5905 | break; |
| 5906 | |
| 5907 | case 'z': |
| 5908 | nzcvq_bits |= (nzcvq_bits & 0x02) ? 0x20 : 0x02; |
| 5909 | break; |
| 5910 | |
| 5911 | case 'c': |
| 5912 | nzcvq_bits |= (nzcvq_bits & 0x04) ? 0x20 : 0x04; |
| 5913 | break; |
| 5914 | |
| 5915 | case 'v': |
| 5916 | nzcvq_bits |= (nzcvq_bits & 0x08) ? 0x20 : 0x08; |
| 5917 | break; |
| 5918 | |
| 5919 | case 'q': |
| 5920 | nzcvq_bits |= (nzcvq_bits & 0x10) ? 0x20 : 0x10; |
| 5921 | break; |
| 5922 | |
| 5923 | case 'g': |
| 5924 | g_bit |= (g_bit & 0x1) ? 0x2 : 0x1; |
| 5925 | break; |
| 5926 | |
| 5927 | default: |
| 5928 | inst.error = _("unexpected bit specified after APSR"); |
| 5929 | return FAIL; |
| 5930 | } |
| 5931 | } |
| 5932 | |
| 5933 | if (nzcvq_bits == 0x1f) |
| 5934 | psr_field |= PSR_f; |
| 5935 | |
| 5936 | if (g_bit == 0x1) |
| 5937 | { |
| 5938 | if (!ARM_CPU_HAS_FEATURE (selected_cpu, arm_ext_v6_dsp)) |
| 5939 | { |
| 5940 | inst.error = _("selected processor does not " |
| 5941 | "support DSP extension"); |
| 5942 | return FAIL; |
| 5943 | } |
| 5944 | |
| 5945 | psr_field |= PSR_s; |
| 5946 | } |
| 5947 | |
| 5948 | if ((nzcvq_bits & 0x20) != 0 |
| 5949 | || (nzcvq_bits != 0x1f && nzcvq_bits != 0) |
| 5950 | || (g_bit & 0x2) != 0) |
| 5951 | { |
| 5952 | inst.error = _("bad bitmask specified after APSR"); |
| 5953 | return FAIL; |
| 5954 | } |
| 5955 | } |
| 5956 | else |
| 5957 | { |
| 5958 | psr = (const struct asm_psr *) hash_find_n (arm_psr_hsh, start, |
| 5959 | p - start); |
| 5960 | if (!psr) |
| 5961 | goto error; |
| 5962 | |
| 5963 | psr_field |= psr->field; |
| 5964 | } |
| 5965 | } |
| 5966 | else |
| 5967 | { |
| 5968 | if (ISALNUM (*p)) |
| 5969 | goto error; /* Garbage after "[CS]PSR". */ |
| 5970 | |
| 5971 | /* Unadorned APSR is equivalent to APSR_nzcvq/CPSR_f (for writes). This |
| 5972 | is deprecated, but allow it anyway. */ |
| 5973 | if (is_apsr && lhs) |
| 5974 | { |
| 5975 | psr_field |= PSR_f; |
| 5976 | as_tsktsk (_("writing to APSR without specifying a bitmask is " |
| 5977 | "deprecated")); |
| 5978 | } |
| 5979 | else if (!m_profile) |
| 5980 | /* These bits are never right for M-profile devices: don't set them |
| 5981 | (only code paths which read/write APSR reach here). */ |
| 5982 | psr_field |= (PSR_c | PSR_f); |
| 5983 | } |
| 5984 | *str = p; |
| 5985 | return psr_field; |
| 5986 | |
| 5987 | unsupported_psr: |
| 5988 | inst.error = _("selected processor does not support requested special " |
| 5989 | "purpose register"); |
| 5990 | return FAIL; |
| 5991 | |
| 5992 | error: |
| 5993 | inst.error = _("flag for {c}psr instruction expected"); |
| 5994 | return FAIL; |
| 5995 | } |
| 5996 | |
| 5997 | /* Parse the flags argument to CPSI[ED]. Returns FAIL on error, or a |
| 5998 | value suitable for splatting into the AIF field of the instruction. */ |
| 5999 | |
| 6000 | static int |
| 6001 | parse_cps_flags (char **str) |
| 6002 | { |
| 6003 | int val = 0; |
| 6004 | int saw_a_flag = 0; |
| 6005 | char *s = *str; |
| 6006 | |
| 6007 | for (;;) |
| 6008 | switch (*s++) |
| 6009 | { |
| 6010 | case '\0': case ',': |
| 6011 | goto done; |
| 6012 | |
| 6013 | case 'a': case 'A': saw_a_flag = 1; val |= 0x4; break; |
| 6014 | case 'i': case 'I': saw_a_flag = 1; val |= 0x2; break; |
| 6015 | case 'f': case 'F': saw_a_flag = 1; val |= 0x1; break; |
| 6016 | |
| 6017 | default: |
| 6018 | inst.error = _("unrecognized CPS flag"); |
| 6019 | return FAIL; |
| 6020 | } |
| 6021 | |
| 6022 | done: |
| 6023 | if (saw_a_flag == 0) |
| 6024 | { |
| 6025 | inst.error = _("missing CPS flags"); |
| 6026 | return FAIL; |
| 6027 | } |
| 6028 | |
| 6029 | *str = s - 1; |
| 6030 | return val; |
| 6031 | } |
| 6032 | |
| 6033 | /* Parse an endian specifier ("BE" or "LE", case insensitive); |
| 6034 | returns 0 for big-endian, 1 for little-endian, FAIL for an error. */ |
| 6035 | |
| 6036 | static int |
| 6037 | parse_endian_specifier (char **str) |
| 6038 | { |
| 6039 | int little_endian; |
| 6040 | char *s = *str; |
| 6041 | |
| 6042 | if (strncasecmp (s, "BE", 2)) |
| 6043 | little_endian = 0; |
| 6044 | else if (strncasecmp (s, "LE", 2)) |
| 6045 | little_endian = 1; |
| 6046 | else |
| 6047 | { |
| 6048 | inst.error = _("valid endian specifiers are be or le"); |
| 6049 | return FAIL; |
| 6050 | } |
| 6051 | |
| 6052 | if (ISALNUM (s[2]) || s[2] == '_') |
| 6053 | { |
| 6054 | inst.error = _("valid endian specifiers are be or le"); |
| 6055 | return FAIL; |
| 6056 | } |
| 6057 | |
| 6058 | *str = s + 2; |
| 6059 | return little_endian; |
| 6060 | } |
| 6061 | |
| 6062 | /* Parse a rotation specifier: ROR #0, #8, #16, #24. *val receives a |
| 6063 | value suitable for poking into the rotate field of an sxt or sxta |
| 6064 | instruction, or FAIL on error. */ |
| 6065 | |
| 6066 | static int |
| 6067 | parse_ror (char **str) |
| 6068 | { |
| 6069 | int rot; |
| 6070 | char *s = *str; |
| 6071 | |
| 6072 | if (strncasecmp (s, "ROR", 3) == 0) |
| 6073 | s += 3; |
| 6074 | else |
| 6075 | { |
| 6076 | inst.error = _("missing rotation field after comma"); |
| 6077 | return FAIL; |
| 6078 | } |
| 6079 | |
| 6080 | if (parse_immediate (&s, &rot, 0, 24, FALSE) == FAIL) |
| 6081 | return FAIL; |
| 6082 | |
| 6083 | switch (rot) |
| 6084 | { |
| 6085 | case 0: *str = s; return 0x0; |
| 6086 | case 8: *str = s; return 0x1; |
| 6087 | case 16: *str = s; return 0x2; |
| 6088 | case 24: *str = s; return 0x3; |
| 6089 | |
| 6090 | default: |
| 6091 | inst.error = _("rotation can only be 0, 8, 16, or 24"); |
| 6092 | return FAIL; |
| 6093 | } |
| 6094 | } |
| 6095 | |
| 6096 | /* Parse a conditional code (from conds[] below). The value returned is in the |
| 6097 | range 0 .. 14, or FAIL. */ |
| 6098 | static int |
| 6099 | parse_cond (char **str) |
| 6100 | { |
| 6101 | char *q; |
| 6102 | const struct asm_cond *c; |
| 6103 | int n; |
| 6104 | /* Condition codes are always 2 characters, so matching up to |
| 6105 | 3 characters is sufficient. */ |
| 6106 | char cond[3]; |
| 6107 | |
| 6108 | q = *str; |
| 6109 | n = 0; |
| 6110 | while (ISALPHA (*q) && n < 3) |
| 6111 | { |
| 6112 | cond[n] = TOLOWER (*q); |
| 6113 | q++; |
| 6114 | n++; |
| 6115 | } |
| 6116 | |
| 6117 | c = (const struct asm_cond *) hash_find_n (arm_cond_hsh, cond, n); |
| 6118 | if (!c) |
| 6119 | { |
| 6120 | inst.error = _("condition required"); |
| 6121 | return FAIL; |
| 6122 | } |
| 6123 | |
| 6124 | *str = q; |
| 6125 | return c->value; |
| 6126 | } |
| 6127 | |
| 6128 | /* Record a use of the given feature. */ |
| 6129 | static void |
| 6130 | record_feature_use (const arm_feature_set *feature) |
| 6131 | { |
| 6132 | if (thumb_mode) |
| 6133 | ARM_MERGE_FEATURE_SETS (thumb_arch_used, thumb_arch_used, *feature); |
| 6134 | else |
| 6135 | ARM_MERGE_FEATURE_SETS (arm_arch_used, arm_arch_used, *feature); |
| 6136 | } |
| 6137 | |
| 6138 | /* If the given feature available in the selected CPU, mark it as used. |
| 6139 | Returns TRUE iff feature is available. */ |
| 6140 | static bfd_boolean |
| 6141 | mark_feature_used (const arm_feature_set *feature) |
| 6142 | { |
| 6143 | /* Ensure the option is valid on the current architecture. */ |
| 6144 | if (!ARM_CPU_HAS_FEATURE (cpu_variant, *feature)) |
| 6145 | return FALSE; |
| 6146 | |
| 6147 | /* Add the appropriate architecture feature for the barrier option used. |
| 6148 | */ |
| 6149 | record_feature_use (feature); |
| 6150 | |
| 6151 | return TRUE; |
| 6152 | } |
| 6153 | |
| 6154 | /* Parse an option for a barrier instruction. Returns the encoding for the |
| 6155 | option, or FAIL. */ |
| 6156 | static int |
| 6157 | parse_barrier (char **str) |
| 6158 | { |
| 6159 | char *p, *q; |
| 6160 | const struct asm_barrier_opt *o; |
| 6161 | |
| 6162 | p = q = *str; |
| 6163 | while (ISALPHA (*q)) |
| 6164 | q++; |
| 6165 | |
| 6166 | o = (const struct asm_barrier_opt *) hash_find_n (arm_barrier_opt_hsh, p, |
| 6167 | q - p); |
| 6168 | if (!o) |
| 6169 | return FAIL; |
| 6170 | |
| 6171 | if (!mark_feature_used (&o->arch)) |
| 6172 | return FAIL; |
| 6173 | |
| 6174 | *str = q; |
| 6175 | return o->value; |
| 6176 | } |
| 6177 | |
| 6178 | /* Parse the operands of a table branch instruction. Similar to a memory |
| 6179 | operand. */ |
| 6180 | static int |
| 6181 | parse_tb (char **str) |
| 6182 | { |
| 6183 | char * p = *str; |
| 6184 | int reg; |
| 6185 | |
| 6186 | if (skip_past_char (&p, '[') == FAIL) |
| 6187 | { |
| 6188 | inst.error = _("'[' expected"); |
| 6189 | return FAIL; |
| 6190 | } |
| 6191 | |
| 6192 | if ((reg = arm_reg_parse (&p, REG_TYPE_RN)) == FAIL) |
| 6193 | { |
| 6194 | inst.error = _(reg_expected_msgs[REG_TYPE_RN]); |
| 6195 | return FAIL; |
| 6196 | } |
| 6197 | inst.operands[0].reg = reg; |
| 6198 | |
| 6199 | if (skip_past_comma (&p) == FAIL) |
| 6200 | { |
| 6201 | inst.error = _("',' expected"); |
| 6202 | return FAIL; |
| 6203 | } |
| 6204 | |
| 6205 | if ((reg = arm_reg_parse (&p, REG_TYPE_RN)) == FAIL) |
| 6206 | { |
| 6207 | inst.error = _(reg_expected_msgs[REG_TYPE_RN]); |
| 6208 | return FAIL; |
| 6209 | } |
| 6210 | inst.operands[0].imm = reg; |
| 6211 | |
| 6212 | if (skip_past_comma (&p) == SUCCESS) |
| 6213 | { |
| 6214 | if (parse_shift (&p, 0, SHIFT_LSL_IMMEDIATE) == FAIL) |
| 6215 | return FAIL; |
| 6216 | if (inst.reloc.exp.X_add_number != 1) |
| 6217 | { |
| 6218 | inst.error = _("invalid shift"); |
| 6219 | return FAIL; |
| 6220 | } |
| 6221 | inst.operands[0].shifted = 1; |
| 6222 | } |
| 6223 | |
| 6224 | if (skip_past_char (&p, ']') == FAIL) |
| 6225 | { |
| 6226 | inst.error = _("']' expected"); |
| 6227 | return FAIL; |
| 6228 | } |
| 6229 | *str = p; |
| 6230 | return SUCCESS; |
| 6231 | } |
| 6232 | |
| 6233 | /* Parse the operands of a Neon VMOV instruction. See do_neon_mov for more |
| 6234 | information on the types the operands can take and how they are encoded. |
| 6235 | Up to four operands may be read; this function handles setting the |
| 6236 | ".present" field for each read operand itself. |
| 6237 | Updates STR and WHICH_OPERAND if parsing is successful and returns SUCCESS, |
| 6238 | else returns FAIL. */ |
| 6239 | |
| 6240 | static int |
| 6241 | parse_neon_mov (char **str, int *which_operand) |
| 6242 | { |
| 6243 | int i = *which_operand, val; |
| 6244 | enum arm_reg_type rtype; |
| 6245 | char *ptr = *str; |
| 6246 | struct neon_type_el optype; |
| 6247 | |
| 6248 | if ((val = parse_scalar (&ptr, 8, &optype)) != FAIL) |
| 6249 | { |
| 6250 | /* Case 4: VMOV<c><q>.<size> <Dn[x]>, <Rd>. */ |
| 6251 | inst.operands[i].reg = val; |
| 6252 | inst.operands[i].isscalar = 1; |
| 6253 | inst.operands[i].vectype = optype; |
| 6254 | inst.operands[i++].present = 1; |
| 6255 | |
| 6256 | if (skip_past_comma (&ptr) == FAIL) |
| 6257 | goto wanted_comma; |
| 6258 | |
| 6259 | if ((val = arm_reg_parse (&ptr, REG_TYPE_RN)) == FAIL) |
| 6260 | goto wanted_arm; |
| 6261 | |
| 6262 | inst.operands[i].reg = val; |
| 6263 | inst.operands[i].isreg = 1; |
| 6264 | inst.operands[i].present = 1; |
| 6265 | } |
| 6266 | else if ((val = arm_typed_reg_parse (&ptr, REG_TYPE_NSDQ, &rtype, &optype)) |
| 6267 | != FAIL) |
| 6268 | { |
| 6269 | /* Cases 0, 1, 2, 3, 5 (D only). */ |
| 6270 | if (skip_past_comma (&ptr) == FAIL) |
| 6271 | goto wanted_comma; |
| 6272 | |
| 6273 | inst.operands[i].reg = val; |
| 6274 | inst.operands[i].isreg = 1; |
| 6275 | inst.operands[i].isquad = (rtype == REG_TYPE_NQ); |
| 6276 | inst.operands[i].issingle = (rtype == REG_TYPE_VFS); |
| 6277 | inst.operands[i].isvec = 1; |
| 6278 | inst.operands[i].vectype = optype; |
| 6279 | inst.operands[i++].present = 1; |
| 6280 | |
| 6281 | if ((val = arm_reg_parse (&ptr, REG_TYPE_RN)) != FAIL) |
| 6282 | { |
| 6283 | /* Case 5: VMOV<c><q> <Dm>, <Rd>, <Rn>. |
| 6284 | Case 13: VMOV <Sd>, <Rm> */ |
| 6285 | inst.operands[i].reg = val; |
| 6286 | inst.operands[i].isreg = 1; |
| 6287 | inst.operands[i].present = 1; |
| 6288 | |
| 6289 | if (rtype == REG_TYPE_NQ) |
| 6290 | { |
| 6291 | first_error (_("can't use Neon quad register here")); |
| 6292 | return FAIL; |
| 6293 | } |
| 6294 | else if (rtype != REG_TYPE_VFS) |
| 6295 | { |
| 6296 | i++; |
| 6297 | if (skip_past_comma (&ptr) == FAIL) |
| 6298 | goto wanted_comma; |
| 6299 | if ((val = arm_reg_parse (&ptr, REG_TYPE_RN)) == FAIL) |
| 6300 | goto wanted_arm; |
| 6301 | inst.operands[i].reg = val; |
| 6302 | inst.operands[i].isreg = 1; |
| 6303 | inst.operands[i].present = 1; |
| 6304 | } |
| 6305 | } |
| 6306 | else if ((val = arm_typed_reg_parse (&ptr, REG_TYPE_NSDQ, &rtype, |
| 6307 | &optype)) != FAIL) |
| 6308 | { |
| 6309 | /* Case 0: VMOV<c><q> <Qd>, <Qm> |
| 6310 | Case 1: VMOV<c><q> <Dd>, <Dm> |
| 6311 | Case 8: VMOV.F32 <Sd>, <Sm> |
| 6312 | Case 15: VMOV <Sd>, <Se>, <Rn>, <Rm> */ |
| 6313 | |
| 6314 | inst.operands[i].reg = val; |
| 6315 | inst.operands[i].isreg = 1; |
| 6316 | inst.operands[i].isquad = (rtype == REG_TYPE_NQ); |
| 6317 | inst.operands[i].issingle = (rtype == REG_TYPE_VFS); |
| 6318 | inst.operands[i].isvec = 1; |
| 6319 | inst.operands[i].vectype = optype; |
| 6320 | inst.operands[i].present = 1; |
| 6321 | |
| 6322 | if (skip_past_comma (&ptr) == SUCCESS) |
| 6323 | { |
| 6324 | /* Case 15. */ |
| 6325 | i++; |
| 6326 | |
| 6327 | if ((val = arm_reg_parse (&ptr, REG_TYPE_RN)) == FAIL) |
| 6328 | goto wanted_arm; |
| 6329 | |
| 6330 | inst.operands[i].reg = val; |
| 6331 | inst.operands[i].isreg = 1; |
| 6332 | inst.operands[i++].present = 1; |
| 6333 | |
| 6334 | if (skip_past_comma (&ptr) == FAIL) |
| 6335 | goto wanted_comma; |
| 6336 | |
| 6337 | if ((val = arm_reg_parse (&ptr, REG_TYPE_RN)) == FAIL) |
| 6338 | goto wanted_arm; |
| 6339 | |
| 6340 | inst.operands[i].reg = val; |
| 6341 | inst.operands[i].isreg = 1; |
| 6342 | inst.operands[i].present = 1; |
| 6343 | } |
| 6344 | } |
| 6345 | else if (parse_qfloat_immediate (&ptr, &inst.operands[i].imm) == SUCCESS) |
| 6346 | /* Case 2: VMOV<c><q>.<dt> <Qd>, #<float-imm> |
| 6347 | Case 3: VMOV<c><q>.<dt> <Dd>, #<float-imm> |
| 6348 | Case 10: VMOV.F32 <Sd>, #<imm> |
| 6349 | Case 11: VMOV.F64 <Dd>, #<imm> */ |
| 6350 | inst.operands[i].immisfloat = 1; |
| 6351 | else if (parse_big_immediate (&ptr, i, NULL, /*allow_symbol_p=*/FALSE) |
| 6352 | == SUCCESS) |
| 6353 | /* Case 2: VMOV<c><q>.<dt> <Qd>, #<imm> |
| 6354 | Case 3: VMOV<c><q>.<dt> <Dd>, #<imm> */ |
| 6355 | ; |
| 6356 | else |
| 6357 | { |
| 6358 | first_error (_("expected <Rm> or <Dm> or <Qm> operand")); |
| 6359 | return FAIL; |
| 6360 | } |
| 6361 | } |
| 6362 | else if ((val = arm_reg_parse (&ptr, REG_TYPE_RN)) != FAIL) |
| 6363 | { |
| 6364 | /* Cases 6, 7. */ |
| 6365 | inst.operands[i].reg = val; |
| 6366 | inst.operands[i].isreg = 1; |
| 6367 | inst.operands[i++].present = 1; |
| 6368 | |
| 6369 | if (skip_past_comma (&ptr) == FAIL) |
| 6370 | goto wanted_comma; |
| 6371 | |
| 6372 | if ((val = parse_scalar (&ptr, 8, &optype)) != FAIL) |
| 6373 | { |
| 6374 | /* Case 6: VMOV<c><q>.<dt> <Rd>, <Dn[x]> */ |
| 6375 | inst.operands[i].reg = val; |
| 6376 | inst.operands[i].isscalar = 1; |
| 6377 | inst.operands[i].present = 1; |
| 6378 | inst.operands[i].vectype = optype; |
| 6379 | } |
| 6380 | else if ((val = arm_reg_parse (&ptr, REG_TYPE_RN)) != FAIL) |
| 6381 | { |
| 6382 | /* Case 7: VMOV<c><q> <Rd>, <Rn>, <Dm> */ |
| 6383 | inst.operands[i].reg = val; |
| 6384 | inst.operands[i].isreg = 1; |
| 6385 | inst.operands[i++].present = 1; |
| 6386 | |
| 6387 | if (skip_past_comma (&ptr) == FAIL) |
| 6388 | goto wanted_comma; |
| 6389 | |
| 6390 | if ((val = arm_typed_reg_parse (&ptr, REG_TYPE_VFSD, &rtype, &optype)) |
| 6391 | == FAIL) |
| 6392 | { |
| 6393 | first_error (_(reg_expected_msgs[REG_TYPE_VFSD])); |
| 6394 | return FAIL; |
| 6395 | } |
| 6396 | |
| 6397 | inst.operands[i].reg = val; |
| 6398 | inst.operands[i].isreg = 1; |
| 6399 | inst.operands[i].isvec = 1; |
| 6400 | inst.operands[i].issingle = (rtype == REG_TYPE_VFS); |
| 6401 | inst.operands[i].vectype = optype; |
| 6402 | inst.operands[i].present = 1; |
| 6403 | |
| 6404 | if (rtype == REG_TYPE_VFS) |
| 6405 | { |
| 6406 | /* Case 14. */ |
| 6407 | i++; |
| 6408 | if (skip_past_comma (&ptr) == FAIL) |
| 6409 | goto wanted_comma; |
| 6410 | if ((val = arm_typed_reg_parse (&ptr, REG_TYPE_VFS, NULL, |
| 6411 | &optype)) == FAIL) |
| 6412 | { |
| 6413 | first_error (_(reg_expected_msgs[REG_TYPE_VFS])); |
| 6414 | return FAIL; |
| 6415 | } |
| 6416 | inst.operands[i].reg = val; |
| 6417 | inst.operands[i].isreg = 1; |
| 6418 | inst.operands[i].isvec = 1; |
| 6419 | inst.operands[i].issingle = 1; |
| 6420 | inst.operands[i].vectype = optype; |
| 6421 | inst.operands[i].present = 1; |
| 6422 | } |
| 6423 | } |
| 6424 | else if ((val = arm_typed_reg_parse (&ptr, REG_TYPE_VFS, NULL, &optype)) |
| 6425 | != FAIL) |
| 6426 | { |
| 6427 | /* Case 13. */ |
| 6428 | inst.operands[i].reg = val; |
| 6429 | inst.operands[i].isreg = 1; |
| 6430 | inst.operands[i].isvec = 1; |
| 6431 | inst.operands[i].issingle = 1; |
| 6432 | inst.operands[i].vectype = optype; |
| 6433 | inst.operands[i].present = 1; |
| 6434 | } |
| 6435 | } |
| 6436 | else |
| 6437 | { |
| 6438 | first_error (_("parse error")); |
| 6439 | return FAIL; |
| 6440 | } |
| 6441 | |
| 6442 | /* Successfully parsed the operands. Update args. */ |
| 6443 | *which_operand = i; |
| 6444 | *str = ptr; |
| 6445 | return SUCCESS; |
| 6446 | |
| 6447 | wanted_comma: |
| 6448 | first_error (_("expected comma")); |
| 6449 | return FAIL; |
| 6450 | |
| 6451 | wanted_arm: |
| 6452 | first_error (_(reg_expected_msgs[REG_TYPE_RN])); |
| 6453 | return FAIL; |
| 6454 | } |
| 6455 | |
| 6456 | /* Use this macro when the operand constraints are different |
| 6457 | for ARM and THUMB (e.g. ldrd). */ |
| 6458 | #define MIX_ARM_THUMB_OPERANDS(arm_operand, thumb_operand) \ |
| 6459 | ((arm_operand) | ((thumb_operand) << 16)) |
| 6460 | |
| 6461 | /* Matcher codes for parse_operands. */ |
| 6462 | enum operand_parse_code |
| 6463 | { |
| 6464 | OP_stop, /* end of line */ |
| 6465 | |
| 6466 | OP_RR, /* ARM register */ |
| 6467 | OP_RRnpc, /* ARM register, not r15 */ |
| 6468 | OP_RRnpcsp, /* ARM register, neither r15 nor r13 (a.k.a. 'BadReg') */ |
| 6469 | OP_RRnpcb, /* ARM register, not r15, in square brackets */ |
| 6470 | OP_RRnpctw, /* ARM register, not r15 in Thumb-state or with writeback, |
| 6471 | optional trailing ! */ |
| 6472 | OP_RRw, /* ARM register, not r15, optional trailing ! */ |
| 6473 | OP_RCP, /* Coprocessor number */ |
| 6474 | OP_RCN, /* Coprocessor register */ |
| 6475 | OP_RF, /* FPA register */ |
| 6476 | OP_RVS, /* VFP single precision register */ |
| 6477 | OP_RVD, /* VFP double precision register (0..15) */ |
| 6478 | OP_RND, /* Neon double precision register (0..31) */ |
| 6479 | OP_RNQ, /* Neon quad precision register */ |
| 6480 | OP_RVSD, /* VFP single or double precision register */ |
| 6481 | OP_RNDQ, /* Neon double or quad precision register */ |
| 6482 | OP_RNSDQ, /* Neon single, double or quad precision register */ |
| 6483 | OP_RNSC, /* Neon scalar D[X] */ |
| 6484 | OP_RVC, /* VFP control register */ |
| 6485 | OP_RMF, /* Maverick F register */ |
| 6486 | OP_RMD, /* Maverick D register */ |
| 6487 | OP_RMFX, /* Maverick FX register */ |
| 6488 | OP_RMDX, /* Maverick DX register */ |
| 6489 | OP_RMAX, /* Maverick AX register */ |
| 6490 | OP_RMDS, /* Maverick DSPSC register */ |
| 6491 | OP_RIWR, /* iWMMXt wR register */ |
| 6492 | OP_RIWC, /* iWMMXt wC register */ |
| 6493 | OP_RIWG, /* iWMMXt wCG register */ |
| 6494 | OP_RXA, /* XScale accumulator register */ |
| 6495 | |
| 6496 | OP_REGLST, /* ARM register list */ |
| 6497 | OP_VRSLST, /* VFP single-precision register list */ |
| 6498 | OP_VRDLST, /* VFP double-precision register list */ |
| 6499 | OP_VRSDLST, /* VFP single or double-precision register list (& quad) */ |
| 6500 | OP_NRDLST, /* Neon double-precision register list (d0-d31, qN aliases) */ |
| 6501 | OP_NSTRLST, /* Neon element/structure list */ |
| 6502 | |
| 6503 | OP_RNDQ_I0, /* Neon D or Q reg, or immediate zero. */ |
| 6504 | OP_RVSD_I0, /* VFP S or D reg, or immediate zero. */ |
| 6505 | OP_RSVD_FI0, /* VFP S or D reg, or floating point immediate zero. */ |
| 6506 | OP_RR_RNSC, /* ARM reg or Neon scalar. */ |
| 6507 | OP_RNSDQ_RNSC, /* Vector S, D or Q reg, or Neon scalar. */ |
| 6508 | OP_RNDQ_RNSC, /* Neon D or Q reg, or Neon scalar. */ |
| 6509 | OP_RND_RNSC, /* Neon D reg, or Neon scalar. */ |
| 6510 | OP_VMOV, /* Neon VMOV operands. */ |
| 6511 | OP_RNDQ_Ibig, /* Neon D or Q reg, or big immediate for logic and VMVN. */ |
| 6512 | OP_RNDQ_I63b, /* Neon D or Q reg, or immediate for shift. */ |
| 6513 | OP_RIWR_I32z, /* iWMMXt wR register, or immediate 0 .. 32 for iWMMXt2. */ |
| 6514 | |
| 6515 | OP_I0, /* immediate zero */ |
| 6516 | OP_I7, /* immediate value 0 .. 7 */ |
| 6517 | OP_I15, /* 0 .. 15 */ |
| 6518 | OP_I16, /* 1 .. 16 */ |
| 6519 | OP_I16z, /* 0 .. 16 */ |
| 6520 | OP_I31, /* 0 .. 31 */ |
| 6521 | OP_I31w, /* 0 .. 31, optional trailing ! */ |
| 6522 | OP_I32, /* 1 .. 32 */ |
| 6523 | OP_I32z, /* 0 .. 32 */ |
| 6524 | OP_I63, /* 0 .. 63 */ |
| 6525 | OP_I63s, /* -64 .. 63 */ |
| 6526 | OP_I64, /* 1 .. 64 */ |
| 6527 | OP_I64z, /* 0 .. 64 */ |
| 6528 | OP_I255, /* 0 .. 255 */ |
| 6529 | |
| 6530 | OP_I4b, /* immediate, prefix optional, 1 .. 4 */ |
| 6531 | OP_I7b, /* 0 .. 7 */ |
| 6532 | OP_I15b, /* 0 .. 15 */ |
| 6533 | OP_I31b, /* 0 .. 31 */ |
| 6534 | |
| 6535 | OP_SH, /* shifter operand */ |
| 6536 | OP_SHG, /* shifter operand with possible group relocation */ |
| 6537 | OP_ADDR, /* Memory address expression (any mode) */ |
| 6538 | OP_ADDRGLDR, /* Mem addr expr (any mode) with possible LDR group reloc */ |
| 6539 | OP_ADDRGLDRS, /* Mem addr expr (any mode) with possible LDRS group reloc */ |
| 6540 | OP_ADDRGLDC, /* Mem addr expr (any mode) with possible LDC group reloc */ |
| 6541 | OP_EXP, /* arbitrary expression */ |
| 6542 | OP_EXPi, /* same, with optional immediate prefix */ |
| 6543 | OP_EXPr, /* same, with optional relocation suffix */ |
| 6544 | OP_HALF, /* 0 .. 65535 or low/high reloc. */ |
| 6545 | OP_IROT1, /* VCADD rotate immediate: 90, 270. */ |
| 6546 | OP_IROT2, /* VCMLA rotate immediate: 0, 90, 180, 270. */ |
| 6547 | |
| 6548 | OP_CPSF, /* CPS flags */ |
| 6549 | OP_ENDI, /* Endianness specifier */ |
| 6550 | OP_wPSR, /* CPSR/SPSR/APSR mask for msr (writing). */ |
| 6551 | OP_rPSR, /* CPSR/SPSR/APSR mask for msr (reading). */ |
| 6552 | OP_COND, /* conditional code */ |
| 6553 | OP_TB, /* Table branch. */ |
| 6554 | |
| 6555 | OP_APSR_RR, /* ARM register or "APSR_nzcv". */ |
| 6556 | |
| 6557 | OP_RRnpc_I0, /* ARM register or literal 0 */ |
| 6558 | OP_RR_EXr, /* ARM register or expression with opt. reloc stuff. */ |
| 6559 | OP_RR_EXi, /* ARM register or expression with imm prefix */ |
| 6560 | OP_RF_IF, /* FPA register or immediate */ |
| 6561 | OP_RIWR_RIWC, /* iWMMXt R or C reg */ |
| 6562 | OP_RIWC_RIWG, /* iWMMXt wC or wCG reg */ |
| 6563 | |
| 6564 | /* Optional operands. */ |
| 6565 | OP_oI7b, /* immediate, prefix optional, 0 .. 7 */ |
| 6566 | OP_oI31b, /* 0 .. 31 */ |
| 6567 | OP_oI32b, /* 1 .. 32 */ |
| 6568 | OP_oI32z, /* 0 .. 32 */ |
| 6569 | OP_oIffffb, /* 0 .. 65535 */ |
| 6570 | OP_oI255c, /* curly-brace enclosed, 0 .. 255 */ |
| 6571 | |
| 6572 | OP_oRR, /* ARM register */ |
| 6573 | OP_oRRnpc, /* ARM register, not the PC */ |
| 6574 | OP_oRRnpcsp, /* ARM register, neither the PC nor the SP (a.k.a. BadReg) */ |
| 6575 | OP_oRRw, /* ARM register, not r15, optional trailing ! */ |
| 6576 | OP_oRND, /* Optional Neon double precision register */ |
| 6577 | OP_oRNQ, /* Optional Neon quad precision register */ |
| 6578 | OP_oRNDQ, /* Optional Neon double or quad precision register */ |
| 6579 | OP_oRNSDQ, /* Optional single, double or quad precision vector register */ |
| 6580 | OP_oSHll, /* LSL immediate */ |
| 6581 | OP_oSHar, /* ASR immediate */ |
| 6582 | OP_oSHllar, /* LSL or ASR immediate */ |
| 6583 | OP_oROR, /* ROR 0/8/16/24 */ |
| 6584 | OP_oBARRIER_I15, /* Option argument for a barrier instruction. */ |
| 6585 | |
| 6586 | /* Some pre-defined mixed (ARM/THUMB) operands. */ |
| 6587 | OP_RR_npcsp = MIX_ARM_THUMB_OPERANDS (OP_RR, OP_RRnpcsp), |
| 6588 | OP_RRnpc_npcsp = MIX_ARM_THUMB_OPERANDS (OP_RRnpc, OP_RRnpcsp), |
| 6589 | OP_oRRnpc_npcsp = MIX_ARM_THUMB_OPERANDS (OP_oRRnpc, OP_oRRnpcsp), |
| 6590 | |
| 6591 | OP_FIRST_OPTIONAL = OP_oI7b |
| 6592 | }; |
| 6593 | |
| 6594 | /* Generic instruction operand parser. This does no encoding and no |
| 6595 | semantic validation; it merely squirrels values away in the inst |
| 6596 | structure. Returns SUCCESS or FAIL depending on whether the |
| 6597 | specified grammar matched. */ |
| 6598 | static int |
| 6599 | parse_operands (char *str, const unsigned int *pattern, bfd_boolean thumb) |
| 6600 | { |
| 6601 | unsigned const int *upat = pattern; |
| 6602 | char *backtrack_pos = 0; |
| 6603 | const char *backtrack_error = 0; |
| 6604 | int i, val = 0, backtrack_index = 0; |
| 6605 | enum arm_reg_type rtype; |
| 6606 | parse_operand_result result; |
| 6607 | unsigned int op_parse_code; |
| 6608 | |
| 6609 | #define po_char_or_fail(chr) \ |
| 6610 | do \ |
| 6611 | { \ |
| 6612 | if (skip_past_char (&str, chr) == FAIL) \ |
| 6613 | goto bad_args; \ |
| 6614 | } \ |
| 6615 | while (0) |
| 6616 | |
| 6617 | #define po_reg_or_fail(regtype) \ |
| 6618 | do \ |
| 6619 | { \ |
| 6620 | val = arm_typed_reg_parse (& str, regtype, & rtype, \ |
| 6621 | & inst.operands[i].vectype); \ |
| 6622 | if (val == FAIL) \ |
| 6623 | { \ |
| 6624 | first_error (_(reg_expected_msgs[regtype])); \ |
| 6625 | goto failure; \ |
| 6626 | } \ |
| 6627 | inst.operands[i].reg = val; \ |
| 6628 | inst.operands[i].isreg = 1; \ |
| 6629 | inst.operands[i].isquad = (rtype == REG_TYPE_NQ); \ |
| 6630 | inst.operands[i].issingle = (rtype == REG_TYPE_VFS); \ |
| 6631 | inst.operands[i].isvec = (rtype == REG_TYPE_VFS \ |
| 6632 | || rtype == REG_TYPE_VFD \ |
| 6633 | || rtype == REG_TYPE_NQ); \ |
| 6634 | } \ |
| 6635 | while (0) |
| 6636 | |
| 6637 | #define po_reg_or_goto(regtype, label) \ |
| 6638 | do \ |
| 6639 | { \ |
| 6640 | val = arm_typed_reg_parse (& str, regtype, & rtype, \ |
| 6641 | & inst.operands[i].vectype); \ |
| 6642 | if (val == FAIL) \ |
| 6643 | goto label; \ |
| 6644 | \ |
| 6645 | inst.operands[i].reg = val; \ |
| 6646 | inst.operands[i].isreg = 1; \ |
| 6647 | inst.operands[i].isquad = (rtype == REG_TYPE_NQ); \ |
| 6648 | inst.operands[i].issingle = (rtype == REG_TYPE_VFS); \ |
| 6649 | inst.operands[i].isvec = (rtype == REG_TYPE_VFS \ |
| 6650 | || rtype == REG_TYPE_VFD \ |
| 6651 | || rtype == REG_TYPE_NQ); \ |
| 6652 | } \ |
| 6653 | while (0) |
| 6654 | |
| 6655 | #define po_imm_or_fail(min, max, popt) \ |
| 6656 | do \ |
| 6657 | { \ |
| 6658 | if (parse_immediate (&str, &val, min, max, popt) == FAIL) \ |
| 6659 | goto failure; \ |
| 6660 | inst.operands[i].imm = val; \ |
| 6661 | } \ |
| 6662 | while (0) |
| 6663 | |
| 6664 | #define po_scalar_or_goto(elsz, label) \ |
| 6665 | do \ |
| 6666 | { \ |
| 6667 | val = parse_scalar (& str, elsz, & inst.operands[i].vectype); \ |
| 6668 | if (val == FAIL) \ |
| 6669 | goto label; \ |
| 6670 | inst.operands[i].reg = val; \ |
| 6671 | inst.operands[i].isscalar = 1; \ |
| 6672 | } \ |
| 6673 | while (0) |
| 6674 | |
| 6675 | #define po_misc_or_fail(expr) \ |
| 6676 | do \ |
| 6677 | { \ |
| 6678 | if (expr) \ |
| 6679 | goto failure; \ |
| 6680 | } \ |
| 6681 | while (0) |
| 6682 | |
| 6683 | #define po_misc_or_fail_no_backtrack(expr) \ |
| 6684 | do \ |
| 6685 | { \ |
| 6686 | result = expr; \ |
| 6687 | if (result == PARSE_OPERAND_FAIL_NO_BACKTRACK) \ |
| 6688 | backtrack_pos = 0; \ |
| 6689 | if (result != PARSE_OPERAND_SUCCESS) \ |
| 6690 | goto failure; \ |
| 6691 | } \ |
| 6692 | while (0) |
| 6693 | |
| 6694 | #define po_barrier_or_imm(str) \ |
| 6695 | do \ |
| 6696 | { \ |
| 6697 | val = parse_barrier (&str); \ |
| 6698 | if (val == FAIL && ! ISALPHA (*str)) \ |
| 6699 | goto immediate; \ |
| 6700 | if (val == FAIL \ |
| 6701 | /* ISB can only take SY as an option. */ \ |
| 6702 | || ((inst.instruction & 0xf0) == 0x60 \ |
| 6703 | && val != 0xf)) \ |
| 6704 | { \ |
| 6705 | inst.error = _("invalid barrier type"); \ |
| 6706 | backtrack_pos = 0; \ |
| 6707 | goto failure; \ |
| 6708 | } \ |
| 6709 | } \ |
| 6710 | while (0) |
| 6711 | |
| 6712 | skip_whitespace (str); |
| 6713 | |
| 6714 | for (i = 0; upat[i] != OP_stop; i++) |
| 6715 | { |
| 6716 | op_parse_code = upat[i]; |
| 6717 | if (op_parse_code >= 1<<16) |
| 6718 | op_parse_code = thumb ? (op_parse_code >> 16) |
| 6719 | : (op_parse_code & ((1<<16)-1)); |
| 6720 | |
| 6721 | if (op_parse_code >= OP_FIRST_OPTIONAL) |
| 6722 | { |
| 6723 | /* Remember where we are in case we need to backtrack. */ |
| 6724 | gas_assert (!backtrack_pos); |
| 6725 | backtrack_pos = str; |
| 6726 | backtrack_error = inst.error; |
| 6727 | backtrack_index = i; |
| 6728 | } |
| 6729 | |
| 6730 | if (i > 0 && (i > 1 || inst.operands[0].present)) |
| 6731 | po_char_or_fail (','); |
| 6732 | |
| 6733 | switch (op_parse_code) |
| 6734 | { |
| 6735 | /* Registers */ |
| 6736 | case OP_oRRnpc: |
| 6737 | case OP_oRRnpcsp: |
| 6738 | case OP_RRnpc: |
| 6739 | case OP_RRnpcsp: |
| 6740 | case OP_oRR: |
| 6741 | case OP_RR: po_reg_or_fail (REG_TYPE_RN); break; |
| 6742 | case OP_RCP: po_reg_or_fail (REG_TYPE_CP); break; |
| 6743 | case OP_RCN: po_reg_or_fail (REG_TYPE_CN); break; |
| 6744 | case OP_RF: po_reg_or_fail (REG_TYPE_FN); break; |
| 6745 | case OP_RVS: po_reg_or_fail (REG_TYPE_VFS); break; |
| 6746 | case OP_RVD: po_reg_or_fail (REG_TYPE_VFD); break; |
| 6747 | case OP_oRND: |
| 6748 | case OP_RND: po_reg_or_fail (REG_TYPE_VFD); break; |
| 6749 | case OP_RVC: |
| 6750 | po_reg_or_goto (REG_TYPE_VFC, coproc_reg); |
| 6751 | break; |
| 6752 | /* Also accept generic coprocessor regs for unknown registers. */ |
| 6753 | coproc_reg: |
| 6754 | po_reg_or_fail (REG_TYPE_CN); |
| 6755 | break; |
| 6756 | case OP_RMF: po_reg_or_fail (REG_TYPE_MVF); break; |
| 6757 | case OP_RMD: po_reg_or_fail (REG_TYPE_MVD); break; |
| 6758 | case OP_RMFX: po_reg_or_fail (REG_TYPE_MVFX); break; |
| 6759 | case OP_RMDX: po_reg_or_fail (REG_TYPE_MVDX); break; |
| 6760 | case OP_RMAX: po_reg_or_fail (REG_TYPE_MVAX); break; |
| 6761 | case OP_RMDS: po_reg_or_fail (REG_TYPE_DSPSC); break; |
| 6762 | case OP_RIWR: po_reg_or_fail (REG_TYPE_MMXWR); break; |
| 6763 | case OP_RIWC: po_reg_or_fail (REG_TYPE_MMXWC); break; |
| 6764 | case OP_RIWG: po_reg_or_fail (REG_TYPE_MMXWCG); break; |
| 6765 | case OP_RXA: po_reg_or_fail (REG_TYPE_XSCALE); break; |
| 6766 | case OP_oRNQ: |
| 6767 | case OP_RNQ: po_reg_or_fail (REG_TYPE_NQ); break; |
| 6768 | case OP_oRNDQ: |
| 6769 | case OP_RNDQ: po_reg_or_fail (REG_TYPE_NDQ); break; |
| 6770 | case OP_RVSD: po_reg_or_fail (REG_TYPE_VFSD); break; |
| 6771 | case OP_oRNSDQ: |
| 6772 | case OP_RNSDQ: po_reg_or_fail (REG_TYPE_NSDQ); break; |
| 6773 | |
| 6774 | /* Neon scalar. Using an element size of 8 means that some invalid |
| 6775 | scalars are accepted here, so deal with those in later code. */ |
| 6776 | case OP_RNSC: po_scalar_or_goto (8, failure); break; |
| 6777 | |
| 6778 | case OP_RNDQ_I0: |
| 6779 | { |
| 6780 | po_reg_or_goto (REG_TYPE_NDQ, try_imm0); |
| 6781 | break; |
| 6782 | try_imm0: |
| 6783 | po_imm_or_fail (0, 0, TRUE); |
| 6784 | } |
| 6785 | break; |
| 6786 | |
| 6787 | case OP_RVSD_I0: |
| 6788 | po_reg_or_goto (REG_TYPE_VFSD, try_imm0); |
| 6789 | break; |
| 6790 | |
| 6791 | case OP_RSVD_FI0: |
| 6792 | { |
| 6793 | po_reg_or_goto (REG_TYPE_VFSD, try_ifimm0); |
| 6794 | break; |
| 6795 | try_ifimm0: |
| 6796 | if (parse_ifimm_zero (&str)) |
| 6797 | inst.operands[i].imm = 0; |
| 6798 | else |
| 6799 | { |
| 6800 | inst.error |
| 6801 | = _("only floating point zero is allowed as immediate value"); |
| 6802 | goto failure; |
| 6803 | } |
| 6804 | } |
| 6805 | break; |
| 6806 | |
| 6807 | case OP_RR_RNSC: |
| 6808 | { |
| 6809 | po_scalar_or_goto (8, try_rr); |
| 6810 | break; |
| 6811 | try_rr: |
| 6812 | po_reg_or_fail (REG_TYPE_RN); |
| 6813 | } |
| 6814 | break; |
| 6815 | |
| 6816 | case OP_RNSDQ_RNSC: |
| 6817 | { |
| 6818 | po_scalar_or_goto (8, try_nsdq); |
| 6819 | break; |
| 6820 | try_nsdq: |
| 6821 | po_reg_or_fail (REG_TYPE_NSDQ); |
| 6822 | } |
| 6823 | break; |
| 6824 | |
| 6825 | case OP_RNDQ_RNSC: |
| 6826 | { |
| 6827 | po_scalar_or_goto (8, try_ndq); |
| 6828 | break; |
| 6829 | try_ndq: |
| 6830 | po_reg_or_fail (REG_TYPE_NDQ); |
| 6831 | } |
| 6832 | break; |
| 6833 | |
| 6834 | case OP_RND_RNSC: |
| 6835 | { |
| 6836 | po_scalar_or_goto (8, try_vfd); |
| 6837 | break; |
| 6838 | try_vfd: |
| 6839 | po_reg_or_fail (REG_TYPE_VFD); |
| 6840 | } |
| 6841 | break; |
| 6842 | |
| 6843 | case OP_VMOV: |
| 6844 | /* WARNING: parse_neon_mov can move the operand counter, i. If we're |
| 6845 | not careful then bad things might happen. */ |
| 6846 | po_misc_or_fail (parse_neon_mov (&str, &i) == FAIL); |
| 6847 | break; |
| 6848 | |
| 6849 | case OP_RNDQ_Ibig: |
| 6850 | { |
| 6851 | po_reg_or_goto (REG_TYPE_NDQ, try_immbig); |
| 6852 | break; |
| 6853 | try_immbig: |
| 6854 | /* There's a possibility of getting a 64-bit immediate here, so |
| 6855 | we need special handling. */ |
| 6856 | if (parse_big_immediate (&str, i, NULL, /*allow_symbol_p=*/FALSE) |
| 6857 | == FAIL) |
| 6858 | { |
| 6859 | inst.error = _("immediate value is out of range"); |
| 6860 | goto failure; |
| 6861 | } |
| 6862 | } |
| 6863 | break; |
| 6864 | |
| 6865 | case OP_RNDQ_I63b: |
| 6866 | { |
| 6867 | po_reg_or_goto (REG_TYPE_NDQ, try_shimm); |
| 6868 | break; |
| 6869 | try_shimm: |
| 6870 | po_imm_or_fail (0, 63, TRUE); |
| 6871 | } |
| 6872 | break; |
| 6873 | |
| 6874 | case OP_RRnpcb: |
| 6875 | po_char_or_fail ('['); |
| 6876 | po_reg_or_fail (REG_TYPE_RN); |
| 6877 | po_char_or_fail (']'); |
| 6878 | break; |
| 6879 | |
| 6880 | case OP_RRnpctw: |
| 6881 | case OP_RRw: |
| 6882 | case OP_oRRw: |
| 6883 | po_reg_or_fail (REG_TYPE_RN); |
| 6884 | if (skip_past_char (&str, '!') == SUCCESS) |
| 6885 | inst.operands[i].writeback = 1; |
| 6886 | break; |
| 6887 | |
| 6888 | /* Immediates */ |
| 6889 | case OP_I7: po_imm_or_fail ( 0, 7, FALSE); break; |
| 6890 | case OP_I15: po_imm_or_fail ( 0, 15, FALSE); break; |
| 6891 | case OP_I16: po_imm_or_fail ( 1, 16, FALSE); break; |
| 6892 | case OP_I16z: po_imm_or_fail ( 0, 16, FALSE); break; |
| 6893 | case OP_I31: po_imm_or_fail ( 0, 31, FALSE); break; |
| 6894 | case OP_I32: po_imm_or_fail ( 1, 32, FALSE); break; |
| 6895 | case OP_I32z: po_imm_or_fail ( 0, 32, FALSE); break; |
| 6896 | case OP_I63s: po_imm_or_fail (-64, 63, FALSE); break; |
| 6897 | case OP_I63: po_imm_or_fail ( 0, 63, FALSE); break; |
| 6898 | case OP_I64: po_imm_or_fail ( 1, 64, FALSE); break; |
| 6899 | case OP_I64z: po_imm_or_fail ( 0, 64, FALSE); break; |
| 6900 | case OP_I255: po_imm_or_fail ( 0, 255, FALSE); break; |
| 6901 | |
| 6902 | case OP_I4b: po_imm_or_fail ( 1, 4, TRUE); break; |
| 6903 | case OP_oI7b: |
| 6904 | case OP_I7b: po_imm_or_fail ( 0, 7, TRUE); break; |
| 6905 | case OP_I15b: po_imm_or_fail ( 0, 15, TRUE); break; |
| 6906 | case OP_oI31b: |
| 6907 | case OP_I31b: po_imm_or_fail ( 0, 31, TRUE); break; |
| 6908 | case OP_oI32b: po_imm_or_fail ( 1, 32, TRUE); break; |
| 6909 | case OP_oI32z: po_imm_or_fail ( 0, 32, TRUE); break; |
| 6910 | case OP_oIffffb: po_imm_or_fail ( 0, 0xffff, TRUE); break; |
| 6911 | |
| 6912 | /* Immediate variants */ |
| 6913 | case OP_oI255c: |
| 6914 | po_char_or_fail ('{'); |
| 6915 | po_imm_or_fail (0, 255, TRUE); |
| 6916 | po_char_or_fail ('}'); |
| 6917 | break; |
| 6918 | |
| 6919 | case OP_I31w: |
| 6920 | /* The expression parser chokes on a trailing !, so we have |
| 6921 | to find it first and zap it. */ |
| 6922 | { |
| 6923 | char *s = str; |
| 6924 | while (*s && *s != ',') |
| 6925 | s++; |
| 6926 | if (s[-1] == '!') |
| 6927 | { |
| 6928 | s[-1] = '\0'; |
| 6929 | inst.operands[i].writeback = 1; |
| 6930 | } |
| 6931 | po_imm_or_fail (0, 31, TRUE); |
| 6932 | if (str == s - 1) |
| 6933 | str = s; |
| 6934 | } |
| 6935 | break; |
| 6936 | |
| 6937 | /* Expressions */ |
| 6938 | case OP_EXPi: EXPi: |
| 6939 | po_misc_or_fail (my_get_expression (&inst.reloc.exp, &str, |
| 6940 | GE_OPT_PREFIX)); |
| 6941 | break; |
| 6942 | |
| 6943 | case OP_EXP: |
| 6944 | po_misc_or_fail (my_get_expression (&inst.reloc.exp, &str, |
| 6945 | GE_NO_PREFIX)); |
| 6946 | break; |
| 6947 | |
| 6948 | case OP_EXPr: EXPr: |
| 6949 | po_misc_or_fail (my_get_expression (&inst.reloc.exp, &str, |
| 6950 | GE_NO_PREFIX)); |
| 6951 | if (inst.reloc.exp.X_op == O_symbol) |
| 6952 | { |
| 6953 | val = parse_reloc (&str); |
| 6954 | if (val == -1) |
| 6955 | { |
| 6956 | inst.error = _("unrecognized relocation suffix"); |
| 6957 | goto failure; |
| 6958 | } |
| 6959 | else if (val != BFD_RELOC_UNUSED) |
| 6960 | { |
| 6961 | inst.operands[i].imm = val; |
| 6962 | inst.operands[i].hasreloc = 1; |
| 6963 | } |
| 6964 | } |
| 6965 | break; |
| 6966 | |
| 6967 | /* Operand for MOVW or MOVT. */ |
| 6968 | case OP_HALF: |
| 6969 | po_misc_or_fail (parse_half (&str)); |
| 6970 | break; |
| 6971 | |
| 6972 | /* Register or expression. */ |
| 6973 | case OP_RR_EXr: po_reg_or_goto (REG_TYPE_RN, EXPr); break; |
| 6974 | case OP_RR_EXi: po_reg_or_goto (REG_TYPE_RN, EXPi); break; |
| 6975 | |
| 6976 | /* Register or immediate. */ |
| 6977 | case OP_RRnpc_I0: po_reg_or_goto (REG_TYPE_RN, I0); break; |
| 6978 | I0: po_imm_or_fail (0, 0, FALSE); break; |
| 6979 | |
| 6980 | case OP_RF_IF: po_reg_or_goto (REG_TYPE_FN, IF); break; |
| 6981 | IF: |
| 6982 | if (!is_immediate_prefix (*str)) |
| 6983 | goto bad_args; |
| 6984 | str++; |
| 6985 | val = parse_fpa_immediate (&str); |
| 6986 | if (val == FAIL) |
| 6987 | goto failure; |
| 6988 | /* FPA immediates are encoded as registers 8-15. |
| 6989 | parse_fpa_immediate has already applied the offset. */ |
| 6990 | inst.operands[i].reg = val; |
| 6991 | inst.operands[i].isreg = 1; |
| 6992 | break; |
| 6993 | |
| 6994 | case OP_RIWR_I32z: po_reg_or_goto (REG_TYPE_MMXWR, I32z); break; |
| 6995 | I32z: po_imm_or_fail (0, 32, FALSE); break; |
| 6996 | |
| 6997 | /* Two kinds of register. */ |
| 6998 | case OP_RIWR_RIWC: |
| 6999 | { |
| 7000 | struct reg_entry *rege = arm_reg_parse_multi (&str); |
| 7001 | if (!rege |
| 7002 | || (rege->type != REG_TYPE_MMXWR |
| 7003 | && rege->type != REG_TYPE_MMXWC |
| 7004 | && rege->type != REG_TYPE_MMXWCG)) |
| 7005 | { |
| 7006 | inst.error = _("iWMMXt data or control register expected"); |
| 7007 | goto failure; |
| 7008 | } |
| 7009 | inst.operands[i].reg = rege->number; |
| 7010 | inst.operands[i].isreg = (rege->type == REG_TYPE_MMXWR); |
| 7011 | } |
| 7012 | break; |
| 7013 | |
| 7014 | case OP_RIWC_RIWG: |
| 7015 | { |
| 7016 | struct reg_entry *rege = arm_reg_parse_multi (&str); |
| 7017 | if (!rege |
| 7018 | || (rege->type != REG_TYPE_MMXWC |
| 7019 | && rege->type != REG_TYPE_MMXWCG)) |
| 7020 | { |
| 7021 | inst.error = _("iWMMXt control register expected"); |
| 7022 | goto failure; |
| 7023 | } |
| 7024 | inst.operands[i].reg = rege->number; |
| 7025 | inst.operands[i].isreg = 1; |
| 7026 | } |
| 7027 | break; |
| 7028 | |
| 7029 | /* Misc */ |
| 7030 | case OP_CPSF: val = parse_cps_flags (&str); break; |
| 7031 | case OP_ENDI: val = parse_endian_specifier (&str); break; |
| 7032 | case OP_oROR: val = parse_ror (&str); break; |
| 7033 | case OP_COND: val = parse_cond (&str); break; |
| 7034 | case OP_oBARRIER_I15: |
| 7035 | po_barrier_or_imm (str); break; |
| 7036 | immediate: |
| 7037 | if (parse_immediate (&str, &val, 0, 15, TRUE) == FAIL) |
| 7038 | goto failure; |
| 7039 | break; |
| 7040 | |
| 7041 | case OP_wPSR: |
| 7042 | case OP_rPSR: |
| 7043 | po_reg_or_goto (REG_TYPE_RNB, try_psr); |
| 7044 | if (!ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_virt)) |
| 7045 | { |
| 7046 | inst.error = _("Banked registers are not available with this " |
| 7047 | "architecture."); |
| 7048 | goto failure; |
| 7049 | } |
| 7050 | break; |
| 7051 | try_psr: |
| 7052 | val = parse_psr (&str, op_parse_code == OP_wPSR); |
| 7053 | break; |
| 7054 | |
| 7055 | case OP_APSR_RR: |
| 7056 | po_reg_or_goto (REG_TYPE_RN, try_apsr); |
| 7057 | break; |
| 7058 | try_apsr: |
| 7059 | /* Parse "APSR_nvzc" operand (for FMSTAT-equivalent MRS |
| 7060 | instruction). */ |
| 7061 | if (strncasecmp (str, "APSR_", 5) == 0) |
| 7062 | { |
| 7063 | unsigned found = 0; |
| 7064 | str += 5; |
| 7065 | while (found < 15) |
| 7066 | switch (*str++) |
| 7067 | { |
| 7068 | case 'c': found = (found & 1) ? 16 : found | 1; break; |
| 7069 | case 'n': found = (found & 2) ? 16 : found | 2; break; |
| 7070 | case 'z': found = (found & 4) ? 16 : found | 4; break; |
| 7071 | case 'v': found = (found & 8) ? 16 : found | 8; break; |
| 7072 | default: found = 16; |
| 7073 | } |
| 7074 | if (found != 15) |
| 7075 | goto failure; |
| 7076 | inst.operands[i].isvec = 1; |
| 7077 | /* APSR_nzcv is encoded in instructions as if it were the REG_PC. */ |
| 7078 | inst.operands[i].reg = REG_PC; |
| 7079 | } |
| 7080 | else |
| 7081 | goto failure; |
| 7082 | break; |
| 7083 | |
| 7084 | case OP_TB: |
| 7085 | po_misc_or_fail (parse_tb (&str)); |
| 7086 | break; |
| 7087 | |
| 7088 | /* Register lists. */ |
| 7089 | case OP_REGLST: |
| 7090 | val = parse_reg_list (&str); |
| 7091 | if (*str == '^') |
| 7092 | { |
| 7093 | inst.operands[i].writeback = 1; |
| 7094 | str++; |
| 7095 | } |
| 7096 | break; |
| 7097 | |
| 7098 | case OP_VRSLST: |
| 7099 | val = parse_vfp_reg_list (&str, &inst.operands[i].reg, REGLIST_VFP_S); |
| 7100 | break; |
| 7101 | |
| 7102 | case OP_VRDLST: |
| 7103 | val = parse_vfp_reg_list (&str, &inst.operands[i].reg, REGLIST_VFP_D); |
| 7104 | break; |
| 7105 | |
| 7106 | case OP_VRSDLST: |
| 7107 | /* Allow Q registers too. */ |
| 7108 | val = parse_vfp_reg_list (&str, &inst.operands[i].reg, |
| 7109 | REGLIST_NEON_D); |
| 7110 | if (val == FAIL) |
| 7111 | { |
| 7112 | inst.error = NULL; |
| 7113 | val = parse_vfp_reg_list (&str, &inst.operands[i].reg, |
| 7114 | REGLIST_VFP_S); |
| 7115 | inst.operands[i].issingle = 1; |
| 7116 | } |
| 7117 | break; |
| 7118 | |
| 7119 | case OP_NRDLST: |
| 7120 | val = parse_vfp_reg_list (&str, &inst.operands[i].reg, |
| 7121 | REGLIST_NEON_D); |
| 7122 | break; |
| 7123 | |
| 7124 | case OP_NSTRLST: |
| 7125 | val = parse_neon_el_struct_list (&str, &inst.operands[i].reg, |
| 7126 | &inst.operands[i].vectype); |
| 7127 | break; |
| 7128 | |
| 7129 | /* Addressing modes */ |
| 7130 | case OP_ADDR: |
| 7131 | po_misc_or_fail (parse_address (&str, i)); |
| 7132 | break; |
| 7133 | |
| 7134 | case OP_ADDRGLDR: |
| 7135 | po_misc_or_fail_no_backtrack ( |
| 7136 | parse_address_group_reloc (&str, i, GROUP_LDR)); |
| 7137 | break; |
| 7138 | |
| 7139 | case OP_ADDRGLDRS: |
| 7140 | po_misc_or_fail_no_backtrack ( |
| 7141 | parse_address_group_reloc (&str, i, GROUP_LDRS)); |
| 7142 | break; |
| 7143 | |
| 7144 | case OP_ADDRGLDC: |
| 7145 | po_misc_or_fail_no_backtrack ( |
| 7146 | parse_address_group_reloc (&str, i, GROUP_LDC)); |
| 7147 | break; |
| 7148 | |
| 7149 | case OP_SH: |
| 7150 | po_misc_or_fail (parse_shifter_operand (&str, i)); |
| 7151 | break; |
| 7152 | |
| 7153 | case OP_SHG: |
| 7154 | po_misc_or_fail_no_backtrack ( |
| 7155 | parse_shifter_operand_group_reloc (&str, i)); |
| 7156 | break; |
| 7157 | |
| 7158 | case OP_oSHll: |
| 7159 | po_misc_or_fail (parse_shift (&str, i, SHIFT_LSL_IMMEDIATE)); |
| 7160 | break; |
| 7161 | |
| 7162 | case OP_oSHar: |
| 7163 | po_misc_or_fail (parse_shift (&str, i, SHIFT_ASR_IMMEDIATE)); |
| 7164 | break; |
| 7165 | |
| 7166 | case OP_oSHllar: |
| 7167 | po_misc_or_fail (parse_shift (&str, i, SHIFT_LSL_OR_ASR_IMMEDIATE)); |
| 7168 | break; |
| 7169 | |
| 7170 | default: |
| 7171 | as_fatal (_("unhandled operand code %d"), op_parse_code); |
| 7172 | } |
| 7173 | |
| 7174 | /* Various value-based sanity checks and shared operations. We |
| 7175 | do not signal immediate failures for the register constraints; |
| 7176 | this allows a syntax error to take precedence. */ |
| 7177 | switch (op_parse_code) |
| 7178 | { |
| 7179 | case OP_oRRnpc: |
| 7180 | case OP_RRnpc: |
| 7181 | case OP_RRnpcb: |
| 7182 | case OP_RRw: |
| 7183 | case OP_oRRw: |
| 7184 | case OP_RRnpc_I0: |
| 7185 | if (inst.operands[i].isreg && inst.operands[i].reg == REG_PC) |
| 7186 | inst.error = BAD_PC; |
| 7187 | break; |
| 7188 | |
| 7189 | case OP_oRRnpcsp: |
| 7190 | case OP_RRnpcsp: |
| 7191 | if (inst.operands[i].isreg) |
| 7192 | { |
| 7193 | if (inst.operands[i].reg == REG_PC) |
| 7194 | inst.error = BAD_PC; |
| 7195 | else if (inst.operands[i].reg == REG_SP |
| 7196 | /* The restriction on Rd/Rt/Rt2 on Thumb mode has been |
| 7197 | relaxed since ARMv8-A. */ |
| 7198 | && !ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v8)) |
| 7199 | { |
| 7200 | gas_assert (thumb); |
| 7201 | inst.error = BAD_SP; |
| 7202 | } |
| 7203 | } |
| 7204 | break; |
| 7205 | |
| 7206 | case OP_RRnpctw: |
| 7207 | if (inst.operands[i].isreg |
| 7208 | && inst.operands[i].reg == REG_PC |
| 7209 | && (inst.operands[i].writeback || thumb)) |
| 7210 | inst.error = BAD_PC; |
| 7211 | break; |
| 7212 | |
| 7213 | case OP_CPSF: |
| 7214 | case OP_ENDI: |
| 7215 | case OP_oROR: |
| 7216 | case OP_wPSR: |
| 7217 | case OP_rPSR: |
| 7218 | case OP_COND: |
| 7219 | case OP_oBARRIER_I15: |
| 7220 | case OP_REGLST: |
| 7221 | case OP_VRSLST: |
| 7222 | case OP_VRDLST: |
| 7223 | case OP_VRSDLST: |
| 7224 | case OP_NRDLST: |
| 7225 | case OP_NSTRLST: |
| 7226 | if (val == FAIL) |
| 7227 | goto failure; |
| 7228 | inst.operands[i].imm = val; |
| 7229 | break; |
| 7230 | |
| 7231 | default: |
| 7232 | break; |
| 7233 | } |
| 7234 | |
| 7235 | /* If we get here, this operand was successfully parsed. */ |
| 7236 | inst.operands[i].present = 1; |
| 7237 | continue; |
| 7238 | |
| 7239 | bad_args: |
| 7240 | inst.error = BAD_ARGS; |
| 7241 | |
| 7242 | failure: |
| 7243 | if (!backtrack_pos) |
| 7244 | { |
| 7245 | /* The parse routine should already have set inst.error, but set a |
| 7246 | default here just in case. */ |
| 7247 | if (!inst.error) |
| 7248 | inst.error = _("syntax error"); |
| 7249 | return FAIL; |
| 7250 | } |
| 7251 | |
| 7252 | /* Do not backtrack over a trailing optional argument that |
| 7253 | absorbed some text. We will only fail again, with the |
| 7254 | 'garbage following instruction' error message, which is |
| 7255 | probably less helpful than the current one. */ |
| 7256 | if (backtrack_index == i && backtrack_pos != str |
| 7257 | && upat[i+1] == OP_stop) |
| 7258 | { |
| 7259 | if (!inst.error) |
| 7260 | inst.error = _("syntax error"); |
| 7261 | return FAIL; |
| 7262 | } |
| 7263 | |
| 7264 | /* Try again, skipping the optional argument at backtrack_pos. */ |
| 7265 | str = backtrack_pos; |
| 7266 | inst.error = backtrack_error; |
| 7267 | inst.operands[backtrack_index].present = 0; |
| 7268 | i = backtrack_index; |
| 7269 | backtrack_pos = 0; |
| 7270 | } |
| 7271 | |
| 7272 | /* Check that we have parsed all the arguments. */ |
| 7273 | if (*str != '\0' && !inst.error) |
| 7274 | inst.error = _("garbage following instruction"); |
| 7275 | |
| 7276 | return inst.error ? FAIL : SUCCESS; |
| 7277 | } |
| 7278 | |
| 7279 | #undef po_char_or_fail |
| 7280 | #undef po_reg_or_fail |
| 7281 | #undef po_reg_or_goto |
| 7282 | #undef po_imm_or_fail |
| 7283 | #undef po_scalar_or_fail |
| 7284 | #undef po_barrier_or_imm |
| 7285 | |
| 7286 | /* Shorthand macro for instruction encoding functions issuing errors. */ |
| 7287 | #define constraint(expr, err) \ |
| 7288 | do \ |
| 7289 | { \ |
| 7290 | if (expr) \ |
| 7291 | { \ |
| 7292 | inst.error = err; \ |
| 7293 | return; \ |
| 7294 | } \ |
| 7295 | } \ |
| 7296 | while (0) |
| 7297 | |
| 7298 | /* Reject "bad registers" for Thumb-2 instructions. Many Thumb-2 |
| 7299 | instructions are unpredictable if these registers are used. This |
| 7300 | is the BadReg predicate in ARM's Thumb-2 documentation. |
| 7301 | |
| 7302 | Before ARMv8-A, REG_PC and REG_SP were not allowed in quite a few |
| 7303 | places, while the restriction on REG_SP was relaxed since ARMv8-A. */ |
| 7304 | #define reject_bad_reg(reg) \ |
| 7305 | do \ |
| 7306 | if (reg == REG_PC) \ |
| 7307 | { \ |
| 7308 | inst.error = BAD_PC; \ |
| 7309 | return; \ |
| 7310 | } \ |
| 7311 | else if (reg == REG_SP \ |
| 7312 | && !ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v8)) \ |
| 7313 | { \ |
| 7314 | inst.error = BAD_SP; \ |
| 7315 | return; \ |
| 7316 | } \ |
| 7317 | while (0) |
| 7318 | |
| 7319 | /* If REG is R13 (the stack pointer), warn that its use is |
| 7320 | deprecated. */ |
| 7321 | #define warn_deprecated_sp(reg) \ |
| 7322 | do \ |
| 7323 | if (warn_on_deprecated && reg == REG_SP) \ |
| 7324 | as_tsktsk (_("use of r13 is deprecated")); \ |
| 7325 | while (0) |
| 7326 | |
| 7327 | /* Functions for operand encoding. ARM, then Thumb. */ |
| 7328 | |
| 7329 | #define rotate_left(v, n) (v << (n & 31) | v >> ((32 - n) & 31)) |
| 7330 | |
| 7331 | /* If the current inst is scalar ARMv8.2 fp16 instruction, do special encoding. |
| 7332 | |
| 7333 | The only binary encoding difference is the Coprocessor number. Coprocessor |
| 7334 | 9 is used for half-precision calculations or conversions. The format of the |
| 7335 | instruction is the same as the equivalent Coprocessor 10 instruction that |
| 7336 | exists for Single-Precision operation. */ |
| 7337 | |
| 7338 | static void |
| 7339 | do_scalar_fp16_v82_encode (void) |
| 7340 | { |
| 7341 | if (inst.cond != COND_ALWAYS) |
| 7342 | as_warn (_("ARMv8.2 scalar fp16 instruction cannot be conditional," |
| 7343 | " the behaviour is UNPREDICTABLE")); |
| 7344 | constraint (!ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_fp16), |
| 7345 | _(BAD_FP16)); |
| 7346 | |
| 7347 | inst.instruction = (inst.instruction & 0xfffff0ff) | 0x900; |
| 7348 | mark_feature_used (&arm_ext_fp16); |
| 7349 | } |
| 7350 | |
| 7351 | /* If VAL can be encoded in the immediate field of an ARM instruction, |
| 7352 | return the encoded form. Otherwise, return FAIL. */ |
| 7353 | |
| 7354 | static unsigned int |
| 7355 | encode_arm_immediate (unsigned int val) |
| 7356 | { |
| 7357 | unsigned int a, i; |
| 7358 | |
| 7359 | if (val <= 0xff) |
| 7360 | return val; |
| 7361 | |
| 7362 | for (i = 2; i < 32; i += 2) |
| 7363 | if ((a = rotate_left (val, i)) <= 0xff) |
| 7364 | return a | (i << 7); /* 12-bit pack: [shift-cnt,const]. */ |
| 7365 | |
| 7366 | return FAIL; |
| 7367 | } |
| 7368 | |
| 7369 | /* If VAL can be encoded in the immediate field of a Thumb32 instruction, |
| 7370 | return the encoded form. Otherwise, return FAIL. */ |
| 7371 | static unsigned int |
| 7372 | encode_thumb32_immediate (unsigned int val) |
| 7373 | { |
| 7374 | unsigned int a, i; |
| 7375 | |
| 7376 | if (val <= 0xff) |
| 7377 | return val; |
| 7378 | |
| 7379 | for (i = 1; i <= 24; i++) |
| 7380 | { |
| 7381 | a = val >> i; |
| 7382 | if ((val & ~(0xff << i)) == 0) |
| 7383 | return ((val >> i) & 0x7f) | ((32 - i) << 7); |
| 7384 | } |
| 7385 | |
| 7386 | a = val & 0xff; |
| 7387 | if (val == ((a << 16) | a)) |
| 7388 | return 0x100 | a; |
| 7389 | if (val == ((a << 24) | (a << 16) | (a << 8) | a)) |
| 7390 | return 0x300 | a; |
| 7391 | |
| 7392 | a = val & 0xff00; |
| 7393 | if (val == ((a << 16) | a)) |
| 7394 | return 0x200 | (a >> 8); |
| 7395 | |
| 7396 | return FAIL; |
| 7397 | } |
| 7398 | /* Encode a VFP SP or DP register number into inst.instruction. */ |
| 7399 | |
| 7400 | static void |
| 7401 | encode_arm_vfp_reg (int reg, enum vfp_reg_pos pos) |
| 7402 | { |
| 7403 | if ((pos == VFP_REG_Dd || pos == VFP_REG_Dn || pos == VFP_REG_Dm) |
| 7404 | && reg > 15) |
| 7405 | { |
| 7406 | if (ARM_CPU_HAS_FEATURE (cpu_variant, fpu_vfp_ext_d32)) |
| 7407 | { |
| 7408 | if (thumb_mode) |
| 7409 | ARM_MERGE_FEATURE_SETS (thumb_arch_used, thumb_arch_used, |
| 7410 | fpu_vfp_ext_d32); |
| 7411 | else |
| 7412 | ARM_MERGE_FEATURE_SETS (arm_arch_used, arm_arch_used, |
| 7413 | fpu_vfp_ext_d32); |
| 7414 | } |
| 7415 | else |
| 7416 | { |
| 7417 | first_error (_("D register out of range for selected VFP version")); |
| 7418 | return; |
| 7419 | } |
| 7420 | } |
| 7421 | |
| 7422 | switch (pos) |
| 7423 | { |
| 7424 | case VFP_REG_Sd: |
| 7425 | inst.instruction |= ((reg >> 1) << 12) | ((reg & 1) << 22); |
| 7426 | break; |
| 7427 | |
| 7428 | case VFP_REG_Sn: |
| 7429 | inst.instruction |= ((reg >> 1) << 16) | ((reg & 1) << 7); |
| 7430 | break; |
| 7431 | |
| 7432 | case VFP_REG_Sm: |
| 7433 | inst.instruction |= ((reg >> 1) << 0) | ((reg & 1) << 5); |
| 7434 | break; |
| 7435 | |
| 7436 | case VFP_REG_Dd: |
| 7437 | inst.instruction |= ((reg & 15) << 12) | ((reg >> 4) << 22); |
| 7438 | break; |
| 7439 | |
| 7440 | case VFP_REG_Dn: |
| 7441 | inst.instruction |= ((reg & 15) << 16) | ((reg >> 4) << 7); |
| 7442 | break; |
| 7443 | |
| 7444 | case VFP_REG_Dm: |
| 7445 | inst.instruction |= (reg & 15) | ((reg >> 4) << 5); |
| 7446 | break; |
| 7447 | |
| 7448 | default: |
| 7449 | abort (); |
| 7450 | } |
| 7451 | } |
| 7452 | |
| 7453 | /* Encode a <shift> in an ARM-format instruction. The immediate, |
| 7454 | if any, is handled by md_apply_fix. */ |
| 7455 | static void |
| 7456 | encode_arm_shift (int i) |
| 7457 | { |
| 7458 | /* register-shifted register. */ |
| 7459 | if (inst.operands[i].immisreg) |
| 7460 | { |
| 7461 | int op_index; |
| 7462 | for (op_index = 0; op_index <= i; ++op_index) |
| 7463 | { |
| 7464 | /* Check the operand only when it's presented. In pre-UAL syntax, |
| 7465 | if the destination register is the same as the first operand, two |
| 7466 | register form of the instruction can be used. */ |
| 7467 | if (inst.operands[op_index].present && inst.operands[op_index].isreg |
| 7468 | && inst.operands[op_index].reg == REG_PC) |
| 7469 | as_warn (UNPRED_REG ("r15")); |
| 7470 | } |
| 7471 | |
| 7472 | if (inst.operands[i].imm == REG_PC) |
| 7473 | as_warn (UNPRED_REG ("r15")); |
| 7474 | } |
| 7475 | |
| 7476 | if (inst.operands[i].shift_kind == SHIFT_RRX) |
| 7477 | inst.instruction |= SHIFT_ROR << 5; |
| 7478 | else |
| 7479 | { |
| 7480 | inst.instruction |= inst.operands[i].shift_kind << 5; |
| 7481 | if (inst.operands[i].immisreg) |
| 7482 | { |
| 7483 | inst.instruction |= SHIFT_BY_REG; |
| 7484 | inst.instruction |= inst.operands[i].imm << 8; |
| 7485 | } |
| 7486 | else |
| 7487 | inst.reloc.type = BFD_RELOC_ARM_SHIFT_IMM; |
| 7488 | } |
| 7489 | } |
| 7490 | |
| 7491 | static void |
| 7492 | encode_arm_shifter_operand (int i) |
| 7493 | { |
| 7494 | if (inst.operands[i].isreg) |
| 7495 | { |
| 7496 | inst.instruction |= inst.operands[i].reg; |
| 7497 | encode_arm_shift (i); |
| 7498 | } |
| 7499 | else |
| 7500 | { |
| 7501 | inst.instruction |= INST_IMMEDIATE; |
| 7502 | if (inst.reloc.type != BFD_RELOC_ARM_IMMEDIATE) |
| 7503 | inst.instruction |= inst.operands[i].imm; |
| 7504 | } |
| 7505 | } |
| 7506 | |
| 7507 | /* Subroutine of encode_arm_addr_mode_2 and encode_arm_addr_mode_3. */ |
| 7508 | static void |
| 7509 | encode_arm_addr_mode_common (int i, bfd_boolean is_t) |
| 7510 | { |
| 7511 | /* PR 14260: |
| 7512 | Generate an error if the operand is not a register. */ |
| 7513 | constraint (!inst.operands[i].isreg, |
| 7514 | _("Instruction does not support =N addresses")); |
| 7515 | |
| 7516 | inst.instruction |= inst.operands[i].reg << 16; |
| 7517 | |
| 7518 | if (inst.operands[i].preind) |
| 7519 | { |
| 7520 | if (is_t) |
| 7521 | { |
| 7522 | inst.error = _("instruction does not accept preindexed addressing"); |
| 7523 | return; |
| 7524 | } |
| 7525 | inst.instruction |= PRE_INDEX; |
| 7526 | if (inst.operands[i].writeback) |
| 7527 | inst.instruction |= WRITE_BACK; |
| 7528 | |
| 7529 | } |
| 7530 | else if (inst.operands[i].postind) |
| 7531 | { |
| 7532 | gas_assert (inst.operands[i].writeback); |
| 7533 | if (is_t) |
| 7534 | inst.instruction |= WRITE_BACK; |
| 7535 | } |
| 7536 | else /* unindexed - only for coprocessor */ |
| 7537 | { |
| 7538 | inst.error = _("instruction does not accept unindexed addressing"); |
| 7539 | return; |
| 7540 | } |
| 7541 | |
| 7542 | if (((inst.instruction & WRITE_BACK) || !(inst.instruction & PRE_INDEX)) |
| 7543 | && (((inst.instruction & 0x000f0000) >> 16) |
| 7544 | == ((inst.instruction & 0x0000f000) >> 12))) |
| 7545 | as_warn ((inst.instruction & LOAD_BIT) |
| 7546 | ? _("destination register same as write-back base") |
| 7547 | : _("source register same as write-back base")); |
| 7548 | } |
| 7549 | |
| 7550 | /* inst.operands[i] was set up by parse_address. Encode it into an |
| 7551 | ARM-format mode 2 load or store instruction. If is_t is true, |
| 7552 | reject forms that cannot be used with a T instruction (i.e. not |
| 7553 | post-indexed). */ |
| 7554 | static void |
| 7555 | encode_arm_addr_mode_2 (int i, bfd_boolean is_t) |
| 7556 | { |
| 7557 | const bfd_boolean is_pc = (inst.operands[i].reg == REG_PC); |
| 7558 | |
| 7559 | encode_arm_addr_mode_common (i, is_t); |
| 7560 | |
| 7561 | if (inst.operands[i].immisreg) |
| 7562 | { |
| 7563 | constraint ((inst.operands[i].imm == REG_PC |
| 7564 | || (is_pc && inst.operands[i].writeback)), |
| 7565 | BAD_PC_ADDRESSING); |
| 7566 | inst.instruction |= INST_IMMEDIATE; /* yes, this is backwards */ |
| 7567 | inst.instruction |= inst.operands[i].imm; |
| 7568 | if (!inst.operands[i].negative) |
| 7569 | inst.instruction |= INDEX_UP; |
| 7570 | if (inst.operands[i].shifted) |
| 7571 | { |
| 7572 | if (inst.operands[i].shift_kind == SHIFT_RRX) |
| 7573 | inst.instruction |= SHIFT_ROR << 5; |
| 7574 | else |
| 7575 | { |
| 7576 | inst.instruction |= inst.operands[i].shift_kind << 5; |
| 7577 | inst.reloc.type = BFD_RELOC_ARM_SHIFT_IMM; |
| 7578 | } |
| 7579 | } |
| 7580 | } |
| 7581 | else /* immediate offset in inst.reloc */ |
| 7582 | { |
| 7583 | if (is_pc && !inst.reloc.pc_rel) |
| 7584 | { |
| 7585 | const bfd_boolean is_load = ((inst.instruction & LOAD_BIT) != 0); |
| 7586 | |
| 7587 | /* If is_t is TRUE, it's called from do_ldstt. ldrt/strt |
| 7588 | cannot use PC in addressing. |
| 7589 | PC cannot be used in writeback addressing, either. */ |
| 7590 | constraint ((is_t || inst.operands[i].writeback), |
| 7591 | BAD_PC_ADDRESSING); |
| 7592 | |
| 7593 | /* Use of PC in str is deprecated for ARMv7. */ |
| 7594 | if (warn_on_deprecated |
| 7595 | && !is_load |
| 7596 | && ARM_CPU_HAS_FEATURE (selected_cpu, arm_ext_v7)) |
| 7597 | as_tsktsk (_("use of PC in this instruction is deprecated")); |
| 7598 | } |
| 7599 | |
| 7600 | if (inst.reloc.type == BFD_RELOC_UNUSED) |
| 7601 | { |
| 7602 | /* Prefer + for zero encoded value. */ |
| 7603 | if (!inst.operands[i].negative) |
| 7604 | inst.instruction |= INDEX_UP; |
| 7605 | inst.reloc.type = BFD_RELOC_ARM_OFFSET_IMM; |
| 7606 | } |
| 7607 | } |
| 7608 | } |
| 7609 | |
| 7610 | /* inst.operands[i] was set up by parse_address. Encode it into an |
| 7611 | ARM-format mode 3 load or store instruction. Reject forms that |
| 7612 | cannot be used with such instructions. If is_t is true, reject |
| 7613 | forms that cannot be used with a T instruction (i.e. not |
| 7614 | post-indexed). */ |
| 7615 | static void |
| 7616 | encode_arm_addr_mode_3 (int i, bfd_boolean is_t) |
| 7617 | { |
| 7618 | if (inst.operands[i].immisreg && inst.operands[i].shifted) |
| 7619 | { |
| 7620 | inst.error = _("instruction does not accept scaled register index"); |
| 7621 | return; |
| 7622 | } |
| 7623 | |
| 7624 | encode_arm_addr_mode_common (i, is_t); |
| 7625 | |
| 7626 | if (inst.operands[i].immisreg) |
| 7627 | { |
| 7628 | constraint ((inst.operands[i].imm == REG_PC |
| 7629 | || (is_t && inst.operands[i].reg == REG_PC)), |
| 7630 | BAD_PC_ADDRESSING); |
| 7631 | constraint (inst.operands[i].reg == REG_PC && inst.operands[i].writeback, |
| 7632 | BAD_PC_WRITEBACK); |
| 7633 | inst.instruction |= inst.operands[i].imm; |
| 7634 | if (!inst.operands[i].negative) |
| 7635 | inst.instruction |= INDEX_UP; |
| 7636 | } |
| 7637 | else /* immediate offset in inst.reloc */ |
| 7638 | { |
| 7639 | constraint ((inst.operands[i].reg == REG_PC && !inst.reloc.pc_rel |
| 7640 | && inst.operands[i].writeback), |
| 7641 | BAD_PC_WRITEBACK); |
| 7642 | inst.instruction |= HWOFFSET_IMM; |
| 7643 | if (inst.reloc.type == BFD_RELOC_UNUSED) |
| 7644 | { |
| 7645 | /* Prefer + for zero encoded value. */ |
| 7646 | if (!inst.operands[i].negative) |
| 7647 | inst.instruction |= INDEX_UP; |
| 7648 | |
| 7649 | inst.reloc.type = BFD_RELOC_ARM_OFFSET_IMM8; |
| 7650 | } |
| 7651 | } |
| 7652 | } |
| 7653 | |
| 7654 | /* Write immediate bits [7:0] to the following locations: |
| 7655 | |
| 7656 | |28/24|23 19|18 16|15 4|3 0| |
| 7657 | | a |x x x x x|b c d|x x x x x x x x x x x x|e f g h| |
| 7658 | |
| 7659 | This function is used by VMOV/VMVN/VORR/VBIC. */ |
| 7660 | |
| 7661 | static void |
| 7662 | neon_write_immbits (unsigned immbits) |
| 7663 | { |
| 7664 | inst.instruction |= immbits & 0xf; |
| 7665 | inst.instruction |= ((immbits >> 4) & 0x7) << 16; |
| 7666 | inst.instruction |= ((immbits >> 7) & 0x1) << (thumb_mode ? 28 : 24); |
| 7667 | } |
| 7668 | |
| 7669 | /* Invert low-order SIZE bits of XHI:XLO. */ |
| 7670 | |
| 7671 | static void |
| 7672 | neon_invert_size (unsigned *xlo, unsigned *xhi, int size) |
| 7673 | { |
| 7674 | unsigned immlo = xlo ? *xlo : 0; |
| 7675 | unsigned immhi = xhi ? *xhi : 0; |
| 7676 | |
| 7677 | switch (size) |
| 7678 | { |
| 7679 | case 8: |
| 7680 | immlo = (~immlo) & 0xff; |
| 7681 | break; |
| 7682 | |
| 7683 | case 16: |
| 7684 | immlo = (~immlo) & 0xffff; |
| 7685 | break; |
| 7686 | |
| 7687 | case 64: |
| 7688 | immhi = (~immhi) & 0xffffffff; |
| 7689 | /* fall through. */ |
| 7690 | |
| 7691 | case 32: |
| 7692 | immlo = (~immlo) & 0xffffffff; |
| 7693 | break; |
| 7694 | |
| 7695 | default: |
| 7696 | abort (); |
| 7697 | } |
| 7698 | |
| 7699 | if (xlo) |
| 7700 | *xlo = immlo; |
| 7701 | |
| 7702 | if (xhi) |
| 7703 | *xhi = immhi; |
| 7704 | } |
| 7705 | |
| 7706 | /* True if IMM has form 0bAAAAAAAABBBBBBBBCCCCCCCCDDDDDDDD for bits |
| 7707 | A, B, C, D. */ |
| 7708 | |
| 7709 | static int |
| 7710 | neon_bits_same_in_bytes (unsigned imm) |
| 7711 | { |
| 7712 | return ((imm & 0x000000ff) == 0 || (imm & 0x000000ff) == 0x000000ff) |
| 7713 | && ((imm & 0x0000ff00) == 0 || (imm & 0x0000ff00) == 0x0000ff00) |
| 7714 | && ((imm & 0x00ff0000) == 0 || (imm & 0x00ff0000) == 0x00ff0000) |
| 7715 | && ((imm & 0xff000000) == 0 || (imm & 0xff000000) == 0xff000000); |
| 7716 | } |
| 7717 | |
| 7718 | /* For immediate of above form, return 0bABCD. */ |
| 7719 | |
| 7720 | static unsigned |
| 7721 | neon_squash_bits (unsigned imm) |
| 7722 | { |
| 7723 | return (imm & 0x01) | ((imm & 0x0100) >> 7) | ((imm & 0x010000) >> 14) |
| 7724 | | ((imm & 0x01000000) >> 21); |
| 7725 | } |
| 7726 | |
| 7727 | /* Compress quarter-float representation to 0b...000 abcdefgh. */ |
| 7728 | |
| 7729 | static unsigned |
| 7730 | neon_qfloat_bits (unsigned imm) |
| 7731 | { |
| 7732 | return ((imm >> 19) & 0x7f) | ((imm >> 24) & 0x80); |
| 7733 | } |
| 7734 | |
| 7735 | /* Returns CMODE. IMMBITS [7:0] is set to bits suitable for inserting into |
| 7736 | the instruction. *OP is passed as the initial value of the op field, and |
| 7737 | may be set to a different value depending on the constant (i.e. |
| 7738 | "MOV I64, 0bAAAAAAAABBBB..." which uses OP = 1 despite being MOV not |
| 7739 | MVN). If the immediate looks like a repeated pattern then also |
| 7740 | try smaller element sizes. */ |
| 7741 | |
| 7742 | static int |
| 7743 | neon_cmode_for_move_imm (unsigned immlo, unsigned immhi, int float_p, |
| 7744 | unsigned *immbits, int *op, int size, |
| 7745 | enum neon_el_type type) |
| 7746 | { |
| 7747 | /* Only permit float immediates (including 0.0/-0.0) if the operand type is |
| 7748 | float. */ |
| 7749 | if (type == NT_float && !float_p) |
| 7750 | return FAIL; |
| 7751 | |
| 7752 | if (type == NT_float && is_quarter_float (immlo) && immhi == 0) |
| 7753 | { |
| 7754 | if (size != 32 || *op == 1) |
| 7755 | return FAIL; |
| 7756 | *immbits = neon_qfloat_bits (immlo); |
| 7757 | return 0xf; |
| 7758 | } |
| 7759 | |
| 7760 | if (size == 64) |
| 7761 | { |
| 7762 | if (neon_bits_same_in_bytes (immhi) |
| 7763 | && neon_bits_same_in_bytes (immlo)) |
| 7764 | { |
| 7765 | if (*op == 1) |
| 7766 | return FAIL; |
| 7767 | *immbits = (neon_squash_bits (immhi) << 4) |
| 7768 | | neon_squash_bits (immlo); |
| 7769 | *op = 1; |
| 7770 | return 0xe; |
| 7771 | } |
| 7772 | |
| 7773 | if (immhi != immlo) |
| 7774 | return FAIL; |
| 7775 | } |
| 7776 | |
| 7777 | if (size >= 32) |
| 7778 | { |
| 7779 | if (immlo == (immlo & 0x000000ff)) |
| 7780 | { |
| 7781 | *immbits = immlo; |
| 7782 | return 0x0; |
| 7783 | } |
| 7784 | else if (immlo == (immlo & 0x0000ff00)) |
| 7785 | { |
| 7786 | *immbits = immlo >> 8; |
| 7787 | return 0x2; |
| 7788 | } |
| 7789 | else if (immlo == (immlo & 0x00ff0000)) |
| 7790 | { |
| 7791 | *immbits = immlo >> 16; |
| 7792 | return 0x4; |
| 7793 | } |
| 7794 | else if (immlo == (immlo & 0xff000000)) |
| 7795 | { |
| 7796 | *immbits = immlo >> 24; |
| 7797 | return 0x6; |
| 7798 | } |
| 7799 | else if (immlo == ((immlo & 0x0000ff00) | 0x000000ff)) |
| 7800 | { |
| 7801 | *immbits = (immlo >> 8) & 0xff; |
| 7802 | return 0xc; |
| 7803 | } |
| 7804 | else if (immlo == ((immlo & 0x00ff0000) | 0x0000ffff)) |
| 7805 | { |
| 7806 | *immbits = (immlo >> 16) & 0xff; |
| 7807 | return 0xd; |
| 7808 | } |
| 7809 | |
| 7810 | if ((immlo & 0xffff) != (immlo >> 16)) |
| 7811 | return FAIL; |
| 7812 | immlo &= 0xffff; |
| 7813 | } |
| 7814 | |
| 7815 | if (size >= 16) |
| 7816 | { |
| 7817 | if (immlo == (immlo & 0x000000ff)) |
| 7818 | { |
| 7819 | *immbits = immlo; |
| 7820 | return 0x8; |
| 7821 | } |
| 7822 | else if (immlo == (immlo & 0x0000ff00)) |
| 7823 | { |
| 7824 | *immbits = immlo >> 8; |
| 7825 | return 0xa; |
| 7826 | } |
| 7827 | |
| 7828 | if ((immlo & 0xff) != (immlo >> 8)) |
| 7829 | return FAIL; |
| 7830 | immlo &= 0xff; |
| 7831 | } |
| 7832 | |
| 7833 | if (immlo == (immlo & 0x000000ff)) |
| 7834 | { |
| 7835 | /* Don't allow MVN with 8-bit immediate. */ |
| 7836 | if (*op == 1) |
| 7837 | return FAIL; |
| 7838 | *immbits = immlo; |
| 7839 | return 0xe; |
| 7840 | } |
| 7841 | |
| 7842 | return FAIL; |
| 7843 | } |
| 7844 | |
| 7845 | #if defined BFD_HOST_64_BIT |
| 7846 | /* Returns TRUE if double precision value V may be cast |
| 7847 | to single precision without loss of accuracy. */ |
| 7848 | |
| 7849 | static bfd_boolean |
| 7850 | is_double_a_single (bfd_int64_t v) |
| 7851 | { |
| 7852 | int exp = (int)((v >> 52) & 0x7FF); |
| 7853 | bfd_int64_t mantissa = (v & (bfd_int64_t)0xFFFFFFFFFFFFFULL); |
| 7854 | |
| 7855 | return (exp == 0 || exp == 0x7FF |
| 7856 | || (exp >= 1023 - 126 && exp <= 1023 + 127)) |
| 7857 | && (mantissa & 0x1FFFFFFFl) == 0; |
| 7858 | } |
| 7859 | |
| 7860 | /* Returns a double precision value casted to single precision |
| 7861 | (ignoring the least significant bits in exponent and mantissa). */ |
| 7862 | |
| 7863 | static int |
| 7864 | double_to_single (bfd_int64_t v) |
| 7865 | { |
| 7866 | int sign = (int) ((v >> 63) & 1l); |
| 7867 | int exp = (int) ((v >> 52) & 0x7FF); |
| 7868 | bfd_int64_t mantissa = (v & (bfd_int64_t)0xFFFFFFFFFFFFFULL); |
| 7869 | |
| 7870 | if (exp == 0x7FF) |
| 7871 | exp = 0xFF; |
| 7872 | else |
| 7873 | { |
| 7874 | exp = exp - 1023 + 127; |
| 7875 | if (exp >= 0xFF) |
| 7876 | { |
| 7877 | /* Infinity. */ |
| 7878 | exp = 0x7F; |
| 7879 | mantissa = 0; |
| 7880 | } |
| 7881 | else if (exp < 0) |
| 7882 | { |
| 7883 | /* No denormalized numbers. */ |
| 7884 | exp = 0; |
| 7885 | mantissa = 0; |
| 7886 | } |
| 7887 | } |
| 7888 | mantissa >>= 29; |
| 7889 | return (sign << 31) | (exp << 23) | mantissa; |
| 7890 | } |
| 7891 | #endif /* BFD_HOST_64_BIT */ |
| 7892 | |
| 7893 | enum lit_type |
| 7894 | { |
| 7895 | CONST_THUMB, |
| 7896 | CONST_ARM, |
| 7897 | CONST_VEC |
| 7898 | }; |
| 7899 | |
| 7900 | static void do_vfp_nsyn_opcode (const char *); |
| 7901 | |
| 7902 | /* inst.reloc.exp describes an "=expr" load pseudo-operation. |
| 7903 | Determine whether it can be performed with a move instruction; if |
| 7904 | it can, convert inst.instruction to that move instruction and |
| 7905 | return TRUE; if it can't, convert inst.instruction to a literal-pool |
| 7906 | load and return FALSE. If this is not a valid thing to do in the |
| 7907 | current context, set inst.error and return TRUE. |
| 7908 | |
| 7909 | inst.operands[i] describes the destination register. */ |
| 7910 | |
| 7911 | static bfd_boolean |
| 7912 | move_or_literal_pool (int i, enum lit_type t, bfd_boolean mode_3) |
| 7913 | { |
| 7914 | unsigned long tbit; |
| 7915 | bfd_boolean thumb_p = (t == CONST_THUMB); |
| 7916 | bfd_boolean arm_p = (t == CONST_ARM); |
| 7917 | |
| 7918 | if (thumb_p) |
| 7919 | tbit = (inst.instruction > 0xffff) ? THUMB2_LOAD_BIT : THUMB_LOAD_BIT; |
| 7920 | else |
| 7921 | tbit = LOAD_BIT; |
| 7922 | |
| 7923 | if ((inst.instruction & tbit) == 0) |
| 7924 | { |
| 7925 | inst.error = _("invalid pseudo operation"); |
| 7926 | return TRUE; |
| 7927 | } |
| 7928 | |
| 7929 | if (inst.reloc.exp.X_op != O_constant |
| 7930 | && inst.reloc.exp.X_op != O_symbol |
| 7931 | && inst.reloc.exp.X_op != O_big) |
| 7932 | { |
| 7933 | inst.error = _("constant expression expected"); |
| 7934 | return TRUE; |
| 7935 | } |
| 7936 | |
| 7937 | if (inst.reloc.exp.X_op == O_constant |
| 7938 | || inst.reloc.exp.X_op == O_big) |
| 7939 | { |
| 7940 | #if defined BFD_HOST_64_BIT |
| 7941 | bfd_int64_t v; |
| 7942 | #else |
| 7943 | offsetT v; |
| 7944 | #endif |
| 7945 | if (inst.reloc.exp.X_op == O_big) |
| 7946 | { |
| 7947 | LITTLENUM_TYPE w[X_PRECISION]; |
| 7948 | LITTLENUM_TYPE * l; |
| 7949 | |
| 7950 | if (inst.reloc.exp.X_add_number == -1) |
| 7951 | { |
| 7952 | gen_to_words (w, X_PRECISION, E_PRECISION); |
| 7953 | l = w; |
| 7954 | /* FIXME: Should we check words w[2..5] ? */ |
| 7955 | } |
| 7956 | else |
| 7957 | l = generic_bignum; |
| 7958 | |
| 7959 | #if defined BFD_HOST_64_BIT |
| 7960 | v = |
| 7961 | ((((((((bfd_int64_t) l[3] & LITTLENUM_MASK) |
| 7962 | << LITTLENUM_NUMBER_OF_BITS) |
| 7963 | | ((bfd_int64_t) l[2] & LITTLENUM_MASK)) |
| 7964 | << LITTLENUM_NUMBER_OF_BITS) |
| 7965 | | ((bfd_int64_t) l[1] & LITTLENUM_MASK)) |
| 7966 | << LITTLENUM_NUMBER_OF_BITS) |
| 7967 | | ((bfd_int64_t) l[0] & LITTLENUM_MASK)); |
| 7968 | #else |
| 7969 | v = ((l[1] & LITTLENUM_MASK) << LITTLENUM_NUMBER_OF_BITS) |
| 7970 | | (l[0] & LITTLENUM_MASK); |
| 7971 | #endif |
| 7972 | } |
| 7973 | else |
| 7974 | v = inst.reloc.exp.X_add_number; |
| 7975 | |
| 7976 | if (!inst.operands[i].issingle) |
| 7977 | { |
| 7978 | if (thumb_p) |
| 7979 | { |
| 7980 | /* LDR should not use lead in a flag-setting instruction being |
| 7981 | chosen so we do not check whether movs can be used. */ |
| 7982 | |
| 7983 | if ((ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v6t2) |
| 7984 | || ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v6t2_v8m)) |
| 7985 | && inst.operands[i].reg != 13 |
| 7986 | && inst.operands[i].reg != 15) |
| 7987 | { |
| 7988 | /* Check if on thumb2 it can be done with a mov.w, mvn or |
| 7989 | movw instruction. */ |
| 7990 | unsigned int newimm; |
| 7991 | bfd_boolean isNegated; |
| 7992 | |
| 7993 | newimm = encode_thumb32_immediate (v); |
| 7994 | if (newimm != (unsigned int) FAIL) |
| 7995 | isNegated = FALSE; |
| 7996 | else |
| 7997 | { |
| 7998 | newimm = encode_thumb32_immediate (~v); |
| 7999 | if (newimm != (unsigned int) FAIL) |
| 8000 | isNegated = TRUE; |
| 8001 | } |
| 8002 | |
| 8003 | /* The number can be loaded with a mov.w or mvn |
| 8004 | instruction. */ |
| 8005 | if (newimm != (unsigned int) FAIL |
| 8006 | && ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v6t2)) |
| 8007 | { |
| 8008 | inst.instruction = (0xf04f0000 /* MOV.W. */ |
| 8009 | | (inst.operands[i].reg << 8)); |
| 8010 | /* Change to MOVN. */ |
| 8011 | inst.instruction |= (isNegated ? 0x200000 : 0); |
| 8012 | inst.instruction |= (newimm & 0x800) << 15; |
| 8013 | inst.instruction |= (newimm & 0x700) << 4; |
| 8014 | inst.instruction |= (newimm & 0x0ff); |
| 8015 | return TRUE; |
| 8016 | } |
| 8017 | /* The number can be loaded with a movw instruction. */ |
| 8018 | else if ((v & ~0xFFFF) == 0 |
| 8019 | && ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v6t2_v8m)) |
| 8020 | { |
| 8021 | int imm = v & 0xFFFF; |
| 8022 | |
| 8023 | inst.instruction = 0xf2400000; /* MOVW. */ |
| 8024 | inst.instruction |= (inst.operands[i].reg << 8); |
| 8025 | inst.instruction |= (imm & 0xf000) << 4; |
| 8026 | inst.instruction |= (imm & 0x0800) << 15; |
| 8027 | inst.instruction |= (imm & 0x0700) << 4; |
| 8028 | inst.instruction |= (imm & 0x00ff); |
| 8029 | return TRUE; |
| 8030 | } |
| 8031 | } |
| 8032 | } |
| 8033 | else if (arm_p) |
| 8034 | { |
| 8035 | int value = encode_arm_immediate (v); |
| 8036 | |
| 8037 | if (value != FAIL) |
| 8038 | { |
| 8039 | /* This can be done with a mov instruction. */ |
| 8040 | inst.instruction &= LITERAL_MASK; |
| 8041 | inst.instruction |= INST_IMMEDIATE | (OPCODE_MOV << DATA_OP_SHIFT); |
| 8042 | inst.instruction |= value & 0xfff; |
| 8043 | return TRUE; |
| 8044 | } |
| 8045 | |
| 8046 | value = encode_arm_immediate (~ v); |
| 8047 | if (value != FAIL) |
| 8048 | { |
| 8049 | /* This can be done with a mvn instruction. */ |
| 8050 | inst.instruction &= LITERAL_MASK; |
| 8051 | inst.instruction |= INST_IMMEDIATE | (OPCODE_MVN << DATA_OP_SHIFT); |
| 8052 | inst.instruction |= value & 0xfff; |
| 8053 | return TRUE; |
| 8054 | } |
| 8055 | } |
| 8056 | else if (t == CONST_VEC && ARM_CPU_HAS_FEATURE (cpu_variant, fpu_neon_ext_v1)) |
| 8057 | { |
| 8058 | int op = 0; |
| 8059 | unsigned immbits = 0; |
| 8060 | unsigned immlo = inst.operands[1].imm; |
| 8061 | unsigned immhi = inst.operands[1].regisimm |
| 8062 | ? inst.operands[1].reg |
| 8063 | : inst.reloc.exp.X_unsigned |
| 8064 | ? 0 |
| 8065 | : ((bfd_int64_t)((int) immlo)) >> 32; |
| 8066 | int cmode = neon_cmode_for_move_imm (immlo, immhi, FALSE, &immbits, |
| 8067 | &op, 64, NT_invtype); |
| 8068 | |
| 8069 | if (cmode == FAIL) |
| 8070 | { |
| 8071 | neon_invert_size (&immlo, &immhi, 64); |
| 8072 | op = !op; |
| 8073 | cmode = neon_cmode_for_move_imm (immlo, immhi, FALSE, &immbits, |
| 8074 | &op, 64, NT_invtype); |
| 8075 | } |
| 8076 | |
| 8077 | if (cmode != FAIL) |
| 8078 | { |
| 8079 | inst.instruction = (inst.instruction & VLDR_VMOV_SAME) |
| 8080 | | (1 << 23) |
| 8081 | | (cmode << 8) |
| 8082 | | (op << 5) |
| 8083 | | (1 << 4); |
| 8084 | |
| 8085 | /* Fill other bits in vmov encoding for both thumb and arm. */ |
| 8086 | if (thumb_mode) |
| 8087 | inst.instruction |= (0x7U << 29) | (0xF << 24); |
| 8088 | else |
| 8089 | inst.instruction |= (0xFU << 28) | (0x1 << 25); |
| 8090 | neon_write_immbits (immbits); |
| 8091 | return TRUE; |
| 8092 | } |
| 8093 | } |
| 8094 | } |
| 8095 | |
| 8096 | if (t == CONST_VEC) |
| 8097 | { |
| 8098 | /* Check if vldr Rx, =constant could be optimized to vmov Rx, #constant. */ |
| 8099 | if (inst.operands[i].issingle |
| 8100 | && is_quarter_float (inst.operands[1].imm) |
| 8101 | && ARM_CPU_HAS_FEATURE (cpu_variant, fpu_vfp_ext_v3xd)) |
| 8102 | { |
| 8103 | inst.operands[1].imm = |
| 8104 | neon_qfloat_bits (v); |
| 8105 | do_vfp_nsyn_opcode ("fconsts"); |
| 8106 | return TRUE; |
| 8107 | } |
| 8108 | |
| 8109 | /* If our host does not support a 64-bit type then we cannot perform |
| 8110 | the following optimization. This mean that there will be a |
| 8111 | discrepancy between the output produced by an assembler built for |
| 8112 | a 32-bit-only host and the output produced from a 64-bit host, but |
| 8113 | this cannot be helped. */ |
| 8114 | #if defined BFD_HOST_64_BIT |
| 8115 | else if (!inst.operands[1].issingle |
| 8116 | && ARM_CPU_HAS_FEATURE (cpu_variant, fpu_vfp_ext_v3)) |
| 8117 | { |
| 8118 | if (is_double_a_single (v) |
| 8119 | && is_quarter_float (double_to_single (v))) |
| 8120 | { |
| 8121 | inst.operands[1].imm = |
| 8122 | neon_qfloat_bits (double_to_single (v)); |
| 8123 | do_vfp_nsyn_opcode ("fconstd"); |
| 8124 | return TRUE; |
| 8125 | } |
| 8126 | } |
| 8127 | #endif |
| 8128 | } |
| 8129 | } |
| 8130 | |
| 8131 | if (add_to_lit_pool ((!inst.operands[i].isvec |
| 8132 | || inst.operands[i].issingle) ? 4 : 8) == FAIL) |
| 8133 | return TRUE; |
| 8134 | |
| 8135 | inst.operands[1].reg = REG_PC; |
| 8136 | inst.operands[1].isreg = 1; |
| 8137 | inst.operands[1].preind = 1; |
| 8138 | inst.reloc.pc_rel = 1; |
| 8139 | inst.reloc.type = (thumb_p |
| 8140 | ? BFD_RELOC_ARM_THUMB_OFFSET |
| 8141 | : (mode_3 |
| 8142 | ? BFD_RELOC_ARM_HWLITERAL |
| 8143 | : BFD_RELOC_ARM_LITERAL)); |
| 8144 | return FALSE; |
| 8145 | } |
| 8146 | |
| 8147 | /* inst.operands[i] was set up by parse_address. Encode it into an |
| 8148 | ARM-format instruction. Reject all forms which cannot be encoded |
| 8149 | into a coprocessor load/store instruction. If wb_ok is false, |
| 8150 | reject use of writeback; if unind_ok is false, reject use of |
| 8151 | unindexed addressing. If reloc_override is not 0, use it instead |
| 8152 | of BFD_ARM_CP_OFF_IMM, unless the initial relocation is a group one |
| 8153 | (in which case it is preserved). */ |
| 8154 | |
| 8155 | static int |
| 8156 | encode_arm_cp_address (int i, int wb_ok, int unind_ok, int reloc_override) |
| 8157 | { |
| 8158 | if (!inst.operands[i].isreg) |
| 8159 | { |
| 8160 | /* PR 18256 */ |
| 8161 | if (! inst.operands[0].isvec) |
| 8162 | { |
| 8163 | inst.error = _("invalid co-processor operand"); |
| 8164 | return FAIL; |
| 8165 | } |
| 8166 | if (move_or_literal_pool (0, CONST_VEC, /*mode_3=*/FALSE)) |
| 8167 | return SUCCESS; |
| 8168 | } |
| 8169 | |
| 8170 | inst.instruction |= inst.operands[i].reg << 16; |
| 8171 | |
| 8172 | gas_assert (!(inst.operands[i].preind && inst.operands[i].postind)); |
| 8173 | |
| 8174 | if (!inst.operands[i].preind && !inst.operands[i].postind) /* unindexed */ |
| 8175 | { |
| 8176 | gas_assert (!inst.operands[i].writeback); |
| 8177 | if (!unind_ok) |
| 8178 | { |
| 8179 | inst.error = _("instruction does not support unindexed addressing"); |
| 8180 | return FAIL; |
| 8181 | } |
| 8182 | inst.instruction |= inst.operands[i].imm; |
| 8183 | inst.instruction |= INDEX_UP; |
| 8184 | return SUCCESS; |
| 8185 | } |
| 8186 | |
| 8187 | if (inst.operands[i].preind) |
| 8188 | inst.instruction |= PRE_INDEX; |
| 8189 | |
| 8190 | if (inst.operands[i].writeback) |
| 8191 | { |
| 8192 | if (inst.operands[i].reg == REG_PC) |
| 8193 | { |
| 8194 | inst.error = _("pc may not be used with write-back"); |
| 8195 | return FAIL; |
| 8196 | } |
| 8197 | if (!wb_ok) |
| 8198 | { |
| 8199 | inst.error = _("instruction does not support writeback"); |
| 8200 | return FAIL; |
| 8201 | } |
| 8202 | inst.instruction |= WRITE_BACK; |
| 8203 | } |
| 8204 | |
| 8205 | if (reloc_override) |
| 8206 | inst.reloc.type = (bfd_reloc_code_real_type) reloc_override; |
| 8207 | else if ((inst.reloc.type < BFD_RELOC_ARM_ALU_PC_G0_NC |
| 8208 | || inst.reloc.type > BFD_RELOC_ARM_LDC_SB_G2) |
| 8209 | && inst.reloc.type != BFD_RELOC_ARM_LDR_PC_G0) |
| 8210 | { |
| 8211 | if (thumb_mode) |
| 8212 | inst.reloc.type = BFD_RELOC_ARM_T32_CP_OFF_IMM; |
| 8213 | else |
| 8214 | inst.reloc.type = BFD_RELOC_ARM_CP_OFF_IMM; |
| 8215 | } |
| 8216 | |
| 8217 | /* Prefer + for zero encoded value. */ |
| 8218 | if (!inst.operands[i].negative) |
| 8219 | inst.instruction |= INDEX_UP; |
| 8220 | |
| 8221 | return SUCCESS; |
| 8222 | } |
| 8223 | |
| 8224 | /* Functions for instruction encoding, sorted by sub-architecture. |
| 8225 | First some generics; their names are taken from the conventional |
| 8226 | bit positions for register arguments in ARM format instructions. */ |
| 8227 | |
| 8228 | static void |
| 8229 | do_noargs (void) |
| 8230 | { |
| 8231 | } |
| 8232 | |
| 8233 | static void |
| 8234 | do_rd (void) |
| 8235 | { |
| 8236 | inst.instruction |= inst.operands[0].reg << 12; |
| 8237 | } |
| 8238 | |
| 8239 | static void |
| 8240 | do_rn (void) |
| 8241 | { |
| 8242 | inst.instruction |= inst.operands[0].reg << 16; |
| 8243 | } |
| 8244 | |
| 8245 | static void |
| 8246 | do_rd_rm (void) |
| 8247 | { |
| 8248 | inst.instruction |= inst.operands[0].reg << 12; |
| 8249 | inst.instruction |= inst.operands[1].reg; |
| 8250 | } |
| 8251 | |
| 8252 | static void |
| 8253 | do_rm_rn (void) |
| 8254 | { |
| 8255 | inst.instruction |= inst.operands[0].reg; |
| 8256 | inst.instruction |= inst.operands[1].reg << 16; |
| 8257 | } |
| 8258 | |
| 8259 | static void |
| 8260 | do_rd_rn (void) |
| 8261 | { |
| 8262 | inst.instruction |= inst.operands[0].reg << 12; |
| 8263 | inst.instruction |= inst.operands[1].reg << 16; |
| 8264 | } |
| 8265 | |
| 8266 | static void |
| 8267 | do_rn_rd (void) |
| 8268 | { |
| 8269 | inst.instruction |= inst.operands[0].reg << 16; |
| 8270 | inst.instruction |= inst.operands[1].reg << 12; |
| 8271 | } |
| 8272 | |
| 8273 | static void |
| 8274 | do_tt (void) |
| 8275 | { |
| 8276 | inst.instruction |= inst.operands[0].reg << 8; |
| 8277 | inst.instruction |= inst.operands[1].reg << 16; |
| 8278 | } |
| 8279 | |
| 8280 | static bfd_boolean |
| 8281 | check_obsolete (const arm_feature_set *feature, const char *msg) |
| 8282 | { |
| 8283 | if (ARM_CPU_IS_ANY (cpu_variant)) |
| 8284 | { |
| 8285 | as_tsktsk ("%s", msg); |
| 8286 | return TRUE; |
| 8287 | } |
| 8288 | else if (ARM_CPU_HAS_FEATURE (cpu_variant, *feature)) |
| 8289 | { |
| 8290 | as_bad ("%s", msg); |
| 8291 | return TRUE; |
| 8292 | } |
| 8293 | |
| 8294 | return FALSE; |
| 8295 | } |
| 8296 | |
| 8297 | static void |
| 8298 | do_rd_rm_rn (void) |
| 8299 | { |
| 8300 | unsigned Rn = inst.operands[2].reg; |
| 8301 | /* Enforce restrictions on SWP instruction. */ |
| 8302 | if ((inst.instruction & 0x0fbfffff) == 0x01000090) |
| 8303 | { |
| 8304 | constraint (Rn == inst.operands[0].reg || Rn == inst.operands[1].reg, |
| 8305 | _("Rn must not overlap other operands")); |
| 8306 | |
| 8307 | /* SWP{b} is obsolete for ARMv8-A, and deprecated for ARMv6* and ARMv7. |
| 8308 | */ |
| 8309 | if (!check_obsolete (&arm_ext_v8, |
| 8310 | _("swp{b} use is obsoleted for ARMv8 and later")) |
| 8311 | && warn_on_deprecated |
| 8312 | && ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v6)) |
| 8313 | as_tsktsk (_("swp{b} use is deprecated for ARMv6 and ARMv7")); |
| 8314 | } |
| 8315 | |
| 8316 | inst.instruction |= inst.operands[0].reg << 12; |
| 8317 | inst.instruction |= inst.operands[1].reg; |
| 8318 | inst.instruction |= Rn << 16; |
| 8319 | } |
| 8320 | |
| 8321 | static void |
| 8322 | do_rd_rn_rm (void) |
| 8323 | { |
| 8324 | inst.instruction |= inst.operands[0].reg << 12; |
| 8325 | inst.instruction |= inst.operands[1].reg << 16; |
| 8326 | inst.instruction |= inst.operands[2].reg; |
| 8327 | } |
| 8328 | |
| 8329 | static void |
| 8330 | do_rm_rd_rn (void) |
| 8331 | { |
| 8332 | constraint ((inst.operands[2].reg == REG_PC), BAD_PC); |
| 8333 | constraint (((inst.reloc.exp.X_op != O_constant |
| 8334 | && inst.reloc.exp.X_op != O_illegal) |
| 8335 | || inst.reloc.exp.X_add_number != 0), |
| 8336 | BAD_ADDR_MODE); |
| 8337 | inst.instruction |= inst.operands[0].reg; |
| 8338 | inst.instruction |= inst.operands[1].reg << 12; |
| 8339 | inst.instruction |= inst.operands[2].reg << 16; |
| 8340 | } |
| 8341 | |
| 8342 | static void |
| 8343 | do_imm0 (void) |
| 8344 | { |
| 8345 | inst.instruction |= inst.operands[0].imm; |
| 8346 | } |
| 8347 | |
| 8348 | static void |
| 8349 | do_rd_cpaddr (void) |
| 8350 | { |
| 8351 | inst.instruction |= inst.operands[0].reg << 12; |
| 8352 | encode_arm_cp_address (1, TRUE, TRUE, 0); |
| 8353 | } |
| 8354 | |
| 8355 | /* ARM instructions, in alphabetical order by function name (except |
| 8356 | that wrapper functions appear immediately after the function they |
| 8357 | wrap). */ |
| 8358 | |
| 8359 | /* This is a pseudo-op of the form "adr rd, label" to be converted |
| 8360 | into a relative address of the form "add rd, pc, #label-.-8". */ |
| 8361 | |
| 8362 | static void |
| 8363 | do_adr (void) |
| 8364 | { |
| 8365 | inst.instruction |= (inst.operands[0].reg << 12); /* Rd */ |
| 8366 | |
| 8367 | /* Frag hacking will turn this into a sub instruction if the offset turns |
| 8368 | out to be negative. */ |
| 8369 | inst.reloc.type = BFD_RELOC_ARM_IMMEDIATE; |
| 8370 | inst.reloc.pc_rel = 1; |
| 8371 | inst.reloc.exp.X_add_number -= 8; |
| 8372 | |
| 8373 | if (inst.reloc.exp.X_op == O_symbol |
| 8374 | && inst.reloc.exp.X_add_symbol != NULL |
| 8375 | && S_IS_DEFINED (inst.reloc.exp.X_add_symbol) |
| 8376 | && THUMB_IS_FUNC (inst.reloc.exp.X_add_symbol)) |
| 8377 | inst.reloc.exp.X_add_number += 1; |
| 8378 | } |
| 8379 | |
| 8380 | /* This is a pseudo-op of the form "adrl rd, label" to be converted |
| 8381 | into a relative address of the form: |
| 8382 | add rd, pc, #low(label-.-8)" |
| 8383 | add rd, rd, #high(label-.-8)" */ |
| 8384 | |
| 8385 | static void |
| 8386 | do_adrl (void) |
| 8387 | { |
| 8388 | inst.instruction |= (inst.operands[0].reg << 12); /* Rd */ |
| 8389 | |
| 8390 | /* Frag hacking will turn this into a sub instruction if the offset turns |
| 8391 | out to be negative. */ |
| 8392 | inst.reloc.type = BFD_RELOC_ARM_ADRL_IMMEDIATE; |
| 8393 | inst.reloc.pc_rel = 1; |
| 8394 | inst.size = INSN_SIZE * 2; |
| 8395 | inst.reloc.exp.X_add_number -= 8; |
| 8396 | |
| 8397 | if (inst.reloc.exp.X_op == O_symbol |
| 8398 | && inst.reloc.exp.X_add_symbol != NULL |
| 8399 | && S_IS_DEFINED (inst.reloc.exp.X_add_symbol) |
| 8400 | && THUMB_IS_FUNC (inst.reloc.exp.X_add_symbol)) |
| 8401 | inst.reloc.exp.X_add_number += 1; |
| 8402 | } |
| 8403 | |
| 8404 | static void |
| 8405 | do_arit (void) |
| 8406 | { |
| 8407 | constraint (inst.reloc.type >= BFD_RELOC_ARM_THUMB_ALU_ABS_G0_NC |
| 8408 | && inst.reloc.type <= BFD_RELOC_ARM_THUMB_ALU_ABS_G3_NC , |
| 8409 | THUMB1_RELOC_ONLY); |
| 8410 | if (!inst.operands[1].present) |
| 8411 | inst.operands[1].reg = inst.operands[0].reg; |
| 8412 | inst.instruction |= inst.operands[0].reg << 12; |
| 8413 | inst.instruction |= inst.operands[1].reg << 16; |
| 8414 | encode_arm_shifter_operand (2); |
| 8415 | } |
| 8416 | |
| 8417 | static void |
| 8418 | do_barrier (void) |
| 8419 | { |
| 8420 | if (inst.operands[0].present) |
| 8421 | inst.instruction |= inst.operands[0].imm; |
| 8422 | else |
| 8423 | inst.instruction |= 0xf; |
| 8424 | } |
| 8425 | |
| 8426 | static void |
| 8427 | do_bfc (void) |
| 8428 | { |
| 8429 | unsigned int msb = inst.operands[1].imm + inst.operands[2].imm; |
| 8430 | constraint (msb > 32, _("bit-field extends past end of register")); |
| 8431 | /* The instruction encoding stores the LSB and MSB, |
| 8432 | not the LSB and width. */ |
| 8433 | inst.instruction |= inst.operands[0].reg << 12; |
| 8434 | inst.instruction |= inst.operands[1].imm << 7; |
| 8435 | inst.instruction |= (msb - 1) << 16; |
| 8436 | } |
| 8437 | |
| 8438 | static void |
| 8439 | do_bfi (void) |
| 8440 | { |
| 8441 | unsigned int msb; |
| 8442 | |
| 8443 | /* #0 in second position is alternative syntax for bfc, which is |
| 8444 | the same instruction but with REG_PC in the Rm field. */ |
| 8445 | if (!inst.operands[1].isreg) |
| 8446 | inst.operands[1].reg = REG_PC; |
| 8447 | |
| 8448 | msb = inst.operands[2].imm + inst.operands[3].imm; |
| 8449 | constraint (msb > 32, _("bit-field extends past end of register")); |
| 8450 | /* The instruction encoding stores the LSB and MSB, |
| 8451 | not the LSB and width. */ |
| 8452 | inst.instruction |= inst.operands[0].reg << 12; |
| 8453 | inst.instruction |= inst.operands[1].reg; |
| 8454 | inst.instruction |= inst.operands[2].imm << 7; |
| 8455 | inst.instruction |= (msb - 1) << 16; |
| 8456 | } |
| 8457 | |
| 8458 | static void |
| 8459 | do_bfx (void) |
| 8460 | { |
| 8461 | constraint (inst.operands[2].imm + inst.operands[3].imm > 32, |
| 8462 | _("bit-field extends past end of register")); |
| 8463 | inst.instruction |= inst.operands[0].reg << 12; |
| 8464 | inst.instruction |= inst.operands[1].reg; |
| 8465 | inst.instruction |= inst.operands[2].imm << 7; |
| 8466 | inst.instruction |= (inst.operands[3].imm - 1) << 16; |
| 8467 | } |
| 8468 | |
| 8469 | /* ARM V5 breakpoint instruction (argument parse) |
| 8470 | BKPT <16 bit unsigned immediate> |
| 8471 | Instruction is not conditional. |
| 8472 | The bit pattern given in insns[] has the COND_ALWAYS condition, |
| 8473 | and it is an error if the caller tried to override that. */ |
| 8474 | |
| 8475 | static void |
| 8476 | do_bkpt (void) |
| 8477 | { |
| 8478 | /* Top 12 of 16 bits to bits 19:8. */ |
| 8479 | inst.instruction |= (inst.operands[0].imm & 0xfff0) << 4; |
| 8480 | |
| 8481 | /* Bottom 4 of 16 bits to bits 3:0. */ |
| 8482 | inst.instruction |= inst.operands[0].imm & 0xf; |
| 8483 | } |
| 8484 | |
| 8485 | static void |
| 8486 | encode_branch (int default_reloc) |
| 8487 | { |
| 8488 | if (inst.operands[0].hasreloc) |
| 8489 | { |
| 8490 | constraint (inst.operands[0].imm != BFD_RELOC_ARM_PLT32 |
| 8491 | && inst.operands[0].imm != BFD_RELOC_ARM_TLS_CALL, |
| 8492 | _("the only valid suffixes here are '(plt)' and '(tlscall)'")); |
| 8493 | inst.reloc.type = inst.operands[0].imm == BFD_RELOC_ARM_PLT32 |
| 8494 | ? BFD_RELOC_ARM_PLT32 |
| 8495 | : thumb_mode ? BFD_RELOC_ARM_THM_TLS_CALL : BFD_RELOC_ARM_TLS_CALL; |
| 8496 | } |
| 8497 | else |
| 8498 | inst.reloc.type = (bfd_reloc_code_real_type) default_reloc; |
| 8499 | inst.reloc.pc_rel = 1; |
| 8500 | } |
| 8501 | |
| 8502 | static void |
| 8503 | do_branch (void) |
| 8504 | { |
| 8505 | #ifdef OBJ_ELF |
| 8506 | if (EF_ARM_EABI_VERSION (meabi_flags) >= EF_ARM_EABI_VER4) |
| 8507 | encode_branch (BFD_RELOC_ARM_PCREL_JUMP); |
| 8508 | else |
| 8509 | #endif |
| 8510 | encode_branch (BFD_RELOC_ARM_PCREL_BRANCH); |
| 8511 | } |
| 8512 | |
| 8513 | static void |
| 8514 | do_bl (void) |
| 8515 | { |
| 8516 | #ifdef OBJ_ELF |
| 8517 | if (EF_ARM_EABI_VERSION (meabi_flags) >= EF_ARM_EABI_VER4) |
| 8518 | { |
| 8519 | if (inst.cond == COND_ALWAYS) |
| 8520 | encode_branch (BFD_RELOC_ARM_PCREL_CALL); |
| 8521 | else |
| 8522 | encode_branch (BFD_RELOC_ARM_PCREL_JUMP); |
| 8523 | } |
| 8524 | else |
| 8525 | #endif |
| 8526 | encode_branch (BFD_RELOC_ARM_PCREL_BRANCH); |
| 8527 | } |
| 8528 | |
| 8529 | /* ARM V5 branch-link-exchange instruction (argument parse) |
| 8530 | BLX <target_addr> ie BLX(1) |
| 8531 | BLX{<condition>} <Rm> ie BLX(2) |
| 8532 | Unfortunately, there are two different opcodes for this mnemonic. |
| 8533 | So, the insns[].value is not used, and the code here zaps values |
| 8534 | into inst.instruction. |
| 8535 | Also, the <target_addr> can be 25 bits, hence has its own reloc. */ |
| 8536 | |
| 8537 | static void |
| 8538 | do_blx (void) |
| 8539 | { |
| 8540 | if (inst.operands[0].isreg) |
| 8541 | { |
| 8542 | /* Arg is a register; the opcode provided by insns[] is correct. |
| 8543 | It is not illegal to do "blx pc", just useless. */ |
| 8544 | if (inst.operands[0].reg == REG_PC) |
| 8545 | as_tsktsk (_("use of r15 in blx in ARM mode is not really useful")); |
| 8546 | |
| 8547 | inst.instruction |= inst.operands[0].reg; |
| 8548 | } |
| 8549 | else |
| 8550 | { |
| 8551 | /* Arg is an address; this instruction cannot be executed |
| 8552 | conditionally, and the opcode must be adjusted. |
| 8553 | We retain the BFD_RELOC_ARM_PCREL_BLX till the very end |
| 8554 | where we generate out a BFD_RELOC_ARM_PCREL_CALL instead. */ |
| 8555 | constraint (inst.cond != COND_ALWAYS, BAD_COND); |
| 8556 | inst.instruction = 0xfa000000; |
| 8557 | encode_branch (BFD_RELOC_ARM_PCREL_BLX); |
| 8558 | } |
| 8559 | } |
| 8560 | |
| 8561 | static void |
| 8562 | do_bx (void) |
| 8563 | { |
| 8564 | bfd_boolean want_reloc; |
| 8565 | |
| 8566 | if (inst.operands[0].reg == REG_PC) |
| 8567 | as_tsktsk (_("use of r15 in bx in ARM mode is not really useful")); |
| 8568 | |
| 8569 | inst.instruction |= inst.operands[0].reg; |
| 8570 | /* Output R_ARM_V4BX relocations if is an EABI object that looks like |
| 8571 | it is for ARMv4t or earlier. */ |
| 8572 | want_reloc = !ARM_CPU_HAS_FEATURE (selected_cpu, arm_ext_v5); |
| 8573 | if (object_arch && !ARM_CPU_HAS_FEATURE (*object_arch, arm_ext_v5)) |
| 8574 | want_reloc = TRUE; |
| 8575 | |
| 8576 | #ifdef OBJ_ELF |
| 8577 | if (EF_ARM_EABI_VERSION (meabi_flags) < EF_ARM_EABI_VER4) |
| 8578 | #endif |
| 8579 | want_reloc = FALSE; |
| 8580 | |
| 8581 | if (want_reloc) |
| 8582 | inst.reloc.type = BFD_RELOC_ARM_V4BX; |
| 8583 | } |
| 8584 | |
| 8585 | |
| 8586 | /* ARM v5TEJ. Jump to Jazelle code. */ |
| 8587 | |
| 8588 | static void |
| 8589 | do_bxj (void) |
| 8590 | { |
| 8591 | if (inst.operands[0].reg == REG_PC) |
| 8592 | as_tsktsk (_("use of r15 in bxj is not really useful")); |
| 8593 | |
| 8594 | inst.instruction |= inst.operands[0].reg; |
| 8595 | } |
| 8596 | |
| 8597 | /* Co-processor data operation: |
| 8598 | CDP{cond} <coproc>, <opcode_1>, <CRd>, <CRn>, <CRm>{, <opcode_2>} |
| 8599 | CDP2 <coproc>, <opcode_1>, <CRd>, <CRn>, <CRm>{, <opcode_2>} */ |
| 8600 | static void |
| 8601 | do_cdp (void) |
| 8602 | { |
| 8603 | inst.instruction |= inst.operands[0].reg << 8; |
| 8604 | inst.instruction |= inst.operands[1].imm << 20; |
| 8605 | inst.instruction |= inst.operands[2].reg << 12; |
| 8606 | inst.instruction |= inst.operands[3].reg << 16; |
| 8607 | inst.instruction |= inst.operands[4].reg; |
| 8608 | inst.instruction |= inst.operands[5].imm << 5; |
| 8609 | } |
| 8610 | |
| 8611 | static void |
| 8612 | do_cmp (void) |
| 8613 | { |
| 8614 | inst.instruction |= inst.operands[0].reg << 16; |
| 8615 | encode_arm_shifter_operand (1); |
| 8616 | } |
| 8617 | |
| 8618 | /* Transfer between coprocessor and ARM registers. |
| 8619 | MRC{cond} <coproc>, <opcode_1>, <Rd>, <CRn>, <CRm>{, <opcode_2>} |
| 8620 | MRC2 |
| 8621 | MCR{cond} |
| 8622 | MCR2 |
| 8623 | |
| 8624 | No special properties. */ |
| 8625 | |
| 8626 | struct deprecated_coproc_regs_s |
| 8627 | { |
| 8628 | unsigned cp; |
| 8629 | int opc1; |
| 8630 | unsigned crn; |
| 8631 | unsigned crm; |
| 8632 | int opc2; |
| 8633 | arm_feature_set deprecated; |
| 8634 | arm_feature_set obsoleted; |
| 8635 | const char *dep_msg; |
| 8636 | const char *obs_msg; |
| 8637 | }; |
| 8638 | |
| 8639 | #define DEPR_ACCESS_V8 \ |
| 8640 | N_("This coprocessor register access is deprecated in ARMv8") |
| 8641 | |
| 8642 | /* Table of all deprecated coprocessor registers. */ |
| 8643 | static struct deprecated_coproc_regs_s deprecated_coproc_regs[] = |
| 8644 | { |
| 8645 | {15, 0, 7, 10, 5, /* CP15DMB. */ |
| 8646 | ARM_FEATURE_CORE_LOW (ARM_EXT_V8), ARM_ARCH_NONE, |
| 8647 | DEPR_ACCESS_V8, NULL}, |
| 8648 | {15, 0, 7, 10, 4, /* CP15DSB. */ |
| 8649 | ARM_FEATURE_CORE_LOW (ARM_EXT_V8), ARM_ARCH_NONE, |
| 8650 | DEPR_ACCESS_V8, NULL}, |
| 8651 | {15, 0, 7, 5, 4, /* CP15ISB. */ |
| 8652 | ARM_FEATURE_CORE_LOW (ARM_EXT_V8), ARM_ARCH_NONE, |
| 8653 | DEPR_ACCESS_V8, NULL}, |
| 8654 | {14, 6, 1, 0, 0, /* TEEHBR. */ |
| 8655 | ARM_FEATURE_CORE_LOW (ARM_EXT_V8), ARM_ARCH_NONE, |
| 8656 | DEPR_ACCESS_V8, NULL}, |
| 8657 | {14, 6, 0, 0, 0, /* TEECR. */ |
| 8658 | ARM_FEATURE_CORE_LOW (ARM_EXT_V8), ARM_ARCH_NONE, |
| 8659 | DEPR_ACCESS_V8, NULL}, |
| 8660 | }; |
| 8661 | |
| 8662 | #undef DEPR_ACCESS_V8 |
| 8663 | |
| 8664 | static const size_t deprecated_coproc_reg_count = |
| 8665 | sizeof (deprecated_coproc_regs) / sizeof (deprecated_coproc_regs[0]); |
| 8666 | |
| 8667 | static void |
| 8668 | do_co_reg (void) |
| 8669 | { |
| 8670 | unsigned Rd; |
| 8671 | size_t i; |
| 8672 | |
| 8673 | Rd = inst.operands[2].reg; |
| 8674 | if (thumb_mode) |
| 8675 | { |
| 8676 | if (inst.instruction == 0xee000010 |
| 8677 | || inst.instruction == 0xfe000010) |
| 8678 | /* MCR, MCR2 */ |
| 8679 | reject_bad_reg (Rd); |
| 8680 | else if (!ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v8)) |
| 8681 | /* MRC, MRC2 */ |
| 8682 | constraint (Rd == REG_SP, BAD_SP); |
| 8683 | } |
| 8684 | else |
| 8685 | { |
| 8686 | /* MCR */ |
| 8687 | if (inst.instruction == 0xe000010) |
| 8688 | constraint (Rd == REG_PC, BAD_PC); |
| 8689 | } |
| 8690 | |
| 8691 | for (i = 0; i < deprecated_coproc_reg_count; ++i) |
| 8692 | { |
| 8693 | const struct deprecated_coproc_regs_s *r = |
| 8694 | deprecated_coproc_regs + i; |
| 8695 | |
| 8696 | if (inst.operands[0].reg == r->cp |
| 8697 | && inst.operands[1].imm == r->opc1 |
| 8698 | && inst.operands[3].reg == r->crn |
| 8699 | && inst.operands[4].reg == r->crm |
| 8700 | && inst.operands[5].imm == r->opc2) |
| 8701 | { |
| 8702 | if (! ARM_CPU_IS_ANY (cpu_variant) |
| 8703 | && warn_on_deprecated |
| 8704 | && ARM_CPU_HAS_FEATURE (cpu_variant, r->deprecated)) |
| 8705 | as_tsktsk ("%s", r->dep_msg); |
| 8706 | } |
| 8707 | } |
| 8708 | |
| 8709 | inst.instruction |= inst.operands[0].reg << 8; |
| 8710 | inst.instruction |= inst.operands[1].imm << 21; |
| 8711 | inst.instruction |= Rd << 12; |
| 8712 | inst.instruction |= inst.operands[3].reg << 16; |
| 8713 | inst.instruction |= inst.operands[4].reg; |
| 8714 | inst.instruction |= inst.operands[5].imm << 5; |
| 8715 | } |
| 8716 | |
| 8717 | /* Transfer between coprocessor register and pair of ARM registers. |
| 8718 | MCRR{cond} <coproc>, <opcode>, <Rd>, <Rn>, <CRm>. |
| 8719 | MCRR2 |
| 8720 | MRRC{cond} |
| 8721 | MRRC2 |
| 8722 | |
| 8723 | Two XScale instructions are special cases of these: |
| 8724 | |
| 8725 | MAR{cond} acc0, <RdLo>, <RdHi> == MCRR{cond} p0, #0, <RdLo>, <RdHi>, c0 |
| 8726 | MRA{cond} acc0, <RdLo>, <RdHi> == MRRC{cond} p0, #0, <RdLo>, <RdHi>, c0 |
| 8727 | |
| 8728 | Result unpredictable if Rd or Rn is R15. */ |
| 8729 | |
| 8730 | static void |
| 8731 | do_co_reg2c (void) |
| 8732 | { |
| 8733 | unsigned Rd, Rn; |
| 8734 | |
| 8735 | Rd = inst.operands[2].reg; |
| 8736 | Rn = inst.operands[3].reg; |
| 8737 | |
| 8738 | if (thumb_mode) |
| 8739 | { |
| 8740 | reject_bad_reg (Rd); |
| 8741 | reject_bad_reg (Rn); |
| 8742 | } |
| 8743 | else |
| 8744 | { |
| 8745 | constraint (Rd == REG_PC, BAD_PC); |
| 8746 | constraint (Rn == REG_PC, BAD_PC); |
| 8747 | } |
| 8748 | |
| 8749 | /* Only check the MRRC{2} variants. */ |
| 8750 | if ((inst.instruction & 0x0FF00000) == 0x0C500000) |
| 8751 | { |
| 8752 | /* If Rd == Rn, error that the operation is |
| 8753 | unpredictable (example MRRC p3,#1,r1,r1,c4). */ |
| 8754 | constraint (Rd == Rn, BAD_OVERLAP); |
| 8755 | } |
| 8756 | |
| 8757 | inst.instruction |= inst.operands[0].reg << 8; |
| 8758 | inst.instruction |= inst.operands[1].imm << 4; |
| 8759 | inst.instruction |= Rd << 12; |
| 8760 | inst.instruction |= Rn << 16; |
| 8761 | inst.instruction |= inst.operands[4].reg; |
| 8762 | } |
| 8763 | |
| 8764 | static void |
| 8765 | do_cpsi (void) |
| 8766 | { |
| 8767 | inst.instruction |= inst.operands[0].imm << 6; |
| 8768 | if (inst.operands[1].present) |
| 8769 | { |
| 8770 | inst.instruction |= CPSI_MMOD; |
| 8771 | inst.instruction |= inst.operands[1].imm; |
| 8772 | } |
| 8773 | } |
| 8774 | |
| 8775 | static void |
| 8776 | do_dbg (void) |
| 8777 | { |
| 8778 | inst.instruction |= inst.operands[0].imm; |
| 8779 | } |
| 8780 | |
| 8781 | static void |
| 8782 | do_div (void) |
| 8783 | { |
| 8784 | unsigned Rd, Rn, Rm; |
| 8785 | |
| 8786 | Rd = inst.operands[0].reg; |
| 8787 | Rn = (inst.operands[1].present |
| 8788 | ? inst.operands[1].reg : Rd); |
| 8789 | Rm = inst.operands[2].reg; |
| 8790 | |
| 8791 | constraint ((Rd == REG_PC), BAD_PC); |
| 8792 | constraint ((Rn == REG_PC), BAD_PC); |
| 8793 | constraint ((Rm == REG_PC), BAD_PC); |
| 8794 | |
| 8795 | inst.instruction |= Rd << 16; |
| 8796 | inst.instruction |= Rn << 0; |
| 8797 | inst.instruction |= Rm << 8; |
| 8798 | } |
| 8799 | |
| 8800 | static void |
| 8801 | do_it (void) |
| 8802 | { |
| 8803 | /* There is no IT instruction in ARM mode. We |
| 8804 | process it to do the validation as if in |
| 8805 | thumb mode, just in case the code gets |
| 8806 | assembled for thumb using the unified syntax. */ |
| 8807 | |
| 8808 | inst.size = 0; |
| 8809 | if (unified_syntax) |
| 8810 | { |
| 8811 | set_it_insn_type (IT_INSN); |
| 8812 | now_it.mask = (inst.instruction & 0xf) | 0x10; |
| 8813 | now_it.cc = inst.operands[0].imm; |
| 8814 | } |
| 8815 | } |
| 8816 | |
| 8817 | /* If there is only one register in the register list, |
| 8818 | then return its register number. Otherwise return -1. */ |
| 8819 | static int |
| 8820 | only_one_reg_in_list (int range) |
| 8821 | { |
| 8822 | int i = ffs (range) - 1; |
| 8823 | return (i > 15 || range != (1 << i)) ? -1 : i; |
| 8824 | } |
| 8825 | |
| 8826 | static void |
| 8827 | encode_ldmstm(int from_push_pop_mnem) |
| 8828 | { |
| 8829 | int base_reg = inst.operands[0].reg; |
| 8830 | int range = inst.operands[1].imm; |
| 8831 | int one_reg; |
| 8832 | |
| 8833 | inst.instruction |= base_reg << 16; |
| 8834 | inst.instruction |= range; |
| 8835 | |
| 8836 | if (inst.operands[1].writeback) |
| 8837 | inst.instruction |= LDM_TYPE_2_OR_3; |
| 8838 | |
| 8839 | if (inst.operands[0].writeback) |
| 8840 | { |
| 8841 | inst.instruction |= WRITE_BACK; |
| 8842 | /* Check for unpredictable uses of writeback. */ |
| 8843 | if (inst.instruction & LOAD_BIT) |
| 8844 | { |
| 8845 | /* Not allowed in LDM type 2. */ |
| 8846 | if ((inst.instruction & LDM_TYPE_2_OR_3) |
| 8847 | && ((range & (1 << REG_PC)) == 0)) |
| 8848 | as_warn (_("writeback of base register is UNPREDICTABLE")); |
| 8849 | /* Only allowed if base reg not in list for other types. */ |
| 8850 | else if (range & (1 << base_reg)) |
| 8851 | as_warn (_("writeback of base register when in register list is UNPREDICTABLE")); |
| 8852 | } |
| 8853 | else /* STM. */ |
| 8854 | { |
| 8855 | /* Not allowed for type 2. */ |
| 8856 | if (inst.instruction & LDM_TYPE_2_OR_3) |
| 8857 | as_warn (_("writeback of base register is UNPREDICTABLE")); |
| 8858 | /* Only allowed if base reg not in list, or first in list. */ |
| 8859 | else if ((range & (1 << base_reg)) |
| 8860 | && (range & ((1 << base_reg) - 1))) |
| 8861 | as_warn (_("if writeback register is in list, it must be the lowest reg in the list")); |
| 8862 | } |
| 8863 | } |
| 8864 | |
| 8865 | /* If PUSH/POP has only one register, then use the A2 encoding. */ |
| 8866 | one_reg = only_one_reg_in_list (range); |
| 8867 | if (from_push_pop_mnem && one_reg >= 0) |
| 8868 | { |
| 8869 | int is_push = (inst.instruction & A_PUSH_POP_OP_MASK) == A1_OPCODE_PUSH; |
| 8870 | |
| 8871 | inst.instruction &= A_COND_MASK; |
| 8872 | inst.instruction |= is_push ? A2_OPCODE_PUSH : A2_OPCODE_POP; |
| 8873 | inst.instruction |= one_reg << 12; |
| 8874 | } |
| 8875 | } |
| 8876 | |
| 8877 | static void |
| 8878 | do_ldmstm (void) |
| 8879 | { |
| 8880 | encode_ldmstm (/*from_push_pop_mnem=*/FALSE); |
| 8881 | } |
| 8882 | |
| 8883 | /* ARMv5TE load-consecutive (argument parse) |
| 8884 | Mode is like LDRH. |
| 8885 | |
| 8886 | LDRccD R, mode |
| 8887 | STRccD R, mode. */ |
| 8888 | |
| 8889 | static void |
| 8890 | do_ldrd (void) |
| 8891 | { |
| 8892 | constraint (inst.operands[0].reg % 2 != 0, |
| 8893 | _("first transfer register must be even")); |
| 8894 | constraint (inst.operands[1].present |
| 8895 | && inst.operands[1].reg != inst.operands[0].reg + 1, |
| 8896 | _("can only transfer two consecutive registers")); |
| 8897 | constraint (inst.operands[0].reg == REG_LR, _("r14 not allowed here")); |
| 8898 | constraint (!inst.operands[2].isreg, _("'[' expected")); |
| 8899 | |
| 8900 | if (!inst.operands[1].present) |
| 8901 | inst.operands[1].reg = inst.operands[0].reg + 1; |
| 8902 | |
| 8903 | /* encode_arm_addr_mode_3 will diagnose overlap between the base |
| 8904 | register and the first register written; we have to diagnose |
| 8905 | overlap between the base and the second register written here. */ |
| 8906 | |
| 8907 | if (inst.operands[2].reg == inst.operands[1].reg |
| 8908 | && (inst.operands[2].writeback || inst.operands[2].postind)) |
| 8909 | as_warn (_("base register written back, and overlaps " |
| 8910 | "second transfer register")); |
| 8911 | |
| 8912 | if (!(inst.instruction & V4_STR_BIT)) |
| 8913 | { |
| 8914 | /* For an index-register load, the index register must not overlap the |
| 8915 | destination (even if not write-back). */ |
| 8916 | if (inst.operands[2].immisreg |
| 8917 | && ((unsigned) inst.operands[2].imm == inst.operands[0].reg |
| 8918 | || (unsigned) inst.operands[2].imm == inst.operands[1].reg)) |
| 8919 | as_warn (_("index register overlaps transfer register")); |
| 8920 | } |
| 8921 | inst.instruction |= inst.operands[0].reg << 12; |
| 8922 | encode_arm_addr_mode_3 (2, /*is_t=*/FALSE); |
| 8923 | } |
| 8924 | |
| 8925 | static void |
| 8926 | do_ldrex (void) |
| 8927 | { |
| 8928 | constraint (!inst.operands[1].isreg || !inst.operands[1].preind |
| 8929 | || inst.operands[1].postind || inst.operands[1].writeback |
| 8930 | || inst.operands[1].immisreg || inst.operands[1].shifted |
| 8931 | || inst.operands[1].negative |
| 8932 | /* This can arise if the programmer has written |
| 8933 | strex rN, rM, foo |
| 8934 | or if they have mistakenly used a register name as the last |
| 8935 | operand, eg: |
| 8936 | strex rN, rM, rX |
| 8937 | It is very difficult to distinguish between these two cases |
| 8938 | because "rX" might actually be a label. ie the register |
| 8939 | name has been occluded by a symbol of the same name. So we |
| 8940 | just generate a general 'bad addressing mode' type error |
| 8941 | message and leave it up to the programmer to discover the |
| 8942 | true cause and fix their mistake. */ |
| 8943 | || (inst.operands[1].reg == REG_PC), |
| 8944 | BAD_ADDR_MODE); |
| 8945 | |
| 8946 | constraint (inst.reloc.exp.X_op != O_constant |
| 8947 | || inst.reloc.exp.X_add_number != 0, |
| 8948 | _("offset must be zero in ARM encoding")); |
| 8949 | |
| 8950 | constraint ((inst.operands[1].reg == REG_PC), BAD_PC); |
| 8951 | |
| 8952 | inst.instruction |= inst.operands[0].reg << 12; |
| 8953 | inst.instruction |= inst.operands[1].reg << 16; |
| 8954 | inst.reloc.type = BFD_RELOC_UNUSED; |
| 8955 | } |
| 8956 | |
| 8957 | static void |
| 8958 | do_ldrexd (void) |
| 8959 | { |
| 8960 | constraint (inst.operands[0].reg % 2 != 0, |
| 8961 | _("even register required")); |
| 8962 | constraint (inst.operands[1].present |
| 8963 | && inst.operands[1].reg != inst.operands[0].reg + 1, |
| 8964 | _("can only load two consecutive registers")); |
| 8965 | /* If op 1 were present and equal to PC, this function wouldn't |
| 8966 | have been called in the first place. */ |
| 8967 | constraint (inst.operands[0].reg == REG_LR, _("r14 not allowed here")); |
| 8968 | |
| 8969 | inst.instruction |= inst.operands[0].reg << 12; |
| 8970 | inst.instruction |= inst.operands[2].reg << 16; |
| 8971 | } |
| 8972 | |
| 8973 | /* In both ARM and thumb state 'ldr pc, #imm' with an immediate |
| 8974 | which is not a multiple of four is UNPREDICTABLE. */ |
| 8975 | static void |
| 8976 | check_ldr_r15_aligned (void) |
| 8977 | { |
| 8978 | constraint (!(inst.operands[1].immisreg) |
| 8979 | && (inst.operands[0].reg == REG_PC |
| 8980 | && inst.operands[1].reg == REG_PC |
| 8981 | && (inst.reloc.exp.X_add_number & 0x3)), |
| 8982 | _("ldr to register 15 must be 4-byte aligned")); |
| 8983 | } |
| 8984 | |
| 8985 | static void |
| 8986 | do_ldst (void) |
| 8987 | { |
| 8988 | inst.instruction |= inst.operands[0].reg << 12; |
| 8989 | if (!inst.operands[1].isreg) |
| 8990 | if (move_or_literal_pool (0, CONST_ARM, /*mode_3=*/FALSE)) |
| 8991 | return; |
| 8992 | encode_arm_addr_mode_2 (1, /*is_t=*/FALSE); |
| 8993 | check_ldr_r15_aligned (); |
| 8994 | } |
| 8995 | |
| 8996 | static void |
| 8997 | do_ldstt (void) |
| 8998 | { |
| 8999 | /* ldrt/strt always use post-indexed addressing. Turn [Rn] into [Rn]! and |
| 9000 | reject [Rn,...]. */ |
| 9001 | if (inst.operands[1].preind) |
| 9002 | { |
| 9003 | constraint (inst.reloc.exp.X_op != O_constant |
| 9004 | || inst.reloc.exp.X_add_number != 0, |
| 9005 | _("this instruction requires a post-indexed address")); |
| 9006 | |
| 9007 | inst.operands[1].preind = 0; |
| 9008 | inst.operands[1].postind = 1; |
| 9009 | inst.operands[1].writeback = 1; |
| 9010 | } |
| 9011 | inst.instruction |= inst.operands[0].reg << 12; |
| 9012 | encode_arm_addr_mode_2 (1, /*is_t=*/TRUE); |
| 9013 | } |
| 9014 | |
| 9015 | /* Halfword and signed-byte load/store operations. */ |
| 9016 | |
| 9017 | static void |
| 9018 | do_ldstv4 (void) |
| 9019 | { |
| 9020 | constraint (inst.operands[0].reg == REG_PC, BAD_PC); |
| 9021 | inst.instruction |= inst.operands[0].reg << 12; |
| 9022 | if (!inst.operands[1].isreg) |
| 9023 | if (move_or_literal_pool (0, CONST_ARM, /*mode_3=*/TRUE)) |
| 9024 | return; |
| 9025 | encode_arm_addr_mode_3 (1, /*is_t=*/FALSE); |
| 9026 | } |
| 9027 | |
| 9028 | static void |
| 9029 | do_ldsttv4 (void) |
| 9030 | { |
| 9031 | /* ldrt/strt always use post-indexed addressing. Turn [Rn] into [Rn]! and |
| 9032 | reject [Rn,...]. */ |
| 9033 | if (inst.operands[1].preind) |
| 9034 | { |
| 9035 | constraint (inst.reloc.exp.X_op != O_constant |
| 9036 | || inst.reloc.exp.X_add_number != 0, |
| 9037 | _("this instruction requires a post-indexed address")); |
| 9038 | |
| 9039 | inst.operands[1].preind = 0; |
| 9040 | inst.operands[1].postind = 1; |
| 9041 | inst.operands[1].writeback = 1; |
| 9042 | } |
| 9043 | inst.instruction |= inst.operands[0].reg << 12; |
| 9044 | encode_arm_addr_mode_3 (1, /*is_t=*/TRUE); |
| 9045 | } |
| 9046 | |
| 9047 | /* Co-processor register load/store. |
| 9048 | Format: <LDC|STC>{cond}[L] CP#,CRd,<address> */ |
| 9049 | static void |
| 9050 | do_lstc (void) |
| 9051 | { |
| 9052 | inst.instruction |= inst.operands[0].reg << 8; |
| 9053 | inst.instruction |= inst.operands[1].reg << 12; |
| 9054 | encode_arm_cp_address (2, TRUE, TRUE, 0); |
| 9055 | } |
| 9056 | |
| 9057 | static void |
| 9058 | do_mlas (void) |
| 9059 | { |
| 9060 | /* This restriction does not apply to mls (nor to mla in v6 or later). */ |
| 9061 | if (inst.operands[0].reg == inst.operands[1].reg |
| 9062 | && !ARM_CPU_HAS_FEATURE (selected_cpu, arm_ext_v6) |
| 9063 | && !(inst.instruction & 0x00400000)) |
| 9064 | as_tsktsk (_("Rd and Rm should be different in mla")); |
| 9065 | |
| 9066 | inst.instruction |= inst.operands[0].reg << 16; |
| 9067 | inst.instruction |= inst.operands[1].reg; |
| 9068 | inst.instruction |= inst.operands[2].reg << 8; |
| 9069 | inst.instruction |= inst.operands[3].reg << 12; |
| 9070 | } |
| 9071 | |
| 9072 | static void |
| 9073 | do_mov (void) |
| 9074 | { |
| 9075 | constraint (inst.reloc.type >= BFD_RELOC_ARM_THUMB_ALU_ABS_G0_NC |
| 9076 | && inst.reloc.type <= BFD_RELOC_ARM_THUMB_ALU_ABS_G3_NC , |
| 9077 | THUMB1_RELOC_ONLY); |
| 9078 | inst.instruction |= inst.operands[0].reg << 12; |
| 9079 | encode_arm_shifter_operand (1); |
| 9080 | } |
| 9081 | |
| 9082 | /* ARM V6T2 16-bit immediate register load: MOV[WT]{cond} Rd, #<imm16>. */ |
| 9083 | static void |
| 9084 | do_mov16 (void) |
| 9085 | { |
| 9086 | bfd_vma imm; |
| 9087 | bfd_boolean top; |
| 9088 | |
| 9089 | top = (inst.instruction & 0x00400000) != 0; |
| 9090 | constraint (top && inst.reloc.type == BFD_RELOC_ARM_MOVW, |
| 9091 | _(":lower16: not allowed in this instruction")); |
| 9092 | constraint (!top && inst.reloc.type == BFD_RELOC_ARM_MOVT, |
| 9093 | _(":upper16: not allowed in this instruction")); |
| 9094 | inst.instruction |= inst.operands[0].reg << 12; |
| 9095 | if (inst.reloc.type == BFD_RELOC_UNUSED) |
| 9096 | { |
| 9097 | imm = inst.reloc.exp.X_add_number; |
| 9098 | /* The value is in two pieces: 0:11, 16:19. */ |
| 9099 | inst.instruction |= (imm & 0x00000fff); |
| 9100 | inst.instruction |= (imm & 0x0000f000) << 4; |
| 9101 | } |
| 9102 | } |
| 9103 | |
| 9104 | static int |
| 9105 | do_vfp_nsyn_mrs (void) |
| 9106 | { |
| 9107 | if (inst.operands[0].isvec) |
| 9108 | { |
| 9109 | if (inst.operands[1].reg != 1) |
| 9110 | first_error (_("operand 1 must be FPSCR")); |
| 9111 | memset (&inst.operands[0], '\0', sizeof (inst.operands[0])); |
| 9112 | memset (&inst.operands[1], '\0', sizeof (inst.operands[1])); |
| 9113 | do_vfp_nsyn_opcode ("fmstat"); |
| 9114 | } |
| 9115 | else if (inst.operands[1].isvec) |
| 9116 | do_vfp_nsyn_opcode ("fmrx"); |
| 9117 | else |
| 9118 | return FAIL; |
| 9119 | |
| 9120 | return SUCCESS; |
| 9121 | } |
| 9122 | |
| 9123 | static int |
| 9124 | do_vfp_nsyn_msr (void) |
| 9125 | { |
| 9126 | if (inst.operands[0].isvec) |
| 9127 | do_vfp_nsyn_opcode ("fmxr"); |
| 9128 | else |
| 9129 | return FAIL; |
| 9130 | |
| 9131 | return SUCCESS; |
| 9132 | } |
| 9133 | |
| 9134 | static void |
| 9135 | do_vmrs (void) |
| 9136 | { |
| 9137 | unsigned Rt = inst.operands[0].reg; |
| 9138 | |
| 9139 | if (thumb_mode && Rt == REG_SP) |
| 9140 | { |
| 9141 | inst.error = BAD_SP; |
| 9142 | return; |
| 9143 | } |
| 9144 | |
| 9145 | /* MVFR2 is only valid at ARMv8-A. */ |
| 9146 | if (inst.operands[1].reg == 5) |
| 9147 | constraint (!ARM_CPU_HAS_FEATURE (cpu_variant, fpu_vfp_ext_armv8), |
| 9148 | _(BAD_FPU)); |
| 9149 | |
| 9150 | /* APSR_ sets isvec. All other refs to PC are illegal. */ |
| 9151 | if (!inst.operands[0].isvec && Rt == REG_PC) |
| 9152 | { |
| 9153 | inst.error = BAD_PC; |
| 9154 | return; |
| 9155 | } |
| 9156 | |
| 9157 | /* If we get through parsing the register name, we just insert the number |
| 9158 | generated into the instruction without further validation. */ |
| 9159 | inst.instruction |= (inst.operands[1].reg << 16); |
| 9160 | inst.instruction |= (Rt << 12); |
| 9161 | } |
| 9162 | |
| 9163 | static void |
| 9164 | do_vmsr (void) |
| 9165 | { |
| 9166 | unsigned Rt = inst.operands[1].reg; |
| 9167 | |
| 9168 | if (thumb_mode) |
| 9169 | reject_bad_reg (Rt); |
| 9170 | else if (Rt == REG_PC) |
| 9171 | { |
| 9172 | inst.error = BAD_PC; |
| 9173 | return; |
| 9174 | } |
| 9175 | |
| 9176 | /* MVFR2 is only valid for ARMv8-A. */ |
| 9177 | if (inst.operands[0].reg == 5) |
| 9178 | constraint (!ARM_CPU_HAS_FEATURE (cpu_variant, fpu_vfp_ext_armv8), |
| 9179 | _(BAD_FPU)); |
| 9180 | |
| 9181 | /* If we get through parsing the register name, we just insert the number |
| 9182 | generated into the instruction without further validation. */ |
| 9183 | inst.instruction |= (inst.operands[0].reg << 16); |
| 9184 | inst.instruction |= (Rt << 12); |
| 9185 | } |
| 9186 | |
| 9187 | static void |
| 9188 | do_mrs (void) |
| 9189 | { |
| 9190 | unsigned br; |
| 9191 | |
| 9192 | if (do_vfp_nsyn_mrs () == SUCCESS) |
| 9193 | return; |
| 9194 | |
| 9195 | constraint (inst.operands[0].reg == REG_PC, BAD_PC); |
| 9196 | inst.instruction |= inst.operands[0].reg << 12; |
| 9197 | |
| 9198 | if (inst.operands[1].isreg) |
| 9199 | { |
| 9200 | br = inst.operands[1].reg; |
| 9201 | if (((br & 0x200) == 0) && ((br & 0xf0000) != 0xf000)) |
| 9202 | as_bad (_("bad register for mrs")); |
| 9203 | } |
| 9204 | else |
| 9205 | { |
| 9206 | /* mrs only accepts CPSR/SPSR/CPSR_all/SPSR_all. */ |
| 9207 | constraint ((inst.operands[1].imm & (PSR_c|PSR_x|PSR_s|PSR_f)) |
| 9208 | != (PSR_c|PSR_f), |
| 9209 | _("'APSR', 'CPSR' or 'SPSR' expected")); |
| 9210 | br = (15<<16) | (inst.operands[1].imm & SPSR_BIT); |
| 9211 | } |
| 9212 | |
| 9213 | inst.instruction |= br; |
| 9214 | } |
| 9215 | |
| 9216 | /* Two possible forms: |
| 9217 | "{C|S}PSR_<field>, Rm", |
| 9218 | "{C|S}PSR_f, #expression". */ |
| 9219 | |
| 9220 | static void |
| 9221 | do_msr (void) |
| 9222 | { |
| 9223 | if (do_vfp_nsyn_msr () == SUCCESS) |
| 9224 | return; |
| 9225 | |
| 9226 | inst.instruction |= inst.operands[0].imm; |
| 9227 | if (inst.operands[1].isreg) |
| 9228 | inst.instruction |= inst.operands[1].reg; |
| 9229 | else |
| 9230 | { |
| 9231 | inst.instruction |= INST_IMMEDIATE; |
| 9232 | inst.reloc.type = BFD_RELOC_ARM_IMMEDIATE; |
| 9233 | inst.reloc.pc_rel = 0; |
| 9234 | } |
| 9235 | } |
| 9236 | |
| 9237 | static void |
| 9238 | do_mul (void) |
| 9239 | { |
| 9240 | constraint (inst.operands[2].reg == REG_PC, BAD_PC); |
| 9241 | |
| 9242 | if (!inst.operands[2].present) |
| 9243 | inst.operands[2].reg = inst.operands[0].reg; |
| 9244 | inst.instruction |= inst.operands[0].reg << 16; |
| 9245 | inst.instruction |= inst.operands[1].reg; |
| 9246 | inst.instruction |= inst.operands[2].reg << 8; |
| 9247 | |
| 9248 | if (inst.operands[0].reg == inst.operands[1].reg |
| 9249 | && !ARM_CPU_HAS_FEATURE (selected_cpu, arm_ext_v6)) |
| 9250 | as_tsktsk (_("Rd and Rm should be different in mul")); |
| 9251 | } |
| 9252 | |
| 9253 | /* Long Multiply Parser |
| 9254 | UMULL RdLo, RdHi, Rm, Rs |
| 9255 | SMULL RdLo, RdHi, Rm, Rs |
| 9256 | UMLAL RdLo, RdHi, Rm, Rs |
| 9257 | SMLAL RdLo, RdHi, Rm, Rs. */ |
| 9258 | |
| 9259 | static void |
| 9260 | do_mull (void) |
| 9261 | { |
| 9262 | inst.instruction |= inst.operands[0].reg << 12; |
| 9263 | inst.instruction |= inst.operands[1].reg << 16; |
| 9264 | inst.instruction |= inst.operands[2].reg; |
| 9265 | inst.instruction |= inst.operands[3].reg << 8; |
| 9266 | |
| 9267 | /* rdhi and rdlo must be different. */ |
| 9268 | if (inst.operands[0].reg == inst.operands[1].reg) |
| 9269 | as_tsktsk (_("rdhi and rdlo must be different")); |
| 9270 | |
| 9271 | /* rdhi, rdlo and rm must all be different before armv6. */ |
| 9272 | if ((inst.operands[0].reg == inst.operands[2].reg |
| 9273 | || inst.operands[1].reg == inst.operands[2].reg) |
| 9274 | && !ARM_CPU_HAS_FEATURE (selected_cpu, arm_ext_v6)) |
| 9275 | as_tsktsk (_("rdhi, rdlo and rm must all be different")); |
| 9276 | } |
| 9277 | |
| 9278 | static void |
| 9279 | do_nop (void) |
| 9280 | { |
| 9281 | if (inst.operands[0].present |
| 9282 | || ARM_CPU_HAS_FEATURE (selected_cpu, arm_ext_v6k)) |
| 9283 | { |
| 9284 | /* Architectural NOP hints are CPSR sets with no bits selected. */ |
| 9285 | inst.instruction &= 0xf0000000; |
| 9286 | inst.instruction |= 0x0320f000; |
| 9287 | if (inst.operands[0].present) |
| 9288 | inst.instruction |= inst.operands[0].imm; |
| 9289 | } |
| 9290 | } |
| 9291 | |
| 9292 | /* ARM V6 Pack Halfword Bottom Top instruction (argument parse). |
| 9293 | PKHBT {<cond>} <Rd>, <Rn>, <Rm> {, LSL #<shift_imm>} |
| 9294 | Condition defaults to COND_ALWAYS. |
| 9295 | Error if Rd, Rn or Rm are R15. */ |
| 9296 | |
| 9297 | static void |
| 9298 | do_pkhbt (void) |
| 9299 | { |
| 9300 | inst.instruction |= inst.operands[0].reg << 12; |
| 9301 | inst.instruction |= inst.operands[1].reg << 16; |
| 9302 | inst.instruction |= inst.operands[2].reg; |
| 9303 | if (inst.operands[3].present) |
| 9304 | encode_arm_shift (3); |
| 9305 | } |
| 9306 | |
| 9307 | /* ARM V6 PKHTB (Argument Parse). */ |
| 9308 | |
| 9309 | static void |
| 9310 | do_pkhtb (void) |
| 9311 | { |
| 9312 | if (!inst.operands[3].present) |
| 9313 | { |
| 9314 | /* If the shift specifier is omitted, turn the instruction |
| 9315 | into pkhbt rd, rm, rn. */ |
| 9316 | inst.instruction &= 0xfff00010; |
| 9317 | inst.instruction |= inst.operands[0].reg << 12; |
| 9318 | inst.instruction |= inst.operands[1].reg; |
| 9319 | inst.instruction |= inst.operands[2].reg << 16; |
| 9320 | } |
| 9321 | else |
| 9322 | { |
| 9323 | inst.instruction |= inst.operands[0].reg << 12; |
| 9324 | inst.instruction |= inst.operands[1].reg << 16; |
| 9325 | inst.instruction |= inst.operands[2].reg; |
| 9326 | encode_arm_shift (3); |
| 9327 | } |
| 9328 | } |
| 9329 | |
| 9330 | /* ARMv5TE: Preload-Cache |
| 9331 | MP Extensions: Preload for write |
| 9332 | |
| 9333 | PLD(W) <addr_mode> |
| 9334 | |
| 9335 | Syntactically, like LDR with B=1, W=0, L=1. */ |
| 9336 | |
| 9337 | static void |
| 9338 | do_pld (void) |
| 9339 | { |
| 9340 | constraint (!inst.operands[0].isreg, |
| 9341 | _("'[' expected after PLD mnemonic")); |
| 9342 | constraint (inst.operands[0].postind, |
| 9343 | _("post-indexed expression used in preload instruction")); |
| 9344 | constraint (inst.operands[0].writeback, |
| 9345 | _("writeback used in preload instruction")); |
| 9346 | constraint (!inst.operands[0].preind, |
| 9347 | _("unindexed addressing used in preload instruction")); |
| 9348 | encode_arm_addr_mode_2 (0, /*is_t=*/FALSE); |
| 9349 | } |
| 9350 | |
| 9351 | /* ARMv7: PLI <addr_mode> */ |
| 9352 | static void |
| 9353 | do_pli (void) |
| 9354 | { |
| 9355 | constraint (!inst.operands[0].isreg, |
| 9356 | _("'[' expected after PLI mnemonic")); |
| 9357 | constraint (inst.operands[0].postind, |
| 9358 | _("post-indexed expression used in preload instruction")); |
| 9359 | constraint (inst.operands[0].writeback, |
| 9360 | _("writeback used in preload instruction")); |
| 9361 | constraint (!inst.operands[0].preind, |
| 9362 | _("unindexed addressing used in preload instruction")); |
| 9363 | encode_arm_addr_mode_2 (0, /*is_t=*/FALSE); |
| 9364 | inst.instruction &= ~PRE_INDEX; |
| 9365 | } |
| 9366 | |
| 9367 | static void |
| 9368 | do_push_pop (void) |
| 9369 | { |
| 9370 | constraint (inst.operands[0].writeback, |
| 9371 | _("push/pop do not support {reglist}^")); |
| 9372 | inst.operands[1] = inst.operands[0]; |
| 9373 | memset (&inst.operands[0], 0, sizeof inst.operands[0]); |
| 9374 | inst.operands[0].isreg = 1; |
| 9375 | inst.operands[0].writeback = 1; |
| 9376 | inst.operands[0].reg = REG_SP; |
| 9377 | encode_ldmstm (/*from_push_pop_mnem=*/TRUE); |
| 9378 | } |
| 9379 | |
| 9380 | /* ARM V6 RFE (Return from Exception) loads the PC and CPSR from the |
| 9381 | word at the specified address and the following word |
| 9382 | respectively. |
| 9383 | Unconditionally executed. |
| 9384 | Error if Rn is R15. */ |
| 9385 | |
| 9386 | static void |
| 9387 | do_rfe (void) |
| 9388 | { |
| 9389 | inst.instruction |= inst.operands[0].reg << 16; |
| 9390 | if (inst.operands[0].writeback) |
| 9391 | inst.instruction |= WRITE_BACK; |
| 9392 | } |
| 9393 | |
| 9394 | /* ARM V6 ssat (argument parse). */ |
| 9395 | |
| 9396 | static void |
| 9397 | do_ssat (void) |
| 9398 | { |
| 9399 | inst.instruction |= inst.operands[0].reg << 12; |
| 9400 | inst.instruction |= (inst.operands[1].imm - 1) << 16; |
| 9401 | inst.instruction |= inst.operands[2].reg; |
| 9402 | |
| 9403 | if (inst.operands[3].present) |
| 9404 | encode_arm_shift (3); |
| 9405 | } |
| 9406 | |
| 9407 | /* ARM V6 usat (argument parse). */ |
| 9408 | |
| 9409 | static void |
| 9410 | do_usat (void) |
| 9411 | { |
| 9412 | inst.instruction |= inst.operands[0].reg << 12; |
| 9413 | inst.instruction |= inst.operands[1].imm << 16; |
| 9414 | inst.instruction |= inst.operands[2].reg; |
| 9415 | |
| 9416 | if (inst.operands[3].present) |
| 9417 | encode_arm_shift (3); |
| 9418 | } |
| 9419 | |
| 9420 | /* ARM V6 ssat16 (argument parse). */ |
| 9421 | |
| 9422 | static void |
| 9423 | do_ssat16 (void) |
| 9424 | { |
| 9425 | inst.instruction |= inst.operands[0].reg << 12; |
| 9426 | inst.instruction |= ((inst.operands[1].imm - 1) << 16); |
| 9427 | inst.instruction |= inst.operands[2].reg; |
| 9428 | } |
| 9429 | |
| 9430 | static void |
| 9431 | do_usat16 (void) |
| 9432 | { |
| 9433 | inst.instruction |= inst.operands[0].reg << 12; |
| 9434 | inst.instruction |= inst.operands[1].imm << 16; |
| 9435 | inst.instruction |= inst.operands[2].reg; |
| 9436 | } |
| 9437 | |
| 9438 | /* ARM V6 SETEND (argument parse). Sets the E bit in the CPSR while |
| 9439 | preserving the other bits. |
| 9440 | |
| 9441 | setend <endian_specifier>, where <endian_specifier> is either |
| 9442 | BE or LE. */ |
| 9443 | |
| 9444 | static void |
| 9445 | do_setend (void) |
| 9446 | { |
| 9447 | if (warn_on_deprecated |
| 9448 | && ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v8)) |
| 9449 | as_tsktsk (_("setend use is deprecated for ARMv8")); |
| 9450 | |
| 9451 | if (inst.operands[0].imm) |
| 9452 | inst.instruction |= 0x200; |
| 9453 | } |
| 9454 | |
| 9455 | static void |
| 9456 | do_shift (void) |
| 9457 | { |
| 9458 | unsigned int Rm = (inst.operands[1].present |
| 9459 | ? inst.operands[1].reg |
| 9460 | : inst.operands[0].reg); |
| 9461 | |
| 9462 | inst.instruction |= inst.operands[0].reg << 12; |
| 9463 | inst.instruction |= Rm; |
| 9464 | if (inst.operands[2].isreg) /* Rd, {Rm,} Rs */ |
| 9465 | { |
| 9466 | inst.instruction |= inst.operands[2].reg << 8; |
| 9467 | inst.instruction |= SHIFT_BY_REG; |
| 9468 | /* PR 12854: Error on extraneous shifts. */ |
| 9469 | constraint (inst.operands[2].shifted, |
| 9470 | _("extraneous shift as part of operand to shift insn")); |
| 9471 | } |
| 9472 | else |
| 9473 | inst.reloc.type = BFD_RELOC_ARM_SHIFT_IMM; |
| 9474 | } |
| 9475 | |
| 9476 | static void |
| 9477 | do_smc (void) |
| 9478 | { |
| 9479 | inst.reloc.type = BFD_RELOC_ARM_SMC; |
| 9480 | inst.reloc.pc_rel = 0; |
| 9481 | } |
| 9482 | |
| 9483 | static void |
| 9484 | do_hvc (void) |
| 9485 | { |
| 9486 | inst.reloc.type = BFD_RELOC_ARM_HVC; |
| 9487 | inst.reloc.pc_rel = 0; |
| 9488 | } |
| 9489 | |
| 9490 | static void |
| 9491 | do_swi (void) |
| 9492 | { |
| 9493 | inst.reloc.type = BFD_RELOC_ARM_SWI; |
| 9494 | inst.reloc.pc_rel = 0; |
| 9495 | } |
| 9496 | |
| 9497 | static void |
| 9498 | do_setpan (void) |
| 9499 | { |
| 9500 | constraint (!ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_pan), |
| 9501 | _("selected processor does not support SETPAN instruction")); |
| 9502 | |
| 9503 | inst.instruction |= ((inst.operands[0].imm & 1) << 9); |
| 9504 | } |
| 9505 | |
| 9506 | static void |
| 9507 | do_t_setpan (void) |
| 9508 | { |
| 9509 | constraint (!ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_pan), |
| 9510 | _("selected processor does not support SETPAN instruction")); |
| 9511 | |
| 9512 | inst.instruction |= (inst.operands[0].imm << 3); |
| 9513 | } |
| 9514 | |
| 9515 | /* ARM V5E (El Segundo) signed-multiply-accumulate (argument parse) |
| 9516 | SMLAxy{cond} Rd,Rm,Rs,Rn |
| 9517 | SMLAWy{cond} Rd,Rm,Rs,Rn |
| 9518 | Error if any register is R15. */ |
| 9519 | |
| 9520 | static void |
| 9521 | do_smla (void) |
| 9522 | { |
| 9523 | inst.instruction |= inst.operands[0].reg << 16; |
| 9524 | inst.instruction |= inst.operands[1].reg; |
| 9525 | inst.instruction |= inst.operands[2].reg << 8; |
| 9526 | inst.instruction |= inst.operands[3].reg << 12; |
| 9527 | } |
| 9528 | |
| 9529 | /* ARM V5E (El Segundo) signed-multiply-accumulate-long (argument parse) |
| 9530 | SMLALxy{cond} Rdlo,Rdhi,Rm,Rs |
| 9531 | Error if any register is R15. |
| 9532 | Warning if Rdlo == Rdhi. */ |
| 9533 | |
| 9534 | static void |
| 9535 | do_smlal (void) |
| 9536 | { |
| 9537 | inst.instruction |= inst.operands[0].reg << 12; |
| 9538 | inst.instruction |= inst.operands[1].reg << 16; |
| 9539 | inst.instruction |= inst.operands[2].reg; |
| 9540 | inst.instruction |= inst.operands[3].reg << 8; |
| 9541 | |
| 9542 | if (inst.operands[0].reg == inst.operands[1].reg) |
| 9543 | as_tsktsk (_("rdhi and rdlo must be different")); |
| 9544 | } |
| 9545 | |
| 9546 | /* ARM V5E (El Segundo) signed-multiply (argument parse) |
| 9547 | SMULxy{cond} Rd,Rm,Rs |
| 9548 | Error if any register is R15. */ |
| 9549 | |
| 9550 | static void |
| 9551 | do_smul (void) |
| 9552 | { |
| 9553 | inst.instruction |= inst.operands[0].reg << 16; |
| 9554 | inst.instruction |= inst.operands[1].reg; |
| 9555 | inst.instruction |= inst.operands[2].reg << 8; |
| 9556 | } |
| 9557 | |
| 9558 | /* ARM V6 srs (argument parse). The variable fields in the encoding are |
| 9559 | the same for both ARM and Thumb-2. */ |
| 9560 | |
| 9561 | static void |
| 9562 | do_srs (void) |
| 9563 | { |
| 9564 | int reg; |
| 9565 | |
| 9566 | if (inst.operands[0].present) |
| 9567 | { |
| 9568 | reg = inst.operands[0].reg; |
| 9569 | constraint (reg != REG_SP, _("SRS base register must be r13")); |
| 9570 | } |
| 9571 | else |
| 9572 | reg = REG_SP; |
| 9573 | |
| 9574 | inst.instruction |= reg << 16; |
| 9575 | inst.instruction |= inst.operands[1].imm; |
| 9576 | if (inst.operands[0].writeback || inst.operands[1].writeback) |
| 9577 | inst.instruction |= WRITE_BACK; |
| 9578 | } |
| 9579 | |
| 9580 | /* ARM V6 strex (argument parse). */ |
| 9581 | |
| 9582 | static void |
| 9583 | do_strex (void) |
| 9584 | { |
| 9585 | constraint (!inst.operands[2].isreg || !inst.operands[2].preind |
| 9586 | || inst.operands[2].postind || inst.operands[2].writeback |
| 9587 | || inst.operands[2].immisreg || inst.operands[2].shifted |
| 9588 | || inst.operands[2].negative |
| 9589 | /* See comment in do_ldrex(). */ |
| 9590 | || (inst.operands[2].reg == REG_PC), |
| 9591 | BAD_ADDR_MODE); |
| 9592 | |
| 9593 | constraint (inst.operands[0].reg == inst.operands[1].reg |
| 9594 | || inst.operands[0].reg == inst.operands[2].reg, BAD_OVERLAP); |
| 9595 | |
| 9596 | constraint (inst.reloc.exp.X_op != O_constant |
| 9597 | || inst.reloc.exp.X_add_number != 0, |
| 9598 | _("offset must be zero in ARM encoding")); |
| 9599 | |
| 9600 | inst.instruction |= inst.operands[0].reg << 12; |
| 9601 | inst.instruction |= inst.operands[1].reg; |
| 9602 | inst.instruction |= inst.operands[2].reg << 16; |
| 9603 | inst.reloc.type = BFD_RELOC_UNUSED; |
| 9604 | } |
| 9605 | |
| 9606 | static void |
| 9607 | do_t_strexbh (void) |
| 9608 | { |
| 9609 | constraint (!inst.operands[2].isreg || !inst.operands[2].preind |
| 9610 | || inst.operands[2].postind || inst.operands[2].writeback |
| 9611 | || inst.operands[2].immisreg || inst.operands[2].shifted |
| 9612 | || inst.operands[2].negative, |
| 9613 | BAD_ADDR_MODE); |
| 9614 | |
| 9615 | constraint (inst.operands[0].reg == inst.operands[1].reg |
| 9616 | || inst.operands[0].reg == inst.operands[2].reg, BAD_OVERLAP); |
| 9617 | |
| 9618 | do_rm_rd_rn (); |
| 9619 | } |
| 9620 | |
| 9621 | static void |
| 9622 | do_strexd (void) |
| 9623 | { |
| 9624 | constraint (inst.operands[1].reg % 2 != 0, |
| 9625 | _("even register required")); |
| 9626 | constraint (inst.operands[2].present |
| 9627 | && inst.operands[2].reg != inst.operands[1].reg + 1, |
| 9628 | _("can only store two consecutive registers")); |
| 9629 | /* If op 2 were present and equal to PC, this function wouldn't |
| 9630 | have been called in the first place. */ |
| 9631 | constraint (inst.operands[1].reg == REG_LR, _("r14 not allowed here")); |
| 9632 | |
| 9633 | constraint (inst.operands[0].reg == inst.operands[1].reg |
| 9634 | || inst.operands[0].reg == inst.operands[1].reg + 1 |
| 9635 | || inst.operands[0].reg == inst.operands[3].reg, |
| 9636 | BAD_OVERLAP); |
| 9637 | |
| 9638 | inst.instruction |= inst.operands[0].reg << 12; |
| 9639 | inst.instruction |= inst.operands[1].reg; |
| 9640 | inst.instruction |= inst.operands[3].reg << 16; |
| 9641 | } |
| 9642 | |
| 9643 | /* ARM V8 STRL. */ |
| 9644 | static void |
| 9645 | do_stlex (void) |
| 9646 | { |
| 9647 | constraint (inst.operands[0].reg == inst.operands[1].reg |
| 9648 | || inst.operands[0].reg == inst.operands[2].reg, BAD_OVERLAP); |
| 9649 | |
| 9650 | do_rd_rm_rn (); |
| 9651 | } |
| 9652 | |
| 9653 | static void |
| 9654 | do_t_stlex (void) |
| 9655 | { |
| 9656 | constraint (inst.operands[0].reg == inst.operands[1].reg |
| 9657 | || inst.operands[0].reg == inst.operands[2].reg, BAD_OVERLAP); |
| 9658 | |
| 9659 | do_rm_rd_rn (); |
| 9660 | } |
| 9661 | |
| 9662 | /* ARM V6 SXTAH extracts a 16-bit value from a register, sign |
| 9663 | extends it to 32-bits, and adds the result to a value in another |
| 9664 | register. You can specify a rotation by 0, 8, 16, or 24 bits |
| 9665 | before extracting the 16-bit value. |
| 9666 | SXTAH{<cond>} <Rd>, <Rn>, <Rm>{, <rotation>} |
| 9667 | Condition defaults to COND_ALWAYS. |
| 9668 | Error if any register uses R15. */ |
| 9669 | |
| 9670 | static void |
| 9671 | do_sxtah (void) |
| 9672 | { |
| 9673 | inst.instruction |= inst.operands[0].reg << 12; |
| 9674 | inst.instruction |= inst.operands[1].reg << 16; |
| 9675 | inst.instruction |= inst.operands[2].reg; |
| 9676 | inst.instruction |= inst.operands[3].imm << 10; |
| 9677 | } |
| 9678 | |
| 9679 | /* ARM V6 SXTH. |
| 9680 | |
| 9681 | SXTH {<cond>} <Rd>, <Rm>{, <rotation>} |
| 9682 | Condition defaults to COND_ALWAYS. |
| 9683 | Error if any register uses R15. */ |
| 9684 | |
| 9685 | static void |
| 9686 | do_sxth (void) |
| 9687 | { |
| 9688 | inst.instruction |= inst.operands[0].reg << 12; |
| 9689 | inst.instruction |= inst.operands[1].reg; |
| 9690 | inst.instruction |= inst.operands[2].imm << 10; |
| 9691 | } |
| 9692 | \f |
| 9693 | /* VFP instructions. In a logical order: SP variant first, monad |
| 9694 | before dyad, arithmetic then move then load/store. */ |
| 9695 | |
| 9696 | static void |
| 9697 | do_vfp_sp_monadic (void) |
| 9698 | { |
| 9699 | encode_arm_vfp_reg (inst.operands[0].reg, VFP_REG_Sd); |
| 9700 | encode_arm_vfp_reg (inst.operands[1].reg, VFP_REG_Sm); |
| 9701 | } |
| 9702 | |
| 9703 | static void |
| 9704 | do_vfp_sp_dyadic (void) |
| 9705 | { |
| 9706 | encode_arm_vfp_reg (inst.operands[0].reg, VFP_REG_Sd); |
| 9707 | encode_arm_vfp_reg (inst.operands[1].reg, VFP_REG_Sn); |
| 9708 | encode_arm_vfp_reg (inst.operands[2].reg, VFP_REG_Sm); |
| 9709 | } |
| 9710 | |
| 9711 | static void |
| 9712 | do_vfp_sp_compare_z (void) |
| 9713 | { |
| 9714 | encode_arm_vfp_reg (inst.operands[0].reg, VFP_REG_Sd); |
| 9715 | } |
| 9716 | |
| 9717 | static void |
| 9718 | do_vfp_dp_sp_cvt (void) |
| 9719 | { |
| 9720 | encode_arm_vfp_reg (inst.operands[0].reg, VFP_REG_Dd); |
| 9721 | encode_arm_vfp_reg (inst.operands[1].reg, VFP_REG_Sm); |
| 9722 | } |
| 9723 | |
| 9724 | static void |
| 9725 | do_vfp_sp_dp_cvt (void) |
| 9726 | { |
| 9727 | encode_arm_vfp_reg (inst.operands[0].reg, VFP_REG_Sd); |
| 9728 | encode_arm_vfp_reg (inst.operands[1].reg, VFP_REG_Dm); |
| 9729 | } |
| 9730 | |
| 9731 | static void |
| 9732 | do_vfp_reg_from_sp (void) |
| 9733 | { |
| 9734 | inst.instruction |= inst.operands[0].reg << 12; |
| 9735 | encode_arm_vfp_reg (inst.operands[1].reg, VFP_REG_Sn); |
| 9736 | } |
| 9737 | |
| 9738 | static void |
| 9739 | do_vfp_reg2_from_sp2 (void) |
| 9740 | { |
| 9741 | constraint (inst.operands[2].imm != 2, |
| 9742 | _("only two consecutive VFP SP registers allowed here")); |
| 9743 | inst.instruction |= inst.operands[0].reg << 12; |
| 9744 | inst.instruction |= inst.operands[1].reg << 16; |
| 9745 | encode_arm_vfp_reg (inst.operands[2].reg, VFP_REG_Sm); |
| 9746 | } |
| 9747 | |
| 9748 | static void |
| 9749 | do_vfp_sp_from_reg (void) |
| 9750 | { |
| 9751 | encode_arm_vfp_reg (inst.operands[0].reg, VFP_REG_Sn); |
| 9752 | inst.instruction |= inst.operands[1].reg << 12; |
| 9753 | } |
| 9754 | |
| 9755 | static void |
| 9756 | do_vfp_sp2_from_reg2 (void) |
| 9757 | { |
| 9758 | constraint (inst.operands[0].imm != 2, |
| 9759 | _("only two consecutive VFP SP registers allowed here")); |
| 9760 | encode_arm_vfp_reg (inst.operands[0].reg, VFP_REG_Sm); |
| 9761 | inst.instruction |= inst.operands[1].reg << 12; |
| 9762 | inst.instruction |= inst.operands[2].reg << 16; |
| 9763 | } |
| 9764 | |
| 9765 | static void |
| 9766 | do_vfp_sp_ldst (void) |
| 9767 | { |
| 9768 | encode_arm_vfp_reg (inst.operands[0].reg, VFP_REG_Sd); |
| 9769 | encode_arm_cp_address (1, FALSE, TRUE, 0); |
| 9770 | } |
| 9771 | |
| 9772 | static void |
| 9773 | do_vfp_dp_ldst (void) |
| 9774 | { |
| 9775 | encode_arm_vfp_reg (inst.operands[0].reg, VFP_REG_Dd); |
| 9776 | encode_arm_cp_address (1, FALSE, TRUE, 0); |
| 9777 | } |
| 9778 | |
| 9779 | |
| 9780 | static void |
| 9781 | vfp_sp_ldstm (enum vfp_ldstm_type ldstm_type) |
| 9782 | { |
| 9783 | if (inst.operands[0].writeback) |
| 9784 | inst.instruction |= WRITE_BACK; |
| 9785 | else |
| 9786 | constraint (ldstm_type != VFP_LDSTMIA, |
| 9787 | _("this addressing mode requires base-register writeback")); |
| 9788 | inst.instruction |= inst.operands[0].reg << 16; |
| 9789 | encode_arm_vfp_reg (inst.operands[1].reg, VFP_REG_Sd); |
| 9790 | inst.instruction |= inst.operands[1].imm; |
| 9791 | } |
| 9792 | |
| 9793 | static void |
| 9794 | vfp_dp_ldstm (enum vfp_ldstm_type ldstm_type) |
| 9795 | { |
| 9796 | int count; |
| 9797 | |
| 9798 | if (inst.operands[0].writeback) |
| 9799 | inst.instruction |= WRITE_BACK; |
| 9800 | else |
| 9801 | constraint (ldstm_type != VFP_LDSTMIA && ldstm_type != VFP_LDSTMIAX, |
| 9802 | _("this addressing mode requires base-register writeback")); |
| 9803 | |
| 9804 | inst.instruction |= inst.operands[0].reg << 16; |
| 9805 | encode_arm_vfp_reg (inst.operands[1].reg, VFP_REG_Dd); |
| 9806 | |
| 9807 | count = inst.operands[1].imm << 1; |
| 9808 | if (ldstm_type == VFP_LDSTMIAX || ldstm_type == VFP_LDSTMDBX) |
| 9809 | count += 1; |
| 9810 | |
| 9811 | inst.instruction |= count; |
| 9812 | } |
| 9813 | |
| 9814 | static void |
| 9815 | do_vfp_sp_ldstmia (void) |
| 9816 | { |
| 9817 | vfp_sp_ldstm (VFP_LDSTMIA); |
| 9818 | } |
| 9819 | |
| 9820 | static void |
| 9821 | do_vfp_sp_ldstmdb (void) |
| 9822 | { |
| 9823 | vfp_sp_ldstm (VFP_LDSTMDB); |
| 9824 | } |
| 9825 | |
| 9826 | static void |
| 9827 | do_vfp_dp_ldstmia (void) |
| 9828 | { |
| 9829 | vfp_dp_ldstm (VFP_LDSTMIA); |
| 9830 | } |
| 9831 | |
| 9832 | static void |
| 9833 | do_vfp_dp_ldstmdb (void) |
| 9834 | { |
| 9835 | vfp_dp_ldstm (VFP_LDSTMDB); |
| 9836 | } |
| 9837 | |
| 9838 | static void |
| 9839 | do_vfp_xp_ldstmia (void) |
| 9840 | { |
| 9841 | vfp_dp_ldstm (VFP_LDSTMIAX); |
| 9842 | } |
| 9843 | |
| 9844 | static void |
| 9845 | do_vfp_xp_ldstmdb (void) |
| 9846 | { |
| 9847 | vfp_dp_ldstm (VFP_LDSTMDBX); |
| 9848 | } |
| 9849 | |
| 9850 | static void |
| 9851 | do_vfp_dp_rd_rm (void) |
| 9852 | { |
| 9853 | encode_arm_vfp_reg (inst.operands[0].reg, VFP_REG_Dd); |
| 9854 | encode_arm_vfp_reg (inst.operands[1].reg, VFP_REG_Dm); |
| 9855 | } |
| 9856 | |
| 9857 | static void |
| 9858 | do_vfp_dp_rn_rd (void) |
| 9859 | { |
| 9860 | encode_arm_vfp_reg (inst.operands[0].reg, VFP_REG_Dn); |
| 9861 | encode_arm_vfp_reg (inst.operands[1].reg, VFP_REG_Dd); |
| 9862 | } |
| 9863 | |
| 9864 | static void |
| 9865 | do_vfp_dp_rd_rn (void) |
| 9866 | { |
| 9867 | encode_arm_vfp_reg (inst.operands[0].reg, VFP_REG_Dd); |
| 9868 | encode_arm_vfp_reg (inst.operands[1].reg, VFP_REG_Dn); |
| 9869 | } |
| 9870 | |
| 9871 | static void |
| 9872 | do_vfp_dp_rd_rn_rm (void) |
| 9873 | { |
| 9874 | encode_arm_vfp_reg (inst.operands[0].reg, VFP_REG_Dd); |
| 9875 | encode_arm_vfp_reg (inst.operands[1].reg, VFP_REG_Dn); |
| 9876 | encode_arm_vfp_reg (inst.operands[2].reg, VFP_REG_Dm); |
| 9877 | } |
| 9878 | |
| 9879 | static void |
| 9880 | do_vfp_dp_rd (void) |
| 9881 | { |
| 9882 | encode_arm_vfp_reg (inst.operands[0].reg, VFP_REG_Dd); |
| 9883 | } |
| 9884 | |
| 9885 | static void |
| 9886 | do_vfp_dp_rm_rd_rn (void) |
| 9887 | { |
| 9888 | encode_arm_vfp_reg (inst.operands[0].reg, VFP_REG_Dm); |
| 9889 | encode_arm_vfp_reg (inst.operands[1].reg, VFP_REG_Dd); |
| 9890 | encode_arm_vfp_reg (inst.operands[2].reg, VFP_REG_Dn); |
| 9891 | } |
| 9892 | |
| 9893 | /* VFPv3 instructions. */ |
| 9894 | static void |
| 9895 | do_vfp_sp_const (void) |
| 9896 | { |
| 9897 | encode_arm_vfp_reg (inst.operands[0].reg, VFP_REG_Sd); |
| 9898 | inst.instruction |= (inst.operands[1].imm & 0xf0) << 12; |
| 9899 | inst.instruction |= (inst.operands[1].imm & 0x0f); |
| 9900 | } |
| 9901 | |
| 9902 | static void |
| 9903 | do_vfp_dp_const (void) |
| 9904 | { |
| 9905 | encode_arm_vfp_reg (inst.operands[0].reg, VFP_REG_Dd); |
| 9906 | inst.instruction |= (inst.operands[1].imm & 0xf0) << 12; |
| 9907 | inst.instruction |= (inst.operands[1].imm & 0x0f); |
| 9908 | } |
| 9909 | |
| 9910 | static void |
| 9911 | vfp_conv (int srcsize) |
| 9912 | { |
| 9913 | int immbits = srcsize - inst.operands[1].imm; |
| 9914 | |
| 9915 | if (srcsize == 16 && !(immbits >= 0 && immbits <= srcsize)) |
| 9916 | { |
| 9917 | /* If srcsize is 16, inst.operands[1].imm must be in the range 0-16. |
| 9918 | i.e. immbits must be in range 0 - 16. */ |
| 9919 | inst.error = _("immediate value out of range, expected range [0, 16]"); |
| 9920 | return; |
| 9921 | } |
| 9922 | else if (srcsize == 32 && !(immbits >= 0 && immbits < srcsize)) |
| 9923 | { |
| 9924 | /* If srcsize is 32, inst.operands[1].imm must be in the range 1-32. |
| 9925 | i.e. immbits must be in range 0 - 31. */ |
| 9926 | inst.error = _("immediate value out of range, expected range [1, 32]"); |
| 9927 | return; |
| 9928 | } |
| 9929 | |
| 9930 | inst.instruction |= (immbits & 1) << 5; |
| 9931 | inst.instruction |= (immbits >> 1); |
| 9932 | } |
| 9933 | |
| 9934 | static void |
| 9935 | do_vfp_sp_conv_16 (void) |
| 9936 | { |
| 9937 | encode_arm_vfp_reg (inst.operands[0].reg, VFP_REG_Sd); |
| 9938 | vfp_conv (16); |
| 9939 | } |
| 9940 | |
| 9941 | static void |
| 9942 | do_vfp_dp_conv_16 (void) |
| 9943 | { |
| 9944 | encode_arm_vfp_reg (inst.operands[0].reg, VFP_REG_Dd); |
| 9945 | vfp_conv (16); |
| 9946 | } |
| 9947 | |
| 9948 | static void |
| 9949 | do_vfp_sp_conv_32 (void) |
| 9950 | { |
| 9951 | encode_arm_vfp_reg (inst.operands[0].reg, VFP_REG_Sd); |
| 9952 | vfp_conv (32); |
| 9953 | } |
| 9954 | |
| 9955 | static void |
| 9956 | do_vfp_dp_conv_32 (void) |
| 9957 | { |
| 9958 | encode_arm_vfp_reg (inst.operands[0].reg, VFP_REG_Dd); |
| 9959 | vfp_conv (32); |
| 9960 | } |
| 9961 | \f |
| 9962 | /* FPA instructions. Also in a logical order. */ |
| 9963 | |
| 9964 | static void |
| 9965 | do_fpa_cmp (void) |
| 9966 | { |
| 9967 | inst.instruction |= inst.operands[0].reg << 16; |
| 9968 | inst.instruction |= inst.operands[1].reg; |
| 9969 | } |
| 9970 | |
| 9971 | static void |
| 9972 | do_fpa_ldmstm (void) |
| 9973 | { |
| 9974 | inst.instruction |= inst.operands[0].reg << 12; |
| 9975 | switch (inst.operands[1].imm) |
| 9976 | { |
| 9977 | case 1: inst.instruction |= CP_T_X; break; |
| 9978 | case 2: inst.instruction |= CP_T_Y; break; |
| 9979 | case 3: inst.instruction |= CP_T_Y | CP_T_X; break; |
| 9980 | case 4: break; |
| 9981 | default: abort (); |
| 9982 | } |
| 9983 | |
| 9984 | if (inst.instruction & (PRE_INDEX | INDEX_UP)) |
| 9985 | { |
| 9986 | /* The instruction specified "ea" or "fd", so we can only accept |
| 9987 | [Rn]{!}. The instruction does not really support stacking or |
| 9988 | unstacking, so we have to emulate these by setting appropriate |
| 9989 | bits and offsets. */ |
| 9990 | constraint (inst.reloc.exp.X_op != O_constant |
| 9991 | || inst.reloc.exp.X_add_number != 0, |
| 9992 | _("this instruction does not support indexing")); |
| 9993 | |
| 9994 | if ((inst.instruction & PRE_INDEX) || inst.operands[2].writeback) |
| 9995 | inst.reloc.exp.X_add_number = 12 * inst.operands[1].imm; |
| 9996 | |
| 9997 | if (!(inst.instruction & INDEX_UP)) |
| 9998 | inst.reloc.exp.X_add_number = -inst.reloc.exp.X_add_number; |
| 9999 | |
| 10000 | if (!(inst.instruction & PRE_INDEX) && inst.operands[2].writeback) |
| 10001 | { |
| 10002 | inst.operands[2].preind = 0; |
| 10003 | inst.operands[2].postind = 1; |
| 10004 | } |
| 10005 | } |
| 10006 | |
| 10007 | encode_arm_cp_address (2, TRUE, TRUE, 0); |
| 10008 | } |
| 10009 | \f |
| 10010 | /* iWMMXt instructions: strictly in alphabetical order. */ |
| 10011 | |
| 10012 | static void |
| 10013 | do_iwmmxt_tandorc (void) |
| 10014 | { |
| 10015 | constraint (inst.operands[0].reg != REG_PC, _("only r15 allowed here")); |
| 10016 | } |
| 10017 | |
| 10018 | static void |
| 10019 | do_iwmmxt_textrc (void) |
| 10020 | { |
| 10021 | inst.instruction |= inst.operands[0].reg << 12; |
| 10022 | inst.instruction |= inst.operands[1].imm; |
| 10023 | } |
| 10024 | |
| 10025 | static void |
| 10026 | do_iwmmxt_textrm (void) |
| 10027 | { |
| 10028 | inst.instruction |= inst.operands[0].reg << 12; |
| 10029 | inst.instruction |= inst.operands[1].reg << 16; |
| 10030 | inst.instruction |= inst.operands[2].imm; |
| 10031 | } |
| 10032 | |
| 10033 | static void |
| 10034 | do_iwmmxt_tinsr (void) |
| 10035 | { |
| 10036 | inst.instruction |= inst.operands[0].reg << 16; |
| 10037 | inst.instruction |= inst.operands[1].reg << 12; |
| 10038 | inst.instruction |= inst.operands[2].imm; |
| 10039 | } |
| 10040 | |
| 10041 | static void |
| 10042 | do_iwmmxt_tmia (void) |
| 10043 | { |
| 10044 | inst.instruction |= inst.operands[0].reg << 5; |
| 10045 | inst.instruction |= inst.operands[1].reg; |
| 10046 | inst.instruction |= inst.operands[2].reg << 12; |
| 10047 | } |
| 10048 | |
| 10049 | static void |
| 10050 | do_iwmmxt_waligni (void) |
| 10051 | { |
| 10052 | inst.instruction |= inst.operands[0].reg << 12; |
| 10053 | inst.instruction |= inst.operands[1].reg << 16; |
| 10054 | inst.instruction |= inst.operands[2].reg; |
| 10055 | inst.instruction |= inst.operands[3].imm << 20; |
| 10056 | } |
| 10057 | |
| 10058 | static void |
| 10059 | do_iwmmxt_wmerge (void) |
| 10060 | { |
| 10061 | inst.instruction |= inst.operands[0].reg << 12; |
| 10062 | inst.instruction |= inst.operands[1].reg << 16; |
| 10063 | inst.instruction |= inst.operands[2].reg; |
| 10064 | inst.instruction |= inst.operands[3].imm << 21; |
| 10065 | } |
| 10066 | |
| 10067 | static void |
| 10068 | do_iwmmxt_wmov (void) |
| 10069 | { |
| 10070 | /* WMOV rD, rN is an alias for WOR rD, rN, rN. */ |
| 10071 | inst.instruction |= inst.operands[0].reg << 12; |
| 10072 | inst.instruction |= inst.operands[1].reg << 16; |
| 10073 | inst.instruction |= inst.operands[1].reg; |
| 10074 | } |
| 10075 | |
| 10076 | static void |
| 10077 | do_iwmmxt_wldstbh (void) |
| 10078 | { |
| 10079 | int reloc; |
| 10080 | inst.instruction |= inst.operands[0].reg << 12; |
| 10081 | if (thumb_mode) |
| 10082 | reloc = BFD_RELOC_ARM_T32_CP_OFF_IMM_S2; |
| 10083 | else |
| 10084 | reloc = BFD_RELOC_ARM_CP_OFF_IMM_S2; |
| 10085 | encode_arm_cp_address (1, TRUE, FALSE, reloc); |
| 10086 | } |
| 10087 | |
| 10088 | static void |
| 10089 | do_iwmmxt_wldstw (void) |
| 10090 | { |
| 10091 | /* RIWR_RIWC clears .isreg for a control register. */ |
| 10092 | if (!inst.operands[0].isreg) |
| 10093 | { |
| 10094 | constraint (inst.cond != COND_ALWAYS, BAD_COND); |
| 10095 | inst.instruction |= 0xf0000000; |
| 10096 | } |
| 10097 | |
| 10098 | inst.instruction |= inst.operands[0].reg << 12; |
| 10099 | encode_arm_cp_address (1, TRUE, TRUE, 0); |
| 10100 | } |
| 10101 | |
| 10102 | static void |
| 10103 | do_iwmmxt_wldstd (void) |
| 10104 | { |
| 10105 | inst.instruction |= inst.operands[0].reg << 12; |
| 10106 | if (ARM_CPU_HAS_FEATURE (cpu_variant, arm_cext_iwmmxt2) |
| 10107 | && inst.operands[1].immisreg) |
| 10108 | { |
| 10109 | inst.instruction &= ~0x1a000ff; |
| 10110 | inst.instruction |= (0xfU << 28); |
| 10111 | if (inst.operands[1].preind) |
| 10112 | inst.instruction |= PRE_INDEX; |
| 10113 | if (!inst.operands[1].negative) |
| 10114 | inst.instruction |= INDEX_UP; |
| 10115 | if (inst.operands[1].writeback) |
| 10116 | inst.instruction |= WRITE_BACK; |
| 10117 | inst.instruction |= inst.operands[1].reg << 16; |
| 10118 | inst.instruction |= inst.reloc.exp.X_add_number << 4; |
| 10119 | inst.instruction |= inst.operands[1].imm; |
| 10120 | } |
| 10121 | else |
| 10122 | encode_arm_cp_address (1, TRUE, FALSE, 0); |
| 10123 | } |
| 10124 | |
| 10125 | static void |
| 10126 | do_iwmmxt_wshufh (void) |
| 10127 | { |
| 10128 | inst.instruction |= inst.operands[0].reg << 12; |
| 10129 | inst.instruction |= inst.operands[1].reg << 16; |
| 10130 | inst.instruction |= ((inst.operands[2].imm & 0xf0) << 16); |
| 10131 | inst.instruction |= (inst.operands[2].imm & 0x0f); |
| 10132 | } |
| 10133 | |
| 10134 | static void |
| 10135 | do_iwmmxt_wzero (void) |
| 10136 | { |
| 10137 | /* WZERO reg is an alias for WANDN reg, reg, reg. */ |
| 10138 | inst.instruction |= inst.operands[0].reg; |
| 10139 | inst.instruction |= inst.operands[0].reg << 12; |
| 10140 | inst.instruction |= inst.operands[0].reg << 16; |
| 10141 | } |
| 10142 | |
| 10143 | static void |
| 10144 | do_iwmmxt_wrwrwr_or_imm5 (void) |
| 10145 | { |
| 10146 | if (inst.operands[2].isreg) |
| 10147 | do_rd_rn_rm (); |
| 10148 | else { |
| 10149 | constraint (!ARM_CPU_HAS_FEATURE (cpu_variant, arm_cext_iwmmxt2), |
| 10150 | _("immediate operand requires iWMMXt2")); |
| 10151 | do_rd_rn (); |
| 10152 | if (inst.operands[2].imm == 0) |
| 10153 | { |
| 10154 | switch ((inst.instruction >> 20) & 0xf) |
| 10155 | { |
| 10156 | case 4: |
| 10157 | case 5: |
| 10158 | case 6: |
| 10159 | case 7: |
| 10160 | /* w...h wrd, wrn, #0 -> wrorh wrd, wrn, #16. */ |
| 10161 | inst.operands[2].imm = 16; |
| 10162 | inst.instruction = (inst.instruction & 0xff0fffff) | (0x7 << 20); |
| 10163 | break; |
| 10164 | case 8: |
| 10165 | case 9: |
| 10166 | case 10: |
| 10167 | case 11: |
| 10168 | /* w...w wrd, wrn, #0 -> wrorw wrd, wrn, #32. */ |
| 10169 | inst.operands[2].imm = 32; |
| 10170 | inst.instruction = (inst.instruction & 0xff0fffff) | (0xb << 20); |
| 10171 | break; |
| 10172 | case 12: |
| 10173 | case 13: |
| 10174 | case 14: |
| 10175 | case 15: |
| 10176 | { |
| 10177 | /* w...d wrd, wrn, #0 -> wor wrd, wrn, wrn. */ |
| 10178 | unsigned long wrn; |
| 10179 | wrn = (inst.instruction >> 16) & 0xf; |
| 10180 | inst.instruction &= 0xff0fff0f; |
| 10181 | inst.instruction |= wrn; |
| 10182 | /* Bail out here; the instruction is now assembled. */ |
| 10183 | return; |
| 10184 | } |
| 10185 | } |
| 10186 | } |
| 10187 | /* Map 32 -> 0, etc. */ |
| 10188 | inst.operands[2].imm &= 0x1f; |
| 10189 | inst.instruction |= (0xfU << 28) | ((inst.operands[2].imm & 0x10) << 4) | (inst.operands[2].imm & 0xf); |
| 10190 | } |
| 10191 | } |
| 10192 | \f |
| 10193 | /* Cirrus Maverick instructions. Simple 2-, 3-, and 4-register |
| 10194 | operations first, then control, shift, and load/store. */ |
| 10195 | |
| 10196 | /* Insns like "foo X,Y,Z". */ |
| 10197 | |
| 10198 | static void |
| 10199 | do_mav_triple (void) |
| 10200 | { |
| 10201 | inst.instruction |= inst.operands[0].reg << 16; |
| 10202 | inst.instruction |= inst.operands[1].reg; |
| 10203 | inst.instruction |= inst.operands[2].reg << 12; |
| 10204 | } |
| 10205 | |
| 10206 | /* Insns like "foo W,X,Y,Z". |
| 10207 | where W=MVAX[0:3] and X,Y,Z=MVFX[0:15]. */ |
| 10208 | |
| 10209 | static void |
| 10210 | do_mav_quad (void) |
| 10211 | { |
| 10212 | inst.instruction |= inst.operands[0].reg << 5; |
| 10213 | inst.instruction |= inst.operands[1].reg << 12; |
| 10214 | inst.instruction |= inst.operands[2].reg << 16; |
| 10215 | inst.instruction |= inst.operands[3].reg; |
| 10216 | } |
| 10217 | |
| 10218 | /* cfmvsc32<cond> DSPSC,MVDX[15:0]. */ |
| 10219 | static void |
| 10220 | do_mav_dspsc (void) |
| 10221 | { |
| 10222 | inst.instruction |= inst.operands[1].reg << 12; |
| 10223 | } |
| 10224 | |
| 10225 | /* Maverick shift immediate instructions. |
| 10226 | cfsh32<cond> MVFX[15:0],MVFX[15:0],Shift[6:0]. |
| 10227 | cfsh64<cond> MVDX[15:0],MVDX[15:0],Shift[6:0]. */ |
| 10228 | |
| 10229 | static void |
| 10230 | do_mav_shift (void) |
| 10231 | { |
| 10232 | int imm = inst.operands[2].imm; |
| 10233 | |
| 10234 | inst.instruction |= inst.operands[0].reg << 12; |
| 10235 | inst.instruction |= inst.operands[1].reg << 16; |
| 10236 | |
| 10237 | /* Bits 0-3 of the insn should have bits 0-3 of the immediate. |
| 10238 | Bits 5-7 of the insn should have bits 4-6 of the immediate. |
| 10239 | Bit 4 should be 0. */ |
| 10240 | imm = (imm & 0xf) | ((imm & 0x70) << 1); |
| 10241 | |
| 10242 | inst.instruction |= imm; |
| 10243 | } |
| 10244 | \f |
| 10245 | /* XScale instructions. Also sorted arithmetic before move. */ |
| 10246 | |
| 10247 | /* Xscale multiply-accumulate (argument parse) |
| 10248 | MIAcc acc0,Rm,Rs |
| 10249 | MIAPHcc acc0,Rm,Rs |
| 10250 | MIAxycc acc0,Rm,Rs. */ |
| 10251 | |
| 10252 | static void |
| 10253 | do_xsc_mia (void) |
| 10254 | { |
| 10255 | inst.instruction |= inst.operands[1].reg; |
| 10256 | inst.instruction |= inst.operands[2].reg << 12; |
| 10257 | } |
| 10258 | |
| 10259 | /* Xscale move-accumulator-register (argument parse) |
| 10260 | |
| 10261 | MARcc acc0,RdLo,RdHi. */ |
| 10262 | |
| 10263 | static void |
| 10264 | do_xsc_mar (void) |
| 10265 | { |
| 10266 | inst.instruction |= inst.operands[1].reg << 12; |
| 10267 | inst.instruction |= inst.operands[2].reg << 16; |
| 10268 | } |
| 10269 | |
| 10270 | /* Xscale move-register-accumulator (argument parse) |
| 10271 | |
| 10272 | MRAcc RdLo,RdHi,acc0. */ |
| 10273 | |
| 10274 | static void |
| 10275 | do_xsc_mra (void) |
| 10276 | { |
| 10277 | constraint (inst.operands[0].reg == inst.operands[1].reg, BAD_OVERLAP); |
| 10278 | inst.instruction |= inst.operands[0].reg << 12; |
| 10279 | inst.instruction |= inst.operands[1].reg << 16; |
| 10280 | } |
| 10281 | \f |
| 10282 | /* Encoding functions relevant only to Thumb. */ |
| 10283 | |
| 10284 | /* inst.operands[i] is a shifted-register operand; encode |
| 10285 | it into inst.instruction in the format used by Thumb32. */ |
| 10286 | |
| 10287 | static void |
| 10288 | encode_thumb32_shifted_operand (int i) |
| 10289 | { |
| 10290 | unsigned int value = inst.reloc.exp.X_add_number; |
| 10291 | unsigned int shift = inst.operands[i].shift_kind; |
| 10292 | |
| 10293 | constraint (inst.operands[i].immisreg, |
| 10294 | _("shift by register not allowed in thumb mode")); |
| 10295 | inst.instruction |= inst.operands[i].reg; |
| 10296 | if (shift == SHIFT_RRX) |
| 10297 | inst.instruction |= SHIFT_ROR << 4; |
| 10298 | else |
| 10299 | { |
| 10300 | constraint (inst.reloc.exp.X_op != O_constant, |
| 10301 | _("expression too complex")); |
| 10302 | |
| 10303 | constraint (value > 32 |
| 10304 | || (value == 32 && (shift == SHIFT_LSL |
| 10305 | || shift == SHIFT_ROR)), |
| 10306 | _("shift expression is too large")); |
| 10307 | |
| 10308 | if (value == 0) |
| 10309 | shift = SHIFT_LSL; |
| 10310 | else if (value == 32) |
| 10311 | value = 0; |
| 10312 | |
| 10313 | inst.instruction |= shift << 4; |
| 10314 | inst.instruction |= (value & 0x1c) << 10; |
| 10315 | inst.instruction |= (value & 0x03) << 6; |
| 10316 | } |
| 10317 | } |
| 10318 | |
| 10319 | |
| 10320 | /* inst.operands[i] was set up by parse_address. Encode it into a |
| 10321 | Thumb32 format load or store instruction. Reject forms that cannot |
| 10322 | be used with such instructions. If is_t is true, reject forms that |
| 10323 | cannot be used with a T instruction; if is_d is true, reject forms |
| 10324 | that cannot be used with a D instruction. If it is a store insn, |
| 10325 | reject PC in Rn. */ |
| 10326 | |
| 10327 | static void |
| 10328 | encode_thumb32_addr_mode (int i, bfd_boolean is_t, bfd_boolean is_d) |
| 10329 | { |
| 10330 | const bfd_boolean is_pc = (inst.operands[i].reg == REG_PC); |
| 10331 | |
| 10332 | constraint (!inst.operands[i].isreg, |
| 10333 | _("Instruction does not support =N addresses")); |
| 10334 | |
| 10335 | inst.instruction |= inst.operands[i].reg << 16; |
| 10336 | if (inst.operands[i].immisreg) |
| 10337 | { |
| 10338 | constraint (is_pc, BAD_PC_ADDRESSING); |
| 10339 | constraint (is_t || is_d, _("cannot use register index with this instruction")); |
| 10340 | constraint (inst.operands[i].negative, |
| 10341 | _("Thumb does not support negative register indexing")); |
| 10342 | constraint (inst.operands[i].postind, |
| 10343 | _("Thumb does not support register post-indexing")); |
| 10344 | constraint (inst.operands[i].writeback, |
| 10345 | _("Thumb does not support register indexing with writeback")); |
| 10346 | constraint (inst.operands[i].shifted && inst.operands[i].shift_kind != SHIFT_LSL, |
| 10347 | _("Thumb supports only LSL in shifted register indexing")); |
| 10348 | |
| 10349 | inst.instruction |= inst.operands[i].imm; |
| 10350 | if (inst.operands[i].shifted) |
| 10351 | { |
| 10352 | constraint (inst.reloc.exp.X_op != O_constant, |
| 10353 | _("expression too complex")); |
| 10354 | constraint (inst.reloc.exp.X_add_number < 0 |
| 10355 | || inst.reloc.exp.X_add_number > 3, |
| 10356 | _("shift out of range")); |
| 10357 | inst.instruction |= inst.reloc.exp.X_add_number << 4; |
| 10358 | } |
| 10359 | inst.reloc.type = BFD_RELOC_UNUSED; |
| 10360 | } |
| 10361 | else if (inst.operands[i].preind) |
| 10362 | { |
| 10363 | constraint (is_pc && inst.operands[i].writeback, BAD_PC_WRITEBACK); |
| 10364 | constraint (is_t && inst.operands[i].writeback, |
| 10365 | _("cannot use writeback with this instruction")); |
| 10366 | constraint (is_pc && ((inst.instruction & THUMB2_LOAD_BIT) == 0), |
| 10367 | BAD_PC_ADDRESSING); |
| 10368 | |
| 10369 | if (is_d) |
| 10370 | { |
| 10371 | inst.instruction |= 0x01000000; |
| 10372 | if (inst.operands[i].writeback) |
| 10373 | inst.instruction |= 0x00200000; |
| 10374 | } |
| 10375 | else |
| 10376 | { |
| 10377 | inst.instruction |= 0x00000c00; |
| 10378 | if (inst.operands[i].writeback) |
| 10379 | inst.instruction |= 0x00000100; |
| 10380 | } |
| 10381 | inst.reloc.type = BFD_RELOC_ARM_T32_OFFSET_IMM; |
| 10382 | } |
| 10383 | else if (inst.operands[i].postind) |
| 10384 | { |
| 10385 | gas_assert (inst.operands[i].writeback); |
| 10386 | constraint (is_pc, _("cannot use post-indexing with PC-relative addressing")); |
| 10387 | constraint (is_t, _("cannot use post-indexing with this instruction")); |
| 10388 | |
| 10389 | if (is_d) |
| 10390 | inst.instruction |= 0x00200000; |
| 10391 | else |
| 10392 | inst.instruction |= 0x00000900; |
| 10393 | inst.reloc.type = BFD_RELOC_ARM_T32_OFFSET_IMM; |
| 10394 | } |
| 10395 | else /* unindexed - only for coprocessor */ |
| 10396 | inst.error = _("instruction does not accept unindexed addressing"); |
| 10397 | } |
| 10398 | |
| 10399 | /* Table of Thumb instructions which exist in both 16- and 32-bit |
| 10400 | encodings (the latter only in post-V6T2 cores). The index is the |
| 10401 | value used in the insns table below. When there is more than one |
| 10402 | possible 16-bit encoding for the instruction, this table always |
| 10403 | holds variant (1). |
| 10404 | Also contains several pseudo-instructions used during relaxation. */ |
| 10405 | #define T16_32_TAB \ |
| 10406 | X(_adc, 4140, eb400000), \ |
| 10407 | X(_adcs, 4140, eb500000), \ |
| 10408 | X(_add, 1c00, eb000000), \ |
| 10409 | X(_adds, 1c00, eb100000), \ |
| 10410 | X(_addi, 0000, f1000000), \ |
| 10411 | X(_addis, 0000, f1100000), \ |
| 10412 | X(_add_pc,000f, f20f0000), \ |
| 10413 | X(_add_sp,000d, f10d0000), \ |
| 10414 | X(_adr, 000f, f20f0000), \ |
| 10415 | X(_and, 4000, ea000000), \ |
| 10416 | X(_ands, 4000, ea100000), \ |
| 10417 | X(_asr, 1000, fa40f000), \ |
| 10418 | X(_asrs, 1000, fa50f000), \ |
| 10419 | X(_b, e000, f000b000), \ |
| 10420 | X(_bcond, d000, f0008000), \ |
| 10421 | X(_bic, 4380, ea200000), \ |
| 10422 | X(_bics, 4380, ea300000), \ |
| 10423 | X(_cmn, 42c0, eb100f00), \ |
| 10424 | X(_cmp, 2800, ebb00f00), \ |
| 10425 | X(_cpsie, b660, f3af8400), \ |
| 10426 | X(_cpsid, b670, f3af8600), \ |
| 10427 | X(_cpy, 4600, ea4f0000), \ |
| 10428 | X(_dec_sp,80dd, f1ad0d00), \ |
| 10429 | X(_eor, 4040, ea800000), \ |
| 10430 | X(_eors, 4040, ea900000), \ |
| 10431 | X(_inc_sp,00dd, f10d0d00), \ |
| 10432 | X(_ldmia, c800, e8900000), \ |
| 10433 | X(_ldr, 6800, f8500000), \ |
| 10434 | X(_ldrb, 7800, f8100000), \ |
| 10435 | X(_ldrh, 8800, f8300000), \ |
| 10436 | X(_ldrsb, 5600, f9100000), \ |
| 10437 | X(_ldrsh, 5e00, f9300000), \ |
| 10438 | X(_ldr_pc,4800, f85f0000), \ |
| 10439 | X(_ldr_pc2,4800, f85f0000), \ |
| 10440 | X(_ldr_sp,9800, f85d0000), \ |
| 10441 | X(_lsl, 0000, fa00f000), \ |
| 10442 | X(_lsls, 0000, fa10f000), \ |
| 10443 | X(_lsr, 0800, fa20f000), \ |
| 10444 | X(_lsrs, 0800, fa30f000), \ |
| 10445 | X(_mov, 2000, ea4f0000), \ |
| 10446 | X(_movs, 2000, ea5f0000), \ |
| 10447 | X(_mul, 4340, fb00f000), \ |
| 10448 | X(_muls, 4340, ffffffff), /* no 32b muls */ \ |
| 10449 | X(_mvn, 43c0, ea6f0000), \ |
| 10450 | X(_mvns, 43c0, ea7f0000), \ |
| 10451 | X(_neg, 4240, f1c00000), /* rsb #0 */ \ |
| 10452 | X(_negs, 4240, f1d00000), /* rsbs #0 */ \ |
| 10453 | X(_orr, 4300, ea400000), \ |
| 10454 | X(_orrs, 4300, ea500000), \ |
| 10455 | X(_pop, bc00, e8bd0000), /* ldmia sp!,... */ \ |
| 10456 | X(_push, b400, e92d0000), /* stmdb sp!,... */ \ |
| 10457 | X(_rev, ba00, fa90f080), \ |
| 10458 | X(_rev16, ba40, fa90f090), \ |
| 10459 | X(_revsh, bac0, fa90f0b0), \ |
| 10460 | X(_ror, 41c0, fa60f000), \ |
| 10461 | X(_rors, 41c0, fa70f000), \ |
| 10462 | X(_sbc, 4180, eb600000), \ |
| 10463 | X(_sbcs, 4180, eb700000), \ |
| 10464 | X(_stmia, c000, e8800000), \ |
| 10465 | X(_str, 6000, f8400000), \ |
| 10466 | X(_strb, 7000, f8000000), \ |
| 10467 | X(_strh, 8000, f8200000), \ |
| 10468 | X(_str_sp,9000, f84d0000), \ |
| 10469 | X(_sub, 1e00, eba00000), \ |
| 10470 | X(_subs, 1e00, ebb00000), \ |
| 10471 | X(_subi, 8000, f1a00000), \ |
| 10472 | X(_subis, 8000, f1b00000), \ |
| 10473 | X(_sxtb, b240, fa4ff080), \ |
| 10474 | X(_sxth, b200, fa0ff080), \ |
| 10475 | X(_tst, 4200, ea100f00), \ |
| 10476 | X(_uxtb, b2c0, fa5ff080), \ |
| 10477 | X(_uxth, b280, fa1ff080), \ |
| 10478 | X(_nop, bf00, f3af8000), \ |
| 10479 | X(_yield, bf10, f3af8001), \ |
| 10480 | X(_wfe, bf20, f3af8002), \ |
| 10481 | X(_wfi, bf30, f3af8003), \ |
| 10482 | X(_sev, bf40, f3af8004), \ |
| 10483 | X(_sevl, bf50, f3af8005), \ |
| 10484 | X(_udf, de00, f7f0a000) |
| 10485 | |
| 10486 | /* To catch errors in encoding functions, the codes are all offset by |
| 10487 | 0xF800, putting them in one of the 32-bit prefix ranges, ergo undefined |
| 10488 | as 16-bit instructions. */ |
| 10489 | #define X(a,b,c) T_MNEM##a |
| 10490 | enum t16_32_codes { T16_32_OFFSET = 0xF7FF, T16_32_TAB }; |
| 10491 | #undef X |
| 10492 | |
| 10493 | #define X(a,b,c) 0x##b |
| 10494 | static const unsigned short thumb_op16[] = { T16_32_TAB }; |
| 10495 | #define THUMB_OP16(n) (thumb_op16[(n) - (T16_32_OFFSET + 1)]) |
| 10496 | #undef X |
| 10497 | |
| 10498 | #define X(a,b,c) 0x##c |
| 10499 | static const unsigned int thumb_op32[] = { T16_32_TAB }; |
| 10500 | #define THUMB_OP32(n) (thumb_op32[(n) - (T16_32_OFFSET + 1)]) |
| 10501 | #define THUMB_SETS_FLAGS(n) (THUMB_OP32 (n) & 0x00100000) |
| 10502 | #undef X |
| 10503 | #undef T16_32_TAB |
| 10504 | |
| 10505 | /* Thumb instruction encoders, in alphabetical order. */ |
| 10506 | |
| 10507 | /* ADDW or SUBW. */ |
| 10508 | |
| 10509 | static void |
| 10510 | do_t_add_sub_w (void) |
| 10511 | { |
| 10512 | int Rd, Rn; |
| 10513 | |
| 10514 | Rd = inst.operands[0].reg; |
| 10515 | Rn = inst.operands[1].reg; |
| 10516 | |
| 10517 | /* If Rn is REG_PC, this is ADR; if Rn is REG_SP, then this |
| 10518 | is the SP-{plus,minus}-immediate form of the instruction. */ |
| 10519 | if (Rn == REG_SP) |
| 10520 | constraint (Rd == REG_PC, BAD_PC); |
| 10521 | else |
| 10522 | reject_bad_reg (Rd); |
| 10523 | |
| 10524 | inst.instruction |= (Rn << 16) | (Rd << 8); |
| 10525 | inst.reloc.type = BFD_RELOC_ARM_T32_IMM12; |
| 10526 | } |
| 10527 | |
| 10528 | /* Parse an add or subtract instruction. We get here with inst.instruction |
| 10529 | equaling any of THUMB_OPCODE_add, adds, sub, or subs. */ |
| 10530 | |
| 10531 | static void |
| 10532 | do_t_add_sub (void) |
| 10533 | { |
| 10534 | int Rd, Rs, Rn; |
| 10535 | |
| 10536 | Rd = inst.operands[0].reg; |
| 10537 | Rs = (inst.operands[1].present |
| 10538 | ? inst.operands[1].reg /* Rd, Rs, foo */ |
| 10539 | : inst.operands[0].reg); /* Rd, foo -> Rd, Rd, foo */ |
| 10540 | |
| 10541 | if (Rd == REG_PC) |
| 10542 | set_it_insn_type_last (); |
| 10543 | |
| 10544 | if (unified_syntax) |
| 10545 | { |
| 10546 | bfd_boolean flags; |
| 10547 | bfd_boolean narrow; |
| 10548 | int opcode; |
| 10549 | |
| 10550 | flags = (inst.instruction == T_MNEM_adds |
| 10551 | || inst.instruction == T_MNEM_subs); |
| 10552 | if (flags) |
| 10553 | narrow = !in_it_block (); |
| 10554 | else |
| 10555 | narrow = in_it_block (); |
| 10556 | if (!inst.operands[2].isreg) |
| 10557 | { |
| 10558 | int add; |
| 10559 | |
| 10560 | if (!ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v8)) |
| 10561 | constraint (Rd == REG_SP && Rs != REG_SP, BAD_SP); |
| 10562 | |
| 10563 | add = (inst.instruction == T_MNEM_add |
| 10564 | || inst.instruction == T_MNEM_adds); |
| 10565 | opcode = 0; |
| 10566 | if (inst.size_req != 4) |
| 10567 | { |
| 10568 | /* Attempt to use a narrow opcode, with relaxation if |
| 10569 | appropriate. */ |
| 10570 | if (Rd == REG_SP && Rs == REG_SP && !flags) |
| 10571 | opcode = add ? T_MNEM_inc_sp : T_MNEM_dec_sp; |
| 10572 | else if (Rd <= 7 && Rs == REG_SP && add && !flags) |
| 10573 | opcode = T_MNEM_add_sp; |
| 10574 | else if (Rd <= 7 && Rs == REG_PC && add && !flags) |
| 10575 | opcode = T_MNEM_add_pc; |
| 10576 | else if (Rd <= 7 && Rs <= 7 && narrow) |
| 10577 | { |
| 10578 | if (flags) |
| 10579 | opcode = add ? T_MNEM_addis : T_MNEM_subis; |
| 10580 | else |
| 10581 | opcode = add ? T_MNEM_addi : T_MNEM_subi; |
| 10582 | } |
| 10583 | if (opcode) |
| 10584 | { |
| 10585 | inst.instruction = THUMB_OP16(opcode); |
| 10586 | inst.instruction |= (Rd << 4) | Rs; |
| 10587 | if (inst.reloc.type < BFD_RELOC_ARM_THUMB_ALU_ABS_G0_NC |
| 10588 | || inst.reloc.type > BFD_RELOC_ARM_THUMB_ALU_ABS_G3_NC) |
| 10589 | { |
| 10590 | if (inst.size_req == 2) |
| 10591 | inst.reloc.type = BFD_RELOC_ARM_THUMB_ADD; |
| 10592 | else |
| 10593 | inst.relax = opcode; |
| 10594 | } |
| 10595 | } |
| 10596 | else |
| 10597 | constraint (inst.size_req == 2, BAD_HIREG); |
| 10598 | } |
| 10599 | if (inst.size_req == 4 |
| 10600 | || (inst.size_req != 2 && !opcode)) |
| 10601 | { |
| 10602 | constraint (inst.reloc.type >= BFD_RELOC_ARM_THUMB_ALU_ABS_G0_NC |
| 10603 | && inst.reloc.type <= BFD_RELOC_ARM_THUMB_ALU_ABS_G3_NC , |
| 10604 | THUMB1_RELOC_ONLY); |
| 10605 | if (Rd == REG_PC) |
| 10606 | { |
| 10607 | constraint (add, BAD_PC); |
| 10608 | constraint (Rs != REG_LR || inst.instruction != T_MNEM_subs, |
| 10609 | _("only SUBS PC, LR, #const allowed")); |
| 10610 | constraint (inst.reloc.exp.X_op != O_constant, |
| 10611 | _("expression too complex")); |
| 10612 | constraint (inst.reloc.exp.X_add_number < 0 |
| 10613 | || inst.reloc.exp.X_add_number > 0xff, |
| 10614 | _("immediate value out of range")); |
| 10615 | inst.instruction = T2_SUBS_PC_LR |
| 10616 | | inst.reloc.exp.X_add_number; |
| 10617 | inst.reloc.type = BFD_RELOC_UNUSED; |
| 10618 | return; |
| 10619 | } |
| 10620 | else if (Rs == REG_PC) |
| 10621 | { |
| 10622 | /* Always use addw/subw. */ |
| 10623 | inst.instruction = add ? 0xf20f0000 : 0xf2af0000; |
| 10624 | inst.reloc.type = BFD_RELOC_ARM_T32_IMM12; |
| 10625 | } |
| 10626 | else |
| 10627 | { |
| 10628 | inst.instruction = THUMB_OP32 (inst.instruction); |
| 10629 | inst.instruction = (inst.instruction & 0xe1ffffff) |
| 10630 | | 0x10000000; |
| 10631 | if (flags) |
| 10632 | inst.reloc.type = BFD_RELOC_ARM_T32_IMMEDIATE; |
| 10633 | else |
| 10634 | inst.reloc.type = BFD_RELOC_ARM_T32_ADD_IMM; |
| 10635 | } |
| 10636 | inst.instruction |= Rd << 8; |
| 10637 | inst.instruction |= Rs << 16; |
| 10638 | } |
| 10639 | } |
| 10640 | else |
| 10641 | { |
| 10642 | unsigned int value = inst.reloc.exp.X_add_number; |
| 10643 | unsigned int shift = inst.operands[2].shift_kind; |
| 10644 | |
| 10645 | Rn = inst.operands[2].reg; |
| 10646 | /* See if we can do this with a 16-bit instruction. */ |
| 10647 | if (!inst.operands[2].shifted && inst.size_req != 4) |
| 10648 | { |
| 10649 | if (Rd > 7 || Rs > 7 || Rn > 7) |
| 10650 | narrow = FALSE; |
| 10651 | |
| 10652 | if (narrow) |
| 10653 | { |
| 10654 | inst.instruction = ((inst.instruction == T_MNEM_adds |
| 10655 | || inst.instruction == T_MNEM_add) |
| 10656 | ? T_OPCODE_ADD_R3 |
| 10657 | : T_OPCODE_SUB_R3); |
| 10658 | inst.instruction |= Rd | (Rs << 3) | (Rn << 6); |
| 10659 | return; |
| 10660 | } |
| 10661 | |
| 10662 | if (inst.instruction == T_MNEM_add && (Rd == Rs || Rd == Rn)) |
| 10663 | { |
| 10664 | /* Thumb-1 cores (except v6-M) require at least one high |
| 10665 | register in a narrow non flag setting add. */ |
| 10666 | if (Rd > 7 || Rn > 7 |
| 10667 | || ARM_CPU_HAS_FEATURE (selected_cpu, arm_ext_v6t2) |
| 10668 | || ARM_CPU_HAS_FEATURE (selected_cpu, arm_ext_msr)) |
| 10669 | { |
| 10670 | if (Rd == Rn) |
| 10671 | { |
| 10672 | Rn = Rs; |
| 10673 | Rs = Rd; |
| 10674 | } |
| 10675 | inst.instruction = T_OPCODE_ADD_HI; |
| 10676 | inst.instruction |= (Rd & 8) << 4; |
| 10677 | inst.instruction |= (Rd & 7); |
| 10678 | inst.instruction |= Rn << 3; |
| 10679 | return; |
| 10680 | } |
| 10681 | } |
| 10682 | } |
| 10683 | |
| 10684 | constraint (Rd == REG_PC, BAD_PC); |
| 10685 | if (!ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v8)) |
| 10686 | constraint (Rd == REG_SP && Rs != REG_SP, BAD_SP); |
| 10687 | constraint (Rs == REG_PC, BAD_PC); |
| 10688 | reject_bad_reg (Rn); |
| 10689 | |
| 10690 | /* If we get here, it can't be done in 16 bits. */ |
| 10691 | constraint (inst.operands[2].shifted && inst.operands[2].immisreg, |
| 10692 | _("shift must be constant")); |
| 10693 | inst.instruction = THUMB_OP32 (inst.instruction); |
| 10694 | inst.instruction |= Rd << 8; |
| 10695 | inst.instruction |= Rs << 16; |
| 10696 | constraint (Rd == REG_SP && Rs == REG_SP && value > 3, |
| 10697 | _("shift value over 3 not allowed in thumb mode")); |
| 10698 | constraint (Rd == REG_SP && Rs == REG_SP && shift != SHIFT_LSL, |
| 10699 | _("only LSL shift allowed in thumb mode")); |
| 10700 | encode_thumb32_shifted_operand (2); |
| 10701 | } |
| 10702 | } |
| 10703 | else |
| 10704 | { |
| 10705 | constraint (inst.instruction == T_MNEM_adds |
| 10706 | || inst.instruction == T_MNEM_subs, |
| 10707 | BAD_THUMB32); |
| 10708 | |
| 10709 | if (!inst.operands[2].isreg) /* Rd, Rs, #imm */ |
| 10710 | { |
| 10711 | constraint ((Rd > 7 && (Rd != REG_SP || Rs != REG_SP)) |
| 10712 | || (Rs > 7 && Rs != REG_SP && Rs != REG_PC), |
| 10713 | BAD_HIREG); |
| 10714 | |
| 10715 | inst.instruction = (inst.instruction == T_MNEM_add |
| 10716 | ? 0x0000 : 0x8000); |
| 10717 | inst.instruction |= (Rd << 4) | Rs; |
| 10718 | inst.reloc.type = BFD_RELOC_ARM_THUMB_ADD; |
| 10719 | return; |
| 10720 | } |
| 10721 | |
| 10722 | Rn = inst.operands[2].reg; |
| 10723 | constraint (inst.operands[2].shifted, _("unshifted register required")); |
| 10724 | |
| 10725 | /* We now have Rd, Rs, and Rn set to registers. */ |
| 10726 | if (Rd > 7 || Rs > 7 || Rn > 7) |
| 10727 | { |
| 10728 | /* Can't do this for SUB. */ |
| 10729 | constraint (inst.instruction == T_MNEM_sub, BAD_HIREG); |
| 10730 | inst.instruction = T_OPCODE_ADD_HI; |
| 10731 | inst.instruction |= (Rd & 8) << 4; |
| 10732 | inst.instruction |= (Rd & 7); |
| 10733 | if (Rs == Rd) |
| 10734 | inst.instruction |= Rn << 3; |
| 10735 | else if (Rn == Rd) |
| 10736 | inst.instruction |= Rs << 3; |
| 10737 | else |
| 10738 | constraint (1, _("dest must overlap one source register")); |
| 10739 | } |
| 10740 | else |
| 10741 | { |
| 10742 | inst.instruction = (inst.instruction == T_MNEM_add |
| 10743 | ? T_OPCODE_ADD_R3 : T_OPCODE_SUB_R3); |
| 10744 | inst.instruction |= Rd | (Rs << 3) | (Rn << 6); |
| 10745 | } |
| 10746 | } |
| 10747 | } |
| 10748 | |
| 10749 | static void |
| 10750 | do_t_adr (void) |
| 10751 | { |
| 10752 | unsigned Rd; |
| 10753 | |
| 10754 | Rd = inst.operands[0].reg; |
| 10755 | reject_bad_reg (Rd); |
| 10756 | |
| 10757 | if (unified_syntax && inst.size_req == 0 && Rd <= 7) |
| 10758 | { |
| 10759 | /* Defer to section relaxation. */ |
| 10760 | inst.relax = inst.instruction; |
| 10761 | inst.instruction = THUMB_OP16 (inst.instruction); |
| 10762 | inst.instruction |= Rd << 4; |
| 10763 | } |
| 10764 | else if (unified_syntax && inst.size_req != 2) |
| 10765 | { |
| 10766 | /* Generate a 32-bit opcode. */ |
| 10767 | inst.instruction = THUMB_OP32 (inst.instruction); |
| 10768 | inst.instruction |= Rd << 8; |
| 10769 | inst.reloc.type = BFD_RELOC_ARM_T32_ADD_PC12; |
| 10770 | inst.reloc.pc_rel = 1; |
| 10771 | } |
| 10772 | else |
| 10773 | { |
| 10774 | /* Generate a 16-bit opcode. */ |
| 10775 | inst.instruction = THUMB_OP16 (inst.instruction); |
| 10776 | inst.reloc.type = BFD_RELOC_ARM_THUMB_ADD; |
| 10777 | inst.reloc.exp.X_add_number -= 4; /* PC relative adjust. */ |
| 10778 | inst.reloc.pc_rel = 1; |
| 10779 | inst.instruction |= Rd << 4; |
| 10780 | } |
| 10781 | |
| 10782 | if (inst.reloc.exp.X_op == O_symbol |
| 10783 | && inst.reloc.exp.X_add_symbol != NULL |
| 10784 | && S_IS_DEFINED (inst.reloc.exp.X_add_symbol) |
| 10785 | && THUMB_IS_FUNC (inst.reloc.exp.X_add_symbol)) |
| 10786 | inst.reloc.exp.X_add_number += 1; |
| 10787 | } |
| 10788 | |
| 10789 | /* Arithmetic instructions for which there is just one 16-bit |
| 10790 | instruction encoding, and it allows only two low registers. |
| 10791 | For maximal compatibility with ARM syntax, we allow three register |
| 10792 | operands even when Thumb-32 instructions are not available, as long |
| 10793 | as the first two are identical. For instance, both "sbc r0,r1" and |
| 10794 | "sbc r0,r0,r1" are allowed. */ |
| 10795 | static void |
| 10796 | do_t_arit3 (void) |
| 10797 | { |
| 10798 | int Rd, Rs, Rn; |
| 10799 | |
| 10800 | Rd = inst.operands[0].reg; |
| 10801 | Rs = (inst.operands[1].present |
| 10802 | ? inst.operands[1].reg /* Rd, Rs, foo */ |
| 10803 | : inst.operands[0].reg); /* Rd, foo -> Rd, Rd, foo */ |
| 10804 | Rn = inst.operands[2].reg; |
| 10805 | |
| 10806 | reject_bad_reg (Rd); |
| 10807 | reject_bad_reg (Rs); |
| 10808 | if (inst.operands[2].isreg) |
| 10809 | reject_bad_reg (Rn); |
| 10810 | |
| 10811 | if (unified_syntax) |
| 10812 | { |
| 10813 | if (!inst.operands[2].isreg) |
| 10814 | { |
| 10815 | /* For an immediate, we always generate a 32-bit opcode; |
| 10816 | section relaxation will shrink it later if possible. */ |
| 10817 | inst.instruction = THUMB_OP32 (inst.instruction); |
| 10818 | inst.instruction = (inst.instruction & 0xe1ffffff) | 0x10000000; |
| 10819 | inst.instruction |= Rd << 8; |
| 10820 | inst.instruction |= Rs << 16; |
| 10821 | inst.reloc.type = BFD_RELOC_ARM_T32_IMMEDIATE; |
| 10822 | } |
| 10823 | else |
| 10824 | { |
| 10825 | bfd_boolean narrow; |
| 10826 | |
| 10827 | /* See if we can do this with a 16-bit instruction. */ |
| 10828 | if (THUMB_SETS_FLAGS (inst.instruction)) |
| 10829 | narrow = !in_it_block (); |
| 10830 | else |
| 10831 | narrow = in_it_block (); |
| 10832 | |
| 10833 | if (Rd > 7 || Rn > 7 || Rs > 7) |
| 10834 | narrow = FALSE; |
| 10835 | if (inst.operands[2].shifted) |
| 10836 | narrow = FALSE; |
| 10837 | if (inst.size_req == 4) |
| 10838 | narrow = FALSE; |
| 10839 | |
| 10840 | if (narrow |
| 10841 | && Rd == Rs) |
| 10842 | { |
| 10843 | inst.instruction = THUMB_OP16 (inst.instruction); |
| 10844 | inst.instruction |= Rd; |
| 10845 | inst.instruction |= Rn << 3; |
| 10846 | return; |
| 10847 | } |
| 10848 | |
| 10849 | /* If we get here, it can't be done in 16 bits. */ |
| 10850 | constraint (inst.operands[2].shifted |
| 10851 | && inst.operands[2].immisreg, |
| 10852 | _("shift must be constant")); |
| 10853 | inst.instruction = THUMB_OP32 (inst.instruction); |
| 10854 | inst.instruction |= Rd << 8; |
| 10855 | inst.instruction |= Rs << 16; |
| 10856 | encode_thumb32_shifted_operand (2); |
| 10857 | } |
| 10858 | } |
| 10859 | else |
| 10860 | { |
| 10861 | /* On its face this is a lie - the instruction does set the |
| 10862 | flags. However, the only supported mnemonic in this mode |
| 10863 | says it doesn't. */ |
| 10864 | constraint (THUMB_SETS_FLAGS (inst.instruction), BAD_THUMB32); |
| 10865 | |
| 10866 | constraint (!inst.operands[2].isreg || inst.operands[2].shifted, |
| 10867 | _("unshifted register required")); |
| 10868 | constraint (Rd > 7 || Rs > 7 || Rn > 7, BAD_HIREG); |
| 10869 | constraint (Rd != Rs, |
| 10870 | _("dest and source1 must be the same register")); |
| 10871 | |
| 10872 | inst.instruction = THUMB_OP16 (inst.instruction); |
| 10873 | inst.instruction |= Rd; |
| 10874 | inst.instruction |= Rn << 3; |
| 10875 | } |
| 10876 | } |
| 10877 | |
| 10878 | /* Similarly, but for instructions where the arithmetic operation is |
| 10879 | commutative, so we can allow either of them to be different from |
| 10880 | the destination operand in a 16-bit instruction. For instance, all |
| 10881 | three of "adc r0,r1", "adc r0,r0,r1", and "adc r0,r1,r0" are |
| 10882 | accepted. */ |
| 10883 | static void |
| 10884 | do_t_arit3c (void) |
| 10885 | { |
| 10886 | int Rd, Rs, Rn; |
| 10887 | |
| 10888 | Rd = inst.operands[0].reg; |
| 10889 | Rs = (inst.operands[1].present |
| 10890 | ? inst.operands[1].reg /* Rd, Rs, foo */ |
| 10891 | : inst.operands[0].reg); /* Rd, foo -> Rd, Rd, foo */ |
| 10892 | Rn = inst.operands[2].reg; |
| 10893 | |
| 10894 | reject_bad_reg (Rd); |
| 10895 | reject_bad_reg (Rs); |
| 10896 | if (inst.operands[2].isreg) |
| 10897 | reject_bad_reg (Rn); |
| 10898 | |
| 10899 | if (unified_syntax) |
| 10900 | { |
| 10901 | if (!inst.operands[2].isreg) |
| 10902 | { |
| 10903 | /* For an immediate, we always generate a 32-bit opcode; |
| 10904 | section relaxation will shrink it later if possible. */ |
| 10905 | inst.instruction = THUMB_OP32 (inst.instruction); |
| 10906 | inst.instruction = (inst.instruction & 0xe1ffffff) | 0x10000000; |
| 10907 | inst.instruction |= Rd << 8; |
| 10908 | inst.instruction |= Rs << 16; |
| 10909 | inst.reloc.type = BFD_RELOC_ARM_T32_IMMEDIATE; |
| 10910 | } |
| 10911 | else |
| 10912 | { |
| 10913 | bfd_boolean narrow; |
| 10914 | |
| 10915 | /* See if we can do this with a 16-bit instruction. */ |
| 10916 | if (THUMB_SETS_FLAGS (inst.instruction)) |
| 10917 | narrow = !in_it_block (); |
| 10918 | else |
| 10919 | narrow = in_it_block (); |
| 10920 | |
| 10921 | if (Rd > 7 || Rn > 7 || Rs > 7) |
| 10922 | narrow = FALSE; |
| 10923 | if (inst.operands[2].shifted) |
| 10924 | narrow = FALSE; |
| 10925 | if (inst.size_req == 4) |
| 10926 | narrow = FALSE; |
| 10927 | |
| 10928 | if (narrow) |
| 10929 | { |
| 10930 | if (Rd == Rs) |
| 10931 | { |
| 10932 | inst.instruction = THUMB_OP16 (inst.instruction); |
| 10933 | inst.instruction |= Rd; |
| 10934 | inst.instruction |= Rn << 3; |
| 10935 | return; |
| 10936 | } |
| 10937 | if (Rd == Rn) |
| 10938 | { |
| 10939 | inst.instruction = THUMB_OP16 (inst.instruction); |
| 10940 | inst.instruction |= Rd; |
| 10941 | inst.instruction |= Rs << 3; |
| 10942 | return; |
| 10943 | } |
| 10944 | } |
| 10945 | |
| 10946 | /* If we get here, it can't be done in 16 bits. */ |
| 10947 | constraint (inst.operands[2].shifted |
| 10948 | && inst.operands[2].immisreg, |
| 10949 | _("shift must be constant")); |
| 10950 | inst.instruction = THUMB_OP32 (inst.instruction); |
| 10951 | inst.instruction |= Rd << 8; |
| 10952 | inst.instruction |= Rs << 16; |
| 10953 | encode_thumb32_shifted_operand (2); |
| 10954 | } |
| 10955 | } |
| 10956 | else |
| 10957 | { |
| 10958 | /* On its face this is a lie - the instruction does set the |
| 10959 | flags. However, the only supported mnemonic in this mode |
| 10960 | says it doesn't. */ |
| 10961 | constraint (THUMB_SETS_FLAGS (inst.instruction), BAD_THUMB32); |
| 10962 | |
| 10963 | constraint (!inst.operands[2].isreg || inst.operands[2].shifted, |
| 10964 | _("unshifted register required")); |
| 10965 | constraint (Rd > 7 || Rs > 7 || Rn > 7, BAD_HIREG); |
| 10966 | |
| 10967 | inst.instruction = THUMB_OP16 (inst.instruction); |
| 10968 | inst.instruction |= Rd; |
| 10969 | |
| 10970 | if (Rd == Rs) |
| 10971 | inst.instruction |= Rn << 3; |
| 10972 | else if (Rd == Rn) |
| 10973 | inst.instruction |= Rs << 3; |
| 10974 | else |
| 10975 | constraint (1, _("dest must overlap one source register")); |
| 10976 | } |
| 10977 | } |
| 10978 | |
| 10979 | static void |
| 10980 | do_t_bfc (void) |
| 10981 | { |
| 10982 | unsigned Rd; |
| 10983 | unsigned int msb = inst.operands[1].imm + inst.operands[2].imm; |
| 10984 | constraint (msb > 32, _("bit-field extends past end of register")); |
| 10985 | /* The instruction encoding stores the LSB and MSB, |
| 10986 | not the LSB and width. */ |
| 10987 | Rd = inst.operands[0].reg; |
| 10988 | reject_bad_reg (Rd); |
| 10989 | inst.instruction |= Rd << 8; |
| 10990 | inst.instruction |= (inst.operands[1].imm & 0x1c) << 10; |
| 10991 | inst.instruction |= (inst.operands[1].imm & 0x03) << 6; |
| 10992 | inst.instruction |= msb - 1; |
| 10993 | } |
| 10994 | |
| 10995 | static void |
| 10996 | do_t_bfi (void) |
| 10997 | { |
| 10998 | int Rd, Rn; |
| 10999 | unsigned int msb; |
| 11000 | |
| 11001 | Rd = inst.operands[0].reg; |
| 11002 | reject_bad_reg (Rd); |
| 11003 | |
| 11004 | /* #0 in second position is alternative syntax for bfc, which is |
| 11005 | the same instruction but with REG_PC in the Rm field. */ |
| 11006 | if (!inst.operands[1].isreg) |
| 11007 | Rn = REG_PC; |
| 11008 | else |
| 11009 | { |
| 11010 | Rn = inst.operands[1].reg; |
| 11011 | reject_bad_reg (Rn); |
| 11012 | } |
| 11013 | |
| 11014 | msb = inst.operands[2].imm + inst.operands[3].imm; |
| 11015 | constraint (msb > 32, _("bit-field extends past end of register")); |
| 11016 | /* The instruction encoding stores the LSB and MSB, |
| 11017 | not the LSB and width. */ |
| 11018 | inst.instruction |= Rd << 8; |
| 11019 | inst.instruction |= Rn << 16; |
| 11020 | inst.instruction |= (inst.operands[2].imm & 0x1c) << 10; |
| 11021 | inst.instruction |= (inst.operands[2].imm & 0x03) << 6; |
| 11022 | inst.instruction |= msb - 1; |
| 11023 | } |
| 11024 | |
| 11025 | static void |
| 11026 | do_t_bfx (void) |
| 11027 | { |
| 11028 | unsigned Rd, Rn; |
| 11029 | |
| 11030 | Rd = inst.operands[0].reg; |
| 11031 | Rn = inst.operands[1].reg; |
| 11032 | |
| 11033 | reject_bad_reg (Rd); |
| 11034 | reject_bad_reg (Rn); |
| 11035 | |
| 11036 | constraint (inst.operands[2].imm + inst.operands[3].imm > 32, |
| 11037 | _("bit-field extends past end of register")); |
| 11038 | inst.instruction |= Rd << 8; |
| 11039 | inst.instruction |= Rn << 16; |
| 11040 | inst.instruction |= (inst.operands[2].imm & 0x1c) << 10; |
| 11041 | inst.instruction |= (inst.operands[2].imm & 0x03) << 6; |
| 11042 | inst.instruction |= inst.operands[3].imm - 1; |
| 11043 | } |
| 11044 | |
| 11045 | /* ARM V5 Thumb BLX (argument parse) |
| 11046 | BLX <target_addr> which is BLX(1) |
| 11047 | BLX <Rm> which is BLX(2) |
| 11048 | Unfortunately, there are two different opcodes for this mnemonic. |
| 11049 | So, the insns[].value is not used, and the code here zaps values |
| 11050 | into inst.instruction. |
| 11051 | |
| 11052 | ??? How to take advantage of the additional two bits of displacement |
| 11053 | available in Thumb32 mode? Need new relocation? */ |
| 11054 | |
| 11055 | static void |
| 11056 | do_t_blx (void) |
| 11057 | { |
| 11058 | set_it_insn_type_last (); |
| 11059 | |
| 11060 | if (inst.operands[0].isreg) |
| 11061 | { |
| 11062 | constraint (inst.operands[0].reg == REG_PC, BAD_PC); |
| 11063 | /* We have a register, so this is BLX(2). */ |
| 11064 | inst.instruction |= inst.operands[0].reg << 3; |
| 11065 | } |
| 11066 | else |
| 11067 | { |
| 11068 | /* No register. This must be BLX(1). */ |
| 11069 | inst.instruction = 0xf000e800; |
| 11070 | encode_branch (BFD_RELOC_THUMB_PCREL_BLX); |
| 11071 | } |
| 11072 | } |
| 11073 | |
| 11074 | static void |
| 11075 | do_t_branch (void) |
| 11076 | { |
| 11077 | int opcode; |
| 11078 | int cond; |
| 11079 | bfd_reloc_code_real_type reloc; |
| 11080 | |
| 11081 | cond = inst.cond; |
| 11082 | set_it_insn_type (IF_INSIDE_IT_LAST_INSN); |
| 11083 | |
| 11084 | if (in_it_block ()) |
| 11085 | { |
| 11086 | /* Conditional branches inside IT blocks are encoded as unconditional |
| 11087 | branches. */ |
| 11088 | cond = COND_ALWAYS; |
| 11089 | } |
| 11090 | else |
| 11091 | cond = inst.cond; |
| 11092 | |
| 11093 | if (cond != COND_ALWAYS) |
| 11094 | opcode = T_MNEM_bcond; |
| 11095 | else |
| 11096 | opcode = inst.instruction; |
| 11097 | |
| 11098 | if (unified_syntax |
| 11099 | && (inst.size_req == 4 |
| 11100 | || (inst.size_req != 2 |
| 11101 | && (inst.operands[0].hasreloc |
| 11102 | || inst.reloc.exp.X_op == O_constant)))) |
| 11103 | { |
| 11104 | inst.instruction = THUMB_OP32(opcode); |
| 11105 | if (cond == COND_ALWAYS) |
| 11106 | reloc = BFD_RELOC_THUMB_PCREL_BRANCH25; |
| 11107 | else |
| 11108 | { |
| 11109 | constraint (!ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v6t2), |
| 11110 | _("selected architecture does not support " |
| 11111 | "wide conditional branch instruction")); |
| 11112 | |
| 11113 | gas_assert (cond != 0xF); |
| 11114 | inst.instruction |= cond << 22; |
| 11115 | reloc = BFD_RELOC_THUMB_PCREL_BRANCH20; |
| 11116 | } |
| 11117 | } |
| 11118 | else |
| 11119 | { |
| 11120 | inst.instruction = THUMB_OP16(opcode); |
| 11121 | if (cond == COND_ALWAYS) |
| 11122 | reloc = BFD_RELOC_THUMB_PCREL_BRANCH12; |
| 11123 | else |
| 11124 | { |
| 11125 | inst.instruction |= cond << 8; |
| 11126 | reloc = BFD_RELOC_THUMB_PCREL_BRANCH9; |
| 11127 | } |
| 11128 | /* Allow section relaxation. */ |
| 11129 | if (unified_syntax && inst.size_req != 2) |
| 11130 | inst.relax = opcode; |
| 11131 | } |
| 11132 | inst.reloc.type = reloc; |
| 11133 | inst.reloc.pc_rel = 1; |
| 11134 | } |
| 11135 | |
| 11136 | /* Actually do the work for Thumb state bkpt and hlt. The only difference |
| 11137 | between the two is the maximum immediate allowed - which is passed in |
| 11138 | RANGE. */ |
| 11139 | static void |
| 11140 | do_t_bkpt_hlt1 (int range) |
| 11141 | { |
| 11142 | constraint (inst.cond != COND_ALWAYS, |
| 11143 | _("instruction is always unconditional")); |
| 11144 | if (inst.operands[0].present) |
| 11145 | { |
| 11146 | constraint (inst.operands[0].imm > range, |
| 11147 | _("immediate value out of range")); |
| 11148 | inst.instruction |= inst.operands[0].imm; |
| 11149 | } |
| 11150 | |
| 11151 | set_it_insn_type (NEUTRAL_IT_INSN); |
| 11152 | } |
| 11153 | |
| 11154 | static void |
| 11155 | do_t_hlt (void) |
| 11156 | { |
| 11157 | do_t_bkpt_hlt1 (63); |
| 11158 | } |
| 11159 | |
| 11160 | static void |
| 11161 | do_t_bkpt (void) |
| 11162 | { |
| 11163 | do_t_bkpt_hlt1 (255); |
| 11164 | } |
| 11165 | |
| 11166 | static void |
| 11167 | do_t_branch23 (void) |
| 11168 | { |
| 11169 | set_it_insn_type_last (); |
| 11170 | encode_branch (BFD_RELOC_THUMB_PCREL_BRANCH23); |
| 11171 | |
| 11172 | /* md_apply_fix blows up with 'bl foo(PLT)' where foo is defined in |
| 11173 | this file. We used to simply ignore the PLT reloc type here -- |
| 11174 | the branch encoding is now needed to deal with TLSCALL relocs. |
| 11175 | So if we see a PLT reloc now, put it back to how it used to be to |
| 11176 | keep the preexisting behaviour. */ |
| 11177 | if (inst.reloc.type == BFD_RELOC_ARM_PLT32) |
| 11178 | inst.reloc.type = BFD_RELOC_THUMB_PCREL_BRANCH23; |
| 11179 | |
| 11180 | #if defined(OBJ_COFF) |
| 11181 | /* If the destination of the branch is a defined symbol which does not have |
| 11182 | the THUMB_FUNC attribute, then we must be calling a function which has |
| 11183 | the (interfacearm) attribute. We look for the Thumb entry point to that |
| 11184 | function and change the branch to refer to that function instead. */ |
| 11185 | if ( inst.reloc.exp.X_op == O_symbol |
| 11186 | && inst.reloc.exp.X_add_symbol != NULL |
| 11187 | && S_IS_DEFINED (inst.reloc.exp.X_add_symbol) |
| 11188 | && ! THUMB_IS_FUNC (inst.reloc.exp.X_add_symbol)) |
| 11189 | inst.reloc.exp.X_add_symbol = |
| 11190 | find_real_start (inst.reloc.exp.X_add_symbol); |
| 11191 | #endif |
| 11192 | } |
| 11193 | |
| 11194 | static void |
| 11195 | do_t_bx (void) |
| 11196 | { |
| 11197 | set_it_insn_type_last (); |
| 11198 | inst.instruction |= inst.operands[0].reg << 3; |
| 11199 | /* ??? FIXME: Should add a hacky reloc here if reg is REG_PC. The reloc |
| 11200 | should cause the alignment to be checked once it is known. This is |
| 11201 | because BX PC only works if the instruction is word aligned. */ |
| 11202 | } |
| 11203 | |
| 11204 | static void |
| 11205 | do_t_bxj (void) |
| 11206 | { |
| 11207 | int Rm; |
| 11208 | |
| 11209 | set_it_insn_type_last (); |
| 11210 | Rm = inst.operands[0].reg; |
| 11211 | reject_bad_reg (Rm); |
| 11212 | inst.instruction |= Rm << 16; |
| 11213 | } |
| 11214 | |
| 11215 | static void |
| 11216 | do_t_clz (void) |
| 11217 | { |
| 11218 | unsigned Rd; |
| 11219 | unsigned Rm; |
| 11220 | |
| 11221 | Rd = inst.operands[0].reg; |
| 11222 | Rm = inst.operands[1].reg; |
| 11223 | |
| 11224 | reject_bad_reg (Rd); |
| 11225 | reject_bad_reg (Rm); |
| 11226 | |
| 11227 | inst.instruction |= Rd << 8; |
| 11228 | inst.instruction |= Rm << 16; |
| 11229 | inst.instruction |= Rm; |
| 11230 | } |
| 11231 | |
| 11232 | static void |
| 11233 | do_t_cps (void) |
| 11234 | { |
| 11235 | set_it_insn_type (OUTSIDE_IT_INSN); |
| 11236 | inst.instruction |= inst.operands[0].imm; |
| 11237 | } |
| 11238 | |
| 11239 | static void |
| 11240 | do_t_cpsi (void) |
| 11241 | { |
| 11242 | set_it_insn_type (OUTSIDE_IT_INSN); |
| 11243 | if (unified_syntax |
| 11244 | && (inst.operands[1].present || inst.size_req == 4) |
| 11245 | && ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v6_notm)) |
| 11246 | { |
| 11247 | unsigned int imod = (inst.instruction & 0x0030) >> 4; |
| 11248 | inst.instruction = 0xf3af8000; |
| 11249 | inst.instruction |= imod << 9; |
| 11250 | inst.instruction |= inst.operands[0].imm << 5; |
| 11251 | if (inst.operands[1].present) |
| 11252 | inst.instruction |= 0x100 | inst.operands[1].imm; |
| 11253 | } |
| 11254 | else |
| 11255 | { |
| 11256 | constraint (!ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v1) |
| 11257 | && (inst.operands[0].imm & 4), |
| 11258 | _("selected processor does not support 'A' form " |
| 11259 | "of this instruction")); |
| 11260 | constraint (inst.operands[1].present || inst.size_req == 4, |
| 11261 | _("Thumb does not support the 2-argument " |
| 11262 | "form of this instruction")); |
| 11263 | inst.instruction |= inst.operands[0].imm; |
| 11264 | } |
| 11265 | } |
| 11266 | |
| 11267 | /* THUMB CPY instruction (argument parse). */ |
| 11268 | |
| 11269 | static void |
| 11270 | do_t_cpy (void) |
| 11271 | { |
| 11272 | if (inst.size_req == 4) |
| 11273 | { |
| 11274 | inst.instruction = THUMB_OP32 (T_MNEM_mov); |
| 11275 | inst.instruction |= inst.operands[0].reg << 8; |
| 11276 | inst.instruction |= inst.operands[1].reg; |
| 11277 | } |
| 11278 | else |
| 11279 | { |
| 11280 | inst.instruction |= (inst.operands[0].reg & 0x8) << 4; |
| 11281 | inst.instruction |= (inst.operands[0].reg & 0x7); |
| 11282 | inst.instruction |= inst.operands[1].reg << 3; |
| 11283 | } |
| 11284 | } |
| 11285 | |
| 11286 | static void |
| 11287 | do_t_cbz (void) |
| 11288 | { |
| 11289 | set_it_insn_type (OUTSIDE_IT_INSN); |
| 11290 | constraint (inst.operands[0].reg > 7, BAD_HIREG); |
| 11291 | inst.instruction |= inst.operands[0].reg; |
| 11292 | inst.reloc.pc_rel = 1; |
| 11293 | inst.reloc.type = BFD_RELOC_THUMB_PCREL_BRANCH7; |
| 11294 | } |
| 11295 | |
| 11296 | static void |
| 11297 | do_t_dbg (void) |
| 11298 | { |
| 11299 | inst.instruction |= inst.operands[0].imm; |
| 11300 | } |
| 11301 | |
| 11302 | static void |
| 11303 | do_t_div (void) |
| 11304 | { |
| 11305 | unsigned Rd, Rn, Rm; |
| 11306 | |
| 11307 | Rd = inst.operands[0].reg; |
| 11308 | Rn = (inst.operands[1].present |
| 11309 | ? inst.operands[1].reg : Rd); |
| 11310 | Rm = inst.operands[2].reg; |
| 11311 | |
| 11312 | reject_bad_reg (Rd); |
| 11313 | reject_bad_reg (Rn); |
| 11314 | reject_bad_reg (Rm); |
| 11315 | |
| 11316 | inst.instruction |= Rd << 8; |
| 11317 | inst.instruction |= Rn << 16; |
| 11318 | inst.instruction |= Rm; |
| 11319 | } |
| 11320 | |
| 11321 | static void |
| 11322 | do_t_hint (void) |
| 11323 | { |
| 11324 | if (unified_syntax && inst.size_req == 4) |
| 11325 | inst.instruction = THUMB_OP32 (inst.instruction); |
| 11326 | else |
| 11327 | inst.instruction = THUMB_OP16 (inst.instruction); |
| 11328 | } |
| 11329 | |
| 11330 | static void |
| 11331 | do_t_it (void) |
| 11332 | { |
| 11333 | unsigned int cond = inst.operands[0].imm; |
| 11334 | |
| 11335 | set_it_insn_type (IT_INSN); |
| 11336 | now_it.mask = (inst.instruction & 0xf) | 0x10; |
| 11337 | now_it.cc = cond; |
| 11338 | now_it.warn_deprecated = FALSE; |
| 11339 | |
| 11340 | /* If the condition is a negative condition, invert the mask. */ |
| 11341 | if ((cond & 0x1) == 0x0) |
| 11342 | { |
| 11343 | unsigned int mask = inst.instruction & 0x000f; |
| 11344 | |
| 11345 | if ((mask & 0x7) == 0) |
| 11346 | { |
| 11347 | /* No conversion needed. */ |
| 11348 | now_it.block_length = 1; |
| 11349 | } |
| 11350 | else if ((mask & 0x3) == 0) |
| 11351 | { |
| 11352 | mask ^= 0x8; |
| 11353 | now_it.block_length = 2; |
| 11354 | } |
| 11355 | else if ((mask & 0x1) == 0) |
| 11356 | { |
| 11357 | mask ^= 0xC; |
| 11358 | now_it.block_length = 3; |
| 11359 | } |
| 11360 | else |
| 11361 | { |
| 11362 | mask ^= 0xE; |
| 11363 | now_it.block_length = 4; |
| 11364 | } |
| 11365 | |
| 11366 | inst.instruction &= 0xfff0; |
| 11367 | inst.instruction |= mask; |
| 11368 | } |
| 11369 | |
| 11370 | inst.instruction |= cond << 4; |
| 11371 | } |
| 11372 | |
| 11373 | /* Helper function used for both push/pop and ldm/stm. */ |
| 11374 | static void |
| 11375 | encode_thumb2_ldmstm (int base, unsigned mask, bfd_boolean writeback) |
| 11376 | { |
| 11377 | bfd_boolean load; |
| 11378 | |
| 11379 | load = (inst.instruction & (1 << 20)) != 0; |
| 11380 | |
| 11381 | if (mask & (1 << 13)) |
| 11382 | inst.error = _("SP not allowed in register list"); |
| 11383 | |
| 11384 | if ((mask & (1 << base)) != 0 |
| 11385 | && writeback) |
| 11386 | inst.error = _("having the base register in the register list when " |
| 11387 | "using write back is UNPREDICTABLE"); |
| 11388 | |
| 11389 | if (load) |
| 11390 | { |
| 11391 | if (mask & (1 << 15)) |
| 11392 | { |
| 11393 | if (mask & (1 << 14)) |
| 11394 | inst.error = _("LR and PC should not both be in register list"); |
| 11395 | else |
| 11396 | set_it_insn_type_last (); |
| 11397 | } |
| 11398 | } |
| 11399 | else |
| 11400 | { |
| 11401 | if (mask & (1 << 15)) |
| 11402 | inst.error = _("PC not allowed in register list"); |
| 11403 | } |
| 11404 | |
| 11405 | if ((mask & (mask - 1)) == 0) |
| 11406 | { |
| 11407 | /* Single register transfers implemented as str/ldr. */ |
| 11408 | if (writeback) |
| 11409 | { |
| 11410 | if (inst.instruction & (1 << 23)) |
| 11411 | inst.instruction = 0x00000b04; /* ia! -> [base], #4 */ |
| 11412 | else |
| 11413 | inst.instruction = 0x00000d04; /* db! -> [base, #-4]! */ |
| 11414 | } |
| 11415 | else |
| 11416 | { |
| 11417 | if (inst.instruction & (1 << 23)) |
| 11418 | inst.instruction = 0x00800000; /* ia -> [base] */ |
| 11419 | else |
| 11420 | inst.instruction = 0x00000c04; /* db -> [base, #-4] */ |
| 11421 | } |
| 11422 | |
| 11423 | inst.instruction |= 0xf8400000; |
| 11424 | if (load) |
| 11425 | inst.instruction |= 0x00100000; |
| 11426 | |
| 11427 | mask = ffs (mask) - 1; |
| 11428 | mask <<= 12; |
| 11429 | } |
| 11430 | else if (writeback) |
| 11431 | inst.instruction |= WRITE_BACK; |
| 11432 | |
| 11433 | inst.instruction |= mask; |
| 11434 | inst.instruction |= base << 16; |
| 11435 | } |
| 11436 | |
| 11437 | static void |
| 11438 | do_t_ldmstm (void) |
| 11439 | { |
| 11440 | /* This really doesn't seem worth it. */ |
| 11441 | constraint (inst.reloc.type != BFD_RELOC_UNUSED, |
| 11442 | _("expression too complex")); |
| 11443 | constraint (inst.operands[1].writeback, |
| 11444 | _("Thumb load/store multiple does not support {reglist}^")); |
| 11445 | |
| 11446 | if (unified_syntax) |
| 11447 | { |
| 11448 | bfd_boolean narrow; |
| 11449 | unsigned mask; |
| 11450 | |
| 11451 | narrow = FALSE; |
| 11452 | /* See if we can use a 16-bit instruction. */ |
| 11453 | if (inst.instruction < 0xffff /* not ldmdb/stmdb */ |
| 11454 | && inst.size_req != 4 |
| 11455 | && !(inst.operands[1].imm & ~0xff)) |
| 11456 | { |
| 11457 | mask = 1 << inst.operands[0].reg; |
| 11458 | |
| 11459 | if (inst.operands[0].reg <= 7) |
| 11460 | { |
| 11461 | if (inst.instruction == T_MNEM_stmia |
| 11462 | ? inst.operands[0].writeback |
| 11463 | : (inst.operands[0].writeback |
| 11464 | == !(inst.operands[1].imm & mask))) |
| 11465 | { |
| 11466 | if (inst.instruction == T_MNEM_stmia |
| 11467 | && (inst.operands[1].imm & mask) |
| 11468 | && (inst.operands[1].imm & (mask - 1))) |
| 11469 | as_warn (_("value stored for r%d is UNKNOWN"), |
| 11470 | inst.operands[0].reg); |
| 11471 | |
| 11472 | inst.instruction = THUMB_OP16 (inst.instruction); |
| 11473 | inst.instruction |= inst.operands[0].reg << 8; |
| 11474 | inst.instruction |= inst.operands[1].imm; |
| 11475 | narrow = TRUE; |
| 11476 | } |
| 11477 | else if ((inst.operands[1].imm & (inst.operands[1].imm-1)) == 0) |
| 11478 | { |
| 11479 | /* This means 1 register in reg list one of 3 situations: |
| 11480 | 1. Instruction is stmia, but without writeback. |
| 11481 | 2. lmdia without writeback, but with Rn not in |
| 11482 | reglist. |
| 11483 | 3. ldmia with writeback, but with Rn in reglist. |
| 11484 | Case 3 is UNPREDICTABLE behaviour, so we handle |
| 11485 | case 1 and 2 which can be converted into a 16-bit |
| 11486 | str or ldr. The SP cases are handled below. */ |
| 11487 | unsigned long opcode; |
| 11488 | /* First, record an error for Case 3. */ |
| 11489 | if (inst.operands[1].imm & mask |
| 11490 | && inst.operands[0].writeback) |
| 11491 | inst.error = |
| 11492 | _("having the base register in the register list when " |
| 11493 | "using write back is UNPREDICTABLE"); |
| 11494 | |
| 11495 | opcode = (inst.instruction == T_MNEM_stmia ? T_MNEM_str |
| 11496 | : T_MNEM_ldr); |
| 11497 | inst.instruction = THUMB_OP16 (opcode); |
| 11498 | inst.instruction |= inst.operands[0].reg << 3; |
| 11499 | inst.instruction |= (ffs (inst.operands[1].imm)-1); |
| 11500 | narrow = TRUE; |
| 11501 | } |
| 11502 | } |
| 11503 | else if (inst.operands[0] .reg == REG_SP) |
| 11504 | { |
| 11505 | if (inst.operands[0].writeback) |
| 11506 | { |
| 11507 | inst.instruction = |
| 11508 | THUMB_OP16 (inst.instruction == T_MNEM_stmia |
| 11509 | ? T_MNEM_push : T_MNEM_pop); |
| 11510 | inst.instruction |= inst.operands[1].imm; |
| 11511 | narrow = TRUE; |
| 11512 | } |
| 11513 | else if ((inst.operands[1].imm & (inst.operands[1].imm-1)) == 0) |
| 11514 | { |
| 11515 | inst.instruction = |
| 11516 | THUMB_OP16 (inst.instruction == T_MNEM_stmia |
| 11517 | ? T_MNEM_str_sp : T_MNEM_ldr_sp); |
| 11518 | inst.instruction |= ((ffs (inst.operands[1].imm)-1) << 8); |
| 11519 | narrow = TRUE; |
| 11520 | } |
| 11521 | } |
| 11522 | } |
| 11523 | |
| 11524 | if (!narrow) |
| 11525 | { |
| 11526 | if (inst.instruction < 0xffff) |
| 11527 | inst.instruction = THUMB_OP32 (inst.instruction); |
| 11528 | |
| 11529 | encode_thumb2_ldmstm (inst.operands[0].reg, inst.operands[1].imm, |
| 11530 | inst.operands[0].writeback); |
| 11531 | } |
| 11532 | } |
| 11533 | else |
| 11534 | { |
| 11535 | constraint (inst.operands[0].reg > 7 |
| 11536 | || (inst.operands[1].imm & ~0xff), BAD_HIREG); |
| 11537 | constraint (inst.instruction != T_MNEM_ldmia |
| 11538 | && inst.instruction != T_MNEM_stmia, |
| 11539 | _("Thumb-2 instruction only valid in unified syntax")); |
| 11540 | if (inst.instruction == T_MNEM_stmia) |
| 11541 | { |
| 11542 | if (!inst.operands[0].writeback) |
| 11543 | as_warn (_("this instruction will write back the base register")); |
| 11544 | if ((inst.operands[1].imm & (1 << inst.operands[0].reg)) |
| 11545 | && (inst.operands[1].imm & ((1 << inst.operands[0].reg) - 1))) |
| 11546 | as_warn (_("value stored for r%d is UNKNOWN"), |
| 11547 | inst.operands[0].reg); |
| 11548 | } |
| 11549 | else |
| 11550 | { |
| 11551 | if (!inst.operands[0].writeback |
| 11552 | && !(inst.operands[1].imm & (1 << inst.operands[0].reg))) |
| 11553 | as_warn (_("this instruction will write back the base register")); |
| 11554 | else if (inst.operands[0].writeback |
| 11555 | && (inst.operands[1].imm & (1 << inst.operands[0].reg))) |
| 11556 | as_warn (_("this instruction will not write back the base register")); |
| 11557 | } |
| 11558 | |
| 11559 | inst.instruction = THUMB_OP16 (inst.instruction); |
| 11560 | inst.instruction |= inst.operands[0].reg << 8; |
| 11561 | inst.instruction |= inst.operands[1].imm; |
| 11562 | } |
| 11563 | } |
| 11564 | |
| 11565 | static void |
| 11566 | do_t_ldrex (void) |
| 11567 | { |
| 11568 | constraint (!inst.operands[1].isreg || !inst.operands[1].preind |
| 11569 | || inst.operands[1].postind || inst.operands[1].writeback |
| 11570 | || inst.operands[1].immisreg || inst.operands[1].shifted |
| 11571 | || inst.operands[1].negative, |
| 11572 | BAD_ADDR_MODE); |
| 11573 | |
| 11574 | constraint ((inst.operands[1].reg == REG_PC), BAD_PC); |
| 11575 | |
| 11576 | inst.instruction |= inst.operands[0].reg << 12; |
| 11577 | inst.instruction |= inst.operands[1].reg << 16; |
| 11578 | inst.reloc.type = BFD_RELOC_ARM_T32_OFFSET_U8; |
| 11579 | } |
| 11580 | |
| 11581 | static void |
| 11582 | do_t_ldrexd (void) |
| 11583 | { |
| 11584 | if (!inst.operands[1].present) |
| 11585 | { |
| 11586 | constraint (inst.operands[0].reg == REG_LR, |
| 11587 | _("r14 not allowed as first register " |
| 11588 | "when second register is omitted")); |
| 11589 | inst.operands[1].reg = inst.operands[0].reg + 1; |
| 11590 | } |
| 11591 | constraint (inst.operands[0].reg == inst.operands[1].reg, |
| 11592 | BAD_OVERLAP); |
| 11593 | |
| 11594 | inst.instruction |= inst.operands[0].reg << 12; |
| 11595 | inst.instruction |= inst.operands[1].reg << 8; |
| 11596 | inst.instruction |= inst.operands[2].reg << 16; |
| 11597 | } |
| 11598 | |
| 11599 | static void |
| 11600 | do_t_ldst (void) |
| 11601 | { |
| 11602 | unsigned long opcode; |
| 11603 | int Rn; |
| 11604 | |
| 11605 | if (inst.operands[0].isreg |
| 11606 | && !inst.operands[0].preind |
| 11607 | && inst.operands[0].reg == REG_PC) |
| 11608 | set_it_insn_type_last (); |
| 11609 | |
| 11610 | opcode = inst.instruction; |
| 11611 | if (unified_syntax) |
| 11612 | { |
| 11613 | if (!inst.operands[1].isreg) |
| 11614 | { |
| 11615 | if (opcode <= 0xffff) |
| 11616 | inst.instruction = THUMB_OP32 (opcode); |
| 11617 | if (move_or_literal_pool (0, CONST_THUMB, /*mode_3=*/FALSE)) |
| 11618 | return; |
| 11619 | } |
| 11620 | if (inst.operands[1].isreg |
| 11621 | && !inst.operands[1].writeback |
| 11622 | && !inst.operands[1].shifted && !inst.operands[1].postind |
| 11623 | && !inst.operands[1].negative && inst.operands[0].reg <= 7 |
| 11624 | && opcode <= 0xffff |
| 11625 | && inst.size_req != 4) |
| 11626 | { |
| 11627 | /* Insn may have a 16-bit form. */ |
| 11628 | Rn = inst.operands[1].reg; |
| 11629 | if (inst.operands[1].immisreg) |
| 11630 | { |
| 11631 | inst.instruction = THUMB_OP16 (opcode); |
| 11632 | /* [Rn, Rik] */ |
| 11633 | if (Rn <= 7 && inst.operands[1].imm <= 7) |
| 11634 | goto op16; |
| 11635 | else if (opcode != T_MNEM_ldr && opcode != T_MNEM_str) |
| 11636 | reject_bad_reg (inst.operands[1].imm); |
| 11637 | } |
| 11638 | else if ((Rn <= 7 && opcode != T_MNEM_ldrsh |
| 11639 | && opcode != T_MNEM_ldrsb) |
| 11640 | || ((Rn == REG_PC || Rn == REG_SP) && opcode == T_MNEM_ldr) |
| 11641 | || (Rn == REG_SP && opcode == T_MNEM_str)) |
| 11642 | { |
| 11643 | /* [Rn, #const] */ |
| 11644 | if (Rn > 7) |
| 11645 | { |
| 11646 | if (Rn == REG_PC) |
| 11647 | { |
| 11648 | if (inst.reloc.pc_rel) |
| 11649 | opcode = T_MNEM_ldr_pc2; |
| 11650 | else |
| 11651 | opcode = T_MNEM_ldr_pc; |
| 11652 | } |
| 11653 | else |
| 11654 | { |
| 11655 | if (opcode == T_MNEM_ldr) |
| 11656 | opcode = T_MNEM_ldr_sp; |
| 11657 | else |
| 11658 | opcode = T_MNEM_str_sp; |
| 11659 | } |
| 11660 | inst.instruction = inst.operands[0].reg << 8; |
| 11661 | } |
| 11662 | else |
| 11663 | { |
| 11664 | inst.instruction = inst.operands[0].reg; |
| 11665 | inst.instruction |= inst.operands[1].reg << 3; |
| 11666 | } |
| 11667 | inst.instruction |= THUMB_OP16 (opcode); |
| 11668 | if (inst.size_req == 2) |
| 11669 | inst.reloc.type = BFD_RELOC_ARM_THUMB_OFFSET; |
| 11670 | else |
| 11671 | inst.relax = opcode; |
| 11672 | return; |
| 11673 | } |
| 11674 | } |
| 11675 | /* Definitely a 32-bit variant. */ |
| 11676 | |
| 11677 | /* Warning for Erratum 752419. */ |
| 11678 | if (opcode == T_MNEM_ldr |
| 11679 | && inst.operands[0].reg == REG_SP |
| 11680 | && inst.operands[1].writeback == 1 |
| 11681 | && !inst.operands[1].immisreg) |
| 11682 | { |
| 11683 | if (no_cpu_selected () |
| 11684 | || (ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v7) |
| 11685 | && !ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v7a) |
| 11686 | && !ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v7r))) |
| 11687 | as_warn (_("This instruction may be unpredictable " |
| 11688 | "if executed on M-profile cores " |
| 11689 | "with interrupts enabled.")); |
| 11690 | } |
| 11691 | |
| 11692 | /* Do some validations regarding addressing modes. */ |
| 11693 | if (inst.operands[1].immisreg) |
| 11694 | reject_bad_reg (inst.operands[1].imm); |
| 11695 | |
| 11696 | constraint (inst.operands[1].writeback == 1 |
| 11697 | && inst.operands[0].reg == inst.operands[1].reg, |
| 11698 | BAD_OVERLAP); |
| 11699 | |
| 11700 | inst.instruction = THUMB_OP32 (opcode); |
| 11701 | inst.instruction |= inst.operands[0].reg << 12; |
| 11702 | encode_thumb32_addr_mode (1, /*is_t=*/FALSE, /*is_d=*/FALSE); |
| 11703 | check_ldr_r15_aligned (); |
| 11704 | return; |
| 11705 | } |
| 11706 | |
| 11707 | constraint (inst.operands[0].reg > 7, BAD_HIREG); |
| 11708 | |
| 11709 | if (inst.instruction == T_MNEM_ldrsh || inst.instruction == T_MNEM_ldrsb) |
| 11710 | { |
| 11711 | /* Only [Rn,Rm] is acceptable. */ |
| 11712 | constraint (inst.operands[1].reg > 7 || inst.operands[1].imm > 7, BAD_HIREG); |
| 11713 | constraint (!inst.operands[1].isreg || !inst.operands[1].immisreg |
| 11714 | || inst.operands[1].postind || inst.operands[1].shifted |
| 11715 | || inst.operands[1].negative, |
| 11716 | _("Thumb does not support this addressing mode")); |
| 11717 | inst.instruction = THUMB_OP16 (inst.instruction); |
| 11718 | goto op16; |
| 11719 | } |
| 11720 | |
| 11721 | inst.instruction = THUMB_OP16 (inst.instruction); |
| 11722 | if (!inst.operands[1].isreg) |
| 11723 | if (move_or_literal_pool (0, CONST_THUMB, /*mode_3=*/FALSE)) |
| 11724 | return; |
| 11725 | |
| 11726 | constraint (!inst.operands[1].preind |
| 11727 | || inst.operands[1].shifted |
| 11728 | || inst.operands[1].writeback, |
| 11729 | _("Thumb does not support this addressing mode")); |
| 11730 | if (inst.operands[1].reg == REG_PC || inst.operands[1].reg == REG_SP) |
| 11731 | { |
| 11732 | constraint (inst.instruction & 0x0600, |
| 11733 | _("byte or halfword not valid for base register")); |
| 11734 | constraint (inst.operands[1].reg == REG_PC |
| 11735 | && !(inst.instruction & THUMB_LOAD_BIT), |
| 11736 | _("r15 based store not allowed")); |
| 11737 | constraint (inst.operands[1].immisreg, |
| 11738 | _("invalid base register for register offset")); |
| 11739 | |
| 11740 | if (inst.operands[1].reg == REG_PC) |
| 11741 | inst.instruction = T_OPCODE_LDR_PC; |
| 11742 | else if (inst.instruction & THUMB_LOAD_BIT) |
| 11743 | inst.instruction = T_OPCODE_LDR_SP; |
| 11744 | else |
| 11745 | inst.instruction = T_OPCODE_STR_SP; |
| 11746 | |
| 11747 | inst.instruction |= inst.operands[0].reg << 8; |
| 11748 | inst.reloc.type = BFD_RELOC_ARM_THUMB_OFFSET; |
| 11749 | return; |
| 11750 | } |
| 11751 | |
| 11752 | constraint (inst.operands[1].reg > 7, BAD_HIREG); |
| 11753 | if (!inst.operands[1].immisreg) |
| 11754 | { |
| 11755 | /* Immediate offset. */ |
| 11756 | inst.instruction |= inst.operands[0].reg; |
| 11757 | inst.instruction |= inst.operands[1].reg << 3; |
| 11758 | inst.reloc.type = BFD_RELOC_ARM_THUMB_OFFSET; |
| 11759 | return; |
| 11760 | } |
| 11761 | |
| 11762 | /* Register offset. */ |
| 11763 | constraint (inst.operands[1].imm > 7, BAD_HIREG); |
| 11764 | constraint (inst.operands[1].negative, |
| 11765 | _("Thumb does not support this addressing mode")); |
| 11766 | |
| 11767 | op16: |
| 11768 | switch (inst.instruction) |
| 11769 | { |
| 11770 | case T_OPCODE_STR_IW: inst.instruction = T_OPCODE_STR_RW; break; |
| 11771 | case T_OPCODE_STR_IH: inst.instruction = T_OPCODE_STR_RH; break; |
| 11772 | case T_OPCODE_STR_IB: inst.instruction = T_OPCODE_STR_RB; break; |
| 11773 | case T_OPCODE_LDR_IW: inst.instruction = T_OPCODE_LDR_RW; break; |
| 11774 | case T_OPCODE_LDR_IH: inst.instruction = T_OPCODE_LDR_RH; break; |
| 11775 | case T_OPCODE_LDR_IB: inst.instruction = T_OPCODE_LDR_RB; break; |
| 11776 | case 0x5600 /* ldrsb */: |
| 11777 | case 0x5e00 /* ldrsh */: break; |
| 11778 | default: abort (); |
| 11779 | } |
| 11780 | |
| 11781 | inst.instruction |= inst.operands[0].reg; |
| 11782 | inst.instruction |= inst.operands[1].reg << 3; |
| 11783 | inst.instruction |= inst.operands[1].imm << 6; |
| 11784 | } |
| 11785 | |
| 11786 | static void |
| 11787 | do_t_ldstd (void) |
| 11788 | { |
| 11789 | if (!inst.operands[1].present) |
| 11790 | { |
| 11791 | inst.operands[1].reg = inst.operands[0].reg + 1; |
| 11792 | constraint (inst.operands[0].reg == REG_LR, |
| 11793 | _("r14 not allowed here")); |
| 11794 | constraint (inst.operands[0].reg == REG_R12, |
| 11795 | _("r12 not allowed here")); |
| 11796 | } |
| 11797 | |
| 11798 | if (inst.operands[2].writeback |
| 11799 | && (inst.operands[0].reg == inst.operands[2].reg |
| 11800 | || inst.operands[1].reg == inst.operands[2].reg)) |
| 11801 | as_warn (_("base register written back, and overlaps " |
| 11802 | "one of transfer registers")); |
| 11803 | |
| 11804 | inst.instruction |= inst.operands[0].reg << 12; |
| 11805 | inst.instruction |= inst.operands[1].reg << 8; |
| 11806 | encode_thumb32_addr_mode (2, /*is_t=*/FALSE, /*is_d=*/TRUE); |
| 11807 | } |
| 11808 | |
| 11809 | static void |
| 11810 | do_t_ldstt (void) |
| 11811 | { |
| 11812 | inst.instruction |= inst.operands[0].reg << 12; |
| 11813 | encode_thumb32_addr_mode (1, /*is_t=*/TRUE, /*is_d=*/FALSE); |
| 11814 | } |
| 11815 | |
| 11816 | static void |
| 11817 | do_t_mla (void) |
| 11818 | { |
| 11819 | unsigned Rd, Rn, Rm, Ra; |
| 11820 | |
| 11821 | Rd = inst.operands[0].reg; |
| 11822 | Rn = inst.operands[1].reg; |
| 11823 | Rm = inst.operands[2].reg; |
| 11824 | Ra = inst.operands[3].reg; |
| 11825 | |
| 11826 | reject_bad_reg (Rd); |
| 11827 | reject_bad_reg (Rn); |
| 11828 | reject_bad_reg (Rm); |
| 11829 | reject_bad_reg (Ra); |
| 11830 | |
| 11831 | inst.instruction |= Rd << 8; |
| 11832 | inst.instruction |= Rn << 16; |
| 11833 | inst.instruction |= Rm; |
| 11834 | inst.instruction |= Ra << 12; |
| 11835 | } |
| 11836 | |
| 11837 | static void |
| 11838 | do_t_mlal (void) |
| 11839 | { |
| 11840 | unsigned RdLo, RdHi, Rn, Rm; |
| 11841 | |
| 11842 | RdLo = inst.operands[0].reg; |
| 11843 | RdHi = inst.operands[1].reg; |
| 11844 | Rn = inst.operands[2].reg; |
| 11845 | Rm = inst.operands[3].reg; |
| 11846 | |
| 11847 | reject_bad_reg (RdLo); |
| 11848 | reject_bad_reg (RdHi); |
| 11849 | reject_bad_reg (Rn); |
| 11850 | reject_bad_reg (Rm); |
| 11851 | |
| 11852 | inst.instruction |= RdLo << 12; |
| 11853 | inst.instruction |= RdHi << 8; |
| 11854 | inst.instruction |= Rn << 16; |
| 11855 | inst.instruction |= Rm; |
| 11856 | } |
| 11857 | |
| 11858 | static void |
| 11859 | do_t_mov_cmp (void) |
| 11860 | { |
| 11861 | unsigned Rn, Rm; |
| 11862 | |
| 11863 | Rn = inst.operands[0].reg; |
| 11864 | Rm = inst.operands[1].reg; |
| 11865 | |
| 11866 | if (Rn == REG_PC) |
| 11867 | set_it_insn_type_last (); |
| 11868 | |
| 11869 | if (unified_syntax) |
| 11870 | { |
| 11871 | int r0off = (inst.instruction == T_MNEM_mov |
| 11872 | || inst.instruction == T_MNEM_movs) ? 8 : 16; |
| 11873 | unsigned long opcode; |
| 11874 | bfd_boolean narrow; |
| 11875 | bfd_boolean low_regs; |
| 11876 | |
| 11877 | low_regs = (Rn <= 7 && Rm <= 7); |
| 11878 | opcode = inst.instruction; |
| 11879 | if (in_it_block ()) |
| 11880 | narrow = opcode != T_MNEM_movs; |
| 11881 | else |
| 11882 | narrow = opcode != T_MNEM_movs || low_regs; |
| 11883 | if (inst.size_req == 4 |
| 11884 | || inst.operands[1].shifted) |
| 11885 | narrow = FALSE; |
| 11886 | |
| 11887 | /* MOVS PC, LR is encoded as SUBS PC, LR, #0. */ |
| 11888 | if (opcode == T_MNEM_movs && inst.operands[1].isreg |
| 11889 | && !inst.operands[1].shifted |
| 11890 | && Rn == REG_PC |
| 11891 | && Rm == REG_LR) |
| 11892 | { |
| 11893 | inst.instruction = T2_SUBS_PC_LR; |
| 11894 | return; |
| 11895 | } |
| 11896 | |
| 11897 | if (opcode == T_MNEM_cmp) |
| 11898 | { |
| 11899 | constraint (Rn == REG_PC, BAD_PC); |
| 11900 | if (narrow) |
| 11901 | { |
| 11902 | /* In the Thumb-2 ISA, use of R13 as Rm is deprecated, |
| 11903 | but valid. */ |
| 11904 | warn_deprecated_sp (Rm); |
| 11905 | /* R15 was documented as a valid choice for Rm in ARMv6, |
| 11906 | but as UNPREDICTABLE in ARMv7. ARM's proprietary |
| 11907 | tools reject R15, so we do too. */ |
| 11908 | constraint (Rm == REG_PC, BAD_PC); |
| 11909 | } |
| 11910 | else |
| 11911 | reject_bad_reg (Rm); |
| 11912 | } |
| 11913 | else if (opcode == T_MNEM_mov |
| 11914 | || opcode == T_MNEM_movs) |
| 11915 | { |
| 11916 | if (inst.operands[1].isreg) |
| 11917 | { |
| 11918 | if (opcode == T_MNEM_movs) |
| 11919 | { |
| 11920 | reject_bad_reg (Rn); |
| 11921 | reject_bad_reg (Rm); |
| 11922 | } |
| 11923 | else if (narrow) |
| 11924 | { |
| 11925 | /* This is mov.n. */ |
| 11926 | if ((Rn == REG_SP || Rn == REG_PC) |
| 11927 | && (Rm == REG_SP || Rm == REG_PC)) |
| 11928 | { |
| 11929 | as_tsktsk (_("Use of r%u as a source register is " |
| 11930 | "deprecated when r%u is the destination " |
| 11931 | "register."), Rm, Rn); |
| 11932 | } |
| 11933 | } |
| 11934 | else |
| 11935 | { |
| 11936 | /* This is mov.w. */ |
| 11937 | constraint (Rn == REG_PC, BAD_PC); |
| 11938 | constraint (Rm == REG_PC, BAD_PC); |
| 11939 | if (!ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v8)) |
| 11940 | constraint (Rn == REG_SP && Rm == REG_SP, BAD_SP); |
| 11941 | } |
| 11942 | } |
| 11943 | else |
| 11944 | reject_bad_reg (Rn); |
| 11945 | } |
| 11946 | |
| 11947 | if (!inst.operands[1].isreg) |
| 11948 | { |
| 11949 | /* Immediate operand. */ |
| 11950 | if (!in_it_block () && opcode == T_MNEM_mov) |
| 11951 | narrow = 0; |
| 11952 | if (low_regs && narrow) |
| 11953 | { |
| 11954 | inst.instruction = THUMB_OP16 (opcode); |
| 11955 | inst.instruction |= Rn << 8; |
| 11956 | if (inst.reloc.type < BFD_RELOC_ARM_THUMB_ALU_ABS_G0_NC |
| 11957 | || inst.reloc.type > BFD_RELOC_ARM_THUMB_ALU_ABS_G3_NC) |
| 11958 | { |
| 11959 | if (inst.size_req == 2) |
| 11960 | inst.reloc.type = BFD_RELOC_ARM_THUMB_IMM; |
| 11961 | else |
| 11962 | inst.relax = opcode; |
| 11963 | } |
| 11964 | } |
| 11965 | else |
| 11966 | { |
| 11967 | constraint (inst.reloc.type >= BFD_RELOC_ARM_THUMB_ALU_ABS_G0_NC |
| 11968 | && inst.reloc.type <= BFD_RELOC_ARM_THUMB_ALU_ABS_G3_NC , |
| 11969 | THUMB1_RELOC_ONLY); |
| 11970 | |
| 11971 | inst.instruction = THUMB_OP32 (inst.instruction); |
| 11972 | inst.instruction = (inst.instruction & 0xe1ffffff) | 0x10000000; |
| 11973 | inst.instruction |= Rn << r0off; |
| 11974 | inst.reloc.type = BFD_RELOC_ARM_T32_IMMEDIATE; |
| 11975 | } |
| 11976 | } |
| 11977 | else if (inst.operands[1].shifted && inst.operands[1].immisreg |
| 11978 | && (inst.instruction == T_MNEM_mov |
| 11979 | || inst.instruction == T_MNEM_movs)) |
| 11980 | { |
| 11981 | /* Register shifts are encoded as separate shift instructions. */ |
| 11982 | bfd_boolean flags = (inst.instruction == T_MNEM_movs); |
| 11983 | |
| 11984 | if (in_it_block ()) |
| 11985 | narrow = !flags; |
| 11986 | else |
| 11987 | narrow = flags; |
| 11988 | |
| 11989 | if (inst.size_req == 4) |
| 11990 | narrow = FALSE; |
| 11991 | |
| 11992 | if (!low_regs || inst.operands[1].imm > 7) |
| 11993 | narrow = FALSE; |
| 11994 | |
| 11995 | if (Rn != Rm) |
| 11996 | narrow = FALSE; |
| 11997 | |
| 11998 | switch (inst.operands[1].shift_kind) |
| 11999 | { |
| 12000 | case SHIFT_LSL: |
| 12001 | opcode = narrow ? T_OPCODE_LSL_R : THUMB_OP32 (T_MNEM_lsl); |
| 12002 | break; |
| 12003 | case SHIFT_ASR: |
| 12004 | opcode = narrow ? T_OPCODE_ASR_R : THUMB_OP32 (T_MNEM_asr); |
| 12005 | break; |
| 12006 | case SHIFT_LSR: |
| 12007 | opcode = narrow ? T_OPCODE_LSR_R : THUMB_OP32 (T_MNEM_lsr); |
| 12008 | break; |
| 12009 | case SHIFT_ROR: |
| 12010 | opcode = narrow ? T_OPCODE_ROR_R : THUMB_OP32 (T_MNEM_ror); |
| 12011 | break; |
| 12012 | default: |
| 12013 | abort (); |
| 12014 | } |
| 12015 | |
| 12016 | inst.instruction = opcode; |
| 12017 | if (narrow) |
| 12018 | { |
| 12019 | inst.instruction |= Rn; |
| 12020 | inst.instruction |= inst.operands[1].imm << 3; |
| 12021 | } |
| 12022 | else |
| 12023 | { |
| 12024 | if (flags) |
| 12025 | inst.instruction |= CONDS_BIT; |
| 12026 | |
| 12027 | inst.instruction |= Rn << 8; |
| 12028 | inst.instruction |= Rm << 16; |
| 12029 | inst.instruction |= inst.operands[1].imm; |
| 12030 | } |
| 12031 | } |
| 12032 | else if (!narrow) |
| 12033 | { |
| 12034 | /* Some mov with immediate shift have narrow variants. |
| 12035 | Register shifts are handled above. */ |
| 12036 | if (low_regs && inst.operands[1].shifted |
| 12037 | && (inst.instruction == T_MNEM_mov |
| 12038 | || inst.instruction == T_MNEM_movs)) |
| 12039 | { |
| 12040 | if (in_it_block ()) |
| 12041 | narrow = (inst.instruction == T_MNEM_mov); |
| 12042 | else |
| 12043 | narrow = (inst.instruction == T_MNEM_movs); |
| 12044 | } |
| 12045 | |
| 12046 | if (narrow) |
| 12047 | { |
| 12048 | switch (inst.operands[1].shift_kind) |
| 12049 | { |
| 12050 | case SHIFT_LSL: inst.instruction = T_OPCODE_LSL_I; break; |
| 12051 | case SHIFT_LSR: inst.instruction = T_OPCODE_LSR_I; break; |
| 12052 | case SHIFT_ASR: inst.instruction = T_OPCODE_ASR_I; break; |
| 12053 | default: narrow = FALSE; break; |
| 12054 | } |
| 12055 | } |
| 12056 | |
| 12057 | if (narrow) |
| 12058 | { |
| 12059 | inst.instruction |= Rn; |
| 12060 | inst.instruction |= Rm << 3; |
| 12061 | inst.reloc.type = BFD_RELOC_ARM_THUMB_SHIFT; |
| 12062 | } |
| 12063 | else |
| 12064 | { |
| 12065 | inst.instruction = THUMB_OP32 (inst.instruction); |
| 12066 | inst.instruction |= Rn << r0off; |
| 12067 | encode_thumb32_shifted_operand (1); |
| 12068 | } |
| 12069 | } |
| 12070 | else |
| 12071 | switch (inst.instruction) |
| 12072 | { |
| 12073 | case T_MNEM_mov: |
| 12074 | /* In v4t or v5t a move of two lowregs produces unpredictable |
| 12075 | results. Don't allow this. */ |
| 12076 | if (low_regs) |
| 12077 | { |
| 12078 | constraint (!ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v6), |
| 12079 | "MOV Rd, Rs with two low registers is not " |
| 12080 | "permitted on this architecture"); |
| 12081 | ARM_MERGE_FEATURE_SETS (thumb_arch_used, thumb_arch_used, |
| 12082 | arm_ext_v6); |
| 12083 | } |
| 12084 | |
| 12085 | inst.instruction = T_OPCODE_MOV_HR; |
| 12086 | inst.instruction |= (Rn & 0x8) << 4; |
| 12087 | inst.instruction |= (Rn & 0x7); |
| 12088 | inst.instruction |= Rm << 3; |
| 12089 | break; |
| 12090 | |
| 12091 | case T_MNEM_movs: |
| 12092 | /* We know we have low registers at this point. |
| 12093 | Generate LSLS Rd, Rs, #0. */ |
| 12094 | inst.instruction = T_OPCODE_LSL_I; |
| 12095 | inst.instruction |= Rn; |
| 12096 | inst.instruction |= Rm << 3; |
| 12097 | break; |
| 12098 | |
| 12099 | case T_MNEM_cmp: |
| 12100 | if (low_regs) |
| 12101 | { |
| 12102 | inst.instruction = T_OPCODE_CMP_LR; |
| 12103 | inst.instruction |= Rn; |
| 12104 | inst.instruction |= Rm << 3; |
| 12105 | } |
| 12106 | else |
| 12107 | { |
| 12108 | inst.instruction = T_OPCODE_CMP_HR; |
| 12109 | inst.instruction |= (Rn & 0x8) << 4; |
| 12110 | inst.instruction |= (Rn & 0x7); |
| 12111 | inst.instruction |= Rm << 3; |
| 12112 | } |
| 12113 | break; |
| 12114 | } |
| 12115 | return; |
| 12116 | } |
| 12117 | |
| 12118 | inst.instruction = THUMB_OP16 (inst.instruction); |
| 12119 | |
| 12120 | /* PR 10443: Do not silently ignore shifted operands. */ |
| 12121 | constraint (inst.operands[1].shifted, |
| 12122 | _("shifts in CMP/MOV instructions are only supported in unified syntax")); |
| 12123 | |
| 12124 | if (inst.operands[1].isreg) |
| 12125 | { |
| 12126 | if (Rn < 8 && Rm < 8) |
| 12127 | { |
| 12128 | /* A move of two lowregs is encoded as ADD Rd, Rs, #0 |
| 12129 | since a MOV instruction produces unpredictable results. */ |
| 12130 | if (inst.instruction == T_OPCODE_MOV_I8) |
| 12131 | inst.instruction = T_OPCODE_ADD_I3; |
| 12132 | else |
| 12133 | inst.instruction = T_OPCODE_CMP_LR; |
| 12134 | |
| 12135 | inst.instruction |= Rn; |
| 12136 | inst.instruction |= Rm << 3; |
| 12137 | } |
| 12138 | else |
| 12139 | { |
| 12140 | if (inst.instruction == T_OPCODE_MOV_I8) |
| 12141 | inst.instruction = T_OPCODE_MOV_HR; |
| 12142 | else |
| 12143 | inst.instruction = T_OPCODE_CMP_HR; |
| 12144 | do_t_cpy (); |
| 12145 | } |
| 12146 | } |
| 12147 | else |
| 12148 | { |
| 12149 | constraint (Rn > 7, |
| 12150 | _("only lo regs allowed with immediate")); |
| 12151 | inst.instruction |= Rn << 8; |
| 12152 | inst.reloc.type = BFD_RELOC_ARM_THUMB_IMM; |
| 12153 | } |
| 12154 | } |
| 12155 | |
| 12156 | static void |
| 12157 | do_t_mov16 (void) |
| 12158 | { |
| 12159 | unsigned Rd; |
| 12160 | bfd_vma imm; |
| 12161 | bfd_boolean top; |
| 12162 | |
| 12163 | top = (inst.instruction & 0x00800000) != 0; |
| 12164 | if (inst.reloc.type == BFD_RELOC_ARM_MOVW) |
| 12165 | { |
| 12166 | constraint (top, _(":lower16: not allowed in this instruction")); |
| 12167 | inst.reloc.type = BFD_RELOC_ARM_THUMB_MOVW; |
| 12168 | } |
| 12169 | else if (inst.reloc.type == BFD_RELOC_ARM_MOVT) |
| 12170 | { |
| 12171 | constraint (!top, _(":upper16: not allowed in this instruction")); |
| 12172 | inst.reloc.type = BFD_RELOC_ARM_THUMB_MOVT; |
| 12173 | } |
| 12174 | |
| 12175 | Rd = inst.operands[0].reg; |
| 12176 | reject_bad_reg (Rd); |
| 12177 | |
| 12178 | inst.instruction |= Rd << 8; |
| 12179 | if (inst.reloc.type == BFD_RELOC_UNUSED) |
| 12180 | { |
| 12181 | imm = inst.reloc.exp.X_add_number; |
| 12182 | inst.instruction |= (imm & 0xf000) << 4; |
| 12183 | inst.instruction |= (imm & 0x0800) << 15; |
| 12184 | inst.instruction |= (imm & 0x0700) << 4; |
| 12185 | inst.instruction |= (imm & 0x00ff); |
| 12186 | } |
| 12187 | } |
| 12188 | |
| 12189 | static void |
| 12190 | do_t_mvn_tst (void) |
| 12191 | { |
| 12192 | unsigned Rn, Rm; |
| 12193 | |
| 12194 | Rn = inst.operands[0].reg; |
| 12195 | Rm = inst.operands[1].reg; |
| 12196 | |
| 12197 | if (inst.instruction == T_MNEM_cmp |
| 12198 | || inst.instruction == T_MNEM_cmn) |
| 12199 | constraint (Rn == REG_PC, BAD_PC); |
| 12200 | else |
| 12201 | reject_bad_reg (Rn); |
| 12202 | reject_bad_reg (Rm); |
| 12203 | |
| 12204 | if (unified_syntax) |
| 12205 | { |
| 12206 | int r0off = (inst.instruction == T_MNEM_mvn |
| 12207 | || inst.instruction == T_MNEM_mvns) ? 8 : 16; |
| 12208 | bfd_boolean narrow; |
| 12209 | |
| 12210 | if (inst.size_req == 4 |
| 12211 | || inst.instruction > 0xffff |
| 12212 | || inst.operands[1].shifted |
| 12213 | || Rn > 7 || Rm > 7) |
| 12214 | narrow = FALSE; |
| 12215 | else if (inst.instruction == T_MNEM_cmn |
| 12216 | || inst.instruction == T_MNEM_tst) |
| 12217 | narrow = TRUE; |
| 12218 | else if (THUMB_SETS_FLAGS (inst.instruction)) |
| 12219 | narrow = !in_it_block (); |
| 12220 | else |
| 12221 | narrow = in_it_block (); |
| 12222 | |
| 12223 | if (!inst.operands[1].isreg) |
| 12224 | { |
| 12225 | /* For an immediate, we always generate a 32-bit opcode; |
| 12226 | section relaxation will shrink it later if possible. */ |
| 12227 | if (inst.instruction < 0xffff) |
| 12228 | inst.instruction = THUMB_OP32 (inst.instruction); |
| 12229 | inst.instruction = (inst.instruction & 0xe1ffffff) | 0x10000000; |
| 12230 | inst.instruction |= Rn << r0off; |
| 12231 | inst.reloc.type = BFD_RELOC_ARM_T32_IMMEDIATE; |
| 12232 | } |
| 12233 | else |
| 12234 | { |
| 12235 | /* See if we can do this with a 16-bit instruction. */ |
| 12236 | if (narrow) |
| 12237 | { |
| 12238 | inst.instruction = THUMB_OP16 (inst.instruction); |
| 12239 | inst.instruction |= Rn; |
| 12240 | inst.instruction |= Rm << 3; |
| 12241 | } |
| 12242 | else |
| 12243 | { |
| 12244 | constraint (inst.operands[1].shifted |
| 12245 | && inst.operands[1].immisreg, |
| 12246 | _("shift must be constant")); |
| 12247 | if (inst.instruction < 0xffff) |
| 12248 | inst.instruction = THUMB_OP32 (inst.instruction); |
| 12249 | inst.instruction |= Rn << r0off; |
| 12250 | encode_thumb32_shifted_operand (1); |
| 12251 | } |
| 12252 | } |
| 12253 | } |
| 12254 | else |
| 12255 | { |
| 12256 | constraint (inst.instruction > 0xffff |
| 12257 | || inst.instruction == T_MNEM_mvns, BAD_THUMB32); |
| 12258 | constraint (!inst.operands[1].isreg || inst.operands[1].shifted, |
| 12259 | _("unshifted register required")); |
| 12260 | constraint (Rn > 7 || Rm > 7, |
| 12261 | BAD_HIREG); |
| 12262 | |
| 12263 | inst.instruction = THUMB_OP16 (inst.instruction); |
| 12264 | inst.instruction |= Rn; |
| 12265 | inst.instruction |= Rm << 3; |
| 12266 | } |
| 12267 | } |
| 12268 | |
| 12269 | static void |
| 12270 | do_t_mrs (void) |
| 12271 | { |
| 12272 | unsigned Rd; |
| 12273 | |
| 12274 | if (do_vfp_nsyn_mrs () == SUCCESS) |
| 12275 | return; |
| 12276 | |
| 12277 | Rd = inst.operands[0].reg; |
| 12278 | reject_bad_reg (Rd); |
| 12279 | inst.instruction |= Rd << 8; |
| 12280 | |
| 12281 | if (inst.operands[1].isreg) |
| 12282 | { |
| 12283 | unsigned br = inst.operands[1].reg; |
| 12284 | if (((br & 0x200) == 0) && ((br & 0xf000) != 0xf000)) |
| 12285 | as_bad (_("bad register for mrs")); |
| 12286 | |
| 12287 | inst.instruction |= br & (0xf << 16); |
| 12288 | inst.instruction |= (br & 0x300) >> 4; |
| 12289 | inst.instruction |= (br & SPSR_BIT) >> 2; |
| 12290 | } |
| 12291 | else |
| 12292 | { |
| 12293 | int flags = inst.operands[1].imm & (PSR_c|PSR_x|PSR_s|PSR_f|SPSR_BIT); |
| 12294 | |
| 12295 | if (ARM_CPU_HAS_FEATURE (selected_cpu, arm_ext_m)) |
| 12296 | { |
| 12297 | /* PR gas/12698: The constraint is only applied for m_profile. |
| 12298 | If the user has specified -march=all, we want to ignore it as |
| 12299 | we are building for any CPU type, including non-m variants. */ |
| 12300 | bfd_boolean m_profile = |
| 12301 | !ARM_FEATURE_CORE_EQUAL (selected_cpu, arm_arch_any); |
| 12302 | constraint ((flags != 0) && m_profile, _("selected processor does " |
| 12303 | "not support requested special purpose register")); |
| 12304 | } |
| 12305 | else |
| 12306 | /* mrs only accepts APSR/CPSR/SPSR/CPSR_all/SPSR_all (for non-M profile |
| 12307 | devices). */ |
| 12308 | constraint ((flags & ~SPSR_BIT) != (PSR_c|PSR_f), |
| 12309 | _("'APSR', 'CPSR' or 'SPSR' expected")); |
| 12310 | |
| 12311 | inst.instruction |= (flags & SPSR_BIT) >> 2; |
| 12312 | inst.instruction |= inst.operands[1].imm & 0xff; |
| 12313 | inst.instruction |= 0xf0000; |
| 12314 | } |
| 12315 | } |
| 12316 | |
| 12317 | static void |
| 12318 | do_t_msr (void) |
| 12319 | { |
| 12320 | int flags; |
| 12321 | unsigned Rn; |
| 12322 | |
| 12323 | if (do_vfp_nsyn_msr () == SUCCESS) |
| 12324 | return; |
| 12325 | |
| 12326 | constraint (!inst.operands[1].isreg, |
| 12327 | _("Thumb encoding does not support an immediate here")); |
| 12328 | |
| 12329 | if (inst.operands[0].isreg) |
| 12330 | flags = (int)(inst.operands[0].reg); |
| 12331 | else |
| 12332 | flags = inst.operands[0].imm; |
| 12333 | |
| 12334 | if (ARM_CPU_HAS_FEATURE (selected_cpu, arm_ext_m)) |
| 12335 | { |
| 12336 | int bits = inst.operands[0].imm & (PSR_c|PSR_x|PSR_s|PSR_f|SPSR_BIT); |
| 12337 | |
| 12338 | /* PR gas/12698: The constraint is only applied for m_profile. |
| 12339 | If the user has specified -march=all, we want to ignore it as |
| 12340 | we are building for any CPU type, including non-m variants. */ |
| 12341 | bfd_boolean m_profile = |
| 12342 | !ARM_FEATURE_CORE_EQUAL (selected_cpu, arm_arch_any); |
| 12343 | constraint (((ARM_CPU_HAS_FEATURE (selected_cpu, arm_ext_v6_dsp) |
| 12344 | && (bits & ~(PSR_s | PSR_f)) != 0) |
| 12345 | || (!ARM_CPU_HAS_FEATURE (selected_cpu, arm_ext_v6_dsp) |
| 12346 | && bits != PSR_f)) && m_profile, |
| 12347 | _("selected processor does not support requested special " |
| 12348 | "purpose register")); |
| 12349 | } |
| 12350 | else |
| 12351 | constraint ((flags & 0xff) != 0, _("selected processor does not support " |
| 12352 | "requested special purpose register")); |
| 12353 | |
| 12354 | Rn = inst.operands[1].reg; |
| 12355 | reject_bad_reg (Rn); |
| 12356 | |
| 12357 | inst.instruction |= (flags & SPSR_BIT) >> 2; |
| 12358 | inst.instruction |= (flags & 0xf0000) >> 8; |
| 12359 | inst.instruction |= (flags & 0x300) >> 4; |
| 12360 | inst.instruction |= (flags & 0xff); |
| 12361 | inst.instruction |= Rn << 16; |
| 12362 | } |
| 12363 | |
| 12364 | static void |
| 12365 | do_t_mul (void) |
| 12366 | { |
| 12367 | bfd_boolean narrow; |
| 12368 | unsigned Rd, Rn, Rm; |
| 12369 | |
| 12370 | if (!inst.operands[2].present) |
| 12371 | inst.operands[2].reg = inst.operands[0].reg; |
| 12372 | |
| 12373 | Rd = inst.operands[0].reg; |
| 12374 | Rn = inst.operands[1].reg; |
| 12375 | Rm = inst.operands[2].reg; |
| 12376 | |
| 12377 | if (unified_syntax) |
| 12378 | { |
| 12379 | if (inst.size_req == 4 |
| 12380 | || (Rd != Rn |
| 12381 | && Rd != Rm) |
| 12382 | || Rn > 7 |
| 12383 | || Rm > 7) |
| 12384 | narrow = FALSE; |
| 12385 | else if (inst.instruction == T_MNEM_muls) |
| 12386 | narrow = !in_it_block (); |
| 12387 | else |
| 12388 | narrow = in_it_block (); |
| 12389 | } |
| 12390 | else |
| 12391 | { |
| 12392 | constraint (inst.instruction == T_MNEM_muls, BAD_THUMB32); |
| 12393 | constraint (Rn > 7 || Rm > 7, |
| 12394 | BAD_HIREG); |
| 12395 | narrow = TRUE; |
| 12396 | } |
| 12397 | |
| 12398 | if (narrow) |
| 12399 | { |
| 12400 | /* 16-bit MULS/Conditional MUL. */ |
| 12401 | inst.instruction = THUMB_OP16 (inst.instruction); |
| 12402 | inst.instruction |= Rd; |
| 12403 | |
| 12404 | if (Rd == Rn) |
| 12405 | inst.instruction |= Rm << 3; |
| 12406 | else if (Rd == Rm) |
| 12407 | inst.instruction |= Rn << 3; |
| 12408 | else |
| 12409 | constraint (1, _("dest must overlap one source register")); |
| 12410 | } |
| 12411 | else |
| 12412 | { |
| 12413 | constraint (inst.instruction != T_MNEM_mul, |
| 12414 | _("Thumb-2 MUL must not set flags")); |
| 12415 | /* 32-bit MUL. */ |
| 12416 | inst.instruction = THUMB_OP32 (inst.instruction); |
| 12417 | inst.instruction |= Rd << 8; |
| 12418 | inst.instruction |= Rn << 16; |
| 12419 | inst.instruction |= Rm << 0; |
| 12420 | |
| 12421 | reject_bad_reg (Rd); |
| 12422 | reject_bad_reg (Rn); |
| 12423 | reject_bad_reg (Rm); |
| 12424 | } |
| 12425 | } |
| 12426 | |
| 12427 | static void |
| 12428 | do_t_mull (void) |
| 12429 | { |
| 12430 | unsigned RdLo, RdHi, Rn, Rm; |
| 12431 | |
| 12432 | RdLo = inst.operands[0].reg; |
| 12433 | RdHi = inst.operands[1].reg; |
| 12434 | Rn = inst.operands[2].reg; |
| 12435 | Rm = inst.operands[3].reg; |
| 12436 | |
| 12437 | reject_bad_reg (RdLo); |
| 12438 | reject_bad_reg (RdHi); |
| 12439 | reject_bad_reg (Rn); |
| 12440 | reject_bad_reg (Rm); |
| 12441 | |
| 12442 | inst.instruction |= RdLo << 12; |
| 12443 | inst.instruction |= RdHi << 8; |
| 12444 | inst.instruction |= Rn << 16; |
| 12445 | inst.instruction |= Rm; |
| 12446 | |
| 12447 | if (RdLo == RdHi) |
| 12448 | as_tsktsk (_("rdhi and rdlo must be different")); |
| 12449 | } |
| 12450 | |
| 12451 | static void |
| 12452 | do_t_nop (void) |
| 12453 | { |
| 12454 | set_it_insn_type (NEUTRAL_IT_INSN); |
| 12455 | |
| 12456 | if (unified_syntax) |
| 12457 | { |
| 12458 | if (inst.size_req == 4 || inst.operands[0].imm > 15) |
| 12459 | { |
| 12460 | inst.instruction = THUMB_OP32 (inst.instruction); |
| 12461 | inst.instruction |= inst.operands[0].imm; |
| 12462 | } |
| 12463 | else |
| 12464 | { |
| 12465 | /* PR9722: Check for Thumb2 availability before |
| 12466 | generating a thumb2 nop instruction. */ |
| 12467 | if (ARM_CPU_HAS_FEATURE (selected_cpu, arm_ext_v6t2)) |
| 12468 | { |
| 12469 | inst.instruction = THUMB_OP16 (inst.instruction); |
| 12470 | inst.instruction |= inst.operands[0].imm << 4; |
| 12471 | } |
| 12472 | else |
| 12473 | inst.instruction = 0x46c0; |
| 12474 | } |
| 12475 | } |
| 12476 | else |
| 12477 | { |
| 12478 | constraint (inst.operands[0].present, |
| 12479 | _("Thumb does not support NOP with hints")); |
| 12480 | inst.instruction = 0x46c0; |
| 12481 | } |
| 12482 | } |
| 12483 | |
| 12484 | static void |
| 12485 | do_t_neg (void) |
| 12486 | { |
| 12487 | if (unified_syntax) |
| 12488 | { |
| 12489 | bfd_boolean narrow; |
| 12490 | |
| 12491 | if (THUMB_SETS_FLAGS (inst.instruction)) |
| 12492 | narrow = !in_it_block (); |
| 12493 | else |
| 12494 | narrow = in_it_block (); |
| 12495 | if (inst.operands[0].reg > 7 || inst.operands[1].reg > 7) |
| 12496 | narrow = FALSE; |
| 12497 | if (inst.size_req == 4) |
| 12498 | narrow = FALSE; |
| 12499 | |
| 12500 | if (!narrow) |
| 12501 | { |
| 12502 | inst.instruction = THUMB_OP32 (inst.instruction); |
| 12503 | inst.instruction |= inst.operands[0].reg << 8; |
| 12504 | inst.instruction |= inst.operands[1].reg << 16; |
| 12505 | } |
| 12506 | else |
| 12507 | { |
| 12508 | inst.instruction = THUMB_OP16 (inst.instruction); |
| 12509 | inst.instruction |= inst.operands[0].reg; |
| 12510 | inst.instruction |= inst.operands[1].reg << 3; |
| 12511 | } |
| 12512 | } |
| 12513 | else |
| 12514 | { |
| 12515 | constraint (inst.operands[0].reg > 7 || inst.operands[1].reg > 7, |
| 12516 | BAD_HIREG); |
| 12517 | constraint (THUMB_SETS_FLAGS (inst.instruction), BAD_THUMB32); |
| 12518 | |
| 12519 | inst.instruction = THUMB_OP16 (inst.instruction); |
| 12520 | inst.instruction |= inst.operands[0].reg; |
| 12521 | inst.instruction |= inst.operands[1].reg << 3; |
| 12522 | } |
| 12523 | } |
| 12524 | |
| 12525 | static void |
| 12526 | do_t_orn (void) |
| 12527 | { |
| 12528 | unsigned Rd, Rn; |
| 12529 | |
| 12530 | Rd = inst.operands[0].reg; |
| 12531 | Rn = inst.operands[1].present ? inst.operands[1].reg : Rd; |
| 12532 | |
| 12533 | reject_bad_reg (Rd); |
| 12534 | /* Rn == REG_SP is unpredictable; Rn == REG_PC is MVN. */ |
| 12535 | reject_bad_reg (Rn); |
| 12536 | |
| 12537 | inst.instruction |= Rd << 8; |
| 12538 | inst.instruction |= Rn << 16; |
| 12539 | |
| 12540 | if (!inst.operands[2].isreg) |
| 12541 | { |
| 12542 | inst.instruction = (inst.instruction & 0xe1ffffff) | 0x10000000; |
| 12543 | inst.reloc.type = BFD_RELOC_ARM_T32_IMMEDIATE; |
| 12544 | } |
| 12545 | else |
| 12546 | { |
| 12547 | unsigned Rm; |
| 12548 | |
| 12549 | Rm = inst.operands[2].reg; |
| 12550 | reject_bad_reg (Rm); |
| 12551 | |
| 12552 | constraint (inst.operands[2].shifted |
| 12553 | && inst.operands[2].immisreg, |
| 12554 | _("shift must be constant")); |
| 12555 | encode_thumb32_shifted_operand (2); |
| 12556 | } |
| 12557 | } |
| 12558 | |
| 12559 | static void |
| 12560 | do_t_pkhbt (void) |
| 12561 | { |
| 12562 | unsigned Rd, Rn, Rm; |
| 12563 | |
| 12564 | Rd = inst.operands[0].reg; |
| 12565 | Rn = inst.operands[1].reg; |
| 12566 | Rm = inst.operands[2].reg; |
| 12567 | |
| 12568 | reject_bad_reg (Rd); |
| 12569 | reject_bad_reg (Rn); |
| 12570 | reject_bad_reg (Rm); |
| 12571 | |
| 12572 | inst.instruction |= Rd << 8; |
| 12573 | inst.instruction |= Rn << 16; |
| 12574 | inst.instruction |= Rm; |
| 12575 | if (inst.operands[3].present) |
| 12576 | { |
| 12577 | unsigned int val = inst.reloc.exp.X_add_number; |
| 12578 | constraint (inst.reloc.exp.X_op != O_constant, |
| 12579 | _("expression too complex")); |
| 12580 | inst.instruction |= (val & 0x1c) << 10; |
| 12581 | inst.instruction |= (val & 0x03) << 6; |
| 12582 | } |
| 12583 | } |
| 12584 | |
| 12585 | static void |
| 12586 | do_t_pkhtb (void) |
| 12587 | { |
| 12588 | if (!inst.operands[3].present) |
| 12589 | { |
| 12590 | unsigned Rtmp; |
| 12591 | |
| 12592 | inst.instruction &= ~0x00000020; |
| 12593 | |
| 12594 | /* PR 10168. Swap the Rm and Rn registers. */ |
| 12595 | Rtmp = inst.operands[1].reg; |
| 12596 | inst.operands[1].reg = inst.operands[2].reg; |
| 12597 | inst.operands[2].reg = Rtmp; |
| 12598 | } |
| 12599 | do_t_pkhbt (); |
| 12600 | } |
| 12601 | |
| 12602 | static void |
| 12603 | do_t_pld (void) |
| 12604 | { |
| 12605 | if (inst.operands[0].immisreg) |
| 12606 | reject_bad_reg (inst.operands[0].imm); |
| 12607 | |
| 12608 | encode_thumb32_addr_mode (0, /*is_t=*/FALSE, /*is_d=*/FALSE); |
| 12609 | } |
| 12610 | |
| 12611 | static void |
| 12612 | do_t_push_pop (void) |
| 12613 | { |
| 12614 | unsigned mask; |
| 12615 | |
| 12616 | constraint (inst.operands[0].writeback, |
| 12617 | _("push/pop do not support {reglist}^")); |
| 12618 | constraint (inst.reloc.type != BFD_RELOC_UNUSED, |
| 12619 | _("expression too complex")); |
| 12620 | |
| 12621 | mask = inst.operands[0].imm; |
| 12622 | if (inst.size_req != 4 && (mask & ~0xff) == 0) |
| 12623 | inst.instruction = THUMB_OP16 (inst.instruction) | mask; |
| 12624 | else if (inst.size_req != 4 |
| 12625 | && (mask & ~0xff) == (1U << (inst.instruction == T_MNEM_push |
| 12626 | ? REG_LR : REG_PC))) |
| 12627 | { |
| 12628 | inst.instruction = THUMB_OP16 (inst.instruction); |
| 12629 | inst.instruction |= THUMB_PP_PC_LR; |
| 12630 | inst.instruction |= mask & 0xff; |
| 12631 | } |
| 12632 | else if (unified_syntax) |
| 12633 | { |
| 12634 | inst.instruction = THUMB_OP32 (inst.instruction); |
| 12635 | encode_thumb2_ldmstm (13, mask, TRUE); |
| 12636 | } |
| 12637 | else |
| 12638 | { |
| 12639 | inst.error = _("invalid register list to push/pop instruction"); |
| 12640 | return; |
| 12641 | } |
| 12642 | } |
| 12643 | |
| 12644 | static void |
| 12645 | do_t_rbit (void) |
| 12646 | { |
| 12647 | unsigned Rd, Rm; |
| 12648 | |
| 12649 | Rd = inst.operands[0].reg; |
| 12650 | Rm = inst.operands[1].reg; |
| 12651 | |
| 12652 | reject_bad_reg (Rd); |
| 12653 | reject_bad_reg (Rm); |
| 12654 | |
| 12655 | inst.instruction |= Rd << 8; |
| 12656 | inst.instruction |= Rm << 16; |
| 12657 | inst.instruction |= Rm; |
| 12658 | } |
| 12659 | |
| 12660 | static void |
| 12661 | do_t_rev (void) |
| 12662 | { |
| 12663 | unsigned Rd, Rm; |
| 12664 | |
| 12665 | Rd = inst.operands[0].reg; |
| 12666 | Rm = inst.operands[1].reg; |
| 12667 | |
| 12668 | reject_bad_reg (Rd); |
| 12669 | reject_bad_reg (Rm); |
| 12670 | |
| 12671 | if (Rd <= 7 && Rm <= 7 |
| 12672 | && inst.size_req != 4) |
| 12673 | { |
| 12674 | inst.instruction = THUMB_OP16 (inst.instruction); |
| 12675 | inst.instruction |= Rd; |
| 12676 | inst.instruction |= Rm << 3; |
| 12677 | } |
| 12678 | else if (unified_syntax) |
| 12679 | { |
| 12680 | inst.instruction = THUMB_OP32 (inst.instruction); |
| 12681 | inst.instruction |= Rd << 8; |
| 12682 | inst.instruction |= Rm << 16; |
| 12683 | inst.instruction |= Rm; |
| 12684 | } |
| 12685 | else |
| 12686 | inst.error = BAD_HIREG; |
| 12687 | } |
| 12688 | |
| 12689 | static void |
| 12690 | do_t_rrx (void) |
| 12691 | { |
| 12692 | unsigned Rd, Rm; |
| 12693 | |
| 12694 | Rd = inst.operands[0].reg; |
| 12695 | Rm = inst.operands[1].reg; |
| 12696 | |
| 12697 | reject_bad_reg (Rd); |
| 12698 | reject_bad_reg (Rm); |
| 12699 | |
| 12700 | inst.instruction |= Rd << 8; |
| 12701 | inst.instruction |= Rm; |
| 12702 | } |
| 12703 | |
| 12704 | static void |
| 12705 | do_t_rsb (void) |
| 12706 | { |
| 12707 | unsigned Rd, Rs; |
| 12708 | |
| 12709 | Rd = inst.operands[0].reg; |
| 12710 | Rs = (inst.operands[1].present |
| 12711 | ? inst.operands[1].reg /* Rd, Rs, foo */ |
| 12712 | : inst.operands[0].reg); /* Rd, foo -> Rd, Rd, foo */ |
| 12713 | |
| 12714 | reject_bad_reg (Rd); |
| 12715 | reject_bad_reg (Rs); |
| 12716 | if (inst.operands[2].isreg) |
| 12717 | reject_bad_reg (inst.operands[2].reg); |
| 12718 | |
| 12719 | inst.instruction |= Rd << 8; |
| 12720 | inst.instruction |= Rs << 16; |
| 12721 | if (!inst.operands[2].isreg) |
| 12722 | { |
| 12723 | bfd_boolean narrow; |
| 12724 | |
| 12725 | if ((inst.instruction & 0x00100000) != 0) |
| 12726 | narrow = !in_it_block (); |
| 12727 | else |
| 12728 | narrow = in_it_block (); |
| 12729 | |
| 12730 | if (Rd > 7 || Rs > 7) |
| 12731 | narrow = FALSE; |
| 12732 | |
| 12733 | if (inst.size_req == 4 || !unified_syntax) |
| 12734 | narrow = FALSE; |
| 12735 | |
| 12736 | if (inst.reloc.exp.X_op != O_constant |
| 12737 | || inst.reloc.exp.X_add_number != 0) |
| 12738 | narrow = FALSE; |
| 12739 | |
| 12740 | /* Turn rsb #0 into 16-bit neg. We should probably do this via |
| 12741 | relaxation, but it doesn't seem worth the hassle. */ |
| 12742 | if (narrow) |
| 12743 | { |
| 12744 | inst.reloc.type = BFD_RELOC_UNUSED; |
| 12745 | inst.instruction = THUMB_OP16 (T_MNEM_negs); |
| 12746 | inst.instruction |= Rs << 3; |
| 12747 | inst.instruction |= Rd; |
| 12748 | } |
| 12749 | else |
| 12750 | { |
| 12751 | inst.instruction = (inst.instruction & 0xe1ffffff) | 0x10000000; |
| 12752 | inst.reloc.type = BFD_RELOC_ARM_T32_IMMEDIATE; |
| 12753 | } |
| 12754 | } |
| 12755 | else |
| 12756 | encode_thumb32_shifted_operand (2); |
| 12757 | } |
| 12758 | |
| 12759 | static void |
| 12760 | do_t_setend (void) |
| 12761 | { |
| 12762 | if (warn_on_deprecated |
| 12763 | && ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v8)) |
| 12764 | as_tsktsk (_("setend use is deprecated for ARMv8")); |
| 12765 | |
| 12766 | set_it_insn_type (OUTSIDE_IT_INSN); |
| 12767 | if (inst.operands[0].imm) |
| 12768 | inst.instruction |= 0x8; |
| 12769 | } |
| 12770 | |
| 12771 | static void |
| 12772 | do_t_shift (void) |
| 12773 | { |
| 12774 | if (!inst.operands[1].present) |
| 12775 | inst.operands[1].reg = inst.operands[0].reg; |
| 12776 | |
| 12777 | if (unified_syntax) |
| 12778 | { |
| 12779 | bfd_boolean narrow; |
| 12780 | int shift_kind; |
| 12781 | |
| 12782 | switch (inst.instruction) |
| 12783 | { |
| 12784 | case T_MNEM_asr: |
| 12785 | case T_MNEM_asrs: shift_kind = SHIFT_ASR; break; |
| 12786 | case T_MNEM_lsl: |
| 12787 | case T_MNEM_lsls: shift_kind = SHIFT_LSL; break; |
| 12788 | case T_MNEM_lsr: |
| 12789 | case T_MNEM_lsrs: shift_kind = SHIFT_LSR; break; |
| 12790 | case T_MNEM_ror: |
| 12791 | case T_MNEM_rors: shift_kind = SHIFT_ROR; break; |
| 12792 | default: abort (); |
| 12793 | } |
| 12794 | |
| 12795 | if (THUMB_SETS_FLAGS (inst.instruction)) |
| 12796 | narrow = !in_it_block (); |
| 12797 | else |
| 12798 | narrow = in_it_block (); |
| 12799 | if (inst.operands[0].reg > 7 || inst.operands[1].reg > 7) |
| 12800 | narrow = FALSE; |
| 12801 | if (!inst.operands[2].isreg && shift_kind == SHIFT_ROR) |
| 12802 | narrow = FALSE; |
| 12803 | if (inst.operands[2].isreg |
| 12804 | && (inst.operands[1].reg != inst.operands[0].reg |
| 12805 | || inst.operands[2].reg > 7)) |
| 12806 | narrow = FALSE; |
| 12807 | if (inst.size_req == 4) |
| 12808 | narrow = FALSE; |
| 12809 | |
| 12810 | reject_bad_reg (inst.operands[0].reg); |
| 12811 | reject_bad_reg (inst.operands[1].reg); |
| 12812 | |
| 12813 | if (!narrow) |
| 12814 | { |
| 12815 | if (inst.operands[2].isreg) |
| 12816 | { |
| 12817 | reject_bad_reg (inst.operands[2].reg); |
| 12818 | inst.instruction = THUMB_OP32 (inst.instruction); |
| 12819 | inst.instruction |= inst.operands[0].reg << 8; |
| 12820 | inst.instruction |= inst.operands[1].reg << 16; |
| 12821 | inst.instruction |= inst.operands[2].reg; |
| 12822 | |
| 12823 | /* PR 12854: Error on extraneous shifts. */ |
| 12824 | constraint (inst.operands[2].shifted, |
| 12825 | _("extraneous shift as part of operand to shift insn")); |
| 12826 | } |
| 12827 | else |
| 12828 | { |
| 12829 | inst.operands[1].shifted = 1; |
| 12830 | inst.operands[1].shift_kind = shift_kind; |
| 12831 | inst.instruction = THUMB_OP32 (THUMB_SETS_FLAGS (inst.instruction) |
| 12832 | ? T_MNEM_movs : T_MNEM_mov); |
| 12833 | inst.instruction |= inst.operands[0].reg << 8; |
| 12834 | encode_thumb32_shifted_operand (1); |
| 12835 | /* Prevent the incorrect generation of an ARM_IMMEDIATE fixup. */ |
| 12836 | inst.reloc.type = BFD_RELOC_UNUSED; |
| 12837 | } |
| 12838 | } |
| 12839 | else |
| 12840 | { |
| 12841 | if (inst.operands[2].isreg) |
| 12842 | { |
| 12843 | switch (shift_kind) |
| 12844 | { |
| 12845 | case SHIFT_ASR: inst.instruction = T_OPCODE_ASR_R; break; |
| 12846 | case SHIFT_LSL: inst.instruction = T_OPCODE_LSL_R; break; |
| 12847 | case SHIFT_LSR: inst.instruction = T_OPCODE_LSR_R; break; |
| 12848 | case SHIFT_ROR: inst.instruction = T_OPCODE_ROR_R; break; |
| 12849 | default: abort (); |
| 12850 | } |
| 12851 | |
| 12852 | inst.instruction |= inst.operands[0].reg; |
| 12853 | inst.instruction |= inst.operands[2].reg << 3; |
| 12854 | |
| 12855 | /* PR 12854: Error on extraneous shifts. */ |
| 12856 | constraint (inst.operands[2].shifted, |
| 12857 | _("extraneous shift as part of operand to shift insn")); |
| 12858 | } |
| 12859 | else |
| 12860 | { |
| 12861 | switch (shift_kind) |
| 12862 | { |
| 12863 | case SHIFT_ASR: inst.instruction = T_OPCODE_ASR_I; break; |
| 12864 | case SHIFT_LSL: inst.instruction = T_OPCODE_LSL_I; break; |
| 12865 | case SHIFT_LSR: inst.instruction = T_OPCODE_LSR_I; break; |
| 12866 | default: abort (); |
| 12867 | } |
| 12868 | inst.reloc.type = BFD_RELOC_ARM_THUMB_SHIFT; |
| 12869 | inst.instruction |= inst.operands[0].reg; |
| 12870 | inst.instruction |= inst.operands[1].reg << 3; |
| 12871 | } |
| 12872 | } |
| 12873 | } |
| 12874 | else |
| 12875 | { |
| 12876 | constraint (inst.operands[0].reg > 7 |
| 12877 | || inst.operands[1].reg > 7, BAD_HIREG); |
| 12878 | constraint (THUMB_SETS_FLAGS (inst.instruction), BAD_THUMB32); |
| 12879 | |
| 12880 | if (inst.operands[2].isreg) /* Rd, {Rs,} Rn */ |
| 12881 | { |
| 12882 | constraint (inst.operands[2].reg > 7, BAD_HIREG); |
| 12883 | constraint (inst.operands[0].reg != inst.operands[1].reg, |
| 12884 | _("source1 and dest must be same register")); |
| 12885 | |
| 12886 | switch (inst.instruction) |
| 12887 | { |
| 12888 | case T_MNEM_asr: inst.instruction = T_OPCODE_ASR_R; break; |
| 12889 | case T_MNEM_lsl: inst.instruction = T_OPCODE_LSL_R; break; |
| 12890 | case T_MNEM_lsr: inst.instruction = T_OPCODE_LSR_R; break; |
| 12891 | case T_MNEM_ror: inst.instruction = T_OPCODE_ROR_R; break; |
| 12892 | default: abort (); |
| 12893 | } |
| 12894 | |
| 12895 | inst.instruction |= inst.operands[0].reg; |
| 12896 | inst.instruction |= inst.operands[2].reg << 3; |
| 12897 | |
| 12898 | /* PR 12854: Error on extraneous shifts. */ |
| 12899 | constraint (inst.operands[2].shifted, |
| 12900 | _("extraneous shift as part of operand to shift insn")); |
| 12901 | } |
| 12902 | else |
| 12903 | { |
| 12904 | switch (inst.instruction) |
| 12905 | { |
| 12906 | case T_MNEM_asr: inst.instruction = T_OPCODE_ASR_I; break; |
| 12907 | case T_MNEM_lsl: inst.instruction = T_OPCODE_LSL_I; break; |
| 12908 | case T_MNEM_lsr: inst.instruction = T_OPCODE_LSR_I; break; |
| 12909 | case T_MNEM_ror: inst.error = _("ror #imm not supported"); return; |
| 12910 | default: abort (); |
| 12911 | } |
| 12912 | inst.reloc.type = BFD_RELOC_ARM_THUMB_SHIFT; |
| 12913 | inst.instruction |= inst.operands[0].reg; |
| 12914 | inst.instruction |= inst.operands[1].reg << 3; |
| 12915 | } |
| 12916 | } |
| 12917 | } |
| 12918 | |
| 12919 | static void |
| 12920 | do_t_simd (void) |
| 12921 | { |
| 12922 | unsigned Rd, Rn, Rm; |
| 12923 | |
| 12924 | Rd = inst.operands[0].reg; |
| 12925 | Rn = inst.operands[1].reg; |
| 12926 | Rm = inst.operands[2].reg; |
| 12927 | |
| 12928 | reject_bad_reg (Rd); |
| 12929 | reject_bad_reg (Rn); |
| 12930 | reject_bad_reg (Rm); |
| 12931 | |
| 12932 | inst.instruction |= Rd << 8; |
| 12933 | inst.instruction |= Rn << 16; |
| 12934 | inst.instruction |= Rm; |
| 12935 | } |
| 12936 | |
| 12937 | static void |
| 12938 | do_t_simd2 (void) |
| 12939 | { |
| 12940 | unsigned Rd, Rn, Rm; |
| 12941 | |
| 12942 | Rd = inst.operands[0].reg; |
| 12943 | Rm = inst.operands[1].reg; |
| 12944 | Rn = inst.operands[2].reg; |
| 12945 | |
| 12946 | reject_bad_reg (Rd); |
| 12947 | reject_bad_reg (Rn); |
| 12948 | reject_bad_reg (Rm); |
| 12949 | |
| 12950 | inst.instruction |= Rd << 8; |
| 12951 | inst.instruction |= Rn << 16; |
| 12952 | inst.instruction |= Rm; |
| 12953 | } |
| 12954 | |
| 12955 | static void |
| 12956 | do_t_smc (void) |
| 12957 | { |
| 12958 | unsigned int value = inst.reloc.exp.X_add_number; |
| 12959 | constraint (!ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v7a), |
| 12960 | _("SMC is not permitted on this architecture")); |
| 12961 | constraint (inst.reloc.exp.X_op != O_constant, |
| 12962 | _("expression too complex")); |
| 12963 | inst.reloc.type = BFD_RELOC_UNUSED; |
| 12964 | inst.instruction |= (value & 0xf000) >> 12; |
| 12965 | inst.instruction |= (value & 0x0ff0); |
| 12966 | inst.instruction |= (value & 0x000f) << 16; |
| 12967 | /* PR gas/15623: SMC instructions must be last in an IT block. */ |
| 12968 | set_it_insn_type_last (); |
| 12969 | } |
| 12970 | |
| 12971 | static void |
| 12972 | do_t_hvc (void) |
| 12973 | { |
| 12974 | unsigned int value = inst.reloc.exp.X_add_number; |
| 12975 | |
| 12976 | inst.reloc.type = BFD_RELOC_UNUSED; |
| 12977 | inst.instruction |= (value & 0x0fff); |
| 12978 | inst.instruction |= (value & 0xf000) << 4; |
| 12979 | } |
| 12980 | |
| 12981 | static void |
| 12982 | do_t_ssat_usat (int bias) |
| 12983 | { |
| 12984 | unsigned Rd, Rn; |
| 12985 | |
| 12986 | Rd = inst.operands[0].reg; |
| 12987 | Rn = inst.operands[2].reg; |
| 12988 | |
| 12989 | reject_bad_reg (Rd); |
| 12990 | reject_bad_reg (Rn); |
| 12991 | |
| 12992 | inst.instruction |= Rd << 8; |
| 12993 | inst.instruction |= inst.operands[1].imm - bias; |
| 12994 | inst.instruction |= Rn << 16; |
| 12995 | |
| 12996 | if (inst.operands[3].present) |
| 12997 | { |
| 12998 | offsetT shift_amount = inst.reloc.exp.X_add_number; |
| 12999 | |
| 13000 | inst.reloc.type = BFD_RELOC_UNUSED; |
| 13001 | |
| 13002 | constraint (inst.reloc.exp.X_op != O_constant, |
| 13003 | _("expression too complex")); |
| 13004 | |
| 13005 | if (shift_amount != 0) |
| 13006 | { |
| 13007 | constraint (shift_amount > 31, |
| 13008 | _("shift expression is too large")); |
| 13009 | |
| 13010 | if (inst.operands[3].shift_kind == SHIFT_ASR) |
| 13011 | inst.instruction |= 0x00200000; /* sh bit. */ |
| 13012 | |
| 13013 | inst.instruction |= (shift_amount & 0x1c) << 10; |
| 13014 | inst.instruction |= (shift_amount & 0x03) << 6; |
| 13015 | } |
| 13016 | } |
| 13017 | } |
| 13018 | |
| 13019 | static void |
| 13020 | do_t_ssat (void) |
| 13021 | { |
| 13022 | do_t_ssat_usat (1); |
| 13023 | } |
| 13024 | |
| 13025 | static void |
| 13026 | do_t_ssat16 (void) |
| 13027 | { |
| 13028 | unsigned Rd, Rn; |
| 13029 | |
| 13030 | Rd = inst.operands[0].reg; |
| 13031 | Rn = inst.operands[2].reg; |
| 13032 | |
| 13033 | reject_bad_reg (Rd); |
| 13034 | reject_bad_reg (Rn); |
| 13035 | |
| 13036 | inst.instruction |= Rd << 8; |
| 13037 | inst.instruction |= inst.operands[1].imm - 1; |
| 13038 | inst.instruction |= Rn << 16; |
| 13039 | } |
| 13040 | |
| 13041 | static void |
| 13042 | do_t_strex (void) |
| 13043 | { |
| 13044 | constraint (!inst.operands[2].isreg || !inst.operands[2].preind |
| 13045 | || inst.operands[2].postind || inst.operands[2].writeback |
| 13046 | || inst.operands[2].immisreg || inst.operands[2].shifted |
| 13047 | || inst.operands[2].negative, |
| 13048 | BAD_ADDR_MODE); |
| 13049 | |
| 13050 | constraint (inst.operands[2].reg == REG_PC, BAD_PC); |
| 13051 | |
| 13052 | inst.instruction |= inst.operands[0].reg << 8; |
| 13053 | inst.instruction |= inst.operands[1].reg << 12; |
| 13054 | inst.instruction |= inst.operands[2].reg << 16; |
| 13055 | inst.reloc.type = BFD_RELOC_ARM_T32_OFFSET_U8; |
| 13056 | } |
| 13057 | |
| 13058 | static void |
| 13059 | do_t_strexd (void) |
| 13060 | { |
| 13061 | if (!inst.operands[2].present) |
| 13062 | inst.operands[2].reg = inst.operands[1].reg + 1; |
| 13063 | |
| 13064 | constraint (inst.operands[0].reg == inst.operands[1].reg |
| 13065 | || inst.operands[0].reg == inst.operands[2].reg |
| 13066 | || inst.operands[0].reg == inst.operands[3].reg, |
| 13067 | BAD_OVERLAP); |
| 13068 | |
| 13069 | inst.instruction |= inst.operands[0].reg; |
| 13070 | inst.instruction |= inst.operands[1].reg << 12; |
| 13071 | inst.instruction |= inst.operands[2].reg << 8; |
| 13072 | inst.instruction |= inst.operands[3].reg << 16; |
| 13073 | } |
| 13074 | |
| 13075 | static void |
| 13076 | do_t_sxtah (void) |
| 13077 | { |
| 13078 | unsigned Rd, Rn, Rm; |
| 13079 | |
| 13080 | Rd = inst.operands[0].reg; |
| 13081 | Rn = inst.operands[1].reg; |
| 13082 | Rm = inst.operands[2].reg; |
| 13083 | |
| 13084 | reject_bad_reg (Rd); |
| 13085 | reject_bad_reg (Rn); |
| 13086 | reject_bad_reg (Rm); |
| 13087 | |
| 13088 | inst.instruction |= Rd << 8; |
| 13089 | inst.instruction |= Rn << 16; |
| 13090 | inst.instruction |= Rm; |
| 13091 | inst.instruction |= inst.operands[3].imm << 4; |
| 13092 | } |
| 13093 | |
| 13094 | static void |
| 13095 | do_t_sxth (void) |
| 13096 | { |
| 13097 | unsigned Rd, Rm; |
| 13098 | |
| 13099 | Rd = inst.operands[0].reg; |
| 13100 | Rm = inst.operands[1].reg; |
| 13101 | |
| 13102 | reject_bad_reg (Rd); |
| 13103 | reject_bad_reg (Rm); |
| 13104 | |
| 13105 | if (inst.instruction <= 0xffff |
| 13106 | && inst.size_req != 4 |
| 13107 | && Rd <= 7 && Rm <= 7 |
| 13108 | && (!inst.operands[2].present || inst.operands[2].imm == 0)) |
| 13109 | { |
| 13110 | inst.instruction = THUMB_OP16 (inst.instruction); |
| 13111 | inst.instruction |= Rd; |
| 13112 | inst.instruction |= Rm << 3; |
| 13113 | } |
| 13114 | else if (unified_syntax) |
| 13115 | { |
| 13116 | if (inst.instruction <= 0xffff) |
| 13117 | inst.instruction = THUMB_OP32 (inst.instruction); |
| 13118 | inst.instruction |= Rd << 8; |
| 13119 | inst.instruction |= Rm; |
| 13120 | inst.instruction |= inst.operands[2].imm << 4; |
| 13121 | } |
| 13122 | else |
| 13123 | { |
| 13124 | constraint (inst.operands[2].present && inst.operands[2].imm != 0, |
| 13125 | _("Thumb encoding does not support rotation")); |
| 13126 | constraint (1, BAD_HIREG); |
| 13127 | } |
| 13128 | } |
| 13129 | |
| 13130 | static void |
| 13131 | do_t_swi (void) |
| 13132 | { |
| 13133 | inst.reloc.type = BFD_RELOC_ARM_SWI; |
| 13134 | } |
| 13135 | |
| 13136 | static void |
| 13137 | do_t_tb (void) |
| 13138 | { |
| 13139 | unsigned Rn, Rm; |
| 13140 | int half; |
| 13141 | |
| 13142 | half = (inst.instruction & 0x10) != 0; |
| 13143 | set_it_insn_type_last (); |
| 13144 | constraint (inst.operands[0].immisreg, |
| 13145 | _("instruction requires register index")); |
| 13146 | |
| 13147 | Rn = inst.operands[0].reg; |
| 13148 | Rm = inst.operands[0].imm; |
| 13149 | |
| 13150 | if (!ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v8)) |
| 13151 | constraint (Rn == REG_SP, BAD_SP); |
| 13152 | reject_bad_reg (Rm); |
| 13153 | |
| 13154 | constraint (!half && inst.operands[0].shifted, |
| 13155 | _("instruction does not allow shifted index")); |
| 13156 | inst.instruction |= (Rn << 16) | Rm; |
| 13157 | } |
| 13158 | |
| 13159 | static void |
| 13160 | do_t_udf (void) |
| 13161 | { |
| 13162 | if (!inst.operands[0].present) |
| 13163 | inst.operands[0].imm = 0; |
| 13164 | |
| 13165 | if ((unsigned int) inst.operands[0].imm > 255 || inst.size_req == 4) |
| 13166 | { |
| 13167 | constraint (inst.size_req == 2, |
| 13168 | _("immediate value out of range")); |
| 13169 | inst.instruction = THUMB_OP32 (inst.instruction); |
| 13170 | inst.instruction |= (inst.operands[0].imm & 0xf000u) << 4; |
| 13171 | inst.instruction |= (inst.operands[0].imm & 0x0fffu) << 0; |
| 13172 | } |
| 13173 | else |
| 13174 | { |
| 13175 | inst.instruction = THUMB_OP16 (inst.instruction); |
| 13176 | inst.instruction |= inst.operands[0].imm; |
| 13177 | } |
| 13178 | |
| 13179 | set_it_insn_type (NEUTRAL_IT_INSN); |
| 13180 | } |
| 13181 | |
| 13182 | |
| 13183 | static void |
| 13184 | do_t_usat (void) |
| 13185 | { |
| 13186 | do_t_ssat_usat (0); |
| 13187 | } |
| 13188 | |
| 13189 | static void |
| 13190 | do_t_usat16 (void) |
| 13191 | { |
| 13192 | unsigned Rd, Rn; |
| 13193 | |
| 13194 | Rd = inst.operands[0].reg; |
| 13195 | Rn = inst.operands[2].reg; |
| 13196 | |
| 13197 | reject_bad_reg (Rd); |
| 13198 | reject_bad_reg (Rn); |
| 13199 | |
| 13200 | inst.instruction |= Rd << 8; |
| 13201 | inst.instruction |= inst.operands[1].imm; |
| 13202 | inst.instruction |= Rn << 16; |
| 13203 | } |
| 13204 | |
| 13205 | /* Neon instruction encoder helpers. */ |
| 13206 | |
| 13207 | /* Encodings for the different types for various Neon opcodes. */ |
| 13208 | |
| 13209 | /* An "invalid" code for the following tables. */ |
| 13210 | #define N_INV -1u |
| 13211 | |
| 13212 | struct neon_tab_entry |
| 13213 | { |
| 13214 | unsigned integer; |
| 13215 | unsigned float_or_poly; |
| 13216 | unsigned scalar_or_imm; |
| 13217 | }; |
| 13218 | |
| 13219 | /* Map overloaded Neon opcodes to their respective encodings. */ |
| 13220 | #define NEON_ENC_TAB \ |
| 13221 | X(vabd, 0x0000700, 0x1200d00, N_INV), \ |
| 13222 | X(vmax, 0x0000600, 0x0000f00, N_INV), \ |
| 13223 | X(vmin, 0x0000610, 0x0200f00, N_INV), \ |
| 13224 | X(vpadd, 0x0000b10, 0x1000d00, N_INV), \ |
| 13225 | X(vpmax, 0x0000a00, 0x1000f00, N_INV), \ |
| 13226 | X(vpmin, 0x0000a10, 0x1200f00, N_INV), \ |
| 13227 | X(vadd, 0x0000800, 0x0000d00, N_INV), \ |
| 13228 | X(vsub, 0x1000800, 0x0200d00, N_INV), \ |
| 13229 | X(vceq, 0x1000810, 0x0000e00, 0x1b10100), \ |
| 13230 | X(vcge, 0x0000310, 0x1000e00, 0x1b10080), \ |
| 13231 | X(vcgt, 0x0000300, 0x1200e00, 0x1b10000), \ |
| 13232 | /* Register variants of the following two instructions are encoded as |
| 13233 | vcge / vcgt with the operands reversed. */ \ |
| 13234 | X(vclt, 0x0000300, 0x1200e00, 0x1b10200), \ |
| 13235 | X(vcle, 0x0000310, 0x1000e00, 0x1b10180), \ |
| 13236 | X(vfma, N_INV, 0x0000c10, N_INV), \ |
| 13237 | X(vfms, N_INV, 0x0200c10, N_INV), \ |
| 13238 | X(vmla, 0x0000900, 0x0000d10, 0x0800040), \ |
| 13239 | X(vmls, 0x1000900, 0x0200d10, 0x0800440), \ |
| 13240 | X(vmul, 0x0000910, 0x1000d10, 0x0800840), \ |
| 13241 | X(vmull, 0x0800c00, 0x0800e00, 0x0800a40), /* polynomial not float. */ \ |
| 13242 | X(vmlal, 0x0800800, N_INV, 0x0800240), \ |
| 13243 | X(vmlsl, 0x0800a00, N_INV, 0x0800640), \ |
| 13244 | X(vqdmlal, 0x0800900, N_INV, 0x0800340), \ |
| 13245 | X(vqdmlsl, 0x0800b00, N_INV, 0x0800740), \ |
| 13246 | X(vqdmull, 0x0800d00, N_INV, 0x0800b40), \ |
| 13247 | X(vqdmulh, 0x0000b00, N_INV, 0x0800c40), \ |
| 13248 | X(vqrdmulh, 0x1000b00, N_INV, 0x0800d40), \ |
| 13249 | X(vqrdmlah, 0x3000b10, N_INV, 0x0800e40), \ |
| 13250 | X(vqrdmlsh, 0x3000c10, N_INV, 0x0800f40), \ |
| 13251 | X(vshl, 0x0000400, N_INV, 0x0800510), \ |
| 13252 | X(vqshl, 0x0000410, N_INV, 0x0800710), \ |
| 13253 | X(vand, 0x0000110, N_INV, 0x0800030), \ |
| 13254 | X(vbic, 0x0100110, N_INV, 0x0800030), \ |
| 13255 | X(veor, 0x1000110, N_INV, N_INV), \ |
| 13256 | X(vorn, 0x0300110, N_INV, 0x0800010), \ |
| 13257 | X(vorr, 0x0200110, N_INV, 0x0800010), \ |
| 13258 | X(vmvn, 0x1b00580, N_INV, 0x0800030), \ |
| 13259 | X(vshll, 0x1b20300, N_INV, 0x0800a10), /* max shift, immediate. */ \ |
| 13260 | X(vcvt, 0x1b30600, N_INV, 0x0800e10), /* integer, fixed-point. */ \ |
| 13261 | X(vdup, 0xe800b10, N_INV, 0x1b00c00), /* arm, scalar. */ \ |
| 13262 | X(vld1, 0x0200000, 0x0a00000, 0x0a00c00), /* interlv, lane, dup. */ \ |
| 13263 | X(vst1, 0x0000000, 0x0800000, N_INV), \ |
| 13264 | X(vld2, 0x0200100, 0x0a00100, 0x0a00d00), \ |
| 13265 | X(vst2, 0x0000100, 0x0800100, N_INV), \ |
| 13266 | X(vld3, 0x0200200, 0x0a00200, 0x0a00e00), \ |
| 13267 | X(vst3, 0x0000200, 0x0800200, N_INV), \ |
| 13268 | X(vld4, 0x0200300, 0x0a00300, 0x0a00f00), \ |
| 13269 | X(vst4, 0x0000300, 0x0800300, N_INV), \ |
| 13270 | X(vmovn, 0x1b20200, N_INV, N_INV), \ |
| 13271 | X(vtrn, 0x1b20080, N_INV, N_INV), \ |
| 13272 | X(vqmovn, 0x1b20200, N_INV, N_INV), \ |
| 13273 | X(vqmovun, 0x1b20240, N_INV, N_INV), \ |
| 13274 | X(vnmul, 0xe200a40, 0xe200b40, N_INV), \ |
| 13275 | X(vnmla, 0xe100a40, 0xe100b40, N_INV), \ |
| 13276 | X(vnmls, 0xe100a00, 0xe100b00, N_INV), \ |
| 13277 | X(vfnma, 0xe900a40, 0xe900b40, N_INV), \ |
| 13278 | X(vfnms, 0xe900a00, 0xe900b00, N_INV), \ |
| 13279 | X(vcmp, 0xeb40a40, 0xeb40b40, N_INV), \ |
| 13280 | X(vcmpz, 0xeb50a40, 0xeb50b40, N_INV), \ |
| 13281 | X(vcmpe, 0xeb40ac0, 0xeb40bc0, N_INV), \ |
| 13282 | X(vcmpez, 0xeb50ac0, 0xeb50bc0, N_INV), \ |
| 13283 | X(vseleq, 0xe000a00, N_INV, N_INV), \ |
| 13284 | X(vselvs, 0xe100a00, N_INV, N_INV), \ |
| 13285 | X(vselge, 0xe200a00, N_INV, N_INV), \ |
| 13286 | X(vselgt, 0xe300a00, N_INV, N_INV), \ |
| 13287 | X(vmaxnm, 0xe800a00, 0x3000f10, N_INV), \ |
| 13288 | X(vminnm, 0xe800a40, 0x3200f10, N_INV), \ |
| 13289 | X(vcvta, 0xebc0a40, 0x3bb0000, N_INV), \ |
| 13290 | X(vrintr, 0xeb60a40, 0x3ba0400, N_INV), \ |
| 13291 | X(vrinta, 0xeb80a40, 0x3ba0400, N_INV), \ |
| 13292 | X(aes, 0x3b00300, N_INV, N_INV), \ |
| 13293 | X(sha3op, 0x2000c00, N_INV, N_INV), \ |
| 13294 | X(sha1h, 0x3b902c0, N_INV, N_INV), \ |
| 13295 | X(sha2op, 0x3ba0380, N_INV, N_INV) |
| 13296 | |
| 13297 | enum neon_opc |
| 13298 | { |
| 13299 | #define X(OPC,I,F,S) N_MNEM_##OPC |
| 13300 | NEON_ENC_TAB |
| 13301 | #undef X |
| 13302 | }; |
| 13303 | |
| 13304 | static const struct neon_tab_entry neon_enc_tab[] = |
| 13305 | { |
| 13306 | #define X(OPC,I,F,S) { (I), (F), (S) } |
| 13307 | NEON_ENC_TAB |
| 13308 | #undef X |
| 13309 | }; |
| 13310 | |
| 13311 | /* Do not use these macros; instead, use NEON_ENCODE defined below. */ |
| 13312 | #define NEON_ENC_INTEGER_(X) (neon_enc_tab[(X) & 0x0fffffff].integer) |
| 13313 | #define NEON_ENC_ARMREG_(X) (neon_enc_tab[(X) & 0x0fffffff].integer) |
| 13314 | #define NEON_ENC_POLY_(X) (neon_enc_tab[(X) & 0x0fffffff].float_or_poly) |
| 13315 | #define NEON_ENC_FLOAT_(X) (neon_enc_tab[(X) & 0x0fffffff].float_or_poly) |
| 13316 | #define NEON_ENC_SCALAR_(X) (neon_enc_tab[(X) & 0x0fffffff].scalar_or_imm) |
| 13317 | #define NEON_ENC_IMMED_(X) (neon_enc_tab[(X) & 0x0fffffff].scalar_or_imm) |
| 13318 | #define NEON_ENC_INTERLV_(X) (neon_enc_tab[(X) & 0x0fffffff].integer) |
| 13319 | #define NEON_ENC_LANE_(X) (neon_enc_tab[(X) & 0x0fffffff].float_or_poly) |
| 13320 | #define NEON_ENC_DUP_(X) (neon_enc_tab[(X) & 0x0fffffff].scalar_or_imm) |
| 13321 | #define NEON_ENC_SINGLE_(X) \ |
| 13322 | ((neon_enc_tab[(X) & 0x0fffffff].integer) | ((X) & 0xf0000000)) |
| 13323 | #define NEON_ENC_DOUBLE_(X) \ |
| 13324 | ((neon_enc_tab[(X) & 0x0fffffff].float_or_poly) | ((X) & 0xf0000000)) |
| 13325 | #define NEON_ENC_FPV8_(X) \ |
| 13326 | ((neon_enc_tab[(X) & 0x0fffffff].integer) | ((X) & 0xf000000)) |
| 13327 | |
| 13328 | #define NEON_ENCODE(type, inst) \ |
| 13329 | do \ |
| 13330 | { \ |
| 13331 | inst.instruction = NEON_ENC_##type##_ (inst.instruction); \ |
| 13332 | inst.is_neon = 1; \ |
| 13333 | } \ |
| 13334 | while (0) |
| 13335 | |
| 13336 | #define check_neon_suffixes \ |
| 13337 | do \ |
| 13338 | { \ |
| 13339 | if (!inst.error && inst.vectype.elems > 0 && !inst.is_neon) \ |
| 13340 | { \ |
| 13341 | as_bad (_("invalid neon suffix for non neon instruction")); \ |
| 13342 | return; \ |
| 13343 | } \ |
| 13344 | } \ |
| 13345 | while (0) |
| 13346 | |
| 13347 | /* Define shapes for instruction operands. The following mnemonic characters |
| 13348 | are used in this table: |
| 13349 | |
| 13350 | F - VFP S<n> register |
| 13351 | D - Neon D<n> register |
| 13352 | Q - Neon Q<n> register |
| 13353 | I - Immediate |
| 13354 | S - Scalar |
| 13355 | R - ARM register |
| 13356 | L - D<n> register list |
| 13357 | |
| 13358 | This table is used to generate various data: |
| 13359 | - enumerations of the form NS_DDR to be used as arguments to |
| 13360 | neon_select_shape. |
| 13361 | - a table classifying shapes into single, double, quad, mixed. |
| 13362 | - a table used to drive neon_select_shape. */ |
| 13363 | |
| 13364 | #define NEON_SHAPE_DEF \ |
| 13365 | X(3, (D, D, D), DOUBLE), \ |
| 13366 | X(3, (Q, Q, Q), QUAD), \ |
| 13367 | X(3, (D, D, I), DOUBLE), \ |
| 13368 | X(3, (Q, Q, I), QUAD), \ |
| 13369 | X(3, (D, D, S), DOUBLE), \ |
| 13370 | X(3, (Q, Q, S), QUAD), \ |
| 13371 | X(2, (D, D), DOUBLE), \ |
| 13372 | X(2, (Q, Q), QUAD), \ |
| 13373 | X(2, (D, S), DOUBLE), \ |
| 13374 | X(2, (Q, S), QUAD), \ |
| 13375 | X(2, (D, R), DOUBLE), \ |
| 13376 | X(2, (Q, R), QUAD), \ |
| 13377 | X(2, (D, I), DOUBLE), \ |
| 13378 | X(2, (Q, I), QUAD), \ |
| 13379 | X(3, (D, L, D), DOUBLE), \ |
| 13380 | X(2, (D, Q), MIXED), \ |
| 13381 | X(2, (Q, D), MIXED), \ |
| 13382 | X(3, (D, Q, I), MIXED), \ |
| 13383 | X(3, (Q, D, I), MIXED), \ |
| 13384 | X(3, (Q, D, D), MIXED), \ |
| 13385 | X(3, (D, Q, Q), MIXED), \ |
| 13386 | X(3, (Q, Q, D), MIXED), \ |
| 13387 | X(3, (Q, D, S), MIXED), \ |
| 13388 | X(3, (D, Q, S), MIXED), \ |
| 13389 | X(4, (D, D, D, I), DOUBLE), \ |
| 13390 | X(4, (Q, Q, Q, I), QUAD), \ |
| 13391 | X(4, (D, D, S, I), DOUBLE), \ |
| 13392 | X(4, (Q, Q, S, I), QUAD), \ |
| 13393 | X(2, (F, F), SINGLE), \ |
| 13394 | X(3, (F, F, F), SINGLE), \ |
| 13395 | X(2, (F, I), SINGLE), \ |
| 13396 | X(2, (F, D), MIXED), \ |
| 13397 | X(2, (D, F), MIXED), \ |
| 13398 | X(3, (F, F, I), MIXED), \ |
| 13399 | X(4, (R, R, F, F), SINGLE), \ |
| 13400 | X(4, (F, F, R, R), SINGLE), \ |
| 13401 | X(3, (D, R, R), DOUBLE), \ |
| 13402 | X(3, (R, R, D), DOUBLE), \ |
| 13403 | X(2, (S, R), SINGLE), \ |
| 13404 | X(2, (R, S), SINGLE), \ |
| 13405 | X(2, (F, R), SINGLE), \ |
| 13406 | X(2, (R, F), SINGLE), \ |
| 13407 | /* Half float shape supported so far. */\ |
| 13408 | X (2, (H, D), MIXED), \ |
| 13409 | X (2, (D, H), MIXED), \ |
| 13410 | X (2, (H, F), MIXED), \ |
| 13411 | X (2, (F, H), MIXED), \ |
| 13412 | X (2, (H, H), HALF), \ |
| 13413 | X (2, (H, R), HALF), \ |
| 13414 | X (2, (R, H), HALF), \ |
| 13415 | X (2, (H, I), HALF), \ |
| 13416 | X (3, (H, H, H), HALF), \ |
| 13417 | X (3, (H, F, I), MIXED), \ |
| 13418 | X (3, (F, H, I), MIXED) |
| 13419 | |
| 13420 | #define S2(A,B) NS_##A##B |
| 13421 | #define S3(A,B,C) NS_##A##B##C |
| 13422 | #define S4(A,B,C,D) NS_##A##B##C##D |
| 13423 | |
| 13424 | #define X(N, L, C) S##N L |
| 13425 | |
| 13426 | enum neon_shape |
| 13427 | { |
| 13428 | NEON_SHAPE_DEF, |
| 13429 | NS_NULL |
| 13430 | }; |
| 13431 | |
| 13432 | #undef X |
| 13433 | #undef S2 |
| 13434 | #undef S3 |
| 13435 | #undef S4 |
| 13436 | |
| 13437 | enum neon_shape_class |
| 13438 | { |
| 13439 | SC_HALF, |
| 13440 | SC_SINGLE, |
| 13441 | SC_DOUBLE, |
| 13442 | SC_QUAD, |
| 13443 | SC_MIXED |
| 13444 | }; |
| 13445 | |
| 13446 | #define X(N, L, C) SC_##C |
| 13447 | |
| 13448 | static enum neon_shape_class neon_shape_class[] = |
| 13449 | { |
| 13450 | NEON_SHAPE_DEF |
| 13451 | }; |
| 13452 | |
| 13453 | #undef X |
| 13454 | |
| 13455 | enum neon_shape_el |
| 13456 | { |
| 13457 | SE_H, |
| 13458 | SE_F, |
| 13459 | SE_D, |
| 13460 | SE_Q, |
| 13461 | SE_I, |
| 13462 | SE_S, |
| 13463 | SE_R, |
| 13464 | SE_L |
| 13465 | }; |
| 13466 | |
| 13467 | /* Register widths of above. */ |
| 13468 | static unsigned neon_shape_el_size[] = |
| 13469 | { |
| 13470 | 16, |
| 13471 | 32, |
| 13472 | 64, |
| 13473 | 128, |
| 13474 | 0, |
| 13475 | 32, |
| 13476 | 32, |
| 13477 | 0 |
| 13478 | }; |
| 13479 | |
| 13480 | struct neon_shape_info |
| 13481 | { |
| 13482 | unsigned els; |
| 13483 | enum neon_shape_el el[NEON_MAX_TYPE_ELS]; |
| 13484 | }; |
| 13485 | |
| 13486 | #define S2(A,B) { SE_##A, SE_##B } |
| 13487 | #define S3(A,B,C) { SE_##A, SE_##B, SE_##C } |
| 13488 | #define S4(A,B,C,D) { SE_##A, SE_##B, SE_##C, SE_##D } |
| 13489 | |
| 13490 | #define X(N, L, C) { N, S##N L } |
| 13491 | |
| 13492 | static struct neon_shape_info neon_shape_tab[] = |
| 13493 | { |
| 13494 | NEON_SHAPE_DEF |
| 13495 | }; |
| 13496 | |
| 13497 | #undef X |
| 13498 | #undef S2 |
| 13499 | #undef S3 |
| 13500 | #undef S4 |
| 13501 | |
| 13502 | /* Bit masks used in type checking given instructions. |
| 13503 | 'N_EQK' means the type must be the same as (or based on in some way) the key |
| 13504 | type, which itself is marked with the 'N_KEY' bit. If the 'N_EQK' bit is |
| 13505 | set, various other bits can be set as well in order to modify the meaning of |
| 13506 | the type constraint. */ |
| 13507 | |
| 13508 | enum neon_type_mask |
| 13509 | { |
| 13510 | N_S8 = 0x0000001, |
| 13511 | N_S16 = 0x0000002, |
| 13512 | N_S32 = 0x0000004, |
| 13513 | N_S64 = 0x0000008, |
| 13514 | N_U8 = 0x0000010, |
| 13515 | N_U16 = 0x0000020, |
| 13516 | N_U32 = 0x0000040, |
| 13517 | N_U64 = 0x0000080, |
| 13518 | N_I8 = 0x0000100, |
| 13519 | N_I16 = 0x0000200, |
| 13520 | N_I32 = 0x0000400, |
| 13521 | N_I64 = 0x0000800, |
| 13522 | N_8 = 0x0001000, |
| 13523 | N_16 = 0x0002000, |
| 13524 | N_32 = 0x0004000, |
| 13525 | N_64 = 0x0008000, |
| 13526 | N_P8 = 0x0010000, |
| 13527 | N_P16 = 0x0020000, |
| 13528 | N_F16 = 0x0040000, |
| 13529 | N_F32 = 0x0080000, |
| 13530 | N_F64 = 0x0100000, |
| 13531 | N_P64 = 0x0200000, |
| 13532 | N_KEY = 0x1000000, /* Key element (main type specifier). */ |
| 13533 | N_EQK = 0x2000000, /* Given operand has the same type & size as the key. */ |
| 13534 | N_VFP = 0x4000000, /* VFP mode: operand size must match register width. */ |
| 13535 | N_UNT = 0x8000000, /* Must be explicitly untyped. */ |
| 13536 | N_DBL = 0x0000001, /* If N_EQK, this operand is twice the size. */ |
| 13537 | N_HLF = 0x0000002, /* If N_EQK, this operand is half the size. */ |
| 13538 | N_SGN = 0x0000004, /* If N_EQK, this operand is forced to be signed. */ |
| 13539 | N_UNS = 0x0000008, /* If N_EQK, this operand is forced to be unsigned. */ |
| 13540 | N_INT = 0x0000010, /* If N_EQK, this operand is forced to be integer. */ |
| 13541 | N_FLT = 0x0000020, /* If N_EQK, this operand is forced to be float. */ |
| 13542 | N_SIZ = 0x0000040, /* If N_EQK, this operand is forced to be size-only. */ |
| 13543 | N_UTYP = 0, |
| 13544 | N_MAX_NONSPECIAL = N_P64 |
| 13545 | }; |
| 13546 | |
| 13547 | #define N_ALLMODS (N_DBL | N_HLF | N_SGN | N_UNS | N_INT | N_FLT | N_SIZ) |
| 13548 | |
| 13549 | #define N_SU_ALL (N_S8 | N_S16 | N_S32 | N_S64 | N_U8 | N_U16 | N_U32 | N_U64) |
| 13550 | #define N_SU_32 (N_S8 | N_S16 | N_S32 | N_U8 | N_U16 | N_U32) |
| 13551 | #define N_SU_16_64 (N_S16 | N_S32 | N_S64 | N_U16 | N_U32 | N_U64) |
| 13552 | #define N_S_32 (N_S8 | N_S16 | N_S32) |
| 13553 | #define N_F_16_32 (N_F16 | N_F32) |
| 13554 | #define N_SUF_32 (N_SU_32 | N_F_16_32) |
| 13555 | #define N_I_ALL (N_I8 | N_I16 | N_I32 | N_I64) |
| 13556 | #define N_IF_32 (N_I8 | N_I16 | N_I32 | N_F16 | N_F32) |
| 13557 | #define N_F_ALL (N_F16 | N_F32 | N_F64) |
| 13558 | |
| 13559 | /* Pass this as the first type argument to neon_check_type to ignore types |
| 13560 | altogether. */ |
| 13561 | #define N_IGNORE_TYPE (N_KEY | N_EQK) |
| 13562 | |
| 13563 | /* Select a "shape" for the current instruction (describing register types or |
| 13564 | sizes) from a list of alternatives. Return NS_NULL if the current instruction |
| 13565 | doesn't fit. For non-polymorphic shapes, checking is usually done as a |
| 13566 | function of operand parsing, so this function doesn't need to be called. |
| 13567 | Shapes should be listed in order of decreasing length. */ |
| 13568 | |
| 13569 | static enum neon_shape |
| 13570 | neon_select_shape (enum neon_shape shape, ...) |
| 13571 | { |
| 13572 | va_list ap; |
| 13573 | enum neon_shape first_shape = shape; |
| 13574 | |
| 13575 | /* Fix missing optional operands. FIXME: we don't know at this point how |
| 13576 | many arguments we should have, so this makes the assumption that we have |
| 13577 | > 1. This is true of all current Neon opcodes, I think, but may not be |
| 13578 | true in the future. */ |
| 13579 | if (!inst.operands[1].present) |
| 13580 | inst.operands[1] = inst.operands[0]; |
| 13581 | |
| 13582 | va_start (ap, shape); |
| 13583 | |
| 13584 | for (; shape != NS_NULL; shape = (enum neon_shape) va_arg (ap, int)) |
| 13585 | { |
| 13586 | unsigned j; |
| 13587 | int matches = 1; |
| 13588 | |
| 13589 | for (j = 0; j < neon_shape_tab[shape].els; j++) |
| 13590 | { |
| 13591 | if (!inst.operands[j].present) |
| 13592 | { |
| 13593 | matches = 0; |
| 13594 | break; |
| 13595 | } |
| 13596 | |
| 13597 | switch (neon_shape_tab[shape].el[j]) |
| 13598 | { |
| 13599 | /* If a .f16, .16, .u16, .s16 type specifier is given over |
| 13600 | a VFP single precision register operand, it's essentially |
| 13601 | means only half of the register is used. |
| 13602 | |
| 13603 | If the type specifier is given after the mnemonics, the |
| 13604 | information is stored in inst.vectype. If the type specifier |
| 13605 | is given after register operand, the information is stored |
| 13606 | in inst.operands[].vectype. |
| 13607 | |
| 13608 | When there is only one type specifier, and all the register |
| 13609 | operands are the same type of hardware register, the type |
| 13610 | specifier applies to all register operands. |
| 13611 | |
| 13612 | If no type specifier is given, the shape is inferred from |
| 13613 | operand information. |
| 13614 | |
| 13615 | for example: |
| 13616 | vadd.f16 s0, s1, s2: NS_HHH |
| 13617 | vabs.f16 s0, s1: NS_HH |
| 13618 | vmov.f16 s0, r1: NS_HR |
| 13619 | vmov.f16 r0, s1: NS_RH |
| 13620 | vcvt.f16 r0, s1: NS_RH |
| 13621 | vcvt.f16.s32 s2, s2, #29: NS_HFI |
| 13622 | vcvt.f16.s32 s2, s2: NS_HF |
| 13623 | */ |
| 13624 | case SE_H: |
| 13625 | if (!(inst.operands[j].isreg |
| 13626 | && inst.operands[j].isvec |
| 13627 | && inst.operands[j].issingle |
| 13628 | && !inst.operands[j].isquad |
| 13629 | && ((inst.vectype.elems == 1 |
| 13630 | && inst.vectype.el[0].size == 16) |
| 13631 | || (inst.vectype.elems > 1 |
| 13632 | && inst.vectype.el[j].size == 16) |
| 13633 | || (inst.vectype.elems == 0 |
| 13634 | && inst.operands[j].vectype.type != NT_invtype |
| 13635 | && inst.operands[j].vectype.size == 16)))) |
| 13636 | matches = 0; |
| 13637 | break; |
| 13638 | |
| 13639 | case SE_F: |
| 13640 | if (!(inst.operands[j].isreg |
| 13641 | && inst.operands[j].isvec |
| 13642 | && inst.operands[j].issingle |
| 13643 | && !inst.operands[j].isquad |
| 13644 | && ((inst.vectype.elems == 1 && inst.vectype.el[0].size == 32) |
| 13645 | || (inst.vectype.elems > 1 && inst.vectype.el[j].size == 32) |
| 13646 | || (inst.vectype.elems == 0 |
| 13647 | && (inst.operands[j].vectype.size == 32 |
| 13648 | || inst.operands[j].vectype.type == NT_invtype))))) |
| 13649 | matches = 0; |
| 13650 | break; |
| 13651 | |
| 13652 | case SE_D: |
| 13653 | if (!(inst.operands[j].isreg |
| 13654 | && inst.operands[j].isvec |
| 13655 | && !inst.operands[j].isquad |
| 13656 | && !inst.operands[j].issingle)) |
| 13657 | matches = 0; |
| 13658 | break; |
| 13659 | |
| 13660 | case SE_R: |
| 13661 | if (!(inst.operands[j].isreg |
| 13662 | && !inst.operands[j].isvec)) |
| 13663 | matches = 0; |
| 13664 | break; |
| 13665 | |
| 13666 | case SE_Q: |
| 13667 | if (!(inst.operands[j].isreg |
| 13668 | && inst.operands[j].isvec |
| 13669 | && inst.operands[j].isquad |
| 13670 | && !inst.operands[j].issingle)) |
| 13671 | matches = 0; |
| 13672 | break; |
| 13673 | |
| 13674 | case SE_I: |
| 13675 | if (!(!inst.operands[j].isreg |
| 13676 | && !inst.operands[j].isscalar)) |
| 13677 | matches = 0; |
| 13678 | break; |
| 13679 | |
| 13680 | case SE_S: |
| 13681 | if (!(!inst.operands[j].isreg |
| 13682 | && inst.operands[j].isscalar)) |
| 13683 | matches = 0; |
| 13684 | break; |
| 13685 | |
| 13686 | case SE_L: |
| 13687 | break; |
| 13688 | } |
| 13689 | if (!matches) |
| 13690 | break; |
| 13691 | } |
| 13692 | if (matches && (j >= ARM_IT_MAX_OPERANDS || !inst.operands[j].present)) |
| 13693 | /* We've matched all the entries in the shape table, and we don't |
| 13694 | have any left over operands which have not been matched. */ |
| 13695 | break; |
| 13696 | } |
| 13697 | |
| 13698 | va_end (ap); |
| 13699 | |
| 13700 | if (shape == NS_NULL && first_shape != NS_NULL) |
| 13701 | first_error (_("invalid instruction shape")); |
| 13702 | |
| 13703 | return shape; |
| 13704 | } |
| 13705 | |
| 13706 | /* True if SHAPE is predominantly a quadword operation (most of the time, this |
| 13707 | means the Q bit should be set). */ |
| 13708 | |
| 13709 | static int |
| 13710 | neon_quad (enum neon_shape shape) |
| 13711 | { |
| 13712 | return neon_shape_class[shape] == SC_QUAD; |
| 13713 | } |
| 13714 | |
| 13715 | static void |
| 13716 | neon_modify_type_size (unsigned typebits, enum neon_el_type *g_type, |
| 13717 | unsigned *g_size) |
| 13718 | { |
| 13719 | /* Allow modification to be made to types which are constrained to be |
| 13720 | based on the key element, based on bits set alongside N_EQK. */ |
| 13721 | if ((typebits & N_EQK) != 0) |
| 13722 | { |
| 13723 | if ((typebits & N_HLF) != 0) |
| 13724 | *g_size /= 2; |
| 13725 | else if ((typebits & N_DBL) != 0) |
| 13726 | *g_size *= 2; |
| 13727 | if ((typebits & N_SGN) != 0) |
| 13728 | *g_type = NT_signed; |
| 13729 | else if ((typebits & N_UNS) != 0) |
| 13730 | *g_type = NT_unsigned; |
| 13731 | else if ((typebits & N_INT) != 0) |
| 13732 | *g_type = NT_integer; |
| 13733 | else if ((typebits & N_FLT) != 0) |
| 13734 | *g_type = NT_float; |
| 13735 | else if ((typebits & N_SIZ) != 0) |
| 13736 | *g_type = NT_untyped; |
| 13737 | } |
| 13738 | } |
| 13739 | |
| 13740 | /* Return operand OPNO promoted by bits set in THISARG. KEY should be the "key" |
| 13741 | operand type, i.e. the single type specified in a Neon instruction when it |
| 13742 | is the only one given. */ |
| 13743 | |
| 13744 | static struct neon_type_el |
| 13745 | neon_type_promote (struct neon_type_el *key, unsigned thisarg) |
| 13746 | { |
| 13747 | struct neon_type_el dest = *key; |
| 13748 | |
| 13749 | gas_assert ((thisarg & N_EQK) != 0); |
| 13750 | |
| 13751 | neon_modify_type_size (thisarg, &dest.type, &dest.size); |
| 13752 | |
| 13753 | return dest; |
| 13754 | } |
| 13755 | |
| 13756 | /* Convert Neon type and size into compact bitmask representation. */ |
| 13757 | |
| 13758 | static enum neon_type_mask |
| 13759 | type_chk_of_el_type (enum neon_el_type type, unsigned size) |
| 13760 | { |
| 13761 | switch (type) |
| 13762 | { |
| 13763 | case NT_untyped: |
| 13764 | switch (size) |
| 13765 | { |
| 13766 | case 8: return N_8; |
| 13767 | case 16: return N_16; |
| 13768 | case 32: return N_32; |
| 13769 | case 64: return N_64; |
| 13770 | default: ; |
| 13771 | } |
| 13772 | break; |
| 13773 | |
| 13774 | case NT_integer: |
| 13775 | switch (size) |
| 13776 | { |
| 13777 | case 8: return N_I8; |
| 13778 | case 16: return N_I16; |
| 13779 | case 32: return N_I32; |
| 13780 | case 64: return N_I64; |
| 13781 | default: ; |
| 13782 | } |
| 13783 | break; |
| 13784 | |
| 13785 | case NT_float: |
| 13786 | switch (size) |
| 13787 | { |
| 13788 | case 16: return N_F16; |
| 13789 | case 32: return N_F32; |
| 13790 | case 64: return N_F64; |
| 13791 | default: ; |
| 13792 | } |
| 13793 | break; |
| 13794 | |
| 13795 | case NT_poly: |
| 13796 | switch (size) |
| 13797 | { |
| 13798 | case 8: return N_P8; |
| 13799 | case 16: return N_P16; |
| 13800 | case 64: return N_P64; |
| 13801 | default: ; |
| 13802 | } |
| 13803 | break; |
| 13804 | |
| 13805 | case NT_signed: |
| 13806 | switch (size) |
| 13807 | { |
| 13808 | case 8: return N_S8; |
| 13809 | case 16: return N_S16; |
| 13810 | case 32: return N_S32; |
| 13811 | case 64: return N_S64; |
| 13812 | default: ; |
| 13813 | } |
| 13814 | break; |
| 13815 | |
| 13816 | case NT_unsigned: |
| 13817 | switch (size) |
| 13818 | { |
| 13819 | case 8: return N_U8; |
| 13820 | case 16: return N_U16; |
| 13821 | case 32: return N_U32; |
| 13822 | case 64: return N_U64; |
| 13823 | default: ; |
| 13824 | } |
| 13825 | break; |
| 13826 | |
| 13827 | default: ; |
| 13828 | } |
| 13829 | |
| 13830 | return N_UTYP; |
| 13831 | } |
| 13832 | |
| 13833 | /* Convert compact Neon bitmask type representation to a type and size. Only |
| 13834 | handles the case where a single bit is set in the mask. */ |
| 13835 | |
| 13836 | static int |
| 13837 | el_type_of_type_chk (enum neon_el_type *type, unsigned *size, |
| 13838 | enum neon_type_mask mask) |
| 13839 | { |
| 13840 | if ((mask & N_EQK) != 0) |
| 13841 | return FAIL; |
| 13842 | |
| 13843 | if ((mask & (N_S8 | N_U8 | N_I8 | N_8 | N_P8)) != 0) |
| 13844 | *size = 8; |
| 13845 | else if ((mask & (N_S16 | N_U16 | N_I16 | N_16 | N_F16 | N_P16)) != 0) |
| 13846 | *size = 16; |
| 13847 | else if ((mask & (N_S32 | N_U32 | N_I32 | N_32 | N_F32)) != 0) |
| 13848 | *size = 32; |
| 13849 | else if ((mask & (N_S64 | N_U64 | N_I64 | N_64 | N_F64 | N_P64)) != 0) |
| 13850 | *size = 64; |
| 13851 | else |
| 13852 | return FAIL; |
| 13853 | |
| 13854 | if ((mask & (N_S8 | N_S16 | N_S32 | N_S64)) != 0) |
| 13855 | *type = NT_signed; |
| 13856 | else if ((mask & (N_U8 | N_U16 | N_U32 | N_U64)) != 0) |
| 13857 | *type = NT_unsigned; |
| 13858 | else if ((mask & (N_I8 | N_I16 | N_I32 | N_I64)) != 0) |
| 13859 | *type = NT_integer; |
| 13860 | else if ((mask & (N_8 | N_16 | N_32 | N_64)) != 0) |
| 13861 | *type = NT_untyped; |
| 13862 | else if ((mask & (N_P8 | N_P16 | N_P64)) != 0) |
| 13863 | *type = NT_poly; |
| 13864 | else if ((mask & (N_F_ALL)) != 0) |
| 13865 | *type = NT_float; |
| 13866 | else |
| 13867 | return FAIL; |
| 13868 | |
| 13869 | return SUCCESS; |
| 13870 | } |
| 13871 | |
| 13872 | /* Modify a bitmask of allowed types. This is only needed for type |
| 13873 | relaxation. */ |
| 13874 | |
| 13875 | static unsigned |
| 13876 | modify_types_allowed (unsigned allowed, unsigned mods) |
| 13877 | { |
| 13878 | unsigned size; |
| 13879 | enum neon_el_type type; |
| 13880 | unsigned destmask; |
| 13881 | int i; |
| 13882 | |
| 13883 | destmask = 0; |
| 13884 | |
| 13885 | for (i = 1; i <= N_MAX_NONSPECIAL; i <<= 1) |
| 13886 | { |
| 13887 | if (el_type_of_type_chk (&type, &size, |
| 13888 | (enum neon_type_mask) (allowed & i)) == SUCCESS) |
| 13889 | { |
| 13890 | neon_modify_type_size (mods, &type, &size); |
| 13891 | destmask |= type_chk_of_el_type (type, size); |
| 13892 | } |
| 13893 | } |
| 13894 | |
| 13895 | return destmask; |
| 13896 | } |
| 13897 | |
| 13898 | /* Check type and return type classification. |
| 13899 | The manual states (paraphrase): If one datatype is given, it indicates the |
| 13900 | type given in: |
| 13901 | - the second operand, if there is one |
| 13902 | - the operand, if there is no second operand |
| 13903 | - the result, if there are no operands. |
| 13904 | This isn't quite good enough though, so we use a concept of a "key" datatype |
| 13905 | which is set on a per-instruction basis, which is the one which matters when |
| 13906 | only one data type is written. |
| 13907 | Note: this function has side-effects (e.g. filling in missing operands). All |
| 13908 | Neon instructions should call it before performing bit encoding. */ |
| 13909 | |
| 13910 | static struct neon_type_el |
| 13911 | neon_check_type (unsigned els, enum neon_shape ns, ...) |
| 13912 | { |
| 13913 | va_list ap; |
| 13914 | unsigned i, pass, key_el = 0; |
| 13915 | unsigned types[NEON_MAX_TYPE_ELS]; |
| 13916 | enum neon_el_type k_type = NT_invtype; |
| 13917 | unsigned k_size = -1u; |
| 13918 | struct neon_type_el badtype = {NT_invtype, -1}; |
| 13919 | unsigned key_allowed = 0; |
| 13920 | |
| 13921 | /* Optional registers in Neon instructions are always (not) in operand 1. |
| 13922 | Fill in the missing operand here, if it was omitted. */ |
| 13923 | if (els > 1 && !inst.operands[1].present) |
| 13924 | inst.operands[1] = inst.operands[0]; |
| 13925 | |
| 13926 | /* Suck up all the varargs. */ |
| 13927 | va_start (ap, ns); |
| 13928 | for (i = 0; i < els; i++) |
| 13929 | { |
| 13930 | unsigned thisarg = va_arg (ap, unsigned); |
| 13931 | if (thisarg == N_IGNORE_TYPE) |
| 13932 | { |
| 13933 | va_end (ap); |
| 13934 | return badtype; |
| 13935 | } |
| 13936 | types[i] = thisarg; |
| 13937 | if ((thisarg & N_KEY) != 0) |
| 13938 | key_el = i; |
| 13939 | } |
| 13940 | va_end (ap); |
| 13941 | |
| 13942 | if (inst.vectype.elems > 0) |
| 13943 | for (i = 0; i < els; i++) |
| 13944 | if (inst.operands[i].vectype.type != NT_invtype) |
| 13945 | { |
| 13946 | first_error (_("types specified in both the mnemonic and operands")); |
| 13947 | return badtype; |
| 13948 | } |
| 13949 | |
| 13950 | /* Duplicate inst.vectype elements here as necessary. |
| 13951 | FIXME: No idea if this is exactly the same as the ARM assembler, |
| 13952 | particularly when an insn takes one register and one non-register |
| 13953 | operand. */ |
| 13954 | if (inst.vectype.elems == 1 && els > 1) |
| 13955 | { |
| 13956 | unsigned j; |
| 13957 | inst.vectype.elems = els; |
| 13958 | inst.vectype.el[key_el] = inst.vectype.el[0]; |
| 13959 | for (j = 0; j < els; j++) |
| 13960 | if (j != key_el) |
| 13961 | inst.vectype.el[j] = neon_type_promote (&inst.vectype.el[key_el], |
| 13962 | types[j]); |
| 13963 | } |
| 13964 | else if (inst.vectype.elems == 0 && els > 0) |
| 13965 | { |
| 13966 | unsigned j; |
| 13967 | /* No types were given after the mnemonic, so look for types specified |
| 13968 | after each operand. We allow some flexibility here; as long as the |
| 13969 | "key" operand has a type, we can infer the others. */ |
| 13970 | for (j = 0; j < els; j++) |
| 13971 | if (inst.operands[j].vectype.type != NT_invtype) |
| 13972 | inst.vectype.el[j] = inst.operands[j].vectype; |
| 13973 | |
| 13974 | if (inst.operands[key_el].vectype.type != NT_invtype) |
| 13975 | { |
| 13976 | for (j = 0; j < els; j++) |
| 13977 | if (inst.operands[j].vectype.type == NT_invtype) |
| 13978 | inst.vectype.el[j] = neon_type_promote (&inst.vectype.el[key_el], |
| 13979 | types[j]); |
| 13980 | } |
| 13981 | else |
| 13982 | { |
| 13983 | first_error (_("operand types can't be inferred")); |
| 13984 | return badtype; |
| 13985 | } |
| 13986 | } |
| 13987 | else if (inst.vectype.elems != els) |
| 13988 | { |
| 13989 | first_error (_("type specifier has the wrong number of parts")); |
| 13990 | return badtype; |
| 13991 | } |
| 13992 | |
| 13993 | for (pass = 0; pass < 2; pass++) |
| 13994 | { |
| 13995 | for (i = 0; i < els; i++) |
| 13996 | { |
| 13997 | unsigned thisarg = types[i]; |
| 13998 | unsigned types_allowed = ((thisarg & N_EQK) != 0 && pass != 0) |
| 13999 | ? modify_types_allowed (key_allowed, thisarg) : thisarg; |
| 14000 | enum neon_el_type g_type = inst.vectype.el[i].type; |
| 14001 | unsigned g_size = inst.vectype.el[i].size; |
| 14002 | |
| 14003 | /* Decay more-specific signed & unsigned types to sign-insensitive |
| 14004 | integer types if sign-specific variants are unavailable. */ |
| 14005 | if ((g_type == NT_signed || g_type == NT_unsigned) |
| 14006 | && (types_allowed & N_SU_ALL) == 0) |
| 14007 | g_type = NT_integer; |
| 14008 | |
| 14009 | /* If only untyped args are allowed, decay any more specific types to |
| 14010 | them. Some instructions only care about signs for some element |
| 14011 | sizes, so handle that properly. */ |
| 14012 | if (((types_allowed & N_UNT) == 0) |
| 14013 | && ((g_size == 8 && (types_allowed & N_8) != 0) |
| 14014 | || (g_size == 16 && (types_allowed & N_16) != 0) |
| 14015 | || (g_size == 32 && (types_allowed & N_32) != 0) |
| 14016 | || (g_size == 64 && (types_allowed & N_64) != 0))) |
| 14017 | g_type = NT_untyped; |
| 14018 | |
| 14019 | if (pass == 0) |
| 14020 | { |
| 14021 | if ((thisarg & N_KEY) != 0) |
| 14022 | { |
| 14023 | k_type = g_type; |
| 14024 | k_size = g_size; |
| 14025 | key_allowed = thisarg & ~N_KEY; |
| 14026 | |
| 14027 | /* Check architecture constraint on FP16 extension. */ |
| 14028 | if (k_size == 16 |
| 14029 | && k_type == NT_float |
| 14030 | && ! ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_fp16)) |
| 14031 | { |
| 14032 | inst.error = _(BAD_FP16); |
| 14033 | return badtype; |
| 14034 | } |
| 14035 | } |
| 14036 | } |
| 14037 | else |
| 14038 | { |
| 14039 | if ((thisarg & N_VFP) != 0) |
| 14040 | { |
| 14041 | enum neon_shape_el regshape; |
| 14042 | unsigned regwidth, match; |
| 14043 | |
| 14044 | /* PR 11136: Catch the case where we are passed a shape of NS_NULL. */ |
| 14045 | if (ns == NS_NULL) |
| 14046 | { |
| 14047 | first_error (_("invalid instruction shape")); |
| 14048 | return badtype; |
| 14049 | } |
| 14050 | regshape = neon_shape_tab[ns].el[i]; |
| 14051 | regwidth = neon_shape_el_size[regshape]; |
| 14052 | |
| 14053 | /* In VFP mode, operands must match register widths. If we |
| 14054 | have a key operand, use its width, else use the width of |
| 14055 | the current operand. */ |
| 14056 | if (k_size != -1u) |
| 14057 | match = k_size; |
| 14058 | else |
| 14059 | match = g_size; |
| 14060 | |
| 14061 | /* FP16 will use a single precision register. */ |
| 14062 | if (regwidth == 32 && match == 16) |
| 14063 | { |
| 14064 | if (ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_fp16)) |
| 14065 | match = regwidth; |
| 14066 | else |
| 14067 | { |
| 14068 | inst.error = _(BAD_FP16); |
| 14069 | return badtype; |
| 14070 | } |
| 14071 | } |
| 14072 | |
| 14073 | if (regwidth != match) |
| 14074 | { |
| 14075 | first_error (_("operand size must match register width")); |
| 14076 | return badtype; |
| 14077 | } |
| 14078 | } |
| 14079 | |
| 14080 | if ((thisarg & N_EQK) == 0) |
| 14081 | { |
| 14082 | unsigned given_type = type_chk_of_el_type (g_type, g_size); |
| 14083 | |
| 14084 | if ((given_type & types_allowed) == 0) |
| 14085 | { |
| 14086 | first_error (_("bad type in Neon instruction")); |
| 14087 | return badtype; |
| 14088 | } |
| 14089 | } |
| 14090 | else |
| 14091 | { |
| 14092 | enum neon_el_type mod_k_type = k_type; |
| 14093 | unsigned mod_k_size = k_size; |
| 14094 | neon_modify_type_size (thisarg, &mod_k_type, &mod_k_size); |
| 14095 | if (g_type != mod_k_type || g_size != mod_k_size) |
| 14096 | { |
| 14097 | first_error (_("inconsistent types in Neon instruction")); |
| 14098 | return badtype; |
| 14099 | } |
| 14100 | } |
| 14101 | } |
| 14102 | } |
| 14103 | } |
| 14104 | |
| 14105 | return inst.vectype.el[key_el]; |
| 14106 | } |
| 14107 | |
| 14108 | /* Neon-style VFP instruction forwarding. */ |
| 14109 | |
| 14110 | /* Thumb VFP instructions have 0xE in the condition field. */ |
| 14111 | |
| 14112 | static void |
| 14113 | do_vfp_cond_or_thumb (void) |
| 14114 | { |
| 14115 | inst.is_neon = 1; |
| 14116 | |
| 14117 | if (thumb_mode) |
| 14118 | inst.instruction |= 0xe0000000; |
| 14119 | else |
| 14120 | inst.instruction |= inst.cond << 28; |
| 14121 | } |
| 14122 | |
| 14123 | /* Look up and encode a simple mnemonic, for use as a helper function for the |
| 14124 | Neon-style VFP syntax. This avoids duplication of bits of the insns table, |
| 14125 | etc. It is assumed that operand parsing has already been done, and that the |
| 14126 | operands are in the form expected by the given opcode (this isn't necessarily |
| 14127 | the same as the form in which they were parsed, hence some massaging must |
| 14128 | take place before this function is called). |
| 14129 | Checks current arch version against that in the looked-up opcode. */ |
| 14130 | |
| 14131 | static void |
| 14132 | do_vfp_nsyn_opcode (const char *opname) |
| 14133 | { |
| 14134 | const struct asm_opcode *opcode; |
| 14135 | |
| 14136 | opcode = (const struct asm_opcode *) hash_find (arm_ops_hsh, opname); |
| 14137 | |
| 14138 | if (!opcode) |
| 14139 | abort (); |
| 14140 | |
| 14141 | constraint (!ARM_CPU_HAS_FEATURE (cpu_variant, |
| 14142 | thumb_mode ? *opcode->tvariant : *opcode->avariant), |
| 14143 | _(BAD_FPU)); |
| 14144 | |
| 14145 | inst.is_neon = 1; |
| 14146 | |
| 14147 | if (thumb_mode) |
| 14148 | { |
| 14149 | inst.instruction = opcode->tvalue; |
| 14150 | opcode->tencode (); |
| 14151 | } |
| 14152 | else |
| 14153 | { |
| 14154 | inst.instruction = (inst.cond << 28) | opcode->avalue; |
| 14155 | opcode->aencode (); |
| 14156 | } |
| 14157 | } |
| 14158 | |
| 14159 | static void |
| 14160 | do_vfp_nsyn_add_sub (enum neon_shape rs) |
| 14161 | { |
| 14162 | int is_add = (inst.instruction & 0x0fffffff) == N_MNEM_vadd; |
| 14163 | |
| 14164 | if (rs == NS_FFF || rs == NS_HHH) |
| 14165 | { |
| 14166 | if (is_add) |
| 14167 | do_vfp_nsyn_opcode ("fadds"); |
| 14168 | else |
| 14169 | do_vfp_nsyn_opcode ("fsubs"); |
| 14170 | |
| 14171 | /* ARMv8.2 fp16 instruction. */ |
| 14172 | if (rs == NS_HHH) |
| 14173 | do_scalar_fp16_v82_encode (); |
| 14174 | } |
| 14175 | else |
| 14176 | { |
| 14177 | if (is_add) |
| 14178 | do_vfp_nsyn_opcode ("faddd"); |
| 14179 | else |
| 14180 | do_vfp_nsyn_opcode ("fsubd"); |
| 14181 | } |
| 14182 | } |
| 14183 | |
| 14184 | /* Check operand types to see if this is a VFP instruction, and if so call |
| 14185 | PFN (). */ |
| 14186 | |
| 14187 | static int |
| 14188 | try_vfp_nsyn (int args, void (*pfn) (enum neon_shape)) |
| 14189 | { |
| 14190 | enum neon_shape rs; |
| 14191 | struct neon_type_el et; |
| 14192 | |
| 14193 | switch (args) |
| 14194 | { |
| 14195 | case 2: |
| 14196 | rs = neon_select_shape (NS_HH, NS_FF, NS_DD, NS_NULL); |
| 14197 | et = neon_check_type (2, rs, N_EQK | N_VFP, N_F_ALL | N_KEY | N_VFP); |
| 14198 | break; |
| 14199 | |
| 14200 | case 3: |
| 14201 | rs = neon_select_shape (NS_HHH, NS_FFF, NS_DDD, NS_NULL); |
| 14202 | et = neon_check_type (3, rs, N_EQK | N_VFP, N_EQK | N_VFP, |
| 14203 | N_F_ALL | N_KEY | N_VFP); |
| 14204 | break; |
| 14205 | |
| 14206 | default: |
| 14207 | abort (); |
| 14208 | } |
| 14209 | |
| 14210 | if (et.type != NT_invtype) |
| 14211 | { |
| 14212 | pfn (rs); |
| 14213 | return SUCCESS; |
| 14214 | } |
| 14215 | |
| 14216 | inst.error = NULL; |
| 14217 | return FAIL; |
| 14218 | } |
| 14219 | |
| 14220 | static void |
| 14221 | do_vfp_nsyn_mla_mls (enum neon_shape rs) |
| 14222 | { |
| 14223 | int is_mla = (inst.instruction & 0x0fffffff) == N_MNEM_vmla; |
| 14224 | |
| 14225 | if (rs == NS_FFF || rs == NS_HHH) |
| 14226 | { |
| 14227 | if (is_mla) |
| 14228 | do_vfp_nsyn_opcode ("fmacs"); |
| 14229 | else |
| 14230 | do_vfp_nsyn_opcode ("fnmacs"); |
| 14231 | |
| 14232 | /* ARMv8.2 fp16 instruction. */ |
| 14233 | if (rs == NS_HHH) |
| 14234 | do_scalar_fp16_v82_encode (); |
| 14235 | } |
| 14236 | else |
| 14237 | { |
| 14238 | if (is_mla) |
| 14239 | do_vfp_nsyn_opcode ("fmacd"); |
| 14240 | else |
| 14241 | do_vfp_nsyn_opcode ("fnmacd"); |
| 14242 | } |
| 14243 | } |
| 14244 | |
| 14245 | static void |
| 14246 | do_vfp_nsyn_fma_fms (enum neon_shape rs) |
| 14247 | { |
| 14248 | int is_fma = (inst.instruction & 0x0fffffff) == N_MNEM_vfma; |
| 14249 | |
| 14250 | if (rs == NS_FFF || rs == NS_HHH) |
| 14251 | { |
| 14252 | if (is_fma) |
| 14253 | do_vfp_nsyn_opcode ("ffmas"); |
| 14254 | else |
| 14255 | do_vfp_nsyn_opcode ("ffnmas"); |
| 14256 | |
| 14257 | /* ARMv8.2 fp16 instruction. */ |
| 14258 | if (rs == NS_HHH) |
| 14259 | do_scalar_fp16_v82_encode (); |
| 14260 | } |
| 14261 | else |
| 14262 | { |
| 14263 | if (is_fma) |
| 14264 | do_vfp_nsyn_opcode ("ffmad"); |
| 14265 | else |
| 14266 | do_vfp_nsyn_opcode ("ffnmad"); |
| 14267 | } |
| 14268 | } |
| 14269 | |
| 14270 | static void |
| 14271 | do_vfp_nsyn_mul (enum neon_shape rs) |
| 14272 | { |
| 14273 | if (rs == NS_FFF || rs == NS_HHH) |
| 14274 | { |
| 14275 | do_vfp_nsyn_opcode ("fmuls"); |
| 14276 | |
| 14277 | /* ARMv8.2 fp16 instruction. */ |
| 14278 | if (rs == NS_HHH) |
| 14279 | do_scalar_fp16_v82_encode (); |
| 14280 | } |
| 14281 | else |
| 14282 | do_vfp_nsyn_opcode ("fmuld"); |
| 14283 | } |
| 14284 | |
| 14285 | static void |
| 14286 | do_vfp_nsyn_abs_neg (enum neon_shape rs) |
| 14287 | { |
| 14288 | int is_neg = (inst.instruction & 0x80) != 0; |
| 14289 | neon_check_type (2, rs, N_EQK | N_VFP, N_F_ALL | N_VFP | N_KEY); |
| 14290 | |
| 14291 | if (rs == NS_FF || rs == NS_HH) |
| 14292 | { |
| 14293 | if (is_neg) |
| 14294 | do_vfp_nsyn_opcode ("fnegs"); |
| 14295 | else |
| 14296 | do_vfp_nsyn_opcode ("fabss"); |
| 14297 | |
| 14298 | /* ARMv8.2 fp16 instruction. */ |
| 14299 | if (rs == NS_HH) |
| 14300 | do_scalar_fp16_v82_encode (); |
| 14301 | } |
| 14302 | else |
| 14303 | { |
| 14304 | if (is_neg) |
| 14305 | do_vfp_nsyn_opcode ("fnegd"); |
| 14306 | else |
| 14307 | do_vfp_nsyn_opcode ("fabsd"); |
| 14308 | } |
| 14309 | } |
| 14310 | |
| 14311 | /* Encode single-precision (only!) VFP fldm/fstm instructions. Double precision |
| 14312 | insns belong to Neon, and are handled elsewhere. */ |
| 14313 | |
| 14314 | static void |
| 14315 | do_vfp_nsyn_ldm_stm (int is_dbmode) |
| 14316 | { |
| 14317 | int is_ldm = (inst.instruction & (1 << 20)) != 0; |
| 14318 | if (is_ldm) |
| 14319 | { |
| 14320 | if (is_dbmode) |
| 14321 | do_vfp_nsyn_opcode ("fldmdbs"); |
| 14322 | else |
| 14323 | do_vfp_nsyn_opcode ("fldmias"); |
| 14324 | } |
| 14325 | else |
| 14326 | { |
| 14327 | if (is_dbmode) |
| 14328 | do_vfp_nsyn_opcode ("fstmdbs"); |
| 14329 | else |
| 14330 | do_vfp_nsyn_opcode ("fstmias"); |
| 14331 | } |
| 14332 | } |
| 14333 | |
| 14334 | static void |
| 14335 | do_vfp_nsyn_sqrt (void) |
| 14336 | { |
| 14337 | enum neon_shape rs = neon_select_shape (NS_HH, NS_FF, NS_DD, NS_NULL); |
| 14338 | neon_check_type (2, rs, N_EQK | N_VFP, N_F_ALL | N_KEY | N_VFP); |
| 14339 | |
| 14340 | if (rs == NS_FF || rs == NS_HH) |
| 14341 | { |
| 14342 | do_vfp_nsyn_opcode ("fsqrts"); |
| 14343 | |
| 14344 | /* ARMv8.2 fp16 instruction. */ |
| 14345 | if (rs == NS_HH) |
| 14346 | do_scalar_fp16_v82_encode (); |
| 14347 | } |
| 14348 | else |
| 14349 | do_vfp_nsyn_opcode ("fsqrtd"); |
| 14350 | } |
| 14351 | |
| 14352 | static void |
| 14353 | do_vfp_nsyn_div (void) |
| 14354 | { |
| 14355 | enum neon_shape rs = neon_select_shape (NS_HHH, NS_FFF, NS_DDD, NS_NULL); |
| 14356 | neon_check_type (3, rs, N_EQK | N_VFP, N_EQK | N_VFP, |
| 14357 | N_F_ALL | N_KEY | N_VFP); |
| 14358 | |
| 14359 | if (rs == NS_FFF || rs == NS_HHH) |
| 14360 | { |
| 14361 | do_vfp_nsyn_opcode ("fdivs"); |
| 14362 | |
| 14363 | /* ARMv8.2 fp16 instruction. */ |
| 14364 | if (rs == NS_HHH) |
| 14365 | do_scalar_fp16_v82_encode (); |
| 14366 | } |
| 14367 | else |
| 14368 | do_vfp_nsyn_opcode ("fdivd"); |
| 14369 | } |
| 14370 | |
| 14371 | static void |
| 14372 | do_vfp_nsyn_nmul (void) |
| 14373 | { |
| 14374 | enum neon_shape rs = neon_select_shape (NS_HHH, NS_FFF, NS_DDD, NS_NULL); |
| 14375 | neon_check_type (3, rs, N_EQK | N_VFP, N_EQK | N_VFP, |
| 14376 | N_F_ALL | N_KEY | N_VFP); |
| 14377 | |
| 14378 | if (rs == NS_FFF || rs == NS_HHH) |
| 14379 | { |
| 14380 | NEON_ENCODE (SINGLE, inst); |
| 14381 | do_vfp_sp_dyadic (); |
| 14382 | |
| 14383 | /* ARMv8.2 fp16 instruction. */ |
| 14384 | if (rs == NS_HHH) |
| 14385 | do_scalar_fp16_v82_encode (); |
| 14386 | } |
| 14387 | else |
| 14388 | { |
| 14389 | NEON_ENCODE (DOUBLE, inst); |
| 14390 | do_vfp_dp_rd_rn_rm (); |
| 14391 | } |
| 14392 | do_vfp_cond_or_thumb (); |
| 14393 | |
| 14394 | } |
| 14395 | |
| 14396 | static void |
| 14397 | do_vfp_nsyn_cmp (void) |
| 14398 | { |
| 14399 | enum neon_shape rs; |
| 14400 | if (inst.operands[1].isreg) |
| 14401 | { |
| 14402 | rs = neon_select_shape (NS_HH, NS_FF, NS_DD, NS_NULL); |
| 14403 | neon_check_type (2, rs, N_EQK | N_VFP, N_F_ALL | N_KEY | N_VFP); |
| 14404 | |
| 14405 | if (rs == NS_FF || rs == NS_HH) |
| 14406 | { |
| 14407 | NEON_ENCODE (SINGLE, inst); |
| 14408 | do_vfp_sp_monadic (); |
| 14409 | } |
| 14410 | else |
| 14411 | { |
| 14412 | NEON_ENCODE (DOUBLE, inst); |
| 14413 | do_vfp_dp_rd_rm (); |
| 14414 | } |
| 14415 | } |
| 14416 | else |
| 14417 | { |
| 14418 | rs = neon_select_shape (NS_HI, NS_FI, NS_DI, NS_NULL); |
| 14419 | neon_check_type (2, rs, N_F_ALL | N_KEY | N_VFP, N_EQK); |
| 14420 | |
| 14421 | switch (inst.instruction & 0x0fffffff) |
| 14422 | { |
| 14423 | case N_MNEM_vcmp: |
| 14424 | inst.instruction += N_MNEM_vcmpz - N_MNEM_vcmp; |
| 14425 | break; |
| 14426 | case N_MNEM_vcmpe: |
| 14427 | inst.instruction += N_MNEM_vcmpez - N_MNEM_vcmpe; |
| 14428 | break; |
| 14429 | default: |
| 14430 | abort (); |
| 14431 | } |
| 14432 | |
| 14433 | if (rs == NS_FI || rs == NS_HI) |
| 14434 | { |
| 14435 | NEON_ENCODE (SINGLE, inst); |
| 14436 | do_vfp_sp_compare_z (); |
| 14437 | } |
| 14438 | else |
| 14439 | { |
| 14440 | NEON_ENCODE (DOUBLE, inst); |
| 14441 | do_vfp_dp_rd (); |
| 14442 | } |
| 14443 | } |
| 14444 | do_vfp_cond_or_thumb (); |
| 14445 | |
| 14446 | /* ARMv8.2 fp16 instruction. */ |
| 14447 | if (rs == NS_HI || rs == NS_HH) |
| 14448 | do_scalar_fp16_v82_encode (); |
| 14449 | } |
| 14450 | |
| 14451 | static void |
| 14452 | nsyn_insert_sp (void) |
| 14453 | { |
| 14454 | inst.operands[1] = inst.operands[0]; |
| 14455 | memset (&inst.operands[0], '\0', sizeof (inst.operands[0])); |
| 14456 | inst.operands[0].reg = REG_SP; |
| 14457 | inst.operands[0].isreg = 1; |
| 14458 | inst.operands[0].writeback = 1; |
| 14459 | inst.operands[0].present = 1; |
| 14460 | } |
| 14461 | |
| 14462 | static void |
| 14463 | do_vfp_nsyn_push (void) |
| 14464 | { |
| 14465 | nsyn_insert_sp (); |
| 14466 | |
| 14467 | constraint (inst.operands[1].imm < 1 || inst.operands[1].imm > 16, |
| 14468 | _("register list must contain at least 1 and at most 16 " |
| 14469 | "registers")); |
| 14470 | |
| 14471 | if (inst.operands[1].issingle) |
| 14472 | do_vfp_nsyn_opcode ("fstmdbs"); |
| 14473 | else |
| 14474 | do_vfp_nsyn_opcode ("fstmdbd"); |
| 14475 | } |
| 14476 | |
| 14477 | static void |
| 14478 | do_vfp_nsyn_pop (void) |
| 14479 | { |
| 14480 | nsyn_insert_sp (); |
| 14481 | |
| 14482 | constraint (inst.operands[1].imm < 1 || inst.operands[1].imm > 16, |
| 14483 | _("register list must contain at least 1 and at most 16 " |
| 14484 | "registers")); |
| 14485 | |
| 14486 | if (inst.operands[1].issingle) |
| 14487 | do_vfp_nsyn_opcode ("fldmias"); |
| 14488 | else |
| 14489 | do_vfp_nsyn_opcode ("fldmiad"); |
| 14490 | } |
| 14491 | |
| 14492 | /* Fix up Neon data-processing instructions, ORing in the correct bits for |
| 14493 | ARM mode or Thumb mode and moving the encoded bit 24 to bit 28. */ |
| 14494 | |
| 14495 | static void |
| 14496 | neon_dp_fixup (struct arm_it* insn) |
| 14497 | { |
| 14498 | unsigned int i = insn->instruction; |
| 14499 | insn->is_neon = 1; |
| 14500 | |
| 14501 | if (thumb_mode) |
| 14502 | { |
| 14503 | /* The U bit is at bit 24 by default. Move to bit 28 in Thumb mode. */ |
| 14504 | if (i & (1 << 24)) |
| 14505 | i |= 1 << 28; |
| 14506 | |
| 14507 | i &= ~(1 << 24); |
| 14508 | |
| 14509 | i |= 0xef000000; |
| 14510 | } |
| 14511 | else |
| 14512 | i |= 0xf2000000; |
| 14513 | |
| 14514 | insn->instruction = i; |
| 14515 | } |
| 14516 | |
| 14517 | /* Turn a size (8, 16, 32, 64) into the respective bit number minus 3 |
| 14518 | (0, 1, 2, 3). */ |
| 14519 | |
| 14520 | static unsigned |
| 14521 | neon_logbits (unsigned x) |
| 14522 | { |
| 14523 | return ffs (x) - 4; |
| 14524 | } |
| 14525 | |
| 14526 | #define LOW4(R) ((R) & 0xf) |
| 14527 | #define HI1(R) (((R) >> 4) & 1) |
| 14528 | |
| 14529 | /* Encode insns with bit pattern: |
| 14530 | |
| 14531 | |28/24|23|22 |21 20|19 16|15 12|11 8|7|6|5|4|3 0| |
| 14532 | | U |x |D |size | Rn | Rd |x x x x|N|Q|M|x| Rm | |
| 14533 | |
| 14534 | SIZE is passed in bits. -1 means size field isn't changed, in case it has a |
| 14535 | different meaning for some instruction. */ |
| 14536 | |
| 14537 | static void |
| 14538 | neon_three_same (int isquad, int ubit, int size) |
| 14539 | { |
| 14540 | inst.instruction |= LOW4 (inst.operands[0].reg) << 12; |
| 14541 | inst.instruction |= HI1 (inst.operands[0].reg) << 22; |
| 14542 | inst.instruction |= LOW4 (inst.operands[1].reg) << 16; |
| 14543 | inst.instruction |= HI1 (inst.operands[1].reg) << 7; |
| 14544 | inst.instruction |= LOW4 (inst.operands[2].reg); |
| 14545 | inst.instruction |= HI1 (inst.operands[2].reg) << 5; |
| 14546 | inst.instruction |= (isquad != 0) << 6; |
| 14547 | inst.instruction |= (ubit != 0) << 24; |
| 14548 | if (size != -1) |
| 14549 | inst.instruction |= neon_logbits (size) << 20; |
| 14550 | |
| 14551 | neon_dp_fixup (&inst); |
| 14552 | } |
| 14553 | |
| 14554 | /* Encode instructions of the form: |
| 14555 | |
| 14556 | |28/24|23|22|21 20|19 18|17 16|15 12|11 7|6|5|4|3 0| |
| 14557 | | U |x |D |x x |size |x x | Rd |x x x x x|Q|M|x| Rm | |
| 14558 | |
| 14559 | Don't write size if SIZE == -1. */ |
| 14560 | |
| 14561 | static void |
| 14562 | neon_two_same (int qbit, int ubit, int size) |
| 14563 | { |
| 14564 | inst.instruction |= LOW4 (inst.operands[0].reg) << 12; |
| 14565 | inst.instruction |= HI1 (inst.operands[0].reg) << 22; |
| 14566 | inst.instruction |= LOW4 (inst.operands[1].reg); |
| 14567 | inst.instruction |= HI1 (inst.operands[1].reg) << 5; |
| 14568 | inst.instruction |= (qbit != 0) << 6; |
| 14569 | inst.instruction |= (ubit != 0) << 24; |
| 14570 | |
| 14571 | if (size != -1) |
| 14572 | inst.instruction |= neon_logbits (size) << 18; |
| 14573 | |
| 14574 | neon_dp_fixup (&inst); |
| 14575 | } |
| 14576 | |
| 14577 | /* Neon instruction encoders, in approximate order of appearance. */ |
| 14578 | |
| 14579 | static void |
| 14580 | do_neon_dyadic_i_su (void) |
| 14581 | { |
| 14582 | enum neon_shape rs = neon_select_shape (NS_DDD, NS_QQQ, NS_NULL); |
| 14583 | struct neon_type_el et = neon_check_type (3, rs, |
| 14584 | N_EQK, N_EQK, N_SU_32 | N_KEY); |
| 14585 | neon_three_same (neon_quad (rs), et.type == NT_unsigned, et.size); |
| 14586 | } |
| 14587 | |
| 14588 | static void |
| 14589 | do_neon_dyadic_i64_su (void) |
| 14590 | { |
| 14591 | enum neon_shape rs = neon_select_shape (NS_DDD, NS_QQQ, NS_NULL); |
| 14592 | struct neon_type_el et = neon_check_type (3, rs, |
| 14593 | N_EQK, N_EQK, N_SU_ALL | N_KEY); |
| 14594 | neon_three_same (neon_quad (rs), et.type == NT_unsigned, et.size); |
| 14595 | } |
| 14596 | |
| 14597 | static void |
| 14598 | neon_imm_shift (int write_ubit, int uval, int isquad, struct neon_type_el et, |
| 14599 | unsigned immbits) |
| 14600 | { |
| 14601 | unsigned size = et.size >> 3; |
| 14602 | inst.instruction |= LOW4 (inst.operands[0].reg) << 12; |
| 14603 | inst.instruction |= HI1 (inst.operands[0].reg) << 22; |
| 14604 | inst.instruction |= LOW4 (inst.operands[1].reg); |
| 14605 | inst.instruction |= HI1 (inst.operands[1].reg) << 5; |
| 14606 | inst.instruction |= (isquad != 0) << 6; |
| 14607 | inst.instruction |= immbits << 16; |
| 14608 | inst.instruction |= (size >> 3) << 7; |
| 14609 | inst.instruction |= (size & 0x7) << 19; |
| 14610 | if (write_ubit) |
| 14611 | inst.instruction |= (uval != 0) << 24; |
| 14612 | |
| 14613 | neon_dp_fixup (&inst); |
| 14614 | } |
| 14615 | |
| 14616 | static void |
| 14617 | do_neon_shl_imm (void) |
| 14618 | { |
| 14619 | if (!inst.operands[2].isreg) |
| 14620 | { |
| 14621 | enum neon_shape rs = neon_select_shape (NS_DDI, NS_QQI, NS_NULL); |
| 14622 | struct neon_type_el et = neon_check_type (2, rs, N_EQK, N_KEY | N_I_ALL); |
| 14623 | int imm = inst.operands[2].imm; |
| 14624 | |
| 14625 | constraint (imm < 0 || (unsigned)imm >= et.size, |
| 14626 | _("immediate out of range for shift")); |
| 14627 | NEON_ENCODE (IMMED, inst); |
| 14628 | neon_imm_shift (FALSE, 0, neon_quad (rs), et, imm); |
| 14629 | } |
| 14630 | else |
| 14631 | { |
| 14632 | enum neon_shape rs = neon_select_shape (NS_DDD, NS_QQQ, NS_NULL); |
| 14633 | struct neon_type_el et = neon_check_type (3, rs, |
| 14634 | N_EQK, N_SU_ALL | N_KEY, N_EQK | N_SGN); |
| 14635 | unsigned int tmp; |
| 14636 | |
| 14637 | /* VSHL/VQSHL 3-register variants have syntax such as: |
| 14638 | vshl.xx Dd, Dm, Dn |
| 14639 | whereas other 3-register operations encoded by neon_three_same have |
| 14640 | syntax like: |
| 14641 | vadd.xx Dd, Dn, Dm |
| 14642 | (i.e. with Dn & Dm reversed). Swap operands[1].reg and operands[2].reg |
| 14643 | here. */ |
| 14644 | tmp = inst.operands[2].reg; |
| 14645 | inst.operands[2].reg = inst.operands[1].reg; |
| 14646 | inst.operands[1].reg = tmp; |
| 14647 | NEON_ENCODE (INTEGER, inst); |
| 14648 | neon_three_same (neon_quad (rs), et.type == NT_unsigned, et.size); |
| 14649 | } |
| 14650 | } |
| 14651 | |
| 14652 | static void |
| 14653 | do_neon_qshl_imm (void) |
| 14654 | { |
| 14655 | if (!inst.operands[2].isreg) |
| 14656 | { |
| 14657 | enum neon_shape rs = neon_select_shape (NS_DDI, NS_QQI, NS_NULL); |
| 14658 | struct neon_type_el et = neon_check_type (2, rs, N_EQK, N_SU_ALL | N_KEY); |
| 14659 | int imm = inst.operands[2].imm; |
| 14660 | |
| 14661 | constraint (imm < 0 || (unsigned)imm >= et.size, |
| 14662 | _("immediate out of range for shift")); |
| 14663 | NEON_ENCODE (IMMED, inst); |
| 14664 | neon_imm_shift (TRUE, et.type == NT_unsigned, neon_quad (rs), et, imm); |
| 14665 | } |
| 14666 | else |
| 14667 | { |
| 14668 | enum neon_shape rs = neon_select_shape (NS_DDD, NS_QQQ, NS_NULL); |
| 14669 | struct neon_type_el et = neon_check_type (3, rs, |
| 14670 | N_EQK, N_SU_ALL | N_KEY, N_EQK | N_SGN); |
| 14671 | unsigned int tmp; |
| 14672 | |
| 14673 | /* See note in do_neon_shl_imm. */ |
| 14674 | tmp = inst.operands[2].reg; |
| 14675 | inst.operands[2].reg = inst.operands[1].reg; |
| 14676 | inst.operands[1].reg = tmp; |
| 14677 | NEON_ENCODE (INTEGER, inst); |
| 14678 | neon_three_same (neon_quad (rs), et.type == NT_unsigned, et.size); |
| 14679 | } |
| 14680 | } |
| 14681 | |
| 14682 | static void |
| 14683 | do_neon_rshl (void) |
| 14684 | { |
| 14685 | enum neon_shape rs = neon_select_shape (NS_DDD, NS_QQQ, NS_NULL); |
| 14686 | struct neon_type_el et = neon_check_type (3, rs, |
| 14687 | N_EQK, N_EQK, N_SU_ALL | N_KEY); |
| 14688 | unsigned int tmp; |
| 14689 | |
| 14690 | tmp = inst.operands[2].reg; |
| 14691 | inst.operands[2].reg = inst.operands[1].reg; |
| 14692 | inst.operands[1].reg = tmp; |
| 14693 | neon_three_same (neon_quad (rs), et.type == NT_unsigned, et.size); |
| 14694 | } |
| 14695 | |
| 14696 | static int |
| 14697 | neon_cmode_for_logic_imm (unsigned immediate, unsigned *immbits, int size) |
| 14698 | { |
| 14699 | /* Handle .I8 pseudo-instructions. */ |
| 14700 | if (size == 8) |
| 14701 | { |
| 14702 | /* Unfortunately, this will make everything apart from zero out-of-range. |
| 14703 | FIXME is this the intended semantics? There doesn't seem much point in |
| 14704 | accepting .I8 if so. */ |
| 14705 | immediate |= immediate << 8; |
| 14706 | size = 16; |
| 14707 | } |
| 14708 | |
| 14709 | if (size >= 32) |
| 14710 | { |
| 14711 | if (immediate == (immediate & 0x000000ff)) |
| 14712 | { |
| 14713 | *immbits = immediate; |
| 14714 | return 0x1; |
| 14715 | } |
| 14716 | else if (immediate == (immediate & 0x0000ff00)) |
| 14717 | { |
| 14718 | *immbits = immediate >> 8; |
| 14719 | return 0x3; |
| 14720 | } |
| 14721 | else if (immediate == (immediate & 0x00ff0000)) |
| 14722 | { |
| 14723 | *immbits = immediate >> 16; |
| 14724 | return 0x5; |
| 14725 | } |
| 14726 | else if (immediate == (immediate & 0xff000000)) |
| 14727 | { |
| 14728 | *immbits = immediate >> 24; |
| 14729 | return 0x7; |
| 14730 | } |
| 14731 | if ((immediate & 0xffff) != (immediate >> 16)) |
| 14732 | goto bad_immediate; |
| 14733 | immediate &= 0xffff; |
| 14734 | } |
| 14735 | |
| 14736 | if (immediate == (immediate & 0x000000ff)) |
| 14737 | { |
| 14738 | *immbits = immediate; |
| 14739 | return 0x9; |
| 14740 | } |
| 14741 | else if (immediate == (immediate & 0x0000ff00)) |
| 14742 | { |
| 14743 | *immbits = immediate >> 8; |
| 14744 | return 0xb; |
| 14745 | } |
| 14746 | |
| 14747 | bad_immediate: |
| 14748 | first_error (_("immediate value out of range")); |
| 14749 | return FAIL; |
| 14750 | } |
| 14751 | |
| 14752 | static void |
| 14753 | do_neon_logic (void) |
| 14754 | { |
| 14755 | if (inst.operands[2].present && inst.operands[2].isreg) |
| 14756 | { |
| 14757 | enum neon_shape rs = neon_select_shape (NS_DDD, NS_QQQ, NS_NULL); |
| 14758 | neon_check_type (3, rs, N_IGNORE_TYPE); |
| 14759 | /* U bit and size field were set as part of the bitmask. */ |
| 14760 | NEON_ENCODE (INTEGER, inst); |
| 14761 | neon_three_same (neon_quad (rs), 0, -1); |
| 14762 | } |
| 14763 | else |
| 14764 | { |
| 14765 | const int three_ops_form = (inst.operands[2].present |
| 14766 | && !inst.operands[2].isreg); |
| 14767 | const int immoperand = (three_ops_form ? 2 : 1); |
| 14768 | enum neon_shape rs = (three_ops_form |
| 14769 | ? neon_select_shape (NS_DDI, NS_QQI, NS_NULL) |
| 14770 | : neon_select_shape (NS_DI, NS_QI, NS_NULL)); |
| 14771 | struct neon_type_el et = neon_check_type (2, rs, |
| 14772 | N_I8 | N_I16 | N_I32 | N_I64 | N_F32 | N_KEY, N_EQK); |
| 14773 | enum neon_opc opcode = (enum neon_opc) inst.instruction & 0x0fffffff; |
| 14774 | unsigned immbits; |
| 14775 | int cmode; |
| 14776 | |
| 14777 | if (et.type == NT_invtype) |
| 14778 | return; |
| 14779 | |
| 14780 | if (three_ops_form) |
| 14781 | constraint (inst.operands[0].reg != inst.operands[1].reg, |
| 14782 | _("first and second operands shall be the same register")); |
| 14783 | |
| 14784 | NEON_ENCODE (IMMED, inst); |
| 14785 | |
| 14786 | immbits = inst.operands[immoperand].imm; |
| 14787 | if (et.size == 64) |
| 14788 | { |
| 14789 | /* .i64 is a pseudo-op, so the immediate must be a repeating |
| 14790 | pattern. */ |
| 14791 | if (immbits != (inst.operands[immoperand].regisimm ? |
| 14792 | inst.operands[immoperand].reg : 0)) |
| 14793 | { |
| 14794 | /* Set immbits to an invalid constant. */ |
| 14795 | immbits = 0xdeadbeef; |
| 14796 | } |
| 14797 | } |
| 14798 | |
| 14799 | switch (opcode) |
| 14800 | { |
| 14801 | case N_MNEM_vbic: |
| 14802 | cmode = neon_cmode_for_logic_imm (immbits, &immbits, et.size); |
| 14803 | break; |
| 14804 | |
| 14805 | case N_MNEM_vorr: |
| 14806 | cmode = neon_cmode_for_logic_imm (immbits, &immbits, et.size); |
| 14807 | break; |
| 14808 | |
| 14809 | case N_MNEM_vand: |
| 14810 | /* Pseudo-instruction for VBIC. */ |
| 14811 | neon_invert_size (&immbits, 0, et.size); |
| 14812 | cmode = neon_cmode_for_logic_imm (immbits, &immbits, et.size); |
| 14813 | break; |
| 14814 | |
| 14815 | case N_MNEM_vorn: |
| 14816 | /* Pseudo-instruction for VORR. */ |
| 14817 | neon_invert_size (&immbits, 0, et.size); |
| 14818 | cmode = neon_cmode_for_logic_imm (immbits, &immbits, et.size); |
| 14819 | break; |
| 14820 | |
| 14821 | default: |
| 14822 | abort (); |
| 14823 | } |
| 14824 | |
| 14825 | if (cmode == FAIL) |
| 14826 | return; |
| 14827 | |
| 14828 | inst.instruction |= neon_quad (rs) << 6; |
| 14829 | inst.instruction |= LOW4 (inst.operands[0].reg) << 12; |
| 14830 | inst.instruction |= HI1 (inst.operands[0].reg) << 22; |
| 14831 | inst.instruction |= cmode << 8; |
| 14832 | neon_write_immbits (immbits); |
| 14833 | |
| 14834 | neon_dp_fixup (&inst); |
| 14835 | } |
| 14836 | } |
| 14837 | |
| 14838 | static void |
| 14839 | do_neon_bitfield (void) |
| 14840 | { |
| 14841 | enum neon_shape rs = neon_select_shape (NS_DDD, NS_QQQ, NS_NULL); |
| 14842 | neon_check_type (3, rs, N_IGNORE_TYPE); |
| 14843 | neon_three_same (neon_quad (rs), 0, -1); |
| 14844 | } |
| 14845 | |
| 14846 | static void |
| 14847 | neon_dyadic_misc (enum neon_el_type ubit_meaning, unsigned types, |
| 14848 | unsigned destbits) |
| 14849 | { |
| 14850 | enum neon_shape rs = neon_select_shape (NS_DDD, NS_QQQ, NS_NULL); |
| 14851 | struct neon_type_el et = neon_check_type (3, rs, N_EQK | destbits, N_EQK, |
| 14852 | types | N_KEY); |
| 14853 | if (et.type == NT_float) |
| 14854 | { |
| 14855 | NEON_ENCODE (FLOAT, inst); |
| 14856 | neon_three_same (neon_quad (rs), 0, et.size == 16 ? (int) et.size : -1); |
| 14857 | } |
| 14858 | else |
| 14859 | { |
| 14860 | NEON_ENCODE (INTEGER, inst); |
| 14861 | neon_three_same (neon_quad (rs), et.type == ubit_meaning, et.size); |
| 14862 | } |
| 14863 | } |
| 14864 | |
| 14865 | static void |
| 14866 | do_neon_dyadic_if_su (void) |
| 14867 | { |
| 14868 | neon_dyadic_misc (NT_unsigned, N_SUF_32, 0); |
| 14869 | } |
| 14870 | |
| 14871 | static void |
| 14872 | do_neon_dyadic_if_su_d (void) |
| 14873 | { |
| 14874 | /* This version only allow D registers, but that constraint is enforced during |
| 14875 | operand parsing so we don't need to do anything extra here. */ |
| 14876 | neon_dyadic_misc (NT_unsigned, N_SUF_32, 0); |
| 14877 | } |
| 14878 | |
| 14879 | static void |
| 14880 | do_neon_dyadic_if_i_d (void) |
| 14881 | { |
| 14882 | /* The "untyped" case can't happen. Do this to stop the "U" bit being |
| 14883 | affected if we specify unsigned args. */ |
| 14884 | neon_dyadic_misc (NT_untyped, N_IF_32, 0); |
| 14885 | } |
| 14886 | |
| 14887 | enum vfp_or_neon_is_neon_bits |
| 14888 | { |
| 14889 | NEON_CHECK_CC = 1, |
| 14890 | NEON_CHECK_ARCH = 2, |
| 14891 | NEON_CHECK_ARCH8 = 4 |
| 14892 | }; |
| 14893 | |
| 14894 | /* Call this function if an instruction which may have belonged to the VFP or |
| 14895 | Neon instruction sets, but turned out to be a Neon instruction (due to the |
| 14896 | operand types involved, etc.). We have to check and/or fix-up a couple of |
| 14897 | things: |
| 14898 | |
| 14899 | - Make sure the user hasn't attempted to make a Neon instruction |
| 14900 | conditional. |
| 14901 | - Alter the value in the condition code field if necessary. |
| 14902 | - Make sure that the arch supports Neon instructions. |
| 14903 | |
| 14904 | Which of these operations take place depends on bits from enum |
| 14905 | vfp_or_neon_is_neon_bits. |
| 14906 | |
| 14907 | WARNING: This function has side effects! If NEON_CHECK_CC is used and the |
| 14908 | current instruction's condition is COND_ALWAYS, the condition field is |
| 14909 | changed to inst.uncond_value. This is necessary because instructions shared |
| 14910 | between VFP and Neon may be conditional for the VFP variants only, and the |
| 14911 | unconditional Neon version must have, e.g., 0xF in the condition field. */ |
| 14912 | |
| 14913 | static int |
| 14914 | vfp_or_neon_is_neon (unsigned check) |
| 14915 | { |
| 14916 | /* Conditions are always legal in Thumb mode (IT blocks). */ |
| 14917 | if (!thumb_mode && (check & NEON_CHECK_CC)) |
| 14918 | { |
| 14919 | if (inst.cond != COND_ALWAYS) |
| 14920 | { |
| 14921 | first_error (_(BAD_COND)); |
| 14922 | return FAIL; |
| 14923 | } |
| 14924 | if (inst.uncond_value != -1) |
| 14925 | inst.instruction |= inst.uncond_value << 28; |
| 14926 | } |
| 14927 | |
| 14928 | if ((check & NEON_CHECK_ARCH) |
| 14929 | && !mark_feature_used (&fpu_neon_ext_v1)) |
| 14930 | { |
| 14931 | first_error (_(BAD_FPU)); |
| 14932 | return FAIL; |
| 14933 | } |
| 14934 | |
| 14935 | if ((check & NEON_CHECK_ARCH8) |
| 14936 | && !mark_feature_used (&fpu_neon_ext_armv8)) |
| 14937 | { |
| 14938 | first_error (_(BAD_FPU)); |
| 14939 | return FAIL; |
| 14940 | } |
| 14941 | |
| 14942 | return SUCCESS; |
| 14943 | } |
| 14944 | |
| 14945 | static void |
| 14946 | do_neon_addsub_if_i (void) |
| 14947 | { |
| 14948 | if (try_vfp_nsyn (3, do_vfp_nsyn_add_sub) == SUCCESS) |
| 14949 | return; |
| 14950 | |
| 14951 | if (vfp_or_neon_is_neon (NEON_CHECK_CC | NEON_CHECK_ARCH) == FAIL) |
| 14952 | return; |
| 14953 | |
| 14954 | /* The "untyped" case can't happen. Do this to stop the "U" bit being |
| 14955 | affected if we specify unsigned args. */ |
| 14956 | neon_dyadic_misc (NT_untyped, N_IF_32 | N_I64, 0); |
| 14957 | } |
| 14958 | |
| 14959 | /* Swaps operands 1 and 2. If operand 1 (optional arg) was omitted, we want the |
| 14960 | result to be: |
| 14961 | V<op> A,B (A is operand 0, B is operand 2) |
| 14962 | to mean: |
| 14963 | V<op> A,B,A |
| 14964 | not: |
| 14965 | V<op> A,B,B |
| 14966 | so handle that case specially. */ |
| 14967 | |
| 14968 | static void |
| 14969 | neon_exchange_operands (void) |
| 14970 | { |
| 14971 | if (inst.operands[1].present) |
| 14972 | { |
| 14973 | void *scratch = xmalloc (sizeof (inst.operands[0])); |
| 14974 | |
| 14975 | /* Swap operands[1] and operands[2]. */ |
| 14976 | memcpy (scratch, &inst.operands[1], sizeof (inst.operands[0])); |
| 14977 | inst.operands[1] = inst.operands[2]; |
| 14978 | memcpy (&inst.operands[2], scratch, sizeof (inst.operands[0])); |
| 14979 | free (scratch); |
| 14980 | } |
| 14981 | else |
| 14982 | { |
| 14983 | inst.operands[1] = inst.operands[2]; |
| 14984 | inst.operands[2] = inst.operands[0]; |
| 14985 | } |
| 14986 | } |
| 14987 | |
| 14988 | static void |
| 14989 | neon_compare (unsigned regtypes, unsigned immtypes, int invert) |
| 14990 | { |
| 14991 | if (inst.operands[2].isreg) |
| 14992 | { |
| 14993 | if (invert) |
| 14994 | neon_exchange_operands (); |
| 14995 | neon_dyadic_misc (NT_unsigned, regtypes, N_SIZ); |
| 14996 | } |
| 14997 | else |
| 14998 | { |
| 14999 | enum neon_shape rs = neon_select_shape (NS_DDI, NS_QQI, NS_NULL); |
| 15000 | struct neon_type_el et = neon_check_type (2, rs, |
| 15001 | N_EQK | N_SIZ, immtypes | N_KEY); |
| 15002 | |
| 15003 | NEON_ENCODE (IMMED, inst); |
| 15004 | inst.instruction |= LOW4 (inst.operands[0].reg) << 12; |
| 15005 | inst.instruction |= HI1 (inst.operands[0].reg) << 22; |
| 15006 | inst.instruction |= LOW4 (inst.operands[1].reg); |
| 15007 | inst.instruction |= HI1 (inst.operands[1].reg) << 5; |
| 15008 | inst.instruction |= neon_quad (rs) << 6; |
| 15009 | inst.instruction |= (et.type == NT_float) << 10; |
| 15010 | inst.instruction |= neon_logbits (et.size) << 18; |
| 15011 | |
| 15012 | neon_dp_fixup (&inst); |
| 15013 | } |
| 15014 | } |
| 15015 | |
| 15016 | static void |
| 15017 | do_neon_cmp (void) |
| 15018 | { |
| 15019 | neon_compare (N_SUF_32, N_S_32 | N_F_16_32, FALSE); |
| 15020 | } |
| 15021 | |
| 15022 | static void |
| 15023 | do_neon_cmp_inv (void) |
| 15024 | { |
| 15025 | neon_compare (N_SUF_32, N_S_32 | N_F_16_32, TRUE); |
| 15026 | } |
| 15027 | |
| 15028 | static void |
| 15029 | do_neon_ceq (void) |
| 15030 | { |
| 15031 | neon_compare (N_IF_32, N_IF_32, FALSE); |
| 15032 | } |
| 15033 | |
| 15034 | /* For multiply instructions, we have the possibility of 16-bit or 32-bit |
| 15035 | scalars, which are encoded in 5 bits, M : Rm. |
| 15036 | For 16-bit scalars, the register is encoded in Rm[2:0] and the index in |
| 15037 | M:Rm[3], and for 32-bit scalars, the register is encoded in Rm[3:0] and the |
| 15038 | index in M. |
| 15039 | |
| 15040 | Dot Product instructions are similar to multiply instructions except elsize |
| 15041 | should always be 32. |
| 15042 | |
| 15043 | This function translates SCALAR, which is GAS's internal encoding of indexed |
| 15044 | scalar register, to raw encoding. There is also register and index range |
| 15045 | check based on ELSIZE. */ |
| 15046 | |
| 15047 | static unsigned |
| 15048 | neon_scalar_for_mul (unsigned scalar, unsigned elsize) |
| 15049 | { |
| 15050 | unsigned regno = NEON_SCALAR_REG (scalar); |
| 15051 | unsigned elno = NEON_SCALAR_INDEX (scalar); |
| 15052 | |
| 15053 | switch (elsize) |
| 15054 | { |
| 15055 | case 16: |
| 15056 | if (regno > 7 || elno > 3) |
| 15057 | goto bad_scalar; |
| 15058 | return regno | (elno << 3); |
| 15059 | |
| 15060 | case 32: |
| 15061 | if (regno > 15 || elno > 1) |
| 15062 | goto bad_scalar; |
| 15063 | return regno | (elno << 4); |
| 15064 | |
| 15065 | default: |
| 15066 | bad_scalar: |
| 15067 | first_error (_("scalar out of range for multiply instruction")); |
| 15068 | } |
| 15069 | |
| 15070 | return 0; |
| 15071 | } |
| 15072 | |
| 15073 | /* Encode multiply / multiply-accumulate scalar instructions. */ |
| 15074 | |
| 15075 | static void |
| 15076 | neon_mul_mac (struct neon_type_el et, int ubit) |
| 15077 | { |
| 15078 | unsigned scalar; |
| 15079 | |
| 15080 | /* Give a more helpful error message if we have an invalid type. */ |
| 15081 | if (et.type == NT_invtype) |
| 15082 | return; |
| 15083 | |
| 15084 | scalar = neon_scalar_for_mul (inst.operands[2].reg, et.size); |
| 15085 | inst.instruction |= LOW4 (inst.operands[0].reg) << 12; |
| 15086 | inst.instruction |= HI1 (inst.operands[0].reg) << 22; |
| 15087 | inst.instruction |= LOW4 (inst.operands[1].reg) << 16; |
| 15088 | inst.instruction |= HI1 (inst.operands[1].reg) << 7; |
| 15089 | inst.instruction |= LOW4 (scalar); |
| 15090 | inst.instruction |= HI1 (scalar) << 5; |
| 15091 | inst.instruction |= (et.type == NT_float) << 8; |
| 15092 | inst.instruction |= neon_logbits (et.size) << 20; |
| 15093 | inst.instruction |= (ubit != 0) << 24; |
| 15094 | |
| 15095 | neon_dp_fixup (&inst); |
| 15096 | } |
| 15097 | |
| 15098 | static void |
| 15099 | do_neon_mac_maybe_scalar (void) |
| 15100 | { |
| 15101 | if (try_vfp_nsyn (3, do_vfp_nsyn_mla_mls) == SUCCESS) |
| 15102 | return; |
| 15103 | |
| 15104 | if (vfp_or_neon_is_neon (NEON_CHECK_CC | NEON_CHECK_ARCH) == FAIL) |
| 15105 | return; |
| 15106 | |
| 15107 | if (inst.operands[2].isscalar) |
| 15108 | { |
| 15109 | enum neon_shape rs = neon_select_shape (NS_DDS, NS_QQS, NS_NULL); |
| 15110 | struct neon_type_el et = neon_check_type (3, rs, |
| 15111 | N_EQK, N_EQK, N_I16 | N_I32 | N_F_16_32 | N_KEY); |
| 15112 | NEON_ENCODE (SCALAR, inst); |
| 15113 | neon_mul_mac (et, neon_quad (rs)); |
| 15114 | } |
| 15115 | else |
| 15116 | { |
| 15117 | /* The "untyped" case can't happen. Do this to stop the "U" bit being |
| 15118 | affected if we specify unsigned args. */ |
| 15119 | neon_dyadic_misc (NT_untyped, N_IF_32, 0); |
| 15120 | } |
| 15121 | } |
| 15122 | |
| 15123 | static void |
| 15124 | do_neon_fmac (void) |
| 15125 | { |
| 15126 | if (try_vfp_nsyn (3, do_vfp_nsyn_fma_fms) == SUCCESS) |
| 15127 | return; |
| 15128 | |
| 15129 | if (vfp_or_neon_is_neon (NEON_CHECK_CC | NEON_CHECK_ARCH) == FAIL) |
| 15130 | return; |
| 15131 | |
| 15132 | neon_dyadic_misc (NT_untyped, N_IF_32, 0); |
| 15133 | } |
| 15134 | |
| 15135 | static void |
| 15136 | do_neon_tst (void) |
| 15137 | { |
| 15138 | enum neon_shape rs = neon_select_shape (NS_DDD, NS_QQQ, NS_NULL); |
| 15139 | struct neon_type_el et = neon_check_type (3, rs, |
| 15140 | N_EQK, N_EQK, N_8 | N_16 | N_32 | N_KEY); |
| 15141 | neon_three_same (neon_quad (rs), 0, et.size); |
| 15142 | } |
| 15143 | |
| 15144 | /* VMUL with 3 registers allows the P8 type. The scalar version supports the |
| 15145 | same types as the MAC equivalents. The polynomial type for this instruction |
| 15146 | is encoded the same as the integer type. */ |
| 15147 | |
| 15148 | static void |
| 15149 | do_neon_mul (void) |
| 15150 | { |
| 15151 | if (try_vfp_nsyn (3, do_vfp_nsyn_mul) == SUCCESS) |
| 15152 | return; |
| 15153 | |
| 15154 | if (vfp_or_neon_is_neon (NEON_CHECK_CC | NEON_CHECK_ARCH) == FAIL) |
| 15155 | return; |
| 15156 | |
| 15157 | if (inst.operands[2].isscalar) |
| 15158 | do_neon_mac_maybe_scalar (); |
| 15159 | else |
| 15160 | neon_dyadic_misc (NT_poly, N_I8 | N_I16 | N_I32 | N_F16 | N_F32 | N_P8, 0); |
| 15161 | } |
| 15162 | |
| 15163 | static void |
| 15164 | do_neon_qdmulh (void) |
| 15165 | { |
| 15166 | if (inst.operands[2].isscalar) |
| 15167 | { |
| 15168 | enum neon_shape rs = neon_select_shape (NS_DDS, NS_QQS, NS_NULL); |
| 15169 | struct neon_type_el et = neon_check_type (3, rs, |
| 15170 | N_EQK, N_EQK, N_S16 | N_S32 | N_KEY); |
| 15171 | NEON_ENCODE (SCALAR, inst); |
| 15172 | neon_mul_mac (et, neon_quad (rs)); |
| 15173 | } |
| 15174 | else |
| 15175 | { |
| 15176 | enum neon_shape rs = neon_select_shape (NS_DDD, NS_QQQ, NS_NULL); |
| 15177 | struct neon_type_el et = neon_check_type (3, rs, |
| 15178 | N_EQK, N_EQK, N_S16 | N_S32 | N_KEY); |
| 15179 | NEON_ENCODE (INTEGER, inst); |
| 15180 | /* The U bit (rounding) comes from bit mask. */ |
| 15181 | neon_three_same (neon_quad (rs), 0, et.size); |
| 15182 | } |
| 15183 | } |
| 15184 | |
| 15185 | static void |
| 15186 | do_neon_qrdmlah (void) |
| 15187 | { |
| 15188 | /* Check we're on the correct architecture. */ |
| 15189 | if (!mark_feature_used (&fpu_neon_ext_armv8)) |
| 15190 | inst.error = |
| 15191 | _("instruction form not available on this architecture."); |
| 15192 | else if (!mark_feature_used (&fpu_neon_ext_v8_1)) |
| 15193 | { |
| 15194 | as_warn (_("this instruction implies use of ARMv8.1 AdvSIMD.")); |
| 15195 | record_feature_use (&fpu_neon_ext_v8_1); |
| 15196 | } |
| 15197 | |
| 15198 | if (inst.operands[2].isscalar) |
| 15199 | { |
| 15200 | enum neon_shape rs = neon_select_shape (NS_DDS, NS_QQS, NS_NULL); |
| 15201 | struct neon_type_el et = neon_check_type (3, rs, |
| 15202 | N_EQK, N_EQK, N_S16 | N_S32 | N_KEY); |
| 15203 | NEON_ENCODE (SCALAR, inst); |
| 15204 | neon_mul_mac (et, neon_quad (rs)); |
| 15205 | } |
| 15206 | else |
| 15207 | { |
| 15208 | enum neon_shape rs = neon_select_shape (NS_DDD, NS_QQQ, NS_NULL); |
| 15209 | struct neon_type_el et = neon_check_type (3, rs, |
| 15210 | N_EQK, N_EQK, N_S16 | N_S32 | N_KEY); |
| 15211 | NEON_ENCODE (INTEGER, inst); |
| 15212 | /* The U bit (rounding) comes from bit mask. */ |
| 15213 | neon_three_same (neon_quad (rs), 0, et.size); |
| 15214 | } |
| 15215 | } |
| 15216 | |
| 15217 | static void |
| 15218 | do_neon_fcmp_absolute (void) |
| 15219 | { |
| 15220 | enum neon_shape rs = neon_select_shape (NS_DDD, NS_QQQ, NS_NULL); |
| 15221 | struct neon_type_el et = neon_check_type (3, rs, N_EQK, N_EQK, |
| 15222 | N_F_16_32 | N_KEY); |
| 15223 | /* Size field comes from bit mask. */ |
| 15224 | neon_three_same (neon_quad (rs), 1, et.size == 16 ? (int) et.size : -1); |
| 15225 | } |
| 15226 | |
| 15227 | static void |
| 15228 | do_neon_fcmp_absolute_inv (void) |
| 15229 | { |
| 15230 | neon_exchange_operands (); |
| 15231 | do_neon_fcmp_absolute (); |
| 15232 | } |
| 15233 | |
| 15234 | static void |
| 15235 | do_neon_step (void) |
| 15236 | { |
| 15237 | enum neon_shape rs = neon_select_shape (NS_DDD, NS_QQQ, NS_NULL); |
| 15238 | struct neon_type_el et = neon_check_type (3, rs, N_EQK, N_EQK, |
| 15239 | N_F_16_32 | N_KEY); |
| 15240 | neon_three_same (neon_quad (rs), 0, et.size == 16 ? (int) et.size : -1); |
| 15241 | } |
| 15242 | |
| 15243 | static void |
| 15244 | do_neon_abs_neg (void) |
| 15245 | { |
| 15246 | enum neon_shape rs; |
| 15247 | struct neon_type_el et; |
| 15248 | |
| 15249 | if (try_vfp_nsyn (2, do_vfp_nsyn_abs_neg) == SUCCESS) |
| 15250 | return; |
| 15251 | |
| 15252 | if (vfp_or_neon_is_neon (NEON_CHECK_CC | NEON_CHECK_ARCH) == FAIL) |
| 15253 | return; |
| 15254 | |
| 15255 | rs = neon_select_shape (NS_DD, NS_QQ, NS_NULL); |
| 15256 | et = neon_check_type (2, rs, N_EQK, N_S_32 | N_F_16_32 | N_KEY); |
| 15257 | |
| 15258 | inst.instruction |= LOW4 (inst.operands[0].reg) << 12; |
| 15259 | inst.instruction |= HI1 (inst.operands[0].reg) << 22; |
| 15260 | inst.instruction |= LOW4 (inst.operands[1].reg); |
| 15261 | inst.instruction |= HI1 (inst.operands[1].reg) << 5; |
| 15262 | inst.instruction |= neon_quad (rs) << 6; |
| 15263 | inst.instruction |= (et.type == NT_float) << 10; |
| 15264 | inst.instruction |= neon_logbits (et.size) << 18; |
| 15265 | |
| 15266 | neon_dp_fixup (&inst); |
| 15267 | } |
| 15268 | |
| 15269 | static void |
| 15270 | do_neon_sli (void) |
| 15271 | { |
| 15272 | enum neon_shape rs = neon_select_shape (NS_DDI, NS_QQI, NS_NULL); |
| 15273 | struct neon_type_el et = neon_check_type (2, rs, |
| 15274 | N_EQK, N_8 | N_16 | N_32 | N_64 | N_KEY); |
| 15275 | int imm = inst.operands[2].imm; |
| 15276 | constraint (imm < 0 || (unsigned)imm >= et.size, |
| 15277 | _("immediate out of range for insert")); |
| 15278 | neon_imm_shift (FALSE, 0, neon_quad (rs), et, imm); |
| 15279 | } |
| 15280 | |
| 15281 | static void |
| 15282 | do_neon_sri (void) |
| 15283 | { |
| 15284 | enum neon_shape rs = neon_select_shape (NS_DDI, NS_QQI, NS_NULL); |
| 15285 | struct neon_type_el et = neon_check_type (2, rs, |
| 15286 | N_EQK, N_8 | N_16 | N_32 | N_64 | N_KEY); |
| 15287 | int imm = inst.operands[2].imm; |
| 15288 | constraint (imm < 1 || (unsigned)imm > et.size, |
| 15289 | _("immediate out of range for insert")); |
| 15290 | neon_imm_shift (FALSE, 0, neon_quad (rs), et, et.size - imm); |
| 15291 | } |
| 15292 | |
| 15293 | static void |
| 15294 | do_neon_qshlu_imm (void) |
| 15295 | { |
| 15296 | enum neon_shape rs = neon_select_shape (NS_DDI, NS_QQI, NS_NULL); |
| 15297 | struct neon_type_el et = neon_check_type (2, rs, |
| 15298 | N_EQK | N_UNS, N_S8 | N_S16 | N_S32 | N_S64 | N_KEY); |
| 15299 | int imm = inst.operands[2].imm; |
| 15300 | constraint (imm < 0 || (unsigned)imm >= et.size, |
| 15301 | _("immediate out of range for shift")); |
| 15302 | /* Only encodes the 'U present' variant of the instruction. |
| 15303 | In this case, signed types have OP (bit 8) set to 0. |
| 15304 | Unsigned types have OP set to 1. */ |
| 15305 | inst.instruction |= (et.type == NT_unsigned) << 8; |
| 15306 | /* The rest of the bits are the same as other immediate shifts. */ |
| 15307 | neon_imm_shift (FALSE, 0, neon_quad (rs), et, imm); |
| 15308 | } |
| 15309 | |
| 15310 | static void |
| 15311 | do_neon_qmovn (void) |
| 15312 | { |
| 15313 | struct neon_type_el et = neon_check_type (2, NS_DQ, |
| 15314 | N_EQK | N_HLF, N_SU_16_64 | N_KEY); |
| 15315 | /* Saturating move where operands can be signed or unsigned, and the |
| 15316 | destination has the same signedness. */ |
| 15317 | NEON_ENCODE (INTEGER, inst); |
| 15318 | if (et.type == NT_unsigned) |
| 15319 | inst.instruction |= 0xc0; |
| 15320 | else |
| 15321 | inst.instruction |= 0x80; |
| 15322 | neon_two_same (0, 1, et.size / 2); |
| 15323 | } |
| 15324 | |
| 15325 | static void |
| 15326 | do_neon_qmovun (void) |
| 15327 | { |
| 15328 | struct neon_type_el et = neon_check_type (2, NS_DQ, |
| 15329 | N_EQK | N_HLF | N_UNS, N_S16 | N_S32 | N_S64 | N_KEY); |
| 15330 | /* Saturating move with unsigned results. Operands must be signed. */ |
| 15331 | NEON_ENCODE (INTEGER, inst); |
| 15332 | neon_two_same (0, 1, et.size / 2); |
| 15333 | } |
| 15334 | |
| 15335 | static void |
| 15336 | do_neon_rshift_sat_narrow (void) |
| 15337 | { |
| 15338 | /* FIXME: Types for narrowing. If operands are signed, results can be signed |
| 15339 | or unsigned. If operands are unsigned, results must also be unsigned. */ |
| 15340 | struct neon_type_el et = neon_check_type (2, NS_DQI, |
| 15341 | N_EQK | N_HLF, N_SU_16_64 | N_KEY); |
| 15342 | int imm = inst.operands[2].imm; |
| 15343 | /* This gets the bounds check, size encoding and immediate bits calculation |
| 15344 | right. */ |
| 15345 | et.size /= 2; |
| 15346 | |
| 15347 | /* VQ{R}SHRN.I<size> <Dd>, <Qm>, #0 is a synonym for |
| 15348 | VQMOVN.I<size> <Dd>, <Qm>. */ |
| 15349 | if (imm == 0) |
| 15350 | { |
| 15351 | inst.operands[2].present = 0; |
| 15352 | inst.instruction = N_MNEM_vqmovn; |
| 15353 | do_neon_qmovn (); |
| 15354 | return; |
| 15355 | } |
| 15356 | |
| 15357 | constraint (imm < 1 || (unsigned)imm > et.size, |
| 15358 | _("immediate out of range")); |
| 15359 | neon_imm_shift (TRUE, et.type == NT_unsigned, 0, et, et.size - imm); |
| 15360 | } |
| 15361 | |
| 15362 | static void |
| 15363 | do_neon_rshift_sat_narrow_u (void) |
| 15364 | { |
| 15365 | /* FIXME: Types for narrowing. If operands are signed, results can be signed |
| 15366 | or unsigned. If operands are unsigned, results must also be unsigned. */ |
| 15367 | struct neon_type_el et = neon_check_type (2, NS_DQI, |
| 15368 | N_EQK | N_HLF | N_UNS, N_S16 | N_S32 | N_S64 | N_KEY); |
| 15369 | int imm = inst.operands[2].imm; |
| 15370 | /* This gets the bounds check, size encoding and immediate bits calculation |
| 15371 | right. */ |
| 15372 | et.size /= 2; |
| 15373 | |
| 15374 | /* VQSHRUN.I<size> <Dd>, <Qm>, #0 is a synonym for |
| 15375 | VQMOVUN.I<size> <Dd>, <Qm>. */ |
| 15376 | if (imm == 0) |
| 15377 | { |
| 15378 | inst.operands[2].present = 0; |
| 15379 | inst.instruction = N_MNEM_vqmovun; |
| 15380 | do_neon_qmovun (); |
| 15381 | return; |
| 15382 | } |
| 15383 | |
| 15384 | constraint (imm < 1 || (unsigned)imm > et.size, |
| 15385 | _("immediate out of range")); |
| 15386 | /* FIXME: The manual is kind of unclear about what value U should have in |
| 15387 | VQ{R}SHRUN instructions, but U=0, op=0 definitely encodes VRSHR, so it |
| 15388 | must be 1. */ |
| 15389 | neon_imm_shift (TRUE, 1, 0, et, et.size - imm); |
| 15390 | } |
| 15391 | |
| 15392 | static void |
| 15393 | do_neon_movn (void) |
| 15394 | { |
| 15395 | struct neon_type_el et = neon_check_type (2, NS_DQ, |
| 15396 | N_EQK | N_HLF, N_I16 | N_I32 | N_I64 | N_KEY); |
| 15397 | NEON_ENCODE (INTEGER, inst); |
| 15398 | neon_two_same (0, 1, et.size / 2); |
| 15399 | } |
| 15400 | |
| 15401 | static void |
| 15402 | do_neon_rshift_narrow (void) |
| 15403 | { |
| 15404 | struct neon_type_el et = neon_check_type (2, NS_DQI, |
| 15405 | N_EQK | N_HLF, N_I16 | N_I32 | N_I64 | N_KEY); |
| 15406 | int imm = inst.operands[2].imm; |
| 15407 | /* This gets the bounds check, size encoding and immediate bits calculation |
| 15408 | right. */ |
| 15409 | et.size /= 2; |
| 15410 | |
| 15411 | /* If immediate is zero then we are a pseudo-instruction for |
| 15412 | VMOVN.I<size> <Dd>, <Qm> */ |
| 15413 | if (imm == 0) |
| 15414 | { |
| 15415 | inst.operands[2].present = 0; |
| 15416 | inst.instruction = N_MNEM_vmovn; |
| 15417 | do_neon_movn (); |
| 15418 | return; |
| 15419 | } |
| 15420 | |
| 15421 | constraint (imm < 1 || (unsigned)imm > et.size, |
| 15422 | _("immediate out of range for narrowing operation")); |
| 15423 | neon_imm_shift (FALSE, 0, 0, et, et.size - imm); |
| 15424 | } |
| 15425 | |
| 15426 | static void |
| 15427 | do_neon_shll (void) |
| 15428 | { |
| 15429 | /* FIXME: Type checking when lengthening. */ |
| 15430 | struct neon_type_el et = neon_check_type (2, NS_QDI, |
| 15431 | N_EQK | N_DBL, N_I8 | N_I16 | N_I32 | N_KEY); |
| 15432 | unsigned imm = inst.operands[2].imm; |
| 15433 | |
| 15434 | if (imm == et.size) |
| 15435 | { |
| 15436 | /* Maximum shift variant. */ |
| 15437 | NEON_ENCODE (INTEGER, inst); |
| 15438 | inst.instruction |= LOW4 (inst.operands[0].reg) << 12; |
| 15439 | inst.instruction |= HI1 (inst.operands[0].reg) << 22; |
| 15440 | inst.instruction |= LOW4 (inst.operands[1].reg); |
| 15441 | inst.instruction |= HI1 (inst.operands[1].reg) << 5; |
| 15442 | inst.instruction |= neon_logbits (et.size) << 18; |
| 15443 | |
| 15444 | neon_dp_fixup (&inst); |
| 15445 | } |
| 15446 | else |
| 15447 | { |
| 15448 | /* A more-specific type check for non-max versions. */ |
| 15449 | et = neon_check_type (2, NS_QDI, |
| 15450 | N_EQK | N_DBL, N_SU_32 | N_KEY); |
| 15451 | NEON_ENCODE (IMMED, inst); |
| 15452 | neon_imm_shift (TRUE, et.type == NT_unsigned, 0, et, imm); |
| 15453 | } |
| 15454 | } |
| 15455 | |
| 15456 | /* Check the various types for the VCVT instruction, and return which version |
| 15457 | the current instruction is. */ |
| 15458 | |
| 15459 | #define CVT_FLAVOUR_VAR \ |
| 15460 | CVT_VAR (s32_f32, N_S32, N_F32, whole_reg, "ftosls", "ftosis", "ftosizs") \ |
| 15461 | CVT_VAR (u32_f32, N_U32, N_F32, whole_reg, "ftouls", "ftouis", "ftouizs") \ |
| 15462 | CVT_VAR (f32_s32, N_F32, N_S32, whole_reg, "fsltos", "fsitos", NULL) \ |
| 15463 | CVT_VAR (f32_u32, N_F32, N_U32, whole_reg, "fultos", "fuitos", NULL) \ |
| 15464 | /* Half-precision conversions. */ \ |
| 15465 | CVT_VAR (s16_f16, N_S16, N_F16 | N_KEY, whole_reg, NULL, NULL, NULL) \ |
| 15466 | CVT_VAR (u16_f16, N_U16, N_F16 | N_KEY, whole_reg, NULL, NULL, NULL) \ |
| 15467 | CVT_VAR (f16_s16, N_F16 | N_KEY, N_S16, whole_reg, NULL, NULL, NULL) \ |
| 15468 | CVT_VAR (f16_u16, N_F16 | N_KEY, N_U16, whole_reg, NULL, NULL, NULL) \ |
| 15469 | CVT_VAR (f32_f16, N_F32, N_F16, whole_reg, NULL, NULL, NULL) \ |
| 15470 | CVT_VAR (f16_f32, N_F16, N_F32, whole_reg, NULL, NULL, NULL) \ |
| 15471 | /* New VCVT instructions introduced by ARMv8.2 fp16 extension. \ |
| 15472 | Compared with single/double precision variants, only the co-processor \ |
| 15473 | field is different, so the encoding flow is reused here. */ \ |
| 15474 | CVT_VAR (f16_s32, N_F16 | N_KEY, N_S32, N_VFP, "fsltos", "fsitos", NULL) \ |
| 15475 | CVT_VAR (f16_u32, N_F16 | N_KEY, N_U32, N_VFP, "fultos", "fuitos", NULL) \ |
| 15476 | CVT_VAR (u32_f16, N_U32, N_F16 | N_KEY, N_VFP, "ftouls", "ftouis", "ftouizs")\ |
| 15477 | CVT_VAR (s32_f16, N_S32, N_F16 | N_KEY, N_VFP, "ftosls", "ftosis", "ftosizs")\ |
| 15478 | /* VFP instructions. */ \ |
| 15479 | CVT_VAR (f32_f64, N_F32, N_F64, N_VFP, NULL, "fcvtsd", NULL) \ |
| 15480 | CVT_VAR (f64_f32, N_F64, N_F32, N_VFP, NULL, "fcvtds", NULL) \ |
| 15481 | CVT_VAR (s32_f64, N_S32, N_F64 | key, N_VFP, "ftosld", "ftosid", "ftosizd") \ |
| 15482 | CVT_VAR (u32_f64, N_U32, N_F64 | key, N_VFP, "ftould", "ftouid", "ftouizd") \ |
| 15483 | CVT_VAR (f64_s32, N_F64 | key, N_S32, N_VFP, "fsltod", "fsitod", NULL) \ |
| 15484 | CVT_VAR (f64_u32, N_F64 | key, N_U32, N_VFP, "fultod", "fuitod", NULL) \ |
| 15485 | /* VFP instructions with bitshift. */ \ |
| 15486 | CVT_VAR (f32_s16, N_F32 | key, N_S16, N_VFP, "fshtos", NULL, NULL) \ |
| 15487 | CVT_VAR (f32_u16, N_F32 | key, N_U16, N_VFP, "fuhtos", NULL, NULL) \ |
| 15488 | CVT_VAR (f64_s16, N_F64 | key, N_S16, N_VFP, "fshtod", NULL, NULL) \ |
| 15489 | CVT_VAR (f64_u16, N_F64 | key, N_U16, N_VFP, "fuhtod", NULL, NULL) \ |
| 15490 | CVT_VAR (s16_f32, N_S16, N_F32 | key, N_VFP, "ftoshs", NULL, NULL) \ |
| 15491 | CVT_VAR (u16_f32, N_U16, N_F32 | key, N_VFP, "ftouhs", NULL, NULL) \ |
| 15492 | CVT_VAR (s16_f64, N_S16, N_F64 | key, N_VFP, "ftoshd", NULL, NULL) \ |
| 15493 | CVT_VAR (u16_f64, N_U16, N_F64 | key, N_VFP, "ftouhd", NULL, NULL) |
| 15494 | |
| 15495 | #define CVT_VAR(C, X, Y, R, BSN, CN, ZN) \ |
| 15496 | neon_cvt_flavour_##C, |
| 15497 | |
| 15498 | /* The different types of conversions we can do. */ |
| 15499 | enum neon_cvt_flavour |
| 15500 | { |
| 15501 | CVT_FLAVOUR_VAR |
| 15502 | neon_cvt_flavour_invalid, |
| 15503 | neon_cvt_flavour_first_fp = neon_cvt_flavour_f32_f64 |
| 15504 | }; |
| 15505 | |
| 15506 | #undef CVT_VAR |
| 15507 | |
| 15508 | static enum neon_cvt_flavour |
| 15509 | get_neon_cvt_flavour (enum neon_shape rs) |
| 15510 | { |
| 15511 | #define CVT_VAR(C,X,Y,R,BSN,CN,ZN) \ |
| 15512 | et = neon_check_type (2, rs, (R) | (X), (R) | (Y)); \ |
| 15513 | if (et.type != NT_invtype) \ |
| 15514 | { \ |
| 15515 | inst.error = NULL; \ |
| 15516 | return (neon_cvt_flavour_##C); \ |
| 15517 | } |
| 15518 | |
| 15519 | struct neon_type_el et; |
| 15520 | unsigned whole_reg = (rs == NS_FFI || rs == NS_FD || rs == NS_DF |
| 15521 | || rs == NS_FF) ? N_VFP : 0; |
| 15522 | /* The instruction versions which take an immediate take one register |
| 15523 | argument, which is extended to the width of the full register. Thus the |
| 15524 | "source" and "destination" registers must have the same width. Hack that |
| 15525 | here by making the size equal to the key (wider, in this case) operand. */ |
| 15526 | unsigned key = (rs == NS_QQI || rs == NS_DDI || rs == NS_FFI) ? N_KEY : 0; |
| 15527 | |
| 15528 | CVT_FLAVOUR_VAR; |
| 15529 | |
| 15530 | return neon_cvt_flavour_invalid; |
| 15531 | #undef CVT_VAR |
| 15532 | } |
| 15533 | |
| 15534 | enum neon_cvt_mode |
| 15535 | { |
| 15536 | neon_cvt_mode_a, |
| 15537 | neon_cvt_mode_n, |
| 15538 | neon_cvt_mode_p, |
| 15539 | neon_cvt_mode_m, |
| 15540 | neon_cvt_mode_z, |
| 15541 | neon_cvt_mode_x, |
| 15542 | neon_cvt_mode_r |
| 15543 | }; |
| 15544 | |
| 15545 | /* Neon-syntax VFP conversions. */ |
| 15546 | |
| 15547 | static void |
| 15548 | do_vfp_nsyn_cvt (enum neon_shape rs, enum neon_cvt_flavour flavour) |
| 15549 | { |
| 15550 | const char *opname = 0; |
| 15551 | |
| 15552 | if (rs == NS_DDI || rs == NS_QQI || rs == NS_FFI |
| 15553 | || rs == NS_FHI || rs == NS_HFI) |
| 15554 | { |
| 15555 | /* Conversions with immediate bitshift. */ |
| 15556 | const char *enc[] = |
| 15557 | { |
| 15558 | #define CVT_VAR(C,A,B,R,BSN,CN,ZN) BSN, |
| 15559 | CVT_FLAVOUR_VAR |
| 15560 | NULL |
| 15561 | #undef CVT_VAR |
| 15562 | }; |
| 15563 | |
| 15564 | if (flavour < (int) ARRAY_SIZE (enc)) |
| 15565 | { |
| 15566 | opname = enc[flavour]; |
| 15567 | constraint (inst.operands[0].reg != inst.operands[1].reg, |
| 15568 | _("operands 0 and 1 must be the same register")); |
| 15569 | inst.operands[1] = inst.operands[2]; |
| 15570 | memset (&inst.operands[2], '\0', sizeof (inst.operands[2])); |
| 15571 | } |
| 15572 | } |
| 15573 | else |
| 15574 | { |
| 15575 | /* Conversions without bitshift. */ |
| 15576 | const char *enc[] = |
| 15577 | { |
| 15578 | #define CVT_VAR(C,A,B,R,BSN,CN,ZN) CN, |
| 15579 | CVT_FLAVOUR_VAR |
| 15580 | NULL |
| 15581 | #undef CVT_VAR |
| 15582 | }; |
| 15583 | |
| 15584 | if (flavour < (int) ARRAY_SIZE (enc)) |
| 15585 | opname = enc[flavour]; |
| 15586 | } |
| 15587 | |
| 15588 | if (opname) |
| 15589 | do_vfp_nsyn_opcode (opname); |
| 15590 | |
| 15591 | /* ARMv8.2 fp16 VCVT instruction. */ |
| 15592 | if (flavour == neon_cvt_flavour_s32_f16 |
| 15593 | || flavour == neon_cvt_flavour_u32_f16 |
| 15594 | || flavour == neon_cvt_flavour_f16_u32 |
| 15595 | || flavour == neon_cvt_flavour_f16_s32) |
| 15596 | do_scalar_fp16_v82_encode (); |
| 15597 | } |
| 15598 | |
| 15599 | static void |
| 15600 | do_vfp_nsyn_cvtz (void) |
| 15601 | { |
| 15602 | enum neon_shape rs = neon_select_shape (NS_FH, NS_FF, NS_FD, NS_NULL); |
| 15603 | enum neon_cvt_flavour flavour = get_neon_cvt_flavour (rs); |
| 15604 | const char *enc[] = |
| 15605 | { |
| 15606 | #define CVT_VAR(C,A,B,R,BSN,CN,ZN) ZN, |
| 15607 | CVT_FLAVOUR_VAR |
| 15608 | NULL |
| 15609 | #undef CVT_VAR |
| 15610 | }; |
| 15611 | |
| 15612 | if (flavour < (int) ARRAY_SIZE (enc) && enc[flavour]) |
| 15613 | do_vfp_nsyn_opcode (enc[flavour]); |
| 15614 | } |
| 15615 | |
| 15616 | static void |
| 15617 | do_vfp_nsyn_cvt_fpv8 (enum neon_cvt_flavour flavour, |
| 15618 | enum neon_cvt_mode mode) |
| 15619 | { |
| 15620 | int sz, op; |
| 15621 | int rm; |
| 15622 | |
| 15623 | /* Targets like FPv5-SP-D16 don't support FP v8 instructions with |
| 15624 | D register operands. */ |
| 15625 | if (flavour == neon_cvt_flavour_s32_f64 |
| 15626 | || flavour == neon_cvt_flavour_u32_f64) |
| 15627 | constraint (!ARM_CPU_HAS_FEATURE (cpu_variant, fpu_vfp_ext_armv8), |
| 15628 | _(BAD_FPU)); |
| 15629 | |
| 15630 | if (flavour == neon_cvt_flavour_s32_f16 |
| 15631 | || flavour == neon_cvt_flavour_u32_f16) |
| 15632 | constraint (!ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_fp16), |
| 15633 | _(BAD_FP16)); |
| 15634 | |
| 15635 | set_it_insn_type (OUTSIDE_IT_INSN); |
| 15636 | |
| 15637 | switch (flavour) |
| 15638 | { |
| 15639 | case neon_cvt_flavour_s32_f64: |
| 15640 | sz = 1; |
| 15641 | op = 1; |
| 15642 | break; |
| 15643 | case neon_cvt_flavour_s32_f32: |
| 15644 | sz = 0; |
| 15645 | op = 1; |
| 15646 | break; |
| 15647 | case neon_cvt_flavour_s32_f16: |
| 15648 | sz = 0; |
| 15649 | op = 1; |
| 15650 | break; |
| 15651 | case neon_cvt_flavour_u32_f64: |
| 15652 | sz = 1; |
| 15653 | op = 0; |
| 15654 | break; |
| 15655 | case neon_cvt_flavour_u32_f32: |
| 15656 | sz = 0; |
| 15657 | op = 0; |
| 15658 | break; |
| 15659 | case neon_cvt_flavour_u32_f16: |
| 15660 | sz = 0; |
| 15661 | op = 0; |
| 15662 | break; |
| 15663 | default: |
| 15664 | first_error (_("invalid instruction shape")); |
| 15665 | return; |
| 15666 | } |
| 15667 | |
| 15668 | switch (mode) |
| 15669 | { |
| 15670 | case neon_cvt_mode_a: rm = 0; break; |
| 15671 | case neon_cvt_mode_n: rm = 1; break; |
| 15672 | case neon_cvt_mode_p: rm = 2; break; |
| 15673 | case neon_cvt_mode_m: rm = 3; break; |
| 15674 | default: first_error (_("invalid rounding mode")); return; |
| 15675 | } |
| 15676 | |
| 15677 | NEON_ENCODE (FPV8, inst); |
| 15678 | encode_arm_vfp_reg (inst.operands[0].reg, VFP_REG_Sd); |
| 15679 | encode_arm_vfp_reg (inst.operands[1].reg, sz == 1 ? VFP_REG_Dm : VFP_REG_Sm); |
| 15680 | inst.instruction |= sz << 8; |
| 15681 | |
| 15682 | /* ARMv8.2 fp16 VCVT instruction. */ |
| 15683 | if (flavour == neon_cvt_flavour_s32_f16 |
| 15684 | ||flavour == neon_cvt_flavour_u32_f16) |
| 15685 | do_scalar_fp16_v82_encode (); |
| 15686 | inst.instruction |= op << 7; |
| 15687 | inst.instruction |= rm << 16; |
| 15688 | inst.instruction |= 0xf0000000; |
| 15689 | inst.is_neon = TRUE; |
| 15690 | } |
| 15691 | |
| 15692 | static void |
| 15693 | do_neon_cvt_1 (enum neon_cvt_mode mode) |
| 15694 | { |
| 15695 | enum neon_shape rs = neon_select_shape (NS_DDI, NS_QQI, NS_FFI, NS_DD, NS_QQ, |
| 15696 | NS_FD, NS_DF, NS_FF, NS_QD, NS_DQ, |
| 15697 | NS_FH, NS_HF, NS_FHI, NS_HFI, |
| 15698 | NS_NULL); |
| 15699 | enum neon_cvt_flavour flavour = get_neon_cvt_flavour (rs); |
| 15700 | |
| 15701 | if (flavour == neon_cvt_flavour_invalid) |
| 15702 | return; |
| 15703 | |
| 15704 | /* PR11109: Handle round-to-zero for VCVT conversions. */ |
| 15705 | if (mode == neon_cvt_mode_z |
| 15706 | && ARM_CPU_HAS_FEATURE (cpu_variant, fpu_arch_vfp_v2) |
| 15707 | && (flavour == neon_cvt_flavour_s16_f16 |
| 15708 | || flavour == neon_cvt_flavour_u16_f16 |
| 15709 | || flavour == neon_cvt_flavour_s32_f32 |
| 15710 | || flavour == neon_cvt_flavour_u32_f32 |
| 15711 | || flavour == neon_cvt_flavour_s32_f64 |
| 15712 | || flavour == neon_cvt_flavour_u32_f64) |
| 15713 | && (rs == NS_FD || rs == NS_FF)) |
| 15714 | { |
| 15715 | do_vfp_nsyn_cvtz (); |
| 15716 | return; |
| 15717 | } |
| 15718 | |
| 15719 | /* ARMv8.2 fp16 VCVT conversions. */ |
| 15720 | if (mode == neon_cvt_mode_z |
| 15721 | && ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_fp16) |
| 15722 | && (flavour == neon_cvt_flavour_s32_f16 |
| 15723 | || flavour == neon_cvt_flavour_u32_f16) |
| 15724 | && (rs == NS_FH)) |
| 15725 | { |
| 15726 | do_vfp_nsyn_cvtz (); |
| 15727 | do_scalar_fp16_v82_encode (); |
| 15728 | return; |
| 15729 | } |
| 15730 | |
| 15731 | /* VFP rather than Neon conversions. */ |
| 15732 | if (flavour >= neon_cvt_flavour_first_fp) |
| 15733 | { |
| 15734 | if (mode == neon_cvt_mode_x || mode == neon_cvt_mode_z) |
| 15735 | do_vfp_nsyn_cvt (rs, flavour); |
| 15736 | else |
| 15737 | do_vfp_nsyn_cvt_fpv8 (flavour, mode); |
| 15738 | |
| 15739 | return; |
| 15740 | } |
| 15741 | |
| 15742 | switch (rs) |
| 15743 | { |
| 15744 | case NS_DDI: |
| 15745 | case NS_QQI: |
| 15746 | { |
| 15747 | unsigned immbits; |
| 15748 | unsigned enctab[] = {0x0000100, 0x1000100, 0x0, 0x1000000, |
| 15749 | 0x0000100, 0x1000100, 0x0, 0x1000000}; |
| 15750 | |
| 15751 | if (vfp_or_neon_is_neon (NEON_CHECK_CC | NEON_CHECK_ARCH) == FAIL) |
| 15752 | return; |
| 15753 | |
| 15754 | /* Fixed-point conversion with #0 immediate is encoded as an |
| 15755 | integer conversion. */ |
| 15756 | if (inst.operands[2].present && inst.operands[2].imm == 0) |
| 15757 | goto int_encode; |
| 15758 | NEON_ENCODE (IMMED, inst); |
| 15759 | if (flavour != neon_cvt_flavour_invalid) |
| 15760 | inst.instruction |= enctab[flavour]; |
| 15761 | inst.instruction |= LOW4 (inst.operands[0].reg) << 12; |
| 15762 | inst.instruction |= HI1 (inst.operands[0].reg) << 22; |
| 15763 | inst.instruction |= LOW4 (inst.operands[1].reg); |
| 15764 | inst.instruction |= HI1 (inst.operands[1].reg) << 5; |
| 15765 | inst.instruction |= neon_quad (rs) << 6; |
| 15766 | inst.instruction |= 1 << 21; |
| 15767 | if (flavour < neon_cvt_flavour_s16_f16) |
| 15768 | { |
| 15769 | inst.instruction |= 1 << 21; |
| 15770 | immbits = 32 - inst.operands[2].imm; |
| 15771 | inst.instruction |= immbits << 16; |
| 15772 | } |
| 15773 | else |
| 15774 | { |
| 15775 | inst.instruction |= 3 << 20; |
| 15776 | immbits = 16 - inst.operands[2].imm; |
| 15777 | inst.instruction |= immbits << 16; |
| 15778 | inst.instruction &= ~(1 << 9); |
| 15779 | } |
| 15780 | |
| 15781 | neon_dp_fixup (&inst); |
| 15782 | } |
| 15783 | break; |
| 15784 | |
| 15785 | case NS_DD: |
| 15786 | case NS_QQ: |
| 15787 | if (mode != neon_cvt_mode_x && mode != neon_cvt_mode_z) |
| 15788 | { |
| 15789 | NEON_ENCODE (FLOAT, inst); |
| 15790 | set_it_insn_type (OUTSIDE_IT_INSN); |
| 15791 | |
| 15792 | if (vfp_or_neon_is_neon (NEON_CHECK_CC | NEON_CHECK_ARCH8) == FAIL) |
| 15793 | return; |
| 15794 | |
| 15795 | inst.instruction |= LOW4 (inst.operands[0].reg) << 12; |
| 15796 | inst.instruction |= HI1 (inst.operands[0].reg) << 22; |
| 15797 | inst.instruction |= LOW4 (inst.operands[1].reg); |
| 15798 | inst.instruction |= HI1 (inst.operands[1].reg) << 5; |
| 15799 | inst.instruction |= neon_quad (rs) << 6; |
| 15800 | inst.instruction |= (flavour == neon_cvt_flavour_u16_f16 |
| 15801 | || flavour == neon_cvt_flavour_u32_f32) << 7; |
| 15802 | inst.instruction |= mode << 8; |
| 15803 | if (flavour == neon_cvt_flavour_u16_f16 |
| 15804 | || flavour == neon_cvt_flavour_s16_f16) |
| 15805 | /* Mask off the original size bits and reencode them. */ |
| 15806 | inst.instruction = ((inst.instruction & 0xfff3ffff) | (1 << 18)); |
| 15807 | |
| 15808 | if (thumb_mode) |
| 15809 | inst.instruction |= 0xfc000000; |
| 15810 | else |
| 15811 | inst.instruction |= 0xf0000000; |
| 15812 | } |
| 15813 | else |
| 15814 | { |
| 15815 | int_encode: |
| 15816 | { |
| 15817 | unsigned enctab[] = { 0x100, 0x180, 0x0, 0x080, |
| 15818 | 0x100, 0x180, 0x0, 0x080}; |
| 15819 | |
| 15820 | NEON_ENCODE (INTEGER, inst); |
| 15821 | |
| 15822 | if (vfp_or_neon_is_neon (NEON_CHECK_CC | NEON_CHECK_ARCH) == FAIL) |
| 15823 | return; |
| 15824 | |
| 15825 | if (flavour != neon_cvt_flavour_invalid) |
| 15826 | inst.instruction |= enctab[flavour]; |
| 15827 | |
| 15828 | inst.instruction |= LOW4 (inst.operands[0].reg) << 12; |
| 15829 | inst.instruction |= HI1 (inst.operands[0].reg) << 22; |
| 15830 | inst.instruction |= LOW4 (inst.operands[1].reg); |
| 15831 | inst.instruction |= HI1 (inst.operands[1].reg) << 5; |
| 15832 | inst.instruction |= neon_quad (rs) << 6; |
| 15833 | if (flavour >= neon_cvt_flavour_s16_f16 |
| 15834 | && flavour <= neon_cvt_flavour_f16_u16) |
| 15835 | /* Half precision. */ |
| 15836 | inst.instruction |= 1 << 18; |
| 15837 | else |
| 15838 | inst.instruction |= 2 << 18; |
| 15839 | |
| 15840 | neon_dp_fixup (&inst); |
| 15841 | } |
| 15842 | } |
| 15843 | break; |
| 15844 | |
| 15845 | /* Half-precision conversions for Advanced SIMD -- neon. */ |
| 15846 | case NS_QD: |
| 15847 | case NS_DQ: |
| 15848 | |
| 15849 | if ((rs == NS_DQ) |
| 15850 | && (inst.vectype.el[0].size != 16 || inst.vectype.el[1].size != 32)) |
| 15851 | { |
| 15852 | as_bad (_("operand size must match register width")); |
| 15853 | break; |
| 15854 | } |
| 15855 | |
| 15856 | if ((rs == NS_QD) |
| 15857 | && ((inst.vectype.el[0].size != 32 || inst.vectype.el[1].size != 16))) |
| 15858 | { |
| 15859 | as_bad (_("operand size must match register width")); |
| 15860 | break; |
| 15861 | } |
| 15862 | |
| 15863 | if (rs == NS_DQ) |
| 15864 | inst.instruction = 0x3b60600; |
| 15865 | else |
| 15866 | inst.instruction = 0x3b60700; |
| 15867 | |
| 15868 | inst.instruction |= LOW4 (inst.operands[0].reg) << 12; |
| 15869 | inst.instruction |= HI1 (inst.operands[0].reg) << 22; |
| 15870 | inst.instruction |= LOW4 (inst.operands[1].reg); |
| 15871 | inst.instruction |= HI1 (inst.operands[1].reg) << 5; |
| 15872 | neon_dp_fixup (&inst); |
| 15873 | break; |
| 15874 | |
| 15875 | default: |
| 15876 | /* Some VFP conversions go here (s32 <-> f32, u32 <-> f32). */ |
| 15877 | if (mode == neon_cvt_mode_x || mode == neon_cvt_mode_z) |
| 15878 | do_vfp_nsyn_cvt (rs, flavour); |
| 15879 | else |
| 15880 | do_vfp_nsyn_cvt_fpv8 (flavour, mode); |
| 15881 | } |
| 15882 | } |
| 15883 | |
| 15884 | static void |
| 15885 | do_neon_cvtr (void) |
| 15886 | { |
| 15887 | do_neon_cvt_1 (neon_cvt_mode_x); |
| 15888 | } |
| 15889 | |
| 15890 | static void |
| 15891 | do_neon_cvt (void) |
| 15892 | { |
| 15893 | do_neon_cvt_1 (neon_cvt_mode_z); |
| 15894 | } |
| 15895 | |
| 15896 | static void |
| 15897 | do_neon_cvta (void) |
| 15898 | { |
| 15899 | do_neon_cvt_1 (neon_cvt_mode_a); |
| 15900 | } |
| 15901 | |
| 15902 | static void |
| 15903 | do_neon_cvtn (void) |
| 15904 | { |
| 15905 | do_neon_cvt_1 (neon_cvt_mode_n); |
| 15906 | } |
| 15907 | |
| 15908 | static void |
| 15909 | do_neon_cvtp (void) |
| 15910 | { |
| 15911 | do_neon_cvt_1 (neon_cvt_mode_p); |
| 15912 | } |
| 15913 | |
| 15914 | static void |
| 15915 | do_neon_cvtm (void) |
| 15916 | { |
| 15917 | do_neon_cvt_1 (neon_cvt_mode_m); |
| 15918 | } |
| 15919 | |
| 15920 | static void |
| 15921 | do_neon_cvttb_2 (bfd_boolean t, bfd_boolean to, bfd_boolean is_double) |
| 15922 | { |
| 15923 | if (is_double) |
| 15924 | mark_feature_used (&fpu_vfp_ext_armv8); |
| 15925 | |
| 15926 | encode_arm_vfp_reg (inst.operands[0].reg, |
| 15927 | (is_double && !to) ? VFP_REG_Dd : VFP_REG_Sd); |
| 15928 | encode_arm_vfp_reg (inst.operands[1].reg, |
| 15929 | (is_double && to) ? VFP_REG_Dm : VFP_REG_Sm); |
| 15930 | inst.instruction |= to ? 0x10000 : 0; |
| 15931 | inst.instruction |= t ? 0x80 : 0; |
| 15932 | inst.instruction |= is_double ? 0x100 : 0; |
| 15933 | do_vfp_cond_or_thumb (); |
| 15934 | } |
| 15935 | |
| 15936 | static void |
| 15937 | do_neon_cvttb_1 (bfd_boolean t) |
| 15938 | { |
| 15939 | enum neon_shape rs = neon_select_shape (NS_HF, NS_HD, NS_FH, NS_FF, NS_FD, |
| 15940 | NS_DF, NS_DH, NS_NULL); |
| 15941 | |
| 15942 | if (rs == NS_NULL) |
| 15943 | return; |
| 15944 | else if (neon_check_type (2, rs, N_F16, N_F32 | N_VFP).type != NT_invtype) |
| 15945 | { |
| 15946 | inst.error = NULL; |
| 15947 | do_neon_cvttb_2 (t, /*to=*/TRUE, /*is_double=*/FALSE); |
| 15948 | } |
| 15949 | else if (neon_check_type (2, rs, N_F32 | N_VFP, N_F16).type != NT_invtype) |
| 15950 | { |
| 15951 | inst.error = NULL; |
| 15952 | do_neon_cvttb_2 (t, /*to=*/FALSE, /*is_double=*/FALSE); |
| 15953 | } |
| 15954 | else if (neon_check_type (2, rs, N_F16, N_F64 | N_VFP).type != NT_invtype) |
| 15955 | { |
| 15956 | /* The VCVTB and VCVTT instructions with D-register operands |
| 15957 | don't work for SP only targets. */ |
| 15958 | constraint (!ARM_CPU_HAS_FEATURE (cpu_variant, fpu_vfp_ext_armv8), |
| 15959 | _(BAD_FPU)); |
| 15960 | |
| 15961 | inst.error = NULL; |
| 15962 | do_neon_cvttb_2 (t, /*to=*/TRUE, /*is_double=*/TRUE); |
| 15963 | } |
| 15964 | else if (neon_check_type (2, rs, N_F64 | N_VFP, N_F16).type != NT_invtype) |
| 15965 | { |
| 15966 | /* The VCVTB and VCVTT instructions with D-register operands |
| 15967 | don't work for SP only targets. */ |
| 15968 | constraint (!ARM_CPU_HAS_FEATURE (cpu_variant, fpu_vfp_ext_armv8), |
| 15969 | _(BAD_FPU)); |
| 15970 | |
| 15971 | inst.error = NULL; |
| 15972 | do_neon_cvttb_2 (t, /*to=*/FALSE, /*is_double=*/TRUE); |
| 15973 | } |
| 15974 | else |
| 15975 | return; |
| 15976 | } |
| 15977 | |
| 15978 | static void |
| 15979 | do_neon_cvtb (void) |
| 15980 | { |
| 15981 | do_neon_cvttb_1 (FALSE); |
| 15982 | } |
| 15983 | |
| 15984 | |
| 15985 | static void |
| 15986 | do_neon_cvtt (void) |
| 15987 | { |
| 15988 | do_neon_cvttb_1 (TRUE); |
| 15989 | } |
| 15990 | |
| 15991 | static void |
| 15992 | neon_move_immediate (void) |
| 15993 | { |
| 15994 | enum neon_shape rs = neon_select_shape (NS_DI, NS_QI, NS_NULL); |
| 15995 | struct neon_type_el et = neon_check_type (2, rs, |
| 15996 | N_I8 | N_I16 | N_I32 | N_I64 | N_F32 | N_KEY, N_EQK); |
| 15997 | unsigned immlo, immhi = 0, immbits; |
| 15998 | int op, cmode, float_p; |
| 15999 | |
| 16000 | constraint (et.type == NT_invtype, |
| 16001 | _("operand size must be specified for immediate VMOV")); |
| 16002 | |
| 16003 | /* We start out as an MVN instruction if OP = 1, MOV otherwise. */ |
| 16004 | op = (inst.instruction & (1 << 5)) != 0; |
| 16005 | |
| 16006 | immlo = inst.operands[1].imm; |
| 16007 | if (inst.operands[1].regisimm) |
| 16008 | immhi = inst.operands[1].reg; |
| 16009 | |
| 16010 | constraint (et.size < 32 && (immlo & ~((1 << et.size) - 1)) != 0, |
| 16011 | _("immediate has bits set outside the operand size")); |
| 16012 | |
| 16013 | float_p = inst.operands[1].immisfloat; |
| 16014 | |
| 16015 | if ((cmode = neon_cmode_for_move_imm (immlo, immhi, float_p, &immbits, &op, |
| 16016 | et.size, et.type)) == FAIL) |
| 16017 | { |
| 16018 | /* Invert relevant bits only. */ |
| 16019 | neon_invert_size (&immlo, &immhi, et.size); |
| 16020 | /* Flip from VMOV/VMVN to VMVN/VMOV. Some immediate types are unavailable |
| 16021 | with one or the other; those cases are caught by |
| 16022 | neon_cmode_for_move_imm. */ |
| 16023 | op = !op; |
| 16024 | if ((cmode = neon_cmode_for_move_imm (immlo, immhi, float_p, &immbits, |
| 16025 | &op, et.size, et.type)) == FAIL) |
| 16026 | { |
| 16027 | first_error (_("immediate out of range")); |
| 16028 | return; |
| 16029 | } |
| 16030 | } |
| 16031 | |
| 16032 | inst.instruction &= ~(1 << 5); |
| 16033 | inst.instruction |= op << 5; |
| 16034 | |
| 16035 | inst.instruction |= LOW4 (inst.operands[0].reg) << 12; |
| 16036 | inst.instruction |= HI1 (inst.operands[0].reg) << 22; |
| 16037 | inst.instruction |= neon_quad (rs) << 6; |
| 16038 | inst.instruction |= cmode << 8; |
| 16039 | |
| 16040 | neon_write_immbits (immbits); |
| 16041 | } |
| 16042 | |
| 16043 | static void |
| 16044 | do_neon_mvn (void) |
| 16045 | { |
| 16046 | if (inst.operands[1].isreg) |
| 16047 | { |
| 16048 | enum neon_shape rs = neon_select_shape (NS_DD, NS_QQ, NS_NULL); |
| 16049 | |
| 16050 | NEON_ENCODE (INTEGER, inst); |
| 16051 | inst.instruction |= LOW4 (inst.operands[0].reg) << 12; |
| 16052 | inst.instruction |= HI1 (inst.operands[0].reg) << 22; |
| 16053 | inst.instruction |= LOW4 (inst.operands[1].reg); |
| 16054 | inst.instruction |= HI1 (inst.operands[1].reg) << 5; |
| 16055 | inst.instruction |= neon_quad (rs) << 6; |
| 16056 | } |
| 16057 | else |
| 16058 | { |
| 16059 | NEON_ENCODE (IMMED, inst); |
| 16060 | neon_move_immediate (); |
| 16061 | } |
| 16062 | |
| 16063 | neon_dp_fixup (&inst); |
| 16064 | } |
| 16065 | |
| 16066 | /* Encode instructions of form: |
| 16067 | |
| 16068 | |28/24|23|22|21 20|19 16|15 12|11 8|7|6|5|4|3 0| |
| 16069 | | U |x |D |size | Rn | Rd |x x x x|N|x|M|x| Rm | */ |
| 16070 | |
| 16071 | static void |
| 16072 | neon_mixed_length (struct neon_type_el et, unsigned size) |
| 16073 | { |
| 16074 | inst.instruction |= LOW4 (inst.operands[0].reg) << 12; |
| 16075 | inst.instruction |= HI1 (inst.operands[0].reg) << 22; |
| 16076 | inst.instruction |= LOW4 (inst.operands[1].reg) << 16; |
| 16077 | inst.instruction |= HI1 (inst.operands[1].reg) << 7; |
| 16078 | inst.instruction |= LOW4 (inst.operands[2].reg); |
| 16079 | inst.instruction |= HI1 (inst.operands[2].reg) << 5; |
| 16080 | inst.instruction |= (et.type == NT_unsigned) << 24; |
| 16081 | inst.instruction |= neon_logbits (size) << 20; |
| 16082 | |
| 16083 | neon_dp_fixup (&inst); |
| 16084 | } |
| 16085 | |
| 16086 | static void |
| 16087 | do_neon_dyadic_long (void) |
| 16088 | { |
| 16089 | /* FIXME: Type checking for lengthening op. */ |
| 16090 | struct neon_type_el et = neon_check_type (3, NS_QDD, |
| 16091 | N_EQK | N_DBL, N_EQK, N_SU_32 | N_KEY); |
| 16092 | neon_mixed_length (et, et.size); |
| 16093 | } |
| 16094 | |
| 16095 | static void |
| 16096 | do_neon_abal (void) |
| 16097 | { |
| 16098 | struct neon_type_el et = neon_check_type (3, NS_QDD, |
| 16099 | N_EQK | N_INT | N_DBL, N_EQK, N_SU_32 | N_KEY); |
| 16100 | neon_mixed_length (et, et.size); |
| 16101 | } |
| 16102 | |
| 16103 | static void |
| 16104 | neon_mac_reg_scalar_long (unsigned regtypes, unsigned scalartypes) |
| 16105 | { |
| 16106 | if (inst.operands[2].isscalar) |
| 16107 | { |
| 16108 | struct neon_type_el et = neon_check_type (3, NS_QDS, |
| 16109 | N_EQK | N_DBL, N_EQK, regtypes | N_KEY); |
| 16110 | NEON_ENCODE (SCALAR, inst); |
| 16111 | neon_mul_mac (et, et.type == NT_unsigned); |
| 16112 | } |
| 16113 | else |
| 16114 | { |
| 16115 | struct neon_type_el et = neon_check_type (3, NS_QDD, |
| 16116 | N_EQK | N_DBL, N_EQK, scalartypes | N_KEY); |
| 16117 | NEON_ENCODE (INTEGER, inst); |
| 16118 | neon_mixed_length (et, et.size); |
| 16119 | } |
| 16120 | } |
| 16121 | |
| 16122 | static void |
| 16123 | do_neon_mac_maybe_scalar_long (void) |
| 16124 | { |
| 16125 | neon_mac_reg_scalar_long (N_S16 | N_S32 | N_U16 | N_U32, N_SU_32); |
| 16126 | } |
| 16127 | |
| 16128 | static void |
| 16129 | do_neon_dyadic_wide (void) |
| 16130 | { |
| 16131 | struct neon_type_el et = neon_check_type (3, NS_QQD, |
| 16132 | N_EQK | N_DBL, N_EQK | N_DBL, N_SU_32 | N_KEY); |
| 16133 | neon_mixed_length (et, et.size); |
| 16134 | } |
| 16135 | |
| 16136 | static void |
| 16137 | do_neon_dyadic_narrow (void) |
| 16138 | { |
| 16139 | struct neon_type_el et = neon_check_type (3, NS_QDD, |
| 16140 | N_EQK | N_DBL, N_EQK, N_I16 | N_I32 | N_I64 | N_KEY); |
| 16141 | /* Operand sign is unimportant, and the U bit is part of the opcode, |
| 16142 | so force the operand type to integer. */ |
| 16143 | et.type = NT_integer; |
| 16144 | neon_mixed_length (et, et.size / 2); |
| 16145 | } |
| 16146 | |
| 16147 | static void |
| 16148 | do_neon_mul_sat_scalar_long (void) |
| 16149 | { |
| 16150 | neon_mac_reg_scalar_long (N_S16 | N_S32, N_S16 | N_S32); |
| 16151 | } |
| 16152 | |
| 16153 | static void |
| 16154 | do_neon_vmull (void) |
| 16155 | { |
| 16156 | if (inst.operands[2].isscalar) |
| 16157 | do_neon_mac_maybe_scalar_long (); |
| 16158 | else |
| 16159 | { |
| 16160 | struct neon_type_el et = neon_check_type (3, NS_QDD, |
| 16161 | N_EQK | N_DBL, N_EQK, N_SU_32 | N_P8 | N_P64 | N_KEY); |
| 16162 | |
| 16163 | if (et.type == NT_poly) |
| 16164 | NEON_ENCODE (POLY, inst); |
| 16165 | else |
| 16166 | NEON_ENCODE (INTEGER, inst); |
| 16167 | |
| 16168 | /* For polynomial encoding the U bit must be zero, and the size must |
| 16169 | be 8 (encoded as 0b00) or, on ARMv8 or later 64 (encoded, non |
| 16170 | obviously, as 0b10). */ |
| 16171 | if (et.size == 64) |
| 16172 | { |
| 16173 | /* Check we're on the correct architecture. */ |
| 16174 | if (!mark_feature_used (&fpu_crypto_ext_armv8)) |
| 16175 | inst.error = |
| 16176 | _("Instruction form not available on this architecture."); |
| 16177 | |
| 16178 | et.size = 32; |
| 16179 | } |
| 16180 | |
| 16181 | neon_mixed_length (et, et.size); |
| 16182 | } |
| 16183 | } |
| 16184 | |
| 16185 | static void |
| 16186 | do_neon_ext (void) |
| 16187 | { |
| 16188 | enum neon_shape rs = neon_select_shape (NS_DDDI, NS_QQQI, NS_NULL); |
| 16189 | struct neon_type_el et = neon_check_type (3, rs, |
| 16190 | N_EQK, N_EQK, N_8 | N_16 | N_32 | N_64 | N_KEY); |
| 16191 | unsigned imm = (inst.operands[3].imm * et.size) / 8; |
| 16192 | |
| 16193 | constraint (imm >= (unsigned) (neon_quad (rs) ? 16 : 8), |
| 16194 | _("shift out of range")); |
| 16195 | inst.instruction |= LOW4 (inst.operands[0].reg) << 12; |
| 16196 | inst.instruction |= HI1 (inst.operands[0].reg) << 22; |
| 16197 | inst.instruction |= LOW4 (inst.operands[1].reg) << 16; |
| 16198 | inst.instruction |= HI1 (inst.operands[1].reg) << 7; |
| 16199 | inst.instruction |= LOW4 (inst.operands[2].reg); |
| 16200 | inst.instruction |= HI1 (inst.operands[2].reg) << 5; |
| 16201 | inst.instruction |= neon_quad (rs) << 6; |
| 16202 | inst.instruction |= imm << 8; |
| 16203 | |
| 16204 | neon_dp_fixup (&inst); |
| 16205 | } |
| 16206 | |
| 16207 | static void |
| 16208 | do_neon_rev (void) |
| 16209 | { |
| 16210 | enum neon_shape rs = neon_select_shape (NS_DD, NS_QQ, NS_NULL); |
| 16211 | struct neon_type_el et = neon_check_type (2, rs, |
| 16212 | N_EQK, N_8 | N_16 | N_32 | N_KEY); |
| 16213 | unsigned op = (inst.instruction >> 7) & 3; |
| 16214 | /* N (width of reversed regions) is encoded as part of the bitmask. We |
| 16215 | extract it here to check the elements to be reversed are smaller. |
| 16216 | Otherwise we'd get a reserved instruction. */ |
| 16217 | unsigned elsize = (op == 2) ? 16 : (op == 1) ? 32 : (op == 0) ? 64 : 0; |
| 16218 | gas_assert (elsize != 0); |
| 16219 | constraint (et.size >= elsize, |
| 16220 | _("elements must be smaller than reversal region")); |
| 16221 | neon_two_same (neon_quad (rs), 1, et.size); |
| 16222 | } |
| 16223 | |
| 16224 | static void |
| 16225 | do_neon_dup (void) |
| 16226 | { |
| 16227 | if (inst.operands[1].isscalar) |
| 16228 | { |
| 16229 | enum neon_shape rs = neon_select_shape (NS_DS, NS_QS, NS_NULL); |
| 16230 | struct neon_type_el et = neon_check_type (2, rs, |
| 16231 | N_EQK, N_8 | N_16 | N_32 | N_KEY); |
| 16232 | unsigned sizebits = et.size >> 3; |
| 16233 | unsigned dm = NEON_SCALAR_REG (inst.operands[1].reg); |
| 16234 | int logsize = neon_logbits (et.size); |
| 16235 | unsigned x = NEON_SCALAR_INDEX (inst.operands[1].reg) << logsize; |
| 16236 | |
| 16237 | if (vfp_or_neon_is_neon (NEON_CHECK_CC) == FAIL) |
| 16238 | return; |
| 16239 | |
| 16240 | NEON_ENCODE (SCALAR, inst); |
| 16241 | inst.instruction |= LOW4 (inst.operands[0].reg) << 12; |
| 16242 | inst.instruction |= HI1 (inst.operands[0].reg) << 22; |
| 16243 | inst.instruction |= LOW4 (dm); |
| 16244 | inst.instruction |= HI1 (dm) << 5; |
| 16245 | inst.instruction |= neon_quad (rs) << 6; |
| 16246 | inst.instruction |= x << 17; |
| 16247 | inst.instruction |= sizebits << 16; |
| 16248 | |
| 16249 | neon_dp_fixup (&inst); |
| 16250 | } |
| 16251 | else |
| 16252 | { |
| 16253 | enum neon_shape rs = neon_select_shape (NS_DR, NS_QR, NS_NULL); |
| 16254 | struct neon_type_el et = neon_check_type (2, rs, |
| 16255 | N_8 | N_16 | N_32 | N_KEY, N_EQK); |
| 16256 | /* Duplicate ARM register to lanes of vector. */ |
| 16257 | NEON_ENCODE (ARMREG, inst); |
| 16258 | switch (et.size) |
| 16259 | { |
| 16260 | case 8: inst.instruction |= 0x400000; break; |
| 16261 | case 16: inst.instruction |= 0x000020; break; |
| 16262 | case 32: inst.instruction |= 0x000000; break; |
| 16263 | default: break; |
| 16264 | } |
| 16265 | inst.instruction |= LOW4 (inst.operands[1].reg) << 12; |
| 16266 | inst.instruction |= LOW4 (inst.operands[0].reg) << 16; |
| 16267 | inst.instruction |= HI1 (inst.operands[0].reg) << 7; |
| 16268 | inst.instruction |= neon_quad (rs) << 21; |
| 16269 | /* The encoding for this instruction is identical for the ARM and Thumb |
| 16270 | variants, except for the condition field. */ |
| 16271 | do_vfp_cond_or_thumb (); |
| 16272 | } |
| 16273 | } |
| 16274 | |
| 16275 | /* VMOV has particularly many variations. It can be one of: |
| 16276 | 0. VMOV<c><q> <Qd>, <Qm> |
| 16277 | 1. VMOV<c><q> <Dd>, <Dm> |
| 16278 | (Register operations, which are VORR with Rm = Rn.) |
| 16279 | 2. VMOV<c><q>.<dt> <Qd>, #<imm> |
| 16280 | 3. VMOV<c><q>.<dt> <Dd>, #<imm> |
| 16281 | (Immediate loads.) |
| 16282 | 4. VMOV<c><q>.<size> <Dn[x]>, <Rd> |
| 16283 | (ARM register to scalar.) |
| 16284 | 5. VMOV<c><q> <Dm>, <Rd>, <Rn> |
| 16285 | (Two ARM registers to vector.) |
| 16286 | 6. VMOV<c><q>.<dt> <Rd>, <Dn[x]> |
| 16287 | (Scalar to ARM register.) |
| 16288 | 7. VMOV<c><q> <Rd>, <Rn>, <Dm> |
| 16289 | (Vector to two ARM registers.) |
| 16290 | 8. VMOV.F32 <Sd>, <Sm> |
| 16291 | 9. VMOV.F64 <Dd>, <Dm> |
| 16292 | (VFP register moves.) |
| 16293 | 10. VMOV.F32 <Sd>, #imm |
| 16294 | 11. VMOV.F64 <Dd>, #imm |
| 16295 | (VFP float immediate load.) |
| 16296 | 12. VMOV <Rd>, <Sm> |
| 16297 | (VFP single to ARM reg.) |
| 16298 | 13. VMOV <Sd>, <Rm> |
| 16299 | (ARM reg to VFP single.) |
| 16300 | 14. VMOV <Rd>, <Re>, <Sn>, <Sm> |
| 16301 | (Two ARM regs to two VFP singles.) |
| 16302 | 15. VMOV <Sd>, <Se>, <Rn>, <Rm> |
| 16303 | (Two VFP singles to two ARM regs.) |
| 16304 | |
| 16305 | These cases can be disambiguated using neon_select_shape, except cases 1/9 |
| 16306 | and 3/11 which depend on the operand type too. |
| 16307 | |
| 16308 | All the encoded bits are hardcoded by this function. |
| 16309 | |
| 16310 | Cases 4, 6 may be used with VFPv1 and above (only 32-bit transfers!). |
| 16311 | Cases 5, 7 may be used with VFPv2 and above. |
| 16312 | |
| 16313 | FIXME: Some of the checking may be a bit sloppy (in a couple of cases you |
| 16314 | can specify a type where it doesn't make sense to, and is ignored). */ |
| 16315 | |
| 16316 | static void |
| 16317 | do_neon_mov (void) |
| 16318 | { |
| 16319 | enum neon_shape rs = neon_select_shape (NS_RRFF, NS_FFRR, NS_DRR, NS_RRD, |
| 16320 | NS_QQ, NS_DD, NS_QI, NS_DI, NS_SR, |
| 16321 | NS_RS, NS_FF, NS_FI, NS_RF, NS_FR, |
| 16322 | NS_HR, NS_RH, NS_HI, NS_NULL); |
| 16323 | struct neon_type_el et; |
| 16324 | const char *ldconst = 0; |
| 16325 | |
| 16326 | switch (rs) |
| 16327 | { |
| 16328 | case NS_DD: /* case 1/9. */ |
| 16329 | et = neon_check_type (2, rs, N_EQK, N_F64 | N_KEY); |
| 16330 | /* It is not an error here if no type is given. */ |
| 16331 | inst.error = NULL; |
| 16332 | if (et.type == NT_float && et.size == 64) |
| 16333 | { |
| 16334 | do_vfp_nsyn_opcode ("fcpyd"); |
| 16335 | break; |
| 16336 | } |
| 16337 | /* fall through. */ |
| 16338 | |
| 16339 | case NS_QQ: /* case 0/1. */ |
| 16340 | { |
| 16341 | if (vfp_or_neon_is_neon (NEON_CHECK_CC | NEON_CHECK_ARCH) == FAIL) |
| 16342 | return; |
| 16343 | /* The architecture manual I have doesn't explicitly state which |
| 16344 | value the U bit should have for register->register moves, but |
| 16345 | the equivalent VORR instruction has U = 0, so do that. */ |
| 16346 | inst.instruction = 0x0200110; |
| 16347 | inst.instruction |= LOW4 (inst.operands[0].reg) << 12; |
| 16348 | inst.instruction |= HI1 (inst.operands[0].reg) << 22; |
| 16349 | inst.instruction |= LOW4 (inst.operands[1].reg); |
| 16350 | inst.instruction |= HI1 (inst.operands[1].reg) << 5; |
| 16351 | inst.instruction |= LOW4 (inst.operands[1].reg) << 16; |
| 16352 | inst.instruction |= HI1 (inst.operands[1].reg) << 7; |
| 16353 | inst.instruction |= neon_quad (rs) << 6; |
| 16354 | |
| 16355 | neon_dp_fixup (&inst); |
| 16356 | } |
| 16357 | break; |
| 16358 | |
| 16359 | case NS_DI: /* case 3/11. */ |
| 16360 | et = neon_check_type (2, rs, N_EQK, N_F64 | N_KEY); |
| 16361 | inst.error = NULL; |
| 16362 | if (et.type == NT_float && et.size == 64) |
| 16363 | { |
| 16364 | /* case 11 (fconstd). */ |
| 16365 | ldconst = "fconstd"; |
| 16366 | goto encode_fconstd; |
| 16367 | } |
| 16368 | /* fall through. */ |
| 16369 | |
| 16370 | case NS_QI: /* case 2/3. */ |
| 16371 | if (vfp_or_neon_is_neon (NEON_CHECK_CC | NEON_CHECK_ARCH) == FAIL) |
| 16372 | return; |
| 16373 | inst.instruction = 0x0800010; |
| 16374 | neon_move_immediate (); |
| 16375 | neon_dp_fixup (&inst); |
| 16376 | break; |
| 16377 | |
| 16378 | case NS_SR: /* case 4. */ |
| 16379 | { |
| 16380 | unsigned bcdebits = 0; |
| 16381 | int logsize; |
| 16382 | unsigned dn = NEON_SCALAR_REG (inst.operands[0].reg); |
| 16383 | unsigned x = NEON_SCALAR_INDEX (inst.operands[0].reg); |
| 16384 | |
| 16385 | /* .<size> is optional here, defaulting to .32. */ |
| 16386 | if (inst.vectype.elems == 0 |
| 16387 | && inst.operands[0].vectype.type == NT_invtype |
| 16388 | && inst.operands[1].vectype.type == NT_invtype) |
| 16389 | { |
| 16390 | inst.vectype.el[0].type = NT_untyped; |
| 16391 | inst.vectype.el[0].size = 32; |
| 16392 | inst.vectype.elems = 1; |
| 16393 | } |
| 16394 | |
| 16395 | et = neon_check_type (2, NS_NULL, N_8 | N_16 | N_32 | N_KEY, N_EQK); |
| 16396 | logsize = neon_logbits (et.size); |
| 16397 | |
| 16398 | constraint (!ARM_CPU_HAS_FEATURE (cpu_variant, fpu_vfp_ext_v1), |
| 16399 | _(BAD_FPU)); |
| 16400 | constraint (!ARM_CPU_HAS_FEATURE (cpu_variant, fpu_neon_ext_v1) |
| 16401 | && et.size != 32, _(BAD_FPU)); |
| 16402 | constraint (et.type == NT_invtype, _("bad type for scalar")); |
| 16403 | constraint (x >= 64 / et.size, _("scalar index out of range")); |
| 16404 | |
| 16405 | switch (et.size) |
| 16406 | { |
| 16407 | case 8: bcdebits = 0x8; break; |
| 16408 | case 16: bcdebits = 0x1; break; |
| 16409 | case 32: bcdebits = 0x0; break; |
| 16410 | default: ; |
| 16411 | } |
| 16412 | |
| 16413 | bcdebits |= x << logsize; |
| 16414 | |
| 16415 | inst.instruction = 0xe000b10; |
| 16416 | do_vfp_cond_or_thumb (); |
| 16417 | inst.instruction |= LOW4 (dn) << 16; |
| 16418 | inst.instruction |= HI1 (dn) << 7; |
| 16419 | inst.instruction |= inst.operands[1].reg << 12; |
| 16420 | inst.instruction |= (bcdebits & 3) << 5; |
| 16421 | inst.instruction |= (bcdebits >> 2) << 21; |
| 16422 | } |
| 16423 | break; |
| 16424 | |
| 16425 | case NS_DRR: /* case 5 (fmdrr). */ |
| 16426 | constraint (!ARM_CPU_HAS_FEATURE (cpu_variant, fpu_vfp_ext_v2), |
| 16427 | _(BAD_FPU)); |
| 16428 | |
| 16429 | inst.instruction = 0xc400b10; |
| 16430 | do_vfp_cond_or_thumb (); |
| 16431 | inst.instruction |= LOW4 (inst.operands[0].reg); |
| 16432 | inst.instruction |= HI1 (inst.operands[0].reg) << 5; |
| 16433 | inst.instruction |= inst.operands[1].reg << 12; |
| 16434 | inst.instruction |= inst.operands[2].reg << 16; |
| 16435 | break; |
| 16436 | |
| 16437 | case NS_RS: /* case 6. */ |
| 16438 | { |
| 16439 | unsigned logsize; |
| 16440 | unsigned dn = NEON_SCALAR_REG (inst.operands[1].reg); |
| 16441 | unsigned x = NEON_SCALAR_INDEX (inst.operands[1].reg); |
| 16442 | unsigned abcdebits = 0; |
| 16443 | |
| 16444 | /* .<dt> is optional here, defaulting to .32. */ |
| 16445 | if (inst.vectype.elems == 0 |
| 16446 | && inst.operands[0].vectype.type == NT_invtype |
| 16447 | && inst.operands[1].vectype.type == NT_invtype) |
| 16448 | { |
| 16449 | inst.vectype.el[0].type = NT_untyped; |
| 16450 | inst.vectype.el[0].size = 32; |
| 16451 | inst.vectype.elems = 1; |
| 16452 | } |
| 16453 | |
| 16454 | et = neon_check_type (2, NS_NULL, |
| 16455 | N_EQK, N_S8 | N_S16 | N_U8 | N_U16 | N_32 | N_KEY); |
| 16456 | logsize = neon_logbits (et.size); |
| 16457 | |
| 16458 | constraint (!ARM_CPU_HAS_FEATURE (cpu_variant, fpu_vfp_ext_v1), |
| 16459 | _(BAD_FPU)); |
| 16460 | constraint (!ARM_CPU_HAS_FEATURE (cpu_variant, fpu_neon_ext_v1) |
| 16461 | && et.size != 32, _(BAD_FPU)); |
| 16462 | constraint (et.type == NT_invtype, _("bad type for scalar")); |
| 16463 | constraint (x >= 64 / et.size, _("scalar index out of range")); |
| 16464 | |
| 16465 | switch (et.size) |
| 16466 | { |
| 16467 | case 8: abcdebits = (et.type == NT_signed) ? 0x08 : 0x18; break; |
| 16468 | case 16: abcdebits = (et.type == NT_signed) ? 0x01 : 0x11; break; |
| 16469 | case 32: abcdebits = 0x00; break; |
| 16470 | default: ; |
| 16471 | } |
| 16472 | |
| 16473 | abcdebits |= x << logsize; |
| 16474 | inst.instruction = 0xe100b10; |
| 16475 | do_vfp_cond_or_thumb (); |
| 16476 | inst.instruction |= LOW4 (dn) << 16; |
| 16477 | inst.instruction |= HI1 (dn) << 7; |
| 16478 | inst.instruction |= inst.operands[0].reg << 12; |
| 16479 | inst.instruction |= (abcdebits & 3) << 5; |
| 16480 | inst.instruction |= (abcdebits >> 2) << 21; |
| 16481 | } |
| 16482 | break; |
| 16483 | |
| 16484 | case NS_RRD: /* case 7 (fmrrd). */ |
| 16485 | constraint (!ARM_CPU_HAS_FEATURE (cpu_variant, fpu_vfp_ext_v2), |
| 16486 | _(BAD_FPU)); |
| 16487 | |
| 16488 | inst.instruction = 0xc500b10; |
| 16489 | do_vfp_cond_or_thumb (); |
| 16490 | inst.instruction |= inst.operands[0].reg << 12; |
| 16491 | inst.instruction |= inst.operands[1].reg << 16; |
| 16492 | inst.instruction |= LOW4 (inst.operands[2].reg); |
| 16493 | inst.instruction |= HI1 (inst.operands[2].reg) << 5; |
| 16494 | break; |
| 16495 | |
| 16496 | case NS_FF: /* case 8 (fcpys). */ |
| 16497 | do_vfp_nsyn_opcode ("fcpys"); |
| 16498 | break; |
| 16499 | |
| 16500 | case NS_HI: |
| 16501 | case NS_FI: /* case 10 (fconsts). */ |
| 16502 | ldconst = "fconsts"; |
| 16503 | encode_fconstd: |
| 16504 | if (is_quarter_float (inst.operands[1].imm)) |
| 16505 | { |
| 16506 | inst.operands[1].imm = neon_qfloat_bits (inst.operands[1].imm); |
| 16507 | do_vfp_nsyn_opcode (ldconst); |
| 16508 | |
| 16509 | /* ARMv8.2 fp16 vmov.f16 instruction. */ |
| 16510 | if (rs == NS_HI) |
| 16511 | do_scalar_fp16_v82_encode (); |
| 16512 | } |
| 16513 | else |
| 16514 | first_error (_("immediate out of range")); |
| 16515 | break; |
| 16516 | |
| 16517 | case NS_RH: |
| 16518 | case NS_RF: /* case 12 (fmrs). */ |
| 16519 | do_vfp_nsyn_opcode ("fmrs"); |
| 16520 | /* ARMv8.2 fp16 vmov.f16 instruction. */ |
| 16521 | if (rs == NS_RH) |
| 16522 | do_scalar_fp16_v82_encode (); |
| 16523 | break; |
| 16524 | |
| 16525 | case NS_HR: |
| 16526 | case NS_FR: /* case 13 (fmsr). */ |
| 16527 | do_vfp_nsyn_opcode ("fmsr"); |
| 16528 | /* ARMv8.2 fp16 vmov.f16 instruction. */ |
| 16529 | if (rs == NS_HR) |
| 16530 | do_scalar_fp16_v82_encode (); |
| 16531 | break; |
| 16532 | |
| 16533 | /* The encoders for the fmrrs and fmsrr instructions expect three operands |
| 16534 | (one of which is a list), but we have parsed four. Do some fiddling to |
| 16535 | make the operands what do_vfp_reg2_from_sp2 and do_vfp_sp2_from_reg2 |
| 16536 | expect. */ |
| 16537 | case NS_RRFF: /* case 14 (fmrrs). */ |
| 16538 | constraint (inst.operands[3].reg != inst.operands[2].reg + 1, |
| 16539 | _("VFP registers must be adjacent")); |
| 16540 | inst.operands[2].imm = 2; |
| 16541 | memset (&inst.operands[3], '\0', sizeof (inst.operands[3])); |
| 16542 | do_vfp_nsyn_opcode ("fmrrs"); |
| 16543 | break; |
| 16544 | |
| 16545 | case NS_FFRR: /* case 15 (fmsrr). */ |
| 16546 | constraint (inst.operands[1].reg != inst.operands[0].reg + 1, |
| 16547 | _("VFP registers must be adjacent")); |
| 16548 | inst.operands[1] = inst.operands[2]; |
| 16549 | inst.operands[2] = inst.operands[3]; |
| 16550 | inst.operands[0].imm = 2; |
| 16551 | memset (&inst.operands[3], '\0', sizeof (inst.operands[3])); |
| 16552 | do_vfp_nsyn_opcode ("fmsrr"); |
| 16553 | break; |
| 16554 | |
| 16555 | case NS_NULL: |
| 16556 | /* neon_select_shape has determined that the instruction |
| 16557 | shape is wrong and has already set the error message. */ |
| 16558 | break; |
| 16559 | |
| 16560 | default: |
| 16561 | abort (); |
| 16562 | } |
| 16563 | } |
| 16564 | |
| 16565 | static void |
| 16566 | do_neon_rshift_round_imm (void) |
| 16567 | { |
| 16568 | enum neon_shape rs = neon_select_shape (NS_DDI, NS_QQI, NS_NULL); |
| 16569 | struct neon_type_el et = neon_check_type (2, rs, N_EQK, N_SU_ALL | N_KEY); |
| 16570 | int imm = inst.operands[2].imm; |
| 16571 | |
| 16572 | /* imm == 0 case is encoded as VMOV for V{R}SHR. */ |
| 16573 | if (imm == 0) |
| 16574 | { |
| 16575 | inst.operands[2].present = 0; |
| 16576 | do_neon_mov (); |
| 16577 | return; |
| 16578 | } |
| 16579 | |
| 16580 | constraint (imm < 1 || (unsigned)imm > et.size, |
| 16581 | _("immediate out of range for shift")); |
| 16582 | neon_imm_shift (TRUE, et.type == NT_unsigned, neon_quad (rs), et, |
| 16583 | et.size - imm); |
| 16584 | } |
| 16585 | |
| 16586 | static void |
| 16587 | do_neon_movhf (void) |
| 16588 | { |
| 16589 | enum neon_shape rs = neon_select_shape (NS_HH, NS_NULL); |
| 16590 | constraint (rs != NS_HH, _("invalid suffix")); |
| 16591 | |
| 16592 | constraint (!ARM_CPU_HAS_FEATURE (cpu_variant, fpu_vfp_ext_armv8), |
| 16593 | _(BAD_FPU)); |
| 16594 | |
| 16595 | do_vfp_sp_monadic (); |
| 16596 | |
| 16597 | inst.is_neon = 1; |
| 16598 | inst.instruction |= 0xf0000000; |
| 16599 | } |
| 16600 | |
| 16601 | static void |
| 16602 | do_neon_movl (void) |
| 16603 | { |
| 16604 | struct neon_type_el et = neon_check_type (2, NS_QD, |
| 16605 | N_EQK | N_DBL, N_SU_32 | N_KEY); |
| 16606 | unsigned sizebits = et.size >> 3; |
| 16607 | inst.instruction |= sizebits << 19; |
| 16608 | neon_two_same (0, et.type == NT_unsigned, -1); |
| 16609 | } |
| 16610 | |
| 16611 | static void |
| 16612 | do_neon_trn (void) |
| 16613 | { |
| 16614 | enum neon_shape rs = neon_select_shape (NS_DD, NS_QQ, NS_NULL); |
| 16615 | struct neon_type_el et = neon_check_type (2, rs, |
| 16616 | N_EQK, N_8 | N_16 | N_32 | N_KEY); |
| 16617 | NEON_ENCODE (INTEGER, inst); |
| 16618 | neon_two_same (neon_quad (rs), 1, et.size); |
| 16619 | } |
| 16620 | |
| 16621 | static void |
| 16622 | do_neon_zip_uzp (void) |
| 16623 | { |
| 16624 | enum neon_shape rs = neon_select_shape (NS_DD, NS_QQ, NS_NULL); |
| 16625 | struct neon_type_el et = neon_check_type (2, rs, |
| 16626 | N_EQK, N_8 | N_16 | N_32 | N_KEY); |
| 16627 | if (rs == NS_DD && et.size == 32) |
| 16628 | { |
| 16629 | /* Special case: encode as VTRN.32 <Dd>, <Dm>. */ |
| 16630 | inst.instruction = N_MNEM_vtrn; |
| 16631 | do_neon_trn (); |
| 16632 | return; |
| 16633 | } |
| 16634 | neon_two_same (neon_quad (rs), 1, et.size); |
| 16635 | } |
| 16636 | |
| 16637 | static void |
| 16638 | do_neon_sat_abs_neg (void) |
| 16639 | { |
| 16640 | enum neon_shape rs = neon_select_shape (NS_DD, NS_QQ, NS_NULL); |
| 16641 | struct neon_type_el et = neon_check_type (2, rs, |
| 16642 | N_EQK, N_S8 | N_S16 | N_S32 | N_KEY); |
| 16643 | neon_two_same (neon_quad (rs), 1, et.size); |
| 16644 | } |
| 16645 | |
| 16646 | static void |
| 16647 | do_neon_pair_long (void) |
| 16648 | { |
| 16649 | enum neon_shape rs = neon_select_shape (NS_DD, NS_QQ, NS_NULL); |
| 16650 | struct neon_type_el et = neon_check_type (2, rs, N_EQK, N_SU_32 | N_KEY); |
| 16651 | /* Unsigned is encoded in OP field (bit 7) for these instruction. */ |
| 16652 | inst.instruction |= (et.type == NT_unsigned) << 7; |
| 16653 | neon_two_same (neon_quad (rs), 1, et.size); |
| 16654 | } |
| 16655 | |
| 16656 | static void |
| 16657 | do_neon_recip_est (void) |
| 16658 | { |
| 16659 | enum neon_shape rs = neon_select_shape (NS_DD, NS_QQ, NS_NULL); |
| 16660 | struct neon_type_el et = neon_check_type (2, rs, |
| 16661 | N_EQK | N_FLT, N_F_16_32 | N_U32 | N_KEY); |
| 16662 | inst.instruction |= (et.type == NT_float) << 8; |
| 16663 | neon_two_same (neon_quad (rs), 1, et.size); |
| 16664 | } |
| 16665 | |
| 16666 | static void |
| 16667 | do_neon_cls (void) |
| 16668 | { |
| 16669 | enum neon_shape rs = neon_select_shape (NS_DD, NS_QQ, NS_NULL); |
| 16670 | struct neon_type_el et = neon_check_type (2, rs, |
| 16671 | N_EQK, N_S8 | N_S16 | N_S32 | N_KEY); |
| 16672 | neon_two_same (neon_quad (rs), 1, et.size); |
| 16673 | } |
| 16674 | |
| 16675 | static void |
| 16676 | do_neon_clz (void) |
| 16677 | { |
| 16678 | enum neon_shape rs = neon_select_shape (NS_DD, NS_QQ, NS_NULL); |
| 16679 | struct neon_type_el et = neon_check_type (2, rs, |
| 16680 | N_EQK, N_I8 | N_I16 | N_I32 | N_KEY); |
| 16681 | neon_two_same (neon_quad (rs), 1, et.size); |
| 16682 | } |
| 16683 | |
| 16684 | static void |
| 16685 | do_neon_cnt (void) |
| 16686 | { |
| 16687 | enum neon_shape rs = neon_select_shape (NS_DD, NS_QQ, NS_NULL); |
| 16688 | struct neon_type_el et = neon_check_type (2, rs, |
| 16689 | N_EQK | N_INT, N_8 | N_KEY); |
| 16690 | neon_two_same (neon_quad (rs), 1, et.size); |
| 16691 | } |
| 16692 | |
| 16693 | static void |
| 16694 | do_neon_swp (void) |
| 16695 | { |
| 16696 | enum neon_shape rs = neon_select_shape (NS_DD, NS_QQ, NS_NULL); |
| 16697 | neon_two_same (neon_quad (rs), 1, -1); |
| 16698 | } |
| 16699 | |
| 16700 | static void |
| 16701 | do_neon_tbl_tbx (void) |
| 16702 | { |
| 16703 | unsigned listlenbits; |
| 16704 | neon_check_type (3, NS_DLD, N_EQK, N_EQK, N_8 | N_KEY); |
| 16705 | |
| 16706 | if (inst.operands[1].imm < 1 || inst.operands[1].imm > 4) |
| 16707 | { |
| 16708 | first_error (_("bad list length for table lookup")); |
| 16709 | return; |
| 16710 | } |
| 16711 | |
| 16712 | listlenbits = inst.operands[1].imm - 1; |
| 16713 | inst.instruction |= LOW4 (inst.operands[0].reg) << 12; |
| 16714 | inst.instruction |= HI1 (inst.operands[0].reg) << 22; |
| 16715 | inst.instruction |= LOW4 (inst.operands[1].reg) << 16; |
| 16716 | inst.instruction |= HI1 (inst.operands[1].reg) << 7; |
| 16717 | inst.instruction |= LOW4 (inst.operands[2].reg); |
| 16718 | inst.instruction |= HI1 (inst.operands[2].reg) << 5; |
| 16719 | inst.instruction |= listlenbits << 8; |
| 16720 | |
| 16721 | neon_dp_fixup (&inst); |
| 16722 | } |
| 16723 | |
| 16724 | static void |
| 16725 | do_neon_ldm_stm (void) |
| 16726 | { |
| 16727 | /* P, U and L bits are part of bitmask. */ |
| 16728 | int is_dbmode = (inst.instruction & (1 << 24)) != 0; |
| 16729 | unsigned offsetbits = inst.operands[1].imm * 2; |
| 16730 | |
| 16731 | if (inst.operands[1].issingle) |
| 16732 | { |
| 16733 | do_vfp_nsyn_ldm_stm (is_dbmode); |
| 16734 | return; |
| 16735 | } |
| 16736 | |
| 16737 | constraint (is_dbmode && !inst.operands[0].writeback, |
| 16738 | _("writeback (!) must be used for VLDMDB and VSTMDB")); |
| 16739 | |
| 16740 | constraint (inst.operands[1].imm < 1 || inst.operands[1].imm > 16, |
| 16741 | _("register list must contain at least 1 and at most 16 " |
| 16742 | "registers")); |
| 16743 | |
| 16744 | inst.instruction |= inst.operands[0].reg << 16; |
| 16745 | inst.instruction |= inst.operands[0].writeback << 21; |
| 16746 | inst.instruction |= LOW4 (inst.operands[1].reg) << 12; |
| 16747 | inst.instruction |= HI1 (inst.operands[1].reg) << 22; |
| 16748 | |
| 16749 | inst.instruction |= offsetbits; |
| 16750 | |
| 16751 | do_vfp_cond_or_thumb (); |
| 16752 | } |
| 16753 | |
| 16754 | static void |
| 16755 | do_neon_ldr_str (void) |
| 16756 | { |
| 16757 | int is_ldr = (inst.instruction & (1 << 20)) != 0; |
| 16758 | |
| 16759 | /* Use of PC in vstr in ARM mode is deprecated in ARMv7. |
| 16760 | And is UNPREDICTABLE in thumb mode. */ |
| 16761 | if (!is_ldr |
| 16762 | && inst.operands[1].reg == REG_PC |
| 16763 | && (ARM_CPU_HAS_FEATURE (selected_cpu, arm_ext_v7) || thumb_mode)) |
| 16764 | { |
| 16765 | if (thumb_mode) |
| 16766 | inst.error = _("Use of PC here is UNPREDICTABLE"); |
| 16767 | else if (warn_on_deprecated) |
| 16768 | as_tsktsk (_("Use of PC here is deprecated")); |
| 16769 | } |
| 16770 | |
| 16771 | if (inst.operands[0].issingle) |
| 16772 | { |
| 16773 | if (is_ldr) |
| 16774 | do_vfp_nsyn_opcode ("flds"); |
| 16775 | else |
| 16776 | do_vfp_nsyn_opcode ("fsts"); |
| 16777 | |
| 16778 | /* ARMv8.2 vldr.16/vstr.16 instruction. */ |
| 16779 | if (inst.vectype.el[0].size == 16) |
| 16780 | do_scalar_fp16_v82_encode (); |
| 16781 | } |
| 16782 | else |
| 16783 | { |
| 16784 | if (is_ldr) |
| 16785 | do_vfp_nsyn_opcode ("fldd"); |
| 16786 | else |
| 16787 | do_vfp_nsyn_opcode ("fstd"); |
| 16788 | } |
| 16789 | } |
| 16790 | |
| 16791 | /* "interleave" version also handles non-interleaving register VLD1/VST1 |
| 16792 | instructions. */ |
| 16793 | |
| 16794 | static void |
| 16795 | do_neon_ld_st_interleave (void) |
| 16796 | { |
| 16797 | struct neon_type_el et = neon_check_type (1, NS_NULL, |
| 16798 | N_8 | N_16 | N_32 | N_64); |
| 16799 | unsigned alignbits = 0; |
| 16800 | unsigned idx; |
| 16801 | /* The bits in this table go: |
| 16802 | 0: register stride of one (0) or two (1) |
| 16803 | 1,2: register list length, minus one (1, 2, 3, 4). |
| 16804 | 3,4: <n> in instruction type, minus one (VLD<n> / VST<n>). |
| 16805 | We use -1 for invalid entries. */ |
| 16806 | const int typetable[] = |
| 16807 | { |
| 16808 | 0x7, -1, 0xa, -1, 0x6, -1, 0x2, -1, /* VLD1 / VST1. */ |
| 16809 | -1, -1, 0x8, 0x9, -1, -1, 0x3, -1, /* VLD2 / VST2. */ |
| 16810 | -1, -1, -1, -1, 0x4, 0x5, -1, -1, /* VLD3 / VST3. */ |
| 16811 | -1, -1, -1, -1, -1, -1, 0x0, 0x1 /* VLD4 / VST4. */ |
| 16812 | }; |
| 16813 | int typebits; |
| 16814 | |
| 16815 | if (et.type == NT_invtype) |
| 16816 | return; |
| 16817 | |
| 16818 | if (inst.operands[1].immisalign) |
| 16819 | switch (inst.operands[1].imm >> 8) |
| 16820 | { |
| 16821 | case 64: alignbits = 1; break; |
| 16822 | case 128: |
| 16823 | if (NEON_REGLIST_LENGTH (inst.operands[0].imm) != 2 |
| 16824 | && NEON_REGLIST_LENGTH (inst.operands[0].imm) != 4) |
| 16825 | goto bad_alignment; |
| 16826 | alignbits = 2; |
| 16827 | break; |
| 16828 | case 256: |
| 16829 | if (NEON_REGLIST_LENGTH (inst.operands[0].imm) != 4) |
| 16830 | goto bad_alignment; |
| 16831 | alignbits = 3; |
| 16832 | break; |
| 16833 | default: |
| 16834 | bad_alignment: |
| 16835 | first_error (_("bad alignment")); |
| 16836 | return; |
| 16837 | } |
| 16838 | |
| 16839 | inst.instruction |= alignbits << 4; |
| 16840 | inst.instruction |= neon_logbits (et.size) << 6; |
| 16841 | |
| 16842 | /* Bits [4:6] of the immediate in a list specifier encode register stride |
| 16843 | (minus 1) in bit 4, and list length in bits [5:6]. We put the <n> of |
| 16844 | VLD<n>/VST<n> in bits [9:8] of the initial bitmask. Suck it out here, look |
| 16845 | up the right value for "type" in a table based on this value and the given |
| 16846 | list style, then stick it back. */ |
| 16847 | idx = ((inst.operands[0].imm >> 4) & 7) |
| 16848 | | (((inst.instruction >> 8) & 3) << 3); |
| 16849 | |
| 16850 | typebits = typetable[idx]; |
| 16851 | |
| 16852 | constraint (typebits == -1, _("bad list type for instruction")); |
| 16853 | constraint (((inst.instruction >> 8) & 3) && et.size == 64, |
| 16854 | _("bad element type for instruction")); |
| 16855 | |
| 16856 | inst.instruction &= ~0xf00; |
| 16857 | inst.instruction |= typebits << 8; |
| 16858 | } |
| 16859 | |
| 16860 | /* Check alignment is valid for do_neon_ld_st_lane and do_neon_ld_dup. |
| 16861 | *DO_ALIGN is set to 1 if the relevant alignment bit should be set, 0 |
| 16862 | otherwise. The variable arguments are a list of pairs of legal (size, align) |
| 16863 | values, terminated with -1. */ |
| 16864 | |
| 16865 | static int |
| 16866 | neon_alignment_bit (int size, int align, int *do_alignment, ...) |
| 16867 | { |
| 16868 | va_list ap; |
| 16869 | int result = FAIL, thissize, thisalign; |
| 16870 | |
| 16871 | if (!inst.operands[1].immisalign) |
| 16872 | { |
| 16873 | *do_alignment = 0; |
| 16874 | return SUCCESS; |
| 16875 | } |
| 16876 | |
| 16877 | va_start (ap, do_alignment); |
| 16878 | |
| 16879 | do |
| 16880 | { |
| 16881 | thissize = va_arg (ap, int); |
| 16882 | if (thissize == -1) |
| 16883 | break; |
| 16884 | thisalign = va_arg (ap, int); |
| 16885 | |
| 16886 | if (size == thissize && align == thisalign) |
| 16887 | result = SUCCESS; |
| 16888 | } |
| 16889 | while (result != SUCCESS); |
| 16890 | |
| 16891 | va_end (ap); |
| 16892 | |
| 16893 | if (result == SUCCESS) |
| 16894 | *do_alignment = 1; |
| 16895 | else |
| 16896 | first_error (_("unsupported alignment for instruction")); |
| 16897 | |
| 16898 | return result; |
| 16899 | } |
| 16900 | |
| 16901 | static void |
| 16902 | do_neon_ld_st_lane (void) |
| 16903 | { |
| 16904 | struct neon_type_el et = neon_check_type (1, NS_NULL, N_8 | N_16 | N_32); |
| 16905 | int align_good, do_alignment = 0; |
| 16906 | int logsize = neon_logbits (et.size); |
| 16907 | int align = inst.operands[1].imm >> 8; |
| 16908 | int n = (inst.instruction >> 8) & 3; |
| 16909 | int max_el = 64 / et.size; |
| 16910 | |
| 16911 | if (et.type == NT_invtype) |
| 16912 | return; |
| 16913 | |
| 16914 | constraint (NEON_REGLIST_LENGTH (inst.operands[0].imm) != n + 1, |
| 16915 | _("bad list length")); |
| 16916 | constraint (NEON_LANE (inst.operands[0].imm) >= max_el, |
| 16917 | _("scalar index out of range")); |
| 16918 | constraint (n != 0 && NEON_REG_STRIDE (inst.operands[0].imm) == 2 |
| 16919 | && et.size == 8, |
| 16920 | _("stride of 2 unavailable when element size is 8")); |
| 16921 | |
| 16922 | switch (n) |
| 16923 | { |
| 16924 | case 0: /* VLD1 / VST1. */ |
| 16925 | align_good = neon_alignment_bit (et.size, align, &do_alignment, 16, 16, |
| 16926 | 32, 32, -1); |
| 16927 | if (align_good == FAIL) |
| 16928 | return; |
| 16929 | if (do_alignment) |
| 16930 | { |
| 16931 | unsigned alignbits = 0; |
| 16932 | switch (et.size) |
| 16933 | { |
| 16934 | case 16: alignbits = 0x1; break; |
| 16935 | case 32: alignbits = 0x3; break; |
| 16936 | default: ; |
| 16937 | } |
| 16938 | inst.instruction |= alignbits << 4; |
| 16939 | } |
| 16940 | break; |
| 16941 | |
| 16942 | case 1: /* VLD2 / VST2. */ |
| 16943 | align_good = neon_alignment_bit (et.size, align, &do_alignment, 8, 16, |
| 16944 | 16, 32, 32, 64, -1); |
| 16945 | if (align_good == FAIL) |
| 16946 | return; |
| 16947 | if (do_alignment) |
| 16948 | inst.instruction |= 1 << 4; |
| 16949 | break; |
| 16950 | |
| 16951 | case 2: /* VLD3 / VST3. */ |
| 16952 | constraint (inst.operands[1].immisalign, |
| 16953 | _("can't use alignment with this instruction")); |
| 16954 | break; |
| 16955 | |
| 16956 | case 3: /* VLD4 / VST4. */ |
| 16957 | align_good = neon_alignment_bit (et.size, align, &do_alignment, 8, 32, |
| 16958 | 16, 64, 32, 64, 32, 128, -1); |
| 16959 | if (align_good == FAIL) |
| 16960 | return; |
| 16961 | if (do_alignment) |
| 16962 | { |
| 16963 | unsigned alignbits = 0; |
| 16964 | switch (et.size) |
| 16965 | { |
| 16966 | case 8: alignbits = 0x1; break; |
| 16967 | case 16: alignbits = 0x1; break; |
| 16968 | case 32: alignbits = (align == 64) ? 0x1 : 0x2; break; |
| 16969 | default: ; |
| 16970 | } |
| 16971 | inst.instruction |= alignbits << 4; |
| 16972 | } |
| 16973 | break; |
| 16974 | |
| 16975 | default: ; |
| 16976 | } |
| 16977 | |
| 16978 | /* Reg stride of 2 is encoded in bit 5 when size==16, bit 6 when size==32. */ |
| 16979 | if (n != 0 && NEON_REG_STRIDE (inst.operands[0].imm) == 2) |
| 16980 | inst.instruction |= 1 << (4 + logsize); |
| 16981 | |
| 16982 | inst.instruction |= NEON_LANE (inst.operands[0].imm) << (logsize + 5); |
| 16983 | inst.instruction |= logsize << 10; |
| 16984 | } |
| 16985 | |
| 16986 | /* Encode single n-element structure to all lanes VLD<n> instructions. */ |
| 16987 | |
| 16988 | static void |
| 16989 | do_neon_ld_dup (void) |
| 16990 | { |
| 16991 | struct neon_type_el et = neon_check_type (1, NS_NULL, N_8 | N_16 | N_32); |
| 16992 | int align_good, do_alignment = 0; |
| 16993 | |
| 16994 | if (et.type == NT_invtype) |
| 16995 | return; |
| 16996 | |
| 16997 | switch ((inst.instruction >> 8) & 3) |
| 16998 | { |
| 16999 | case 0: /* VLD1. */ |
| 17000 | gas_assert (NEON_REG_STRIDE (inst.operands[0].imm) != 2); |
| 17001 | align_good = neon_alignment_bit (et.size, inst.operands[1].imm >> 8, |
| 17002 | &do_alignment, 16, 16, 32, 32, -1); |
| 17003 | if (align_good == FAIL) |
| 17004 | return; |
| 17005 | switch (NEON_REGLIST_LENGTH (inst.operands[0].imm)) |
| 17006 | { |
| 17007 | case 1: break; |
| 17008 | case 2: inst.instruction |= 1 << 5; break; |
| 17009 | default: first_error (_("bad list length")); return; |
| 17010 | } |
| 17011 | inst.instruction |= neon_logbits (et.size) << 6; |
| 17012 | break; |
| 17013 | |
| 17014 | case 1: /* VLD2. */ |
| 17015 | align_good = neon_alignment_bit (et.size, inst.operands[1].imm >> 8, |
| 17016 | &do_alignment, 8, 16, 16, 32, 32, 64, |
| 17017 | -1); |
| 17018 | if (align_good == FAIL) |
| 17019 | return; |
| 17020 | constraint (NEON_REGLIST_LENGTH (inst.operands[0].imm) != 2, |
| 17021 | _("bad list length")); |
| 17022 | if (NEON_REG_STRIDE (inst.operands[0].imm) == 2) |
| 17023 | inst.instruction |= 1 << 5; |
| 17024 | inst.instruction |= neon_logbits (et.size) << 6; |
| 17025 | break; |
| 17026 | |
| 17027 | case 2: /* VLD3. */ |
| 17028 | constraint (inst.operands[1].immisalign, |
| 17029 | _("can't use alignment with this instruction")); |
| 17030 | constraint (NEON_REGLIST_LENGTH (inst.operands[0].imm) != 3, |
| 17031 | _("bad list length")); |
| 17032 | if (NEON_REG_STRIDE (inst.operands[0].imm) == 2) |
| 17033 | inst.instruction |= 1 << 5; |
| 17034 | inst.instruction |= neon_logbits (et.size) << 6; |
| 17035 | break; |
| 17036 | |
| 17037 | case 3: /* VLD4. */ |
| 17038 | { |
| 17039 | int align = inst.operands[1].imm >> 8; |
| 17040 | align_good = neon_alignment_bit (et.size, align, &do_alignment, 8, 32, |
| 17041 | 16, 64, 32, 64, 32, 128, -1); |
| 17042 | if (align_good == FAIL) |
| 17043 | return; |
| 17044 | constraint (NEON_REGLIST_LENGTH (inst.operands[0].imm) != 4, |
| 17045 | _("bad list length")); |
| 17046 | if (NEON_REG_STRIDE (inst.operands[0].imm) == 2) |
| 17047 | inst.instruction |= 1 << 5; |
| 17048 | if (et.size == 32 && align == 128) |
| 17049 | inst.instruction |= 0x3 << 6; |
| 17050 | else |
| 17051 | inst.instruction |= neon_logbits (et.size) << 6; |
| 17052 | } |
| 17053 | break; |
| 17054 | |
| 17055 | default: ; |
| 17056 | } |
| 17057 | |
| 17058 | inst.instruction |= do_alignment << 4; |
| 17059 | } |
| 17060 | |
| 17061 | /* Disambiguate VLD<n> and VST<n> instructions, and fill in common bits (those |
| 17062 | apart from bits [11:4]. */ |
| 17063 | |
| 17064 | static void |
| 17065 | do_neon_ldx_stx (void) |
| 17066 | { |
| 17067 | if (inst.operands[1].isreg) |
| 17068 | constraint (inst.operands[1].reg == REG_PC, BAD_PC); |
| 17069 | |
| 17070 | switch (NEON_LANE (inst.operands[0].imm)) |
| 17071 | { |
| 17072 | case NEON_INTERLEAVE_LANES: |
| 17073 | NEON_ENCODE (INTERLV, inst); |
| 17074 | do_neon_ld_st_interleave (); |
| 17075 | break; |
| 17076 | |
| 17077 | case NEON_ALL_LANES: |
| 17078 | NEON_ENCODE (DUP, inst); |
| 17079 | if (inst.instruction == N_INV) |
| 17080 | { |
| 17081 | first_error ("only loads support such operands"); |
| 17082 | break; |
| 17083 | } |
| 17084 | do_neon_ld_dup (); |
| 17085 | break; |
| 17086 | |
| 17087 | default: |
| 17088 | NEON_ENCODE (LANE, inst); |
| 17089 | do_neon_ld_st_lane (); |
| 17090 | } |
| 17091 | |
| 17092 | /* L bit comes from bit mask. */ |
| 17093 | inst.instruction |= LOW4 (inst.operands[0].reg) << 12; |
| 17094 | inst.instruction |= HI1 (inst.operands[0].reg) << 22; |
| 17095 | inst.instruction |= inst.operands[1].reg << 16; |
| 17096 | |
| 17097 | if (inst.operands[1].postind) |
| 17098 | { |
| 17099 | int postreg = inst.operands[1].imm & 0xf; |
| 17100 | constraint (!inst.operands[1].immisreg, |
| 17101 | _("post-index must be a register")); |
| 17102 | constraint (postreg == 0xd || postreg == 0xf, |
| 17103 | _("bad register for post-index")); |
| 17104 | inst.instruction |= postreg; |
| 17105 | } |
| 17106 | else |
| 17107 | { |
| 17108 | constraint (inst.operands[1].immisreg, BAD_ADDR_MODE); |
| 17109 | constraint (inst.reloc.exp.X_op != O_constant |
| 17110 | || inst.reloc.exp.X_add_number != 0, |
| 17111 | BAD_ADDR_MODE); |
| 17112 | |
| 17113 | if (inst.operands[1].writeback) |
| 17114 | { |
| 17115 | inst.instruction |= 0xd; |
| 17116 | } |
| 17117 | else |
| 17118 | inst.instruction |= 0xf; |
| 17119 | } |
| 17120 | |
| 17121 | if (thumb_mode) |
| 17122 | inst.instruction |= 0xf9000000; |
| 17123 | else |
| 17124 | inst.instruction |= 0xf4000000; |
| 17125 | } |
| 17126 | |
| 17127 | /* FP v8. */ |
| 17128 | static void |
| 17129 | do_vfp_nsyn_fpv8 (enum neon_shape rs) |
| 17130 | { |
| 17131 | /* Targets like FPv5-SP-D16 don't support FP v8 instructions with |
| 17132 | D register operands. */ |
| 17133 | if (neon_shape_class[rs] == SC_DOUBLE) |
| 17134 | constraint (!ARM_CPU_HAS_FEATURE (cpu_variant, fpu_vfp_ext_armv8), |
| 17135 | _(BAD_FPU)); |
| 17136 | |
| 17137 | NEON_ENCODE (FPV8, inst); |
| 17138 | |
| 17139 | if (rs == NS_FFF || rs == NS_HHH) |
| 17140 | { |
| 17141 | do_vfp_sp_dyadic (); |
| 17142 | |
| 17143 | /* ARMv8.2 fp16 instruction. */ |
| 17144 | if (rs == NS_HHH) |
| 17145 | do_scalar_fp16_v82_encode (); |
| 17146 | } |
| 17147 | else |
| 17148 | do_vfp_dp_rd_rn_rm (); |
| 17149 | |
| 17150 | if (rs == NS_DDD) |
| 17151 | inst.instruction |= 0x100; |
| 17152 | |
| 17153 | inst.instruction |= 0xf0000000; |
| 17154 | } |
| 17155 | |
| 17156 | static void |
| 17157 | do_vsel (void) |
| 17158 | { |
| 17159 | set_it_insn_type (OUTSIDE_IT_INSN); |
| 17160 | |
| 17161 | if (try_vfp_nsyn (3, do_vfp_nsyn_fpv8) != SUCCESS) |
| 17162 | first_error (_("invalid instruction shape")); |
| 17163 | } |
| 17164 | |
| 17165 | static void |
| 17166 | do_vmaxnm (void) |
| 17167 | { |
| 17168 | set_it_insn_type (OUTSIDE_IT_INSN); |
| 17169 | |
| 17170 | if (try_vfp_nsyn (3, do_vfp_nsyn_fpv8) == SUCCESS) |
| 17171 | return; |
| 17172 | |
| 17173 | if (vfp_or_neon_is_neon (NEON_CHECK_CC | NEON_CHECK_ARCH8) == FAIL) |
| 17174 | return; |
| 17175 | |
| 17176 | neon_dyadic_misc (NT_untyped, N_F_16_32, 0); |
| 17177 | } |
| 17178 | |
| 17179 | static void |
| 17180 | do_vrint_1 (enum neon_cvt_mode mode) |
| 17181 | { |
| 17182 | enum neon_shape rs = neon_select_shape (NS_HH, NS_FF, NS_DD, NS_QQ, NS_NULL); |
| 17183 | struct neon_type_el et; |
| 17184 | |
| 17185 | if (rs == NS_NULL) |
| 17186 | return; |
| 17187 | |
| 17188 | /* Targets like FPv5-SP-D16 don't support FP v8 instructions with |
| 17189 | D register operands. */ |
| 17190 | if (neon_shape_class[rs] == SC_DOUBLE) |
| 17191 | constraint (!ARM_CPU_HAS_FEATURE (cpu_variant, fpu_vfp_ext_armv8), |
| 17192 | _(BAD_FPU)); |
| 17193 | |
| 17194 | et = neon_check_type (2, rs, N_EQK | N_VFP, N_F_ALL | N_KEY |
| 17195 | | N_VFP); |
| 17196 | if (et.type != NT_invtype) |
| 17197 | { |
| 17198 | /* VFP encodings. */ |
| 17199 | if (mode == neon_cvt_mode_a || mode == neon_cvt_mode_n |
| 17200 | || mode == neon_cvt_mode_p || mode == neon_cvt_mode_m) |
| 17201 | set_it_insn_type (OUTSIDE_IT_INSN); |
| 17202 | |
| 17203 | NEON_ENCODE (FPV8, inst); |
| 17204 | if (rs == NS_FF || rs == NS_HH) |
| 17205 | do_vfp_sp_monadic (); |
| 17206 | else |
| 17207 | do_vfp_dp_rd_rm (); |
| 17208 | |
| 17209 | switch (mode) |
| 17210 | { |
| 17211 | case neon_cvt_mode_r: inst.instruction |= 0x00000000; break; |
| 17212 | case neon_cvt_mode_z: inst.instruction |= 0x00000080; break; |
| 17213 | case neon_cvt_mode_x: inst.instruction |= 0x00010000; break; |
| 17214 | case neon_cvt_mode_a: inst.instruction |= 0xf0000000; break; |
| 17215 | case neon_cvt_mode_n: inst.instruction |= 0xf0010000; break; |
| 17216 | case neon_cvt_mode_p: inst.instruction |= 0xf0020000; break; |
| 17217 | case neon_cvt_mode_m: inst.instruction |= 0xf0030000; break; |
| 17218 | default: abort (); |
| 17219 | } |
| 17220 | |
| 17221 | inst.instruction |= (rs == NS_DD) << 8; |
| 17222 | do_vfp_cond_or_thumb (); |
| 17223 | |
| 17224 | /* ARMv8.2 fp16 vrint instruction. */ |
| 17225 | if (rs == NS_HH) |
| 17226 | do_scalar_fp16_v82_encode (); |
| 17227 | } |
| 17228 | else |
| 17229 | { |
| 17230 | /* Neon encodings (or something broken...). */ |
| 17231 | inst.error = NULL; |
| 17232 | et = neon_check_type (2, rs, N_EQK, N_F_16_32 | N_KEY); |
| 17233 | |
| 17234 | if (et.type == NT_invtype) |
| 17235 | return; |
| 17236 | |
| 17237 | set_it_insn_type (OUTSIDE_IT_INSN); |
| 17238 | NEON_ENCODE (FLOAT, inst); |
| 17239 | |
| 17240 | if (vfp_or_neon_is_neon (NEON_CHECK_CC | NEON_CHECK_ARCH8) == FAIL) |
| 17241 | return; |
| 17242 | |
| 17243 | inst.instruction |= LOW4 (inst.operands[0].reg) << 12; |
| 17244 | inst.instruction |= HI1 (inst.operands[0].reg) << 22; |
| 17245 | inst.instruction |= LOW4 (inst.operands[1].reg); |
| 17246 | inst.instruction |= HI1 (inst.operands[1].reg) << 5; |
| 17247 | inst.instruction |= neon_quad (rs) << 6; |
| 17248 | /* Mask off the original size bits and reencode them. */ |
| 17249 | inst.instruction = ((inst.instruction & 0xfff3ffff) |
| 17250 | | neon_logbits (et.size) << 18); |
| 17251 | |
| 17252 | switch (mode) |
| 17253 | { |
| 17254 | case neon_cvt_mode_z: inst.instruction |= 3 << 7; break; |
| 17255 | case neon_cvt_mode_x: inst.instruction |= 1 << 7; break; |
| 17256 | case neon_cvt_mode_a: inst.instruction |= 2 << 7; break; |
| 17257 | case neon_cvt_mode_n: inst.instruction |= 0 << 7; break; |
| 17258 | case neon_cvt_mode_p: inst.instruction |= 7 << 7; break; |
| 17259 | case neon_cvt_mode_m: inst.instruction |= 5 << 7; break; |
| 17260 | case neon_cvt_mode_r: inst.error = _("invalid rounding mode"); break; |
| 17261 | default: abort (); |
| 17262 | } |
| 17263 | |
| 17264 | if (thumb_mode) |
| 17265 | inst.instruction |= 0xfc000000; |
| 17266 | else |
| 17267 | inst.instruction |= 0xf0000000; |
| 17268 | } |
| 17269 | } |
| 17270 | |
| 17271 | static void |
| 17272 | do_vrintx (void) |
| 17273 | { |
| 17274 | do_vrint_1 (neon_cvt_mode_x); |
| 17275 | } |
| 17276 | |
| 17277 | static void |
| 17278 | do_vrintz (void) |
| 17279 | { |
| 17280 | do_vrint_1 (neon_cvt_mode_z); |
| 17281 | } |
| 17282 | |
| 17283 | static void |
| 17284 | do_vrintr (void) |
| 17285 | { |
| 17286 | do_vrint_1 (neon_cvt_mode_r); |
| 17287 | } |
| 17288 | |
| 17289 | static void |
| 17290 | do_vrinta (void) |
| 17291 | { |
| 17292 | do_vrint_1 (neon_cvt_mode_a); |
| 17293 | } |
| 17294 | |
| 17295 | static void |
| 17296 | do_vrintn (void) |
| 17297 | { |
| 17298 | do_vrint_1 (neon_cvt_mode_n); |
| 17299 | } |
| 17300 | |
| 17301 | static void |
| 17302 | do_vrintp (void) |
| 17303 | { |
| 17304 | do_vrint_1 (neon_cvt_mode_p); |
| 17305 | } |
| 17306 | |
| 17307 | static void |
| 17308 | do_vrintm (void) |
| 17309 | { |
| 17310 | do_vrint_1 (neon_cvt_mode_m); |
| 17311 | } |
| 17312 | |
| 17313 | static unsigned |
| 17314 | neon_scalar_for_vcmla (unsigned opnd, unsigned elsize) |
| 17315 | { |
| 17316 | unsigned regno = NEON_SCALAR_REG (opnd); |
| 17317 | unsigned elno = NEON_SCALAR_INDEX (opnd); |
| 17318 | |
| 17319 | if (elsize == 16 && elno < 2 && regno < 16) |
| 17320 | return regno | (elno << 4); |
| 17321 | else if (elsize == 32 && elno == 0) |
| 17322 | return regno; |
| 17323 | |
| 17324 | first_error (_("scalar out of range")); |
| 17325 | return 0; |
| 17326 | } |
| 17327 | |
| 17328 | static void |
| 17329 | do_vcmla (void) |
| 17330 | { |
| 17331 | constraint (!ARM_CPU_HAS_FEATURE (cpu_variant, fpu_neon_ext_armv8), |
| 17332 | _(BAD_FPU)); |
| 17333 | constraint (inst.reloc.exp.X_op != O_constant, _("expression too complex")); |
| 17334 | unsigned rot = inst.reloc.exp.X_add_number; |
| 17335 | constraint (rot != 0 && rot != 90 && rot != 180 && rot != 270, |
| 17336 | _("immediate out of range")); |
| 17337 | rot /= 90; |
| 17338 | if (inst.operands[2].isscalar) |
| 17339 | { |
| 17340 | enum neon_shape rs = neon_select_shape (NS_DDSI, NS_QQSI, NS_NULL); |
| 17341 | unsigned size = neon_check_type (3, rs, N_EQK, N_EQK, |
| 17342 | N_KEY | N_F16 | N_F32).size; |
| 17343 | unsigned m = neon_scalar_for_vcmla (inst.operands[2].reg, size); |
| 17344 | inst.is_neon = 1; |
| 17345 | inst.instruction = 0xfe000800; |
| 17346 | inst.instruction |= LOW4 (inst.operands[0].reg) << 12; |
| 17347 | inst.instruction |= HI1 (inst.operands[0].reg) << 22; |
| 17348 | inst.instruction |= LOW4 (inst.operands[1].reg) << 16; |
| 17349 | inst.instruction |= HI1 (inst.operands[1].reg) << 7; |
| 17350 | inst.instruction |= LOW4 (m); |
| 17351 | inst.instruction |= HI1 (m) << 5; |
| 17352 | inst.instruction |= neon_quad (rs) << 6; |
| 17353 | inst.instruction |= rot << 20; |
| 17354 | inst.instruction |= (size == 32) << 23; |
| 17355 | } |
| 17356 | else |
| 17357 | { |
| 17358 | enum neon_shape rs = neon_select_shape (NS_DDDI, NS_QQQI, NS_NULL); |
| 17359 | unsigned size = neon_check_type (3, rs, N_EQK, N_EQK, |
| 17360 | N_KEY | N_F16 | N_F32).size; |
| 17361 | neon_three_same (neon_quad (rs), 0, -1); |
| 17362 | inst.instruction &= 0x00ffffff; /* Undo neon_dp_fixup. */ |
| 17363 | inst.instruction |= 0xfc200800; |
| 17364 | inst.instruction |= rot << 23; |
| 17365 | inst.instruction |= (size == 32) << 20; |
| 17366 | } |
| 17367 | } |
| 17368 | |
| 17369 | static void |
| 17370 | do_vcadd (void) |
| 17371 | { |
| 17372 | constraint (!ARM_CPU_HAS_FEATURE (cpu_variant, fpu_neon_ext_armv8), |
| 17373 | _(BAD_FPU)); |
| 17374 | constraint (inst.reloc.exp.X_op != O_constant, _("expression too complex")); |
| 17375 | unsigned rot = inst.reloc.exp.X_add_number; |
| 17376 | constraint (rot != 90 && rot != 270, _("immediate out of range")); |
| 17377 | enum neon_shape rs = neon_select_shape (NS_DDDI, NS_QQQI, NS_NULL); |
| 17378 | unsigned size = neon_check_type (3, rs, N_EQK, N_EQK, |
| 17379 | N_KEY | N_F16 | N_F32).size; |
| 17380 | neon_three_same (neon_quad (rs), 0, -1); |
| 17381 | inst.instruction &= 0x00ffffff; /* Undo neon_dp_fixup. */ |
| 17382 | inst.instruction |= 0xfc800800; |
| 17383 | inst.instruction |= (rot == 270) << 24; |
| 17384 | inst.instruction |= (size == 32) << 20; |
| 17385 | } |
| 17386 | |
| 17387 | /* Dot Product instructions encoding support. */ |
| 17388 | |
| 17389 | static void |
| 17390 | do_neon_dotproduct (int unsigned_p) |
| 17391 | { |
| 17392 | enum neon_shape rs; |
| 17393 | unsigned scalar_oprd2 = 0; |
| 17394 | int high8; |
| 17395 | |
| 17396 | if (inst.cond != COND_ALWAYS) |
| 17397 | as_warn (_("Dot Product instructions cannot be conditional, the behaviour " |
| 17398 | "is UNPREDICTABLE")); |
| 17399 | |
| 17400 | constraint (!ARM_CPU_HAS_FEATURE (cpu_variant, fpu_neon_ext_armv8), |
| 17401 | _(BAD_FPU)); |
| 17402 | |
| 17403 | /* Dot Product instructions are in three-same D/Q register format or the third |
| 17404 | operand can be a scalar index register. */ |
| 17405 | if (inst.operands[2].isscalar) |
| 17406 | { |
| 17407 | scalar_oprd2 = neon_scalar_for_mul (inst.operands[2].reg, 32); |
| 17408 | high8 = 0xfe000000; |
| 17409 | rs = neon_select_shape (NS_DDS, NS_QQS, NS_NULL); |
| 17410 | } |
| 17411 | else |
| 17412 | { |
| 17413 | high8 = 0xfc000000; |
| 17414 | rs = neon_select_shape (NS_DDD, NS_QQQ, NS_NULL); |
| 17415 | } |
| 17416 | |
| 17417 | if (unsigned_p) |
| 17418 | neon_check_type (3, rs, N_EQK, N_EQK, N_KEY | N_U8); |
| 17419 | else |
| 17420 | neon_check_type (3, rs, N_EQK, N_EQK, N_KEY | N_S8); |
| 17421 | |
| 17422 | /* The "U" bit in traditional Three Same encoding is fixed to 0 for Dot |
| 17423 | Product instruction, so we pass 0 as the "ubit" parameter. And the |
| 17424 | "Size" field are fixed to 0x2, so we pass 32 as the "size" parameter. */ |
| 17425 | neon_three_same (neon_quad (rs), 0, 32); |
| 17426 | |
| 17427 | /* Undo neon_dp_fixup. Dot Product instructions are using a slightly |
| 17428 | different NEON three-same encoding. */ |
| 17429 | inst.instruction &= 0x00ffffff; |
| 17430 | inst.instruction |= high8; |
| 17431 | /* Encode 'U' bit which indicates signedness. */ |
| 17432 | inst.instruction |= (unsigned_p ? 1 : 0) << 4; |
| 17433 | /* Re-encode operand2 if it's indexed scalar operand. What has been encoded |
| 17434 | from inst.operand[2].reg in neon_three_same is GAS's internal encoding, not |
| 17435 | the instruction encoding. */ |
| 17436 | if (inst.operands[2].isscalar) |
| 17437 | { |
| 17438 | inst.instruction &= 0xffffffd0; |
| 17439 | inst.instruction |= LOW4 (scalar_oprd2); |
| 17440 | inst.instruction |= HI1 (scalar_oprd2) << 5; |
| 17441 | } |
| 17442 | } |
| 17443 | |
| 17444 | /* Dot Product instructions for signed integer. */ |
| 17445 | |
| 17446 | static void |
| 17447 | do_neon_dotproduct_s (void) |
| 17448 | { |
| 17449 | return do_neon_dotproduct (0); |
| 17450 | } |
| 17451 | |
| 17452 | /* Dot Product instructions for unsigned integer. */ |
| 17453 | |
| 17454 | static void |
| 17455 | do_neon_dotproduct_u (void) |
| 17456 | { |
| 17457 | return do_neon_dotproduct (1); |
| 17458 | } |
| 17459 | |
| 17460 | /* Crypto v1 instructions. */ |
| 17461 | static void |
| 17462 | do_crypto_2op_1 (unsigned elttype, int op) |
| 17463 | { |
| 17464 | set_it_insn_type (OUTSIDE_IT_INSN); |
| 17465 | |
| 17466 | if (neon_check_type (2, NS_QQ, N_EQK | N_UNT, elttype | N_UNT | N_KEY).type |
| 17467 | == NT_invtype) |
| 17468 | return; |
| 17469 | |
| 17470 | inst.error = NULL; |
| 17471 | |
| 17472 | NEON_ENCODE (INTEGER, inst); |
| 17473 | inst.instruction |= LOW4 (inst.operands[0].reg) << 12; |
| 17474 | inst.instruction |= HI1 (inst.operands[0].reg) << 22; |
| 17475 | inst.instruction |= LOW4 (inst.operands[1].reg); |
| 17476 | inst.instruction |= HI1 (inst.operands[1].reg) << 5; |
| 17477 | if (op != -1) |
| 17478 | inst.instruction |= op << 6; |
| 17479 | |
| 17480 | if (thumb_mode) |
| 17481 | inst.instruction |= 0xfc000000; |
| 17482 | else |
| 17483 | inst.instruction |= 0xf0000000; |
| 17484 | } |
| 17485 | |
| 17486 | static void |
| 17487 | do_crypto_3op_1 (int u, int op) |
| 17488 | { |
| 17489 | set_it_insn_type (OUTSIDE_IT_INSN); |
| 17490 | |
| 17491 | if (neon_check_type (3, NS_QQQ, N_EQK | N_UNT, N_EQK | N_UNT, |
| 17492 | N_32 | N_UNT | N_KEY).type == NT_invtype) |
| 17493 | return; |
| 17494 | |
| 17495 | inst.error = NULL; |
| 17496 | |
| 17497 | NEON_ENCODE (INTEGER, inst); |
| 17498 | neon_three_same (1, u, 8 << op); |
| 17499 | } |
| 17500 | |
| 17501 | static void |
| 17502 | do_aese (void) |
| 17503 | { |
| 17504 | do_crypto_2op_1 (N_8, 0); |
| 17505 | } |
| 17506 | |
| 17507 | static void |
| 17508 | do_aesd (void) |
| 17509 | { |
| 17510 | do_crypto_2op_1 (N_8, 1); |
| 17511 | } |
| 17512 | |
| 17513 | static void |
| 17514 | do_aesmc (void) |
| 17515 | { |
| 17516 | do_crypto_2op_1 (N_8, 2); |
| 17517 | } |
| 17518 | |
| 17519 | static void |
| 17520 | do_aesimc (void) |
| 17521 | { |
| 17522 | do_crypto_2op_1 (N_8, 3); |
| 17523 | } |
| 17524 | |
| 17525 | static void |
| 17526 | do_sha1c (void) |
| 17527 | { |
| 17528 | do_crypto_3op_1 (0, 0); |
| 17529 | } |
| 17530 | |
| 17531 | static void |
| 17532 | do_sha1p (void) |
| 17533 | { |
| 17534 | do_crypto_3op_1 (0, 1); |
| 17535 | } |
| 17536 | |
| 17537 | static void |
| 17538 | do_sha1m (void) |
| 17539 | { |
| 17540 | do_crypto_3op_1 (0, 2); |
| 17541 | } |
| 17542 | |
| 17543 | static void |
| 17544 | do_sha1su0 (void) |
| 17545 | { |
| 17546 | do_crypto_3op_1 (0, 3); |
| 17547 | } |
| 17548 | |
| 17549 | static void |
| 17550 | do_sha256h (void) |
| 17551 | { |
| 17552 | do_crypto_3op_1 (1, 0); |
| 17553 | } |
| 17554 | |
| 17555 | static void |
| 17556 | do_sha256h2 (void) |
| 17557 | { |
| 17558 | do_crypto_3op_1 (1, 1); |
| 17559 | } |
| 17560 | |
| 17561 | static void |
| 17562 | do_sha256su1 (void) |
| 17563 | { |
| 17564 | do_crypto_3op_1 (1, 2); |
| 17565 | } |
| 17566 | |
| 17567 | static void |
| 17568 | do_sha1h (void) |
| 17569 | { |
| 17570 | do_crypto_2op_1 (N_32, -1); |
| 17571 | } |
| 17572 | |
| 17573 | static void |
| 17574 | do_sha1su1 (void) |
| 17575 | { |
| 17576 | do_crypto_2op_1 (N_32, 0); |
| 17577 | } |
| 17578 | |
| 17579 | static void |
| 17580 | do_sha256su0 (void) |
| 17581 | { |
| 17582 | do_crypto_2op_1 (N_32, 1); |
| 17583 | } |
| 17584 | |
| 17585 | static void |
| 17586 | do_crc32_1 (unsigned int poly, unsigned int sz) |
| 17587 | { |
| 17588 | unsigned int Rd = inst.operands[0].reg; |
| 17589 | unsigned int Rn = inst.operands[1].reg; |
| 17590 | unsigned int Rm = inst.operands[2].reg; |
| 17591 | |
| 17592 | set_it_insn_type (OUTSIDE_IT_INSN); |
| 17593 | inst.instruction |= LOW4 (Rd) << (thumb_mode ? 8 : 12); |
| 17594 | inst.instruction |= LOW4 (Rn) << 16; |
| 17595 | inst.instruction |= LOW4 (Rm); |
| 17596 | inst.instruction |= sz << (thumb_mode ? 4 : 21); |
| 17597 | inst.instruction |= poly << (thumb_mode ? 20 : 9); |
| 17598 | |
| 17599 | if (Rd == REG_PC || Rn == REG_PC || Rm == REG_PC) |
| 17600 | as_warn (UNPRED_REG ("r15")); |
| 17601 | } |
| 17602 | |
| 17603 | static void |
| 17604 | do_crc32b (void) |
| 17605 | { |
| 17606 | do_crc32_1 (0, 0); |
| 17607 | } |
| 17608 | |
| 17609 | static void |
| 17610 | do_crc32h (void) |
| 17611 | { |
| 17612 | do_crc32_1 (0, 1); |
| 17613 | } |
| 17614 | |
| 17615 | static void |
| 17616 | do_crc32w (void) |
| 17617 | { |
| 17618 | do_crc32_1 (0, 2); |
| 17619 | } |
| 17620 | |
| 17621 | static void |
| 17622 | do_crc32cb (void) |
| 17623 | { |
| 17624 | do_crc32_1 (1, 0); |
| 17625 | } |
| 17626 | |
| 17627 | static void |
| 17628 | do_crc32ch (void) |
| 17629 | { |
| 17630 | do_crc32_1 (1, 1); |
| 17631 | } |
| 17632 | |
| 17633 | static void |
| 17634 | do_crc32cw (void) |
| 17635 | { |
| 17636 | do_crc32_1 (1, 2); |
| 17637 | } |
| 17638 | |
| 17639 | static void |
| 17640 | do_vjcvt (void) |
| 17641 | { |
| 17642 | constraint (!ARM_CPU_HAS_FEATURE (cpu_variant, fpu_vfp_ext_armv8), |
| 17643 | _(BAD_FPU)); |
| 17644 | neon_check_type (2, NS_FD, N_S32, N_F64); |
| 17645 | do_vfp_sp_dp_cvt (); |
| 17646 | do_vfp_cond_or_thumb (); |
| 17647 | } |
| 17648 | |
| 17649 | \f |
| 17650 | /* Overall per-instruction processing. */ |
| 17651 | |
| 17652 | /* We need to be able to fix up arbitrary expressions in some statements. |
| 17653 | This is so that we can handle symbols that are an arbitrary distance from |
| 17654 | the pc. The most common cases are of the form ((+/-sym -/+ . - 8) & mask), |
| 17655 | which returns part of an address in a form which will be valid for |
| 17656 | a data instruction. We do this by pushing the expression into a symbol |
| 17657 | in the expr_section, and creating a fix for that. */ |
| 17658 | |
| 17659 | static void |
| 17660 | fix_new_arm (fragS * frag, |
| 17661 | int where, |
| 17662 | short int size, |
| 17663 | expressionS * exp, |
| 17664 | int pc_rel, |
| 17665 | int reloc) |
| 17666 | { |
| 17667 | fixS * new_fix; |
| 17668 | |
| 17669 | switch (exp->X_op) |
| 17670 | { |
| 17671 | case O_constant: |
| 17672 | if (pc_rel) |
| 17673 | { |
| 17674 | /* Create an absolute valued symbol, so we have something to |
| 17675 | refer to in the object file. Unfortunately for us, gas's |
| 17676 | generic expression parsing will already have folded out |
| 17677 | any use of .set foo/.type foo %function that may have |
| 17678 | been used to set type information of the target location, |
| 17679 | that's being specified symbolically. We have to presume |
| 17680 | the user knows what they are doing. */ |
| 17681 | char name[16 + 8]; |
| 17682 | symbolS *symbol; |
| 17683 | |
| 17684 | sprintf (name, "*ABS*0x%lx", (unsigned long)exp->X_add_number); |
| 17685 | |
| 17686 | symbol = symbol_find_or_make (name); |
| 17687 | S_SET_SEGMENT (symbol, absolute_section); |
| 17688 | symbol_set_frag (symbol, &zero_address_frag); |
| 17689 | S_SET_VALUE (symbol, exp->X_add_number); |
| 17690 | exp->X_op = O_symbol; |
| 17691 | exp->X_add_symbol = symbol; |
| 17692 | exp->X_add_number = 0; |
| 17693 | } |
| 17694 | /* FALLTHROUGH */ |
| 17695 | case O_symbol: |
| 17696 | case O_add: |
| 17697 | case O_subtract: |
| 17698 | new_fix = fix_new_exp (frag, where, size, exp, pc_rel, |
| 17699 | (enum bfd_reloc_code_real) reloc); |
| 17700 | break; |
| 17701 | |
| 17702 | default: |
| 17703 | new_fix = (fixS *) fix_new (frag, where, size, make_expr_symbol (exp), 0, |
| 17704 | pc_rel, (enum bfd_reloc_code_real) reloc); |
| 17705 | break; |
| 17706 | } |
| 17707 | |
| 17708 | /* Mark whether the fix is to a THUMB instruction, or an ARM |
| 17709 | instruction. */ |
| 17710 | new_fix->tc_fix_data = thumb_mode; |
| 17711 | } |
| 17712 | |
| 17713 | /* Create a frg for an instruction requiring relaxation. */ |
| 17714 | static void |
| 17715 | output_relax_insn (void) |
| 17716 | { |
| 17717 | char * to; |
| 17718 | symbolS *sym; |
| 17719 | int offset; |
| 17720 | |
| 17721 | /* The size of the instruction is unknown, so tie the debug info to the |
| 17722 | start of the instruction. */ |
| 17723 | dwarf2_emit_insn (0); |
| 17724 | |
| 17725 | switch (inst.reloc.exp.X_op) |
| 17726 | { |
| 17727 | case O_symbol: |
| 17728 | sym = inst.reloc.exp.X_add_symbol; |
| 17729 | offset = inst.reloc.exp.X_add_number; |
| 17730 | break; |
| 17731 | case O_constant: |
| 17732 | sym = NULL; |
| 17733 | offset = inst.reloc.exp.X_add_number; |
| 17734 | break; |
| 17735 | default: |
| 17736 | sym = make_expr_symbol (&inst.reloc.exp); |
| 17737 | offset = 0; |
| 17738 | break; |
| 17739 | } |
| 17740 | to = frag_var (rs_machine_dependent, INSN_SIZE, THUMB_SIZE, |
| 17741 | inst.relax, sym, offset, NULL/*offset, opcode*/); |
| 17742 | md_number_to_chars (to, inst.instruction, THUMB_SIZE); |
| 17743 | } |
| 17744 | |
| 17745 | /* Write a 32-bit thumb instruction to buf. */ |
| 17746 | static void |
| 17747 | put_thumb32_insn (char * buf, unsigned long insn) |
| 17748 | { |
| 17749 | md_number_to_chars (buf, insn >> 16, THUMB_SIZE); |
| 17750 | md_number_to_chars (buf + THUMB_SIZE, insn, THUMB_SIZE); |
| 17751 | } |
| 17752 | |
| 17753 | static void |
| 17754 | output_inst (const char * str) |
| 17755 | { |
| 17756 | char * to = NULL; |
| 17757 | |
| 17758 | if (inst.error) |
| 17759 | { |
| 17760 | as_bad ("%s -- `%s'", inst.error, str); |
| 17761 | return; |
| 17762 | } |
| 17763 | if (inst.relax) |
| 17764 | { |
| 17765 | output_relax_insn (); |
| 17766 | return; |
| 17767 | } |
| 17768 | if (inst.size == 0) |
| 17769 | return; |
| 17770 | |
| 17771 | to = frag_more (inst.size); |
| 17772 | /* PR 9814: Record the thumb mode into the current frag so that we know |
| 17773 | what type of NOP padding to use, if necessary. We override any previous |
| 17774 | setting so that if the mode has changed then the NOPS that we use will |
| 17775 | match the encoding of the last instruction in the frag. */ |
| 17776 | frag_now->tc_frag_data.thumb_mode = thumb_mode | MODE_RECORDED; |
| 17777 | |
| 17778 | if (thumb_mode && (inst.size > THUMB_SIZE)) |
| 17779 | { |
| 17780 | gas_assert (inst.size == (2 * THUMB_SIZE)); |
| 17781 | put_thumb32_insn (to, inst.instruction); |
| 17782 | } |
| 17783 | else if (inst.size > INSN_SIZE) |
| 17784 | { |
| 17785 | gas_assert (inst.size == (2 * INSN_SIZE)); |
| 17786 | md_number_to_chars (to, inst.instruction, INSN_SIZE); |
| 17787 | md_number_to_chars (to + INSN_SIZE, inst.instruction, INSN_SIZE); |
| 17788 | } |
| 17789 | else |
| 17790 | md_number_to_chars (to, inst.instruction, inst.size); |
| 17791 | |
| 17792 | if (inst.reloc.type != BFD_RELOC_UNUSED) |
| 17793 | fix_new_arm (frag_now, to - frag_now->fr_literal, |
| 17794 | inst.size, & inst.reloc.exp, inst.reloc.pc_rel, |
| 17795 | inst.reloc.type); |
| 17796 | |
| 17797 | dwarf2_emit_insn (inst.size); |
| 17798 | } |
| 17799 | |
| 17800 | static char * |
| 17801 | output_it_inst (int cond, int mask, char * to) |
| 17802 | { |
| 17803 | unsigned long instruction = 0xbf00; |
| 17804 | |
| 17805 | mask &= 0xf; |
| 17806 | instruction |= mask; |
| 17807 | instruction |= cond << 4; |
| 17808 | |
| 17809 | if (to == NULL) |
| 17810 | { |
| 17811 | to = frag_more (2); |
| 17812 | #ifdef OBJ_ELF |
| 17813 | dwarf2_emit_insn (2); |
| 17814 | #endif |
| 17815 | } |
| 17816 | |
| 17817 | md_number_to_chars (to, instruction, 2); |
| 17818 | |
| 17819 | return to; |
| 17820 | } |
| 17821 | |
| 17822 | /* Tag values used in struct asm_opcode's tag field. */ |
| 17823 | enum opcode_tag |
| 17824 | { |
| 17825 | OT_unconditional, /* Instruction cannot be conditionalized. |
| 17826 | The ARM condition field is still 0xE. */ |
| 17827 | OT_unconditionalF, /* Instruction cannot be conditionalized |
| 17828 | and carries 0xF in its ARM condition field. */ |
| 17829 | OT_csuffix, /* Instruction takes a conditional suffix. */ |
| 17830 | OT_csuffixF, /* Some forms of the instruction take a conditional |
| 17831 | suffix, others place 0xF where the condition field |
| 17832 | would be. */ |
| 17833 | OT_cinfix3, /* Instruction takes a conditional infix, |
| 17834 | beginning at character index 3. (In |
| 17835 | unified mode, it becomes a suffix.) */ |
| 17836 | OT_cinfix3_deprecated, /* The same as OT_cinfix3. This is used for |
| 17837 | tsts, cmps, cmns, and teqs. */ |
| 17838 | OT_cinfix3_legacy, /* Legacy instruction takes a conditional infix at |
| 17839 | character index 3, even in unified mode. Used for |
| 17840 | legacy instructions where suffix and infix forms |
| 17841 | may be ambiguous. */ |
| 17842 | OT_csuf_or_in3, /* Instruction takes either a conditional |
| 17843 | suffix or an infix at character index 3. */ |
| 17844 | OT_odd_infix_unc, /* This is the unconditional variant of an |
| 17845 | instruction that takes a conditional infix |
| 17846 | at an unusual position. In unified mode, |
| 17847 | this variant will accept a suffix. */ |
| 17848 | OT_odd_infix_0 /* Values greater than or equal to OT_odd_infix_0 |
| 17849 | are the conditional variants of instructions that |
| 17850 | take conditional infixes in unusual positions. |
| 17851 | The infix appears at character index |
| 17852 | (tag - OT_odd_infix_0). These are not accepted |
| 17853 | in unified mode. */ |
| 17854 | }; |
| 17855 | |
| 17856 | /* Subroutine of md_assemble, responsible for looking up the primary |
| 17857 | opcode from the mnemonic the user wrote. STR points to the |
| 17858 | beginning of the mnemonic. |
| 17859 | |
| 17860 | This is not simply a hash table lookup, because of conditional |
| 17861 | variants. Most instructions have conditional variants, which are |
| 17862 | expressed with a _conditional affix_ to the mnemonic. If we were |
| 17863 | to encode each conditional variant as a literal string in the opcode |
| 17864 | table, it would have approximately 20,000 entries. |
| 17865 | |
| 17866 | Most mnemonics take this affix as a suffix, and in unified syntax, |
| 17867 | 'most' is upgraded to 'all'. However, in the divided syntax, some |
| 17868 | instructions take the affix as an infix, notably the s-variants of |
| 17869 | the arithmetic instructions. Of those instructions, all but six |
| 17870 | have the infix appear after the third character of the mnemonic. |
| 17871 | |
| 17872 | Accordingly, the algorithm for looking up primary opcodes given |
| 17873 | an identifier is: |
| 17874 | |
| 17875 | 1. Look up the identifier in the opcode table. |
| 17876 | If we find a match, go to step U. |
| 17877 | |
| 17878 | 2. Look up the last two characters of the identifier in the |
| 17879 | conditions table. If we find a match, look up the first N-2 |
| 17880 | characters of the identifier in the opcode table. If we |
| 17881 | find a match, go to step CE. |
| 17882 | |
| 17883 | 3. Look up the fourth and fifth characters of the identifier in |
| 17884 | the conditions table. If we find a match, extract those |
| 17885 | characters from the identifier, and look up the remaining |
| 17886 | characters in the opcode table. If we find a match, go |
| 17887 | to step CM. |
| 17888 | |
| 17889 | 4. Fail. |
| 17890 | |
| 17891 | U. Examine the tag field of the opcode structure, in case this is |
| 17892 | one of the six instructions with its conditional infix in an |
| 17893 | unusual place. If it is, the tag tells us where to find the |
| 17894 | infix; look it up in the conditions table and set inst.cond |
| 17895 | accordingly. Otherwise, this is an unconditional instruction. |
| 17896 | Again set inst.cond accordingly. Return the opcode structure. |
| 17897 | |
| 17898 | CE. Examine the tag field to make sure this is an instruction that |
| 17899 | should receive a conditional suffix. If it is not, fail. |
| 17900 | Otherwise, set inst.cond from the suffix we already looked up, |
| 17901 | and return the opcode structure. |
| 17902 | |
| 17903 | CM. Examine the tag field to make sure this is an instruction that |
| 17904 | should receive a conditional infix after the third character. |
| 17905 | If it is not, fail. Otherwise, undo the edits to the current |
| 17906 | line of input and proceed as for case CE. */ |
| 17907 | |
| 17908 | static const struct asm_opcode * |
| 17909 | opcode_lookup (char **str) |
| 17910 | { |
| 17911 | char *end, *base; |
| 17912 | char *affix; |
| 17913 | const struct asm_opcode *opcode; |
| 17914 | const struct asm_cond *cond; |
| 17915 | char save[2]; |
| 17916 | |
| 17917 | /* Scan up to the end of the mnemonic, which must end in white space, |
| 17918 | '.' (in unified mode, or for Neon/VFP instructions), or end of string. */ |
| 17919 | for (base = end = *str; *end != '\0'; end++) |
| 17920 | if (*end == ' ' || *end == '.') |
| 17921 | break; |
| 17922 | |
| 17923 | if (end == base) |
| 17924 | return NULL; |
| 17925 | |
| 17926 | /* Handle a possible width suffix and/or Neon type suffix. */ |
| 17927 | if (end[0] == '.') |
| 17928 | { |
| 17929 | int offset = 2; |
| 17930 | |
| 17931 | /* The .w and .n suffixes are only valid if the unified syntax is in |
| 17932 | use. */ |
| 17933 | if (unified_syntax && end[1] == 'w') |
| 17934 | inst.size_req = 4; |
| 17935 | else if (unified_syntax && end[1] == 'n') |
| 17936 | inst.size_req = 2; |
| 17937 | else |
| 17938 | offset = 0; |
| 17939 | |
| 17940 | inst.vectype.elems = 0; |
| 17941 | |
| 17942 | *str = end + offset; |
| 17943 | |
| 17944 | if (end[offset] == '.') |
| 17945 | { |
| 17946 | /* See if we have a Neon type suffix (possible in either unified or |
| 17947 | non-unified ARM syntax mode). */ |
| 17948 | if (parse_neon_type (&inst.vectype, str) == FAIL) |
| 17949 | return NULL; |
| 17950 | } |
| 17951 | else if (end[offset] != '\0' && end[offset] != ' ') |
| 17952 | return NULL; |
| 17953 | } |
| 17954 | else |
| 17955 | *str = end; |
| 17956 | |
| 17957 | /* Look for unaffixed or special-case affixed mnemonic. */ |
| 17958 | opcode = (const struct asm_opcode *) hash_find_n (arm_ops_hsh, base, |
| 17959 | end - base); |
| 17960 | if (opcode) |
| 17961 | { |
| 17962 | /* step U */ |
| 17963 | if (opcode->tag < OT_odd_infix_0) |
| 17964 | { |
| 17965 | inst.cond = COND_ALWAYS; |
| 17966 | return opcode; |
| 17967 | } |
| 17968 | |
| 17969 | if (warn_on_deprecated && unified_syntax) |
| 17970 | as_tsktsk (_("conditional infixes are deprecated in unified syntax")); |
| 17971 | affix = base + (opcode->tag - OT_odd_infix_0); |
| 17972 | cond = (const struct asm_cond *) hash_find_n (arm_cond_hsh, affix, 2); |
| 17973 | gas_assert (cond); |
| 17974 | |
| 17975 | inst.cond = cond->value; |
| 17976 | return opcode; |
| 17977 | } |
| 17978 | |
| 17979 | /* Cannot have a conditional suffix on a mnemonic of less than two |
| 17980 | characters. */ |
| 17981 | if (end - base < 3) |
| 17982 | return NULL; |
| 17983 | |
| 17984 | /* Look for suffixed mnemonic. */ |
| 17985 | affix = end - 2; |
| 17986 | cond = (const struct asm_cond *) hash_find_n (arm_cond_hsh, affix, 2); |
| 17987 | opcode = (const struct asm_opcode *) hash_find_n (arm_ops_hsh, base, |
| 17988 | affix - base); |
| 17989 | if (opcode && cond) |
| 17990 | { |
| 17991 | /* step CE */ |
| 17992 | switch (opcode->tag) |
| 17993 | { |
| 17994 | case OT_cinfix3_legacy: |
| 17995 | /* Ignore conditional suffixes matched on infix only mnemonics. */ |
| 17996 | break; |
| 17997 | |
| 17998 | case OT_cinfix3: |
| 17999 | case OT_cinfix3_deprecated: |
| 18000 | case OT_odd_infix_unc: |
| 18001 | if (!unified_syntax) |
| 18002 | return 0; |
| 18003 | /* Fall through. */ |
| 18004 | |
| 18005 | case OT_csuffix: |
| 18006 | case OT_csuffixF: |
| 18007 | case OT_csuf_or_in3: |
| 18008 | inst.cond = cond->value; |
| 18009 | return opcode; |
| 18010 | |
| 18011 | case OT_unconditional: |
| 18012 | case OT_unconditionalF: |
| 18013 | if (thumb_mode) |
| 18014 | inst.cond = cond->value; |
| 18015 | else |
| 18016 | { |
| 18017 | /* Delayed diagnostic. */ |
| 18018 | inst.error = BAD_COND; |
| 18019 | inst.cond = COND_ALWAYS; |
| 18020 | } |
| 18021 | return opcode; |
| 18022 | |
| 18023 | default: |
| 18024 | return NULL; |
| 18025 | } |
| 18026 | } |
| 18027 | |
| 18028 | /* Cannot have a usual-position infix on a mnemonic of less than |
| 18029 | six characters (five would be a suffix). */ |
| 18030 | if (end - base < 6) |
| 18031 | return NULL; |
| 18032 | |
| 18033 | /* Look for infixed mnemonic in the usual position. */ |
| 18034 | affix = base + 3; |
| 18035 | cond = (const struct asm_cond *) hash_find_n (arm_cond_hsh, affix, 2); |
| 18036 | if (!cond) |
| 18037 | return NULL; |
| 18038 | |
| 18039 | memcpy (save, affix, 2); |
| 18040 | memmove (affix, affix + 2, (end - affix) - 2); |
| 18041 | opcode = (const struct asm_opcode *) hash_find_n (arm_ops_hsh, base, |
| 18042 | (end - base) - 2); |
| 18043 | memmove (affix + 2, affix, (end - affix) - 2); |
| 18044 | memcpy (affix, save, 2); |
| 18045 | |
| 18046 | if (opcode |
| 18047 | && (opcode->tag == OT_cinfix3 |
| 18048 | || opcode->tag == OT_cinfix3_deprecated |
| 18049 | || opcode->tag == OT_csuf_or_in3 |
| 18050 | || opcode->tag == OT_cinfix3_legacy)) |
| 18051 | { |
| 18052 | /* Step CM. */ |
| 18053 | if (warn_on_deprecated && unified_syntax |
| 18054 | && (opcode->tag == OT_cinfix3 |
| 18055 | || opcode->tag == OT_cinfix3_deprecated)) |
| 18056 | as_tsktsk (_("conditional infixes are deprecated in unified syntax")); |
| 18057 | |
| 18058 | inst.cond = cond->value; |
| 18059 | return opcode; |
| 18060 | } |
| 18061 | |
| 18062 | return NULL; |
| 18063 | } |
| 18064 | |
| 18065 | /* This function generates an initial IT instruction, leaving its block |
| 18066 | virtually open for the new instructions. Eventually, |
| 18067 | the mask will be updated by now_it_add_mask () each time |
| 18068 | a new instruction needs to be included in the IT block. |
| 18069 | Finally, the block is closed with close_automatic_it_block (). |
| 18070 | The block closure can be requested either from md_assemble (), |
| 18071 | a tencode (), or due to a label hook. */ |
| 18072 | |
| 18073 | static void |
| 18074 | new_automatic_it_block (int cond) |
| 18075 | { |
| 18076 | now_it.state = AUTOMATIC_IT_BLOCK; |
| 18077 | now_it.mask = 0x18; |
| 18078 | now_it.cc = cond; |
| 18079 | now_it.block_length = 1; |
| 18080 | mapping_state (MAP_THUMB); |
| 18081 | now_it.insn = output_it_inst (cond, now_it.mask, NULL); |
| 18082 | now_it.warn_deprecated = FALSE; |
| 18083 | now_it.insn_cond = TRUE; |
| 18084 | } |
| 18085 | |
| 18086 | /* Close an automatic IT block. |
| 18087 | See comments in new_automatic_it_block (). */ |
| 18088 | |
| 18089 | static void |
| 18090 | close_automatic_it_block (void) |
| 18091 | { |
| 18092 | now_it.mask = 0x10; |
| 18093 | now_it.block_length = 0; |
| 18094 | } |
| 18095 | |
| 18096 | /* Update the mask of the current automatically-generated IT |
| 18097 | instruction. See comments in new_automatic_it_block (). */ |
| 18098 | |
| 18099 | static void |
| 18100 | now_it_add_mask (int cond) |
| 18101 | { |
| 18102 | #define CLEAR_BIT(value, nbit) ((value) & ~(1 << (nbit))) |
| 18103 | #define SET_BIT_VALUE(value, bitvalue, nbit) (CLEAR_BIT (value, nbit) \ |
| 18104 | | ((bitvalue) << (nbit))) |
| 18105 | const int resulting_bit = (cond & 1); |
| 18106 | |
| 18107 | now_it.mask &= 0xf; |
| 18108 | now_it.mask = SET_BIT_VALUE (now_it.mask, |
| 18109 | resulting_bit, |
| 18110 | (5 - now_it.block_length)); |
| 18111 | now_it.mask = SET_BIT_VALUE (now_it.mask, |
| 18112 | 1, |
| 18113 | ((5 - now_it.block_length) - 1) ); |
| 18114 | output_it_inst (now_it.cc, now_it.mask, now_it.insn); |
| 18115 | |
| 18116 | #undef CLEAR_BIT |
| 18117 | #undef SET_BIT_VALUE |
| 18118 | } |
| 18119 | |
| 18120 | /* The IT blocks handling machinery is accessed through the these functions: |
| 18121 | it_fsm_pre_encode () from md_assemble () |
| 18122 | set_it_insn_type () optional, from the tencode functions |
| 18123 | set_it_insn_type_last () ditto |
| 18124 | in_it_block () ditto |
| 18125 | it_fsm_post_encode () from md_assemble () |
| 18126 | force_automatic_it_block_close () from label handling functions |
| 18127 | |
| 18128 | Rationale: |
| 18129 | 1) md_assemble () calls it_fsm_pre_encode () before calling tencode (), |
| 18130 | initializing the IT insn type with a generic initial value depending |
| 18131 | on the inst.condition. |
| 18132 | 2) During the tencode function, two things may happen: |
| 18133 | a) The tencode function overrides the IT insn type by |
| 18134 | calling either set_it_insn_type (type) or set_it_insn_type_last (). |
| 18135 | b) The tencode function queries the IT block state by |
| 18136 | calling in_it_block () (i.e. to determine narrow/not narrow mode). |
| 18137 | |
| 18138 | Both set_it_insn_type and in_it_block run the internal FSM state |
| 18139 | handling function (handle_it_state), because: a) setting the IT insn |
| 18140 | type may incur in an invalid state (exiting the function), |
| 18141 | and b) querying the state requires the FSM to be updated. |
| 18142 | Specifically we want to avoid creating an IT block for conditional |
| 18143 | branches, so it_fsm_pre_encode is actually a guess and we can't |
| 18144 | determine whether an IT block is required until the tencode () routine |
| 18145 | has decided what type of instruction this actually it. |
| 18146 | Because of this, if set_it_insn_type and in_it_block have to be used, |
| 18147 | set_it_insn_type has to be called first. |
| 18148 | |
| 18149 | set_it_insn_type_last () is a wrapper of set_it_insn_type (type), that |
| 18150 | determines the insn IT type depending on the inst.cond code. |
| 18151 | When a tencode () routine encodes an instruction that can be |
| 18152 | either outside an IT block, or, in the case of being inside, has to be |
| 18153 | the last one, set_it_insn_type_last () will determine the proper |
| 18154 | IT instruction type based on the inst.cond code. Otherwise, |
| 18155 | set_it_insn_type can be called for overriding that logic or |
| 18156 | for covering other cases. |
| 18157 | |
| 18158 | Calling handle_it_state () may not transition the IT block state to |
| 18159 | OUTSIDE_IT_BLOCK immediately, since the (current) state could be |
| 18160 | still queried. Instead, if the FSM determines that the state should |
| 18161 | be transitioned to OUTSIDE_IT_BLOCK, a flag is marked to be closed |
| 18162 | after the tencode () function: that's what it_fsm_post_encode () does. |
| 18163 | |
| 18164 | Since in_it_block () calls the state handling function to get an |
| 18165 | updated state, an error may occur (due to invalid insns combination). |
| 18166 | In that case, inst.error is set. |
| 18167 | Therefore, inst.error has to be checked after the execution of |
| 18168 | the tencode () routine. |
| 18169 | |
| 18170 | 3) Back in md_assemble(), it_fsm_post_encode () is called to commit |
| 18171 | any pending state change (if any) that didn't take place in |
| 18172 | handle_it_state () as explained above. */ |
| 18173 | |
| 18174 | static void |
| 18175 | it_fsm_pre_encode (void) |
| 18176 | { |
| 18177 | if (inst.cond != COND_ALWAYS) |
| 18178 | inst.it_insn_type = INSIDE_IT_INSN; |
| 18179 | else |
| 18180 | inst.it_insn_type = OUTSIDE_IT_INSN; |
| 18181 | |
| 18182 | now_it.state_handled = 0; |
| 18183 | } |
| 18184 | |
| 18185 | /* IT state FSM handling function. */ |
| 18186 | |
| 18187 | static int |
| 18188 | handle_it_state (void) |
| 18189 | { |
| 18190 | now_it.state_handled = 1; |
| 18191 | now_it.insn_cond = FALSE; |
| 18192 | |
| 18193 | switch (now_it.state) |
| 18194 | { |
| 18195 | case OUTSIDE_IT_BLOCK: |
| 18196 | switch (inst.it_insn_type) |
| 18197 | { |
| 18198 | case OUTSIDE_IT_INSN: |
| 18199 | break; |
| 18200 | |
| 18201 | case INSIDE_IT_INSN: |
| 18202 | case INSIDE_IT_LAST_INSN: |
| 18203 | if (thumb_mode == 0) |
| 18204 | { |
| 18205 | if (unified_syntax |
| 18206 | && !(implicit_it_mode & IMPLICIT_IT_MODE_ARM)) |
| 18207 | as_tsktsk (_("Warning: conditional outside an IT block"\ |
| 18208 | " for Thumb.")); |
| 18209 | } |
| 18210 | else |
| 18211 | { |
| 18212 | if ((implicit_it_mode & IMPLICIT_IT_MODE_THUMB) |
| 18213 | && ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v6t2)) |
| 18214 | { |
| 18215 | /* Automatically generate the IT instruction. */ |
| 18216 | new_automatic_it_block (inst.cond); |
| 18217 | if (inst.it_insn_type == INSIDE_IT_LAST_INSN) |
| 18218 | close_automatic_it_block (); |
| 18219 | } |
| 18220 | else |
| 18221 | { |
| 18222 | inst.error = BAD_OUT_IT; |
| 18223 | return FAIL; |
| 18224 | } |
| 18225 | } |
| 18226 | break; |
| 18227 | |
| 18228 | case IF_INSIDE_IT_LAST_INSN: |
| 18229 | case NEUTRAL_IT_INSN: |
| 18230 | break; |
| 18231 | |
| 18232 | case IT_INSN: |
| 18233 | now_it.state = MANUAL_IT_BLOCK; |
| 18234 | now_it.block_length = 0; |
| 18235 | break; |
| 18236 | } |
| 18237 | break; |
| 18238 | |
| 18239 | case AUTOMATIC_IT_BLOCK: |
| 18240 | /* Three things may happen now: |
| 18241 | a) We should increment current it block size; |
| 18242 | b) We should close current it block (closing insn or 4 insns); |
| 18243 | c) We should close current it block and start a new one (due |
| 18244 | to incompatible conditions or |
| 18245 | 4 insns-length block reached). */ |
| 18246 | |
| 18247 | switch (inst.it_insn_type) |
| 18248 | { |
| 18249 | case OUTSIDE_IT_INSN: |
| 18250 | /* The closure of the block shall happen immediately, |
| 18251 | so any in_it_block () call reports the block as closed. */ |
| 18252 | force_automatic_it_block_close (); |
| 18253 | break; |
| 18254 | |
| 18255 | case INSIDE_IT_INSN: |
| 18256 | case INSIDE_IT_LAST_INSN: |
| 18257 | case IF_INSIDE_IT_LAST_INSN: |
| 18258 | now_it.block_length++; |
| 18259 | |
| 18260 | if (now_it.block_length > 4 |
| 18261 | || !now_it_compatible (inst.cond)) |
| 18262 | { |
| 18263 | force_automatic_it_block_close (); |
| 18264 | if (inst.it_insn_type != IF_INSIDE_IT_LAST_INSN) |
| 18265 | new_automatic_it_block (inst.cond); |
| 18266 | } |
| 18267 | else |
| 18268 | { |
| 18269 | now_it.insn_cond = TRUE; |
| 18270 | now_it_add_mask (inst.cond); |
| 18271 | } |
| 18272 | |
| 18273 | if (now_it.state == AUTOMATIC_IT_BLOCK |
| 18274 | && (inst.it_insn_type == INSIDE_IT_LAST_INSN |
| 18275 | || inst.it_insn_type == IF_INSIDE_IT_LAST_INSN)) |
| 18276 | close_automatic_it_block (); |
| 18277 | break; |
| 18278 | |
| 18279 | case NEUTRAL_IT_INSN: |
| 18280 | now_it.block_length++; |
| 18281 | now_it.insn_cond = TRUE; |
| 18282 | |
| 18283 | if (now_it.block_length > 4) |
| 18284 | force_automatic_it_block_close (); |
| 18285 | else |
| 18286 | now_it_add_mask (now_it.cc & 1); |
| 18287 | break; |
| 18288 | |
| 18289 | case IT_INSN: |
| 18290 | close_automatic_it_block (); |
| 18291 | now_it.state = MANUAL_IT_BLOCK; |
| 18292 | break; |
| 18293 | } |
| 18294 | break; |
| 18295 | |
| 18296 | case MANUAL_IT_BLOCK: |
| 18297 | { |
| 18298 | /* Check conditional suffixes. */ |
| 18299 | const int cond = now_it.cc ^ ((now_it.mask >> 4) & 1) ^ 1; |
| 18300 | int is_last; |
| 18301 | now_it.mask <<= 1; |
| 18302 | now_it.mask &= 0x1f; |
| 18303 | is_last = (now_it.mask == 0x10); |
| 18304 | now_it.insn_cond = TRUE; |
| 18305 | |
| 18306 | switch (inst.it_insn_type) |
| 18307 | { |
| 18308 | case OUTSIDE_IT_INSN: |
| 18309 | inst.error = BAD_NOT_IT; |
| 18310 | return FAIL; |
| 18311 | |
| 18312 | case INSIDE_IT_INSN: |
| 18313 | if (cond != inst.cond) |
| 18314 | { |
| 18315 | inst.error = BAD_IT_COND; |
| 18316 | return FAIL; |
| 18317 | } |
| 18318 | break; |
| 18319 | |
| 18320 | case INSIDE_IT_LAST_INSN: |
| 18321 | case IF_INSIDE_IT_LAST_INSN: |
| 18322 | if (cond != inst.cond) |
| 18323 | { |
| 18324 | inst.error = BAD_IT_COND; |
| 18325 | return FAIL; |
| 18326 | } |
| 18327 | if (!is_last) |
| 18328 | { |
| 18329 | inst.error = BAD_BRANCH; |
| 18330 | return FAIL; |
| 18331 | } |
| 18332 | break; |
| 18333 | |
| 18334 | case NEUTRAL_IT_INSN: |
| 18335 | /* The BKPT instruction is unconditional even in an IT block. */ |
| 18336 | break; |
| 18337 | |
| 18338 | case IT_INSN: |
| 18339 | inst.error = BAD_IT_IT; |
| 18340 | return FAIL; |
| 18341 | } |
| 18342 | } |
| 18343 | break; |
| 18344 | } |
| 18345 | |
| 18346 | return SUCCESS; |
| 18347 | } |
| 18348 | |
| 18349 | struct depr_insn_mask |
| 18350 | { |
| 18351 | unsigned long pattern; |
| 18352 | unsigned long mask; |
| 18353 | const char* description; |
| 18354 | }; |
| 18355 | |
| 18356 | /* List of 16-bit instruction patterns deprecated in an IT block in |
| 18357 | ARMv8. */ |
| 18358 | static const struct depr_insn_mask depr_it_insns[] = { |
| 18359 | { 0xc000, 0xc000, N_("Short branches, Undefined, SVC, LDM/STM") }, |
| 18360 | { 0xb000, 0xb000, N_("Miscellaneous 16-bit instructions") }, |
| 18361 | { 0xa000, 0xb800, N_("ADR") }, |
| 18362 | { 0x4800, 0xf800, N_("Literal loads") }, |
| 18363 | { 0x4478, 0xf478, N_("Hi-register ADD, MOV, CMP, BX, BLX using pc") }, |
| 18364 | { 0x4487, 0xfc87, N_("Hi-register ADD, MOV, CMP using pc") }, |
| 18365 | /* NOTE: 0x00dd is not the real encoding, instead, it is the 'tvalue' |
| 18366 | field in asm_opcode. 'tvalue' is used at the stage this check happen. */ |
| 18367 | { 0x00dd, 0x7fff, N_("ADD/SUB sp, sp #imm") }, |
| 18368 | { 0, 0, NULL } |
| 18369 | }; |
| 18370 | |
| 18371 | static void |
| 18372 | it_fsm_post_encode (void) |
| 18373 | { |
| 18374 | int is_last; |
| 18375 | |
| 18376 | if (!now_it.state_handled) |
| 18377 | handle_it_state (); |
| 18378 | |
| 18379 | if (now_it.insn_cond |
| 18380 | && !now_it.warn_deprecated |
| 18381 | && warn_on_deprecated |
| 18382 | && ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v8)) |
| 18383 | { |
| 18384 | if (inst.instruction >= 0x10000) |
| 18385 | { |
| 18386 | as_tsktsk (_("IT blocks containing 32-bit Thumb instructions are " |
| 18387 | "deprecated in ARMv8")); |
| 18388 | now_it.warn_deprecated = TRUE; |
| 18389 | } |
| 18390 | else |
| 18391 | { |
| 18392 | const struct depr_insn_mask *p = depr_it_insns; |
| 18393 | |
| 18394 | while (p->mask != 0) |
| 18395 | { |
| 18396 | if ((inst.instruction & p->mask) == p->pattern) |
| 18397 | { |
| 18398 | as_tsktsk (_("IT blocks containing 16-bit Thumb instructions " |
| 18399 | "of the following class are deprecated in ARMv8: " |
| 18400 | "%s"), p->description); |
| 18401 | now_it.warn_deprecated = TRUE; |
| 18402 | break; |
| 18403 | } |
| 18404 | |
| 18405 | ++p; |
| 18406 | } |
| 18407 | } |
| 18408 | |
| 18409 | if (now_it.block_length > 1) |
| 18410 | { |
| 18411 | as_tsktsk (_("IT blocks containing more than one conditional " |
| 18412 | "instruction are deprecated in ARMv8")); |
| 18413 | now_it.warn_deprecated = TRUE; |
| 18414 | } |
| 18415 | } |
| 18416 | |
| 18417 | is_last = (now_it.mask == 0x10); |
| 18418 | if (is_last) |
| 18419 | { |
| 18420 | now_it.state = OUTSIDE_IT_BLOCK; |
| 18421 | now_it.mask = 0; |
| 18422 | } |
| 18423 | } |
| 18424 | |
| 18425 | static void |
| 18426 | force_automatic_it_block_close (void) |
| 18427 | { |
| 18428 | if (now_it.state == AUTOMATIC_IT_BLOCK) |
| 18429 | { |
| 18430 | close_automatic_it_block (); |
| 18431 | now_it.state = OUTSIDE_IT_BLOCK; |
| 18432 | now_it.mask = 0; |
| 18433 | } |
| 18434 | } |
| 18435 | |
| 18436 | static int |
| 18437 | in_it_block (void) |
| 18438 | { |
| 18439 | if (!now_it.state_handled) |
| 18440 | handle_it_state (); |
| 18441 | |
| 18442 | return now_it.state != OUTSIDE_IT_BLOCK; |
| 18443 | } |
| 18444 | |
| 18445 | /* Whether OPCODE only has T32 encoding. Since this function is only used by |
| 18446 | t32_insn_ok, OPCODE enabled by v6t2 extension bit do not need to be listed |
| 18447 | here, hence the "known" in the function name. */ |
| 18448 | |
| 18449 | static bfd_boolean |
| 18450 | known_t32_only_insn (const struct asm_opcode *opcode) |
| 18451 | { |
| 18452 | /* Original Thumb-1 wide instruction. */ |
| 18453 | if (opcode->tencode == do_t_blx |
| 18454 | || opcode->tencode == do_t_branch23 |
| 18455 | || ARM_CPU_HAS_FEATURE (*opcode->tvariant, arm_ext_msr) |
| 18456 | || ARM_CPU_HAS_FEATURE (*opcode->tvariant, arm_ext_barrier)) |
| 18457 | return TRUE; |
| 18458 | |
| 18459 | /* Wide-only instruction added to ARMv8-M Baseline. */ |
| 18460 | if (ARM_CPU_HAS_FEATURE (*opcode->tvariant, arm_ext_v8m_m_only) |
| 18461 | || ARM_CPU_HAS_FEATURE (*opcode->tvariant, arm_ext_atomics) |
| 18462 | || ARM_CPU_HAS_FEATURE (*opcode->tvariant, arm_ext_v6t2_v8m) |
| 18463 | || ARM_CPU_HAS_FEATURE (*opcode->tvariant, arm_ext_div)) |
| 18464 | return TRUE; |
| 18465 | |
| 18466 | return FALSE; |
| 18467 | } |
| 18468 | |
| 18469 | /* Whether wide instruction variant can be used if available for a valid OPCODE |
| 18470 | in ARCH. */ |
| 18471 | |
| 18472 | static bfd_boolean |
| 18473 | t32_insn_ok (arm_feature_set arch, const struct asm_opcode *opcode) |
| 18474 | { |
| 18475 | if (known_t32_only_insn (opcode)) |
| 18476 | return TRUE; |
| 18477 | |
| 18478 | /* Instruction with narrow and wide encoding added to ARMv8-M. Availability |
| 18479 | of variant T3 of B.W is checked in do_t_branch. */ |
| 18480 | if (ARM_CPU_HAS_FEATURE (arch, arm_ext_v8m) |
| 18481 | && opcode->tencode == do_t_branch) |
| 18482 | return TRUE; |
| 18483 | |
| 18484 | /* MOV accepts T1/T3 encodings under Baseline, T3 encoding is 32bit. */ |
| 18485 | if (ARM_CPU_HAS_FEATURE (arch, arm_ext_v8m) |
| 18486 | && opcode->tencode == do_t_mov_cmp |
| 18487 | /* Make sure CMP instruction is not affected. */ |
| 18488 | && opcode->aencode == do_mov) |
| 18489 | return TRUE; |
| 18490 | |
| 18491 | /* Wide instruction variants of all instructions with narrow *and* wide |
| 18492 | variants become available with ARMv6t2. Other opcodes are either |
| 18493 | narrow-only or wide-only and are thus available if OPCODE is valid. */ |
| 18494 | if (ARM_CPU_HAS_FEATURE (arch, arm_ext_v6t2)) |
| 18495 | return TRUE; |
| 18496 | |
| 18497 | /* OPCODE with narrow only instruction variant or wide variant not |
| 18498 | available. */ |
| 18499 | return FALSE; |
| 18500 | } |
| 18501 | |
| 18502 | void |
| 18503 | md_assemble (char *str) |
| 18504 | { |
| 18505 | char *p = str; |
| 18506 | const struct asm_opcode * opcode; |
| 18507 | |
| 18508 | /* Align the previous label if needed. */ |
| 18509 | if (last_label_seen != NULL) |
| 18510 | { |
| 18511 | symbol_set_frag (last_label_seen, frag_now); |
| 18512 | S_SET_VALUE (last_label_seen, (valueT) frag_now_fix ()); |
| 18513 | S_SET_SEGMENT (last_label_seen, now_seg); |
| 18514 | } |
| 18515 | |
| 18516 | memset (&inst, '\0', sizeof (inst)); |
| 18517 | inst.reloc.type = BFD_RELOC_UNUSED; |
| 18518 | |
| 18519 | opcode = opcode_lookup (&p); |
| 18520 | if (!opcode) |
| 18521 | { |
| 18522 | /* It wasn't an instruction, but it might be a register alias of |
| 18523 | the form alias .req reg, or a Neon .dn/.qn directive. */ |
| 18524 | if (! create_register_alias (str, p) |
| 18525 | && ! create_neon_reg_alias (str, p)) |
| 18526 | as_bad (_("bad instruction `%s'"), str); |
| 18527 | |
| 18528 | return; |
| 18529 | } |
| 18530 | |
| 18531 | if (warn_on_deprecated && opcode->tag == OT_cinfix3_deprecated) |
| 18532 | as_tsktsk (_("s suffix on comparison instruction is deprecated")); |
| 18533 | |
| 18534 | /* The value which unconditional instructions should have in place of the |
| 18535 | condition field. */ |
| 18536 | inst.uncond_value = (opcode->tag == OT_csuffixF) ? 0xf : -1; |
| 18537 | |
| 18538 | if (thumb_mode) |
| 18539 | { |
| 18540 | arm_feature_set variant; |
| 18541 | |
| 18542 | variant = cpu_variant; |
| 18543 | /* Only allow coprocessor instructions on Thumb-2 capable devices. */ |
| 18544 | if (!ARM_CPU_HAS_FEATURE (variant, arm_arch_t2)) |
| 18545 | ARM_CLEAR_FEATURE (variant, variant, fpu_any_hard); |
| 18546 | /* Check that this instruction is supported for this CPU. */ |
| 18547 | if (!opcode->tvariant |
| 18548 | || (thumb_mode == 1 |
| 18549 | && !ARM_CPU_HAS_FEATURE (variant, *opcode->tvariant))) |
| 18550 | { |
| 18551 | if (opcode->tencode == do_t_swi) |
| 18552 | as_bad (_("SVC is not permitted on this architecture")); |
| 18553 | else |
| 18554 | as_bad (_("selected processor does not support `%s' in Thumb mode"), str); |
| 18555 | return; |
| 18556 | } |
| 18557 | if (inst.cond != COND_ALWAYS && !unified_syntax |
| 18558 | && opcode->tencode != do_t_branch) |
| 18559 | { |
| 18560 | as_bad (_("Thumb does not support conditional execution")); |
| 18561 | return; |
| 18562 | } |
| 18563 | |
| 18564 | /* Two things are addressed here: |
| 18565 | 1) Implicit require narrow instructions on Thumb-1. |
| 18566 | This avoids relaxation accidentally introducing Thumb-2 |
| 18567 | instructions. |
| 18568 | 2) Reject wide instructions in non Thumb-2 cores. |
| 18569 | |
| 18570 | Only instructions with narrow and wide variants need to be handled |
| 18571 | but selecting all non wide-only instructions is easier. */ |
| 18572 | if (!ARM_CPU_HAS_FEATURE (variant, arm_ext_v6t2) |
| 18573 | && !t32_insn_ok (variant, opcode)) |
| 18574 | { |
| 18575 | if (inst.size_req == 0) |
| 18576 | inst.size_req = 2; |
| 18577 | else if (inst.size_req == 4) |
| 18578 | { |
| 18579 | if (ARM_CPU_HAS_FEATURE (variant, arm_ext_v8m)) |
| 18580 | as_bad (_("selected processor does not support 32bit wide " |
| 18581 | "variant of instruction `%s'"), str); |
| 18582 | else |
| 18583 | as_bad (_("selected processor does not support `%s' in " |
| 18584 | "Thumb-2 mode"), str); |
| 18585 | return; |
| 18586 | } |
| 18587 | } |
| 18588 | |
| 18589 | inst.instruction = opcode->tvalue; |
| 18590 | |
| 18591 | if (!parse_operands (p, opcode->operands, /*thumb=*/TRUE)) |
| 18592 | { |
| 18593 | /* Prepare the it_insn_type for those encodings that don't set |
| 18594 | it. */ |
| 18595 | it_fsm_pre_encode (); |
| 18596 | |
| 18597 | opcode->tencode (); |
| 18598 | |
| 18599 | it_fsm_post_encode (); |
| 18600 | } |
| 18601 | |
| 18602 | if (!(inst.error || inst.relax)) |
| 18603 | { |
| 18604 | gas_assert (inst.instruction < 0xe800 || inst.instruction > 0xffff); |
| 18605 | inst.size = (inst.instruction > 0xffff ? 4 : 2); |
| 18606 | if (inst.size_req && inst.size_req != inst.size) |
| 18607 | { |
| 18608 | as_bad (_("cannot honor width suffix -- `%s'"), str); |
| 18609 | return; |
| 18610 | } |
| 18611 | } |
| 18612 | |
| 18613 | /* Something has gone badly wrong if we try to relax a fixed size |
| 18614 | instruction. */ |
| 18615 | gas_assert (inst.size_req == 0 || !inst.relax); |
| 18616 | |
| 18617 | ARM_MERGE_FEATURE_SETS (thumb_arch_used, thumb_arch_used, |
| 18618 | *opcode->tvariant); |
| 18619 | /* Many Thumb-2 instructions also have Thumb-1 variants, so explicitly |
| 18620 | set those bits when Thumb-2 32-bit instructions are seen. The impact |
| 18621 | of relaxable instructions will be considered later after we finish all |
| 18622 | relaxation. */ |
| 18623 | if (ARM_FEATURE_CORE_EQUAL (cpu_variant, arm_arch_any)) |
| 18624 | variant = arm_arch_none; |
| 18625 | else |
| 18626 | variant = cpu_variant; |
| 18627 | if (inst.size == 4 && !t32_insn_ok (variant, opcode)) |
| 18628 | ARM_MERGE_FEATURE_SETS (thumb_arch_used, thumb_arch_used, |
| 18629 | arm_ext_v6t2); |
| 18630 | |
| 18631 | check_neon_suffixes; |
| 18632 | |
| 18633 | if (!inst.error) |
| 18634 | { |
| 18635 | mapping_state (MAP_THUMB); |
| 18636 | } |
| 18637 | } |
| 18638 | else if (ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v1)) |
| 18639 | { |
| 18640 | bfd_boolean is_bx; |
| 18641 | |
| 18642 | /* bx is allowed on v5 cores, and sometimes on v4 cores. */ |
| 18643 | is_bx = (opcode->aencode == do_bx); |
| 18644 | |
| 18645 | /* Check that this instruction is supported for this CPU. */ |
| 18646 | if (!(is_bx && fix_v4bx) |
| 18647 | && !(opcode->avariant && |
| 18648 | ARM_CPU_HAS_FEATURE (cpu_variant, *opcode->avariant))) |
| 18649 | { |
| 18650 | as_bad (_("selected processor does not support `%s' in ARM mode"), str); |
| 18651 | return; |
| 18652 | } |
| 18653 | if (inst.size_req) |
| 18654 | { |
| 18655 | as_bad (_("width suffixes are invalid in ARM mode -- `%s'"), str); |
| 18656 | return; |
| 18657 | } |
| 18658 | |
| 18659 | inst.instruction = opcode->avalue; |
| 18660 | if (opcode->tag == OT_unconditionalF) |
| 18661 | inst.instruction |= 0xFU << 28; |
| 18662 | else |
| 18663 | inst.instruction |= inst.cond << 28; |
| 18664 | inst.size = INSN_SIZE; |
| 18665 | if (!parse_operands (p, opcode->operands, /*thumb=*/FALSE)) |
| 18666 | { |
| 18667 | it_fsm_pre_encode (); |
| 18668 | opcode->aencode (); |
| 18669 | it_fsm_post_encode (); |
| 18670 | } |
| 18671 | /* Arm mode bx is marked as both v4T and v5 because it's still required |
| 18672 | on a hypothetical non-thumb v5 core. */ |
| 18673 | if (is_bx) |
| 18674 | ARM_MERGE_FEATURE_SETS (arm_arch_used, arm_arch_used, arm_ext_v4t); |
| 18675 | else |
| 18676 | ARM_MERGE_FEATURE_SETS (arm_arch_used, arm_arch_used, |
| 18677 | *opcode->avariant); |
| 18678 | |
| 18679 | check_neon_suffixes; |
| 18680 | |
| 18681 | if (!inst.error) |
| 18682 | { |
| 18683 | mapping_state (MAP_ARM); |
| 18684 | } |
| 18685 | } |
| 18686 | else |
| 18687 | { |
| 18688 | as_bad (_("attempt to use an ARM instruction on a Thumb-only processor " |
| 18689 | "-- `%s'"), str); |
| 18690 | return; |
| 18691 | } |
| 18692 | output_inst (str); |
| 18693 | } |
| 18694 | |
| 18695 | static void |
| 18696 | check_it_blocks_finished (void) |
| 18697 | { |
| 18698 | #ifdef OBJ_ELF |
| 18699 | asection *sect; |
| 18700 | |
| 18701 | for (sect = stdoutput->sections; sect != NULL; sect = sect->next) |
| 18702 | if (seg_info (sect)->tc_segment_info_data.current_it.state |
| 18703 | == MANUAL_IT_BLOCK) |
| 18704 | { |
| 18705 | as_warn (_("section '%s' finished with an open IT block."), |
| 18706 | sect->name); |
| 18707 | } |
| 18708 | #else |
| 18709 | if (now_it.state == MANUAL_IT_BLOCK) |
| 18710 | as_warn (_("file finished with an open IT block.")); |
| 18711 | #endif |
| 18712 | } |
| 18713 | |
| 18714 | /* Various frobbings of labels and their addresses. */ |
| 18715 | |
| 18716 | void |
| 18717 | arm_start_line_hook (void) |
| 18718 | { |
| 18719 | last_label_seen = NULL; |
| 18720 | } |
| 18721 | |
| 18722 | void |
| 18723 | arm_frob_label (symbolS * sym) |
| 18724 | { |
| 18725 | last_label_seen = sym; |
| 18726 | |
| 18727 | ARM_SET_THUMB (sym, thumb_mode); |
| 18728 | |
| 18729 | #if defined OBJ_COFF || defined OBJ_ELF |
| 18730 | ARM_SET_INTERWORK (sym, support_interwork); |
| 18731 | #endif |
| 18732 | |
| 18733 | force_automatic_it_block_close (); |
| 18734 | |
| 18735 | /* Note - do not allow local symbols (.Lxxx) to be labelled |
| 18736 | as Thumb functions. This is because these labels, whilst |
| 18737 | they exist inside Thumb code, are not the entry points for |
| 18738 | possible ARM->Thumb calls. Also, these labels can be used |
| 18739 | as part of a computed goto or switch statement. eg gcc |
| 18740 | can generate code that looks like this: |
| 18741 | |
| 18742 | ldr r2, [pc, .Laaa] |
| 18743 | lsl r3, r3, #2 |
| 18744 | ldr r2, [r3, r2] |
| 18745 | mov pc, r2 |
| 18746 | |
| 18747 | .Lbbb: .word .Lxxx |
| 18748 | .Lccc: .word .Lyyy |
| 18749 | ..etc... |
| 18750 | .Laaa: .word Lbbb |
| 18751 | |
| 18752 | The first instruction loads the address of the jump table. |
| 18753 | The second instruction converts a table index into a byte offset. |
| 18754 | The third instruction gets the jump address out of the table. |
| 18755 | The fourth instruction performs the jump. |
| 18756 | |
| 18757 | If the address stored at .Laaa is that of a symbol which has the |
| 18758 | Thumb_Func bit set, then the linker will arrange for this address |
| 18759 | to have the bottom bit set, which in turn would mean that the |
| 18760 | address computation performed by the third instruction would end |
| 18761 | up with the bottom bit set. Since the ARM is capable of unaligned |
| 18762 | word loads, the instruction would then load the incorrect address |
| 18763 | out of the jump table, and chaos would ensue. */ |
| 18764 | if (label_is_thumb_function_name |
| 18765 | && (S_GET_NAME (sym)[0] != '.' || S_GET_NAME (sym)[1] != 'L') |
| 18766 | && (bfd_get_section_flags (stdoutput, now_seg) & SEC_CODE) != 0) |
| 18767 | { |
| 18768 | /* When the address of a Thumb function is taken the bottom |
| 18769 | bit of that address should be set. This will allow |
| 18770 | interworking between Arm and Thumb functions to work |
| 18771 | correctly. */ |
| 18772 | |
| 18773 | THUMB_SET_FUNC (sym, 1); |
| 18774 | |
| 18775 | label_is_thumb_function_name = FALSE; |
| 18776 | } |
| 18777 | |
| 18778 | dwarf2_emit_label (sym); |
| 18779 | } |
| 18780 | |
| 18781 | bfd_boolean |
| 18782 | arm_data_in_code (void) |
| 18783 | { |
| 18784 | if (thumb_mode && ! strncmp (input_line_pointer + 1, "data:", 5)) |
| 18785 | { |
| 18786 | *input_line_pointer = '/'; |
| 18787 | input_line_pointer += 5; |
| 18788 | *input_line_pointer = 0; |
| 18789 | return TRUE; |
| 18790 | } |
| 18791 | |
| 18792 | return FALSE; |
| 18793 | } |
| 18794 | |
| 18795 | char * |
| 18796 | arm_canonicalize_symbol_name (char * name) |
| 18797 | { |
| 18798 | int len; |
| 18799 | |
| 18800 | if (thumb_mode && (len = strlen (name)) > 5 |
| 18801 | && streq (name + len - 5, "/data")) |
| 18802 | *(name + len - 5) = 0; |
| 18803 | |
| 18804 | return name; |
| 18805 | } |
| 18806 | \f |
| 18807 | /* Table of all register names defined by default. The user can |
| 18808 | define additional names with .req. Note that all register names |
| 18809 | should appear in both upper and lowercase variants. Some registers |
| 18810 | also have mixed-case names. */ |
| 18811 | |
| 18812 | #define REGDEF(s,n,t) { #s, n, REG_TYPE_##t, TRUE, 0 } |
| 18813 | #define REGNUM(p,n,t) REGDEF(p##n, n, t) |
| 18814 | #define REGNUM2(p,n,t) REGDEF(p##n, 2 * n, t) |
| 18815 | #define REGSET(p,t) \ |
| 18816 | REGNUM(p, 0,t), REGNUM(p, 1,t), REGNUM(p, 2,t), REGNUM(p, 3,t), \ |
| 18817 | REGNUM(p, 4,t), REGNUM(p, 5,t), REGNUM(p, 6,t), REGNUM(p, 7,t), \ |
| 18818 | REGNUM(p, 8,t), REGNUM(p, 9,t), REGNUM(p,10,t), REGNUM(p,11,t), \ |
| 18819 | REGNUM(p,12,t), REGNUM(p,13,t), REGNUM(p,14,t), REGNUM(p,15,t) |
| 18820 | #define REGSETH(p,t) \ |
| 18821 | REGNUM(p,16,t), REGNUM(p,17,t), REGNUM(p,18,t), REGNUM(p,19,t), \ |
| 18822 | REGNUM(p,20,t), REGNUM(p,21,t), REGNUM(p,22,t), REGNUM(p,23,t), \ |
| 18823 | REGNUM(p,24,t), REGNUM(p,25,t), REGNUM(p,26,t), REGNUM(p,27,t), \ |
| 18824 | REGNUM(p,28,t), REGNUM(p,29,t), REGNUM(p,30,t), REGNUM(p,31,t) |
| 18825 | #define REGSET2(p,t) \ |
| 18826 | REGNUM2(p, 0,t), REGNUM2(p, 1,t), REGNUM2(p, 2,t), REGNUM2(p, 3,t), \ |
| 18827 | REGNUM2(p, 4,t), REGNUM2(p, 5,t), REGNUM2(p, 6,t), REGNUM2(p, 7,t), \ |
| 18828 | REGNUM2(p, 8,t), REGNUM2(p, 9,t), REGNUM2(p,10,t), REGNUM2(p,11,t), \ |
| 18829 | REGNUM2(p,12,t), REGNUM2(p,13,t), REGNUM2(p,14,t), REGNUM2(p,15,t) |
| 18830 | #define SPLRBANK(base,bank,t) \ |
| 18831 | REGDEF(lr_##bank, 768|((base+0)<<16), t), \ |
| 18832 | REGDEF(sp_##bank, 768|((base+1)<<16), t), \ |
| 18833 | REGDEF(spsr_##bank, 768|(base<<16)|SPSR_BIT, t), \ |
| 18834 | REGDEF(LR_##bank, 768|((base+0)<<16), t), \ |
| 18835 | REGDEF(SP_##bank, 768|((base+1)<<16), t), \ |
| 18836 | REGDEF(SPSR_##bank, 768|(base<<16)|SPSR_BIT, t) |
| 18837 | |
| 18838 | static const struct reg_entry reg_names[] = |
| 18839 | { |
| 18840 | /* ARM integer registers. */ |
| 18841 | REGSET(r, RN), REGSET(R, RN), |
| 18842 | |
| 18843 | /* ATPCS synonyms. */ |
| 18844 | REGDEF(a1,0,RN), REGDEF(a2,1,RN), REGDEF(a3, 2,RN), REGDEF(a4, 3,RN), |
| 18845 | REGDEF(v1,4,RN), REGDEF(v2,5,RN), REGDEF(v3, 6,RN), REGDEF(v4, 7,RN), |
| 18846 | REGDEF(v5,8,RN), REGDEF(v6,9,RN), REGDEF(v7,10,RN), REGDEF(v8,11,RN), |
| 18847 | |
| 18848 | REGDEF(A1,0,RN), REGDEF(A2,1,RN), REGDEF(A3, 2,RN), REGDEF(A4, 3,RN), |
| 18849 | REGDEF(V1,4,RN), REGDEF(V2,5,RN), REGDEF(V3, 6,RN), REGDEF(V4, 7,RN), |
| 18850 | REGDEF(V5,8,RN), REGDEF(V6,9,RN), REGDEF(V7,10,RN), REGDEF(V8,11,RN), |
| 18851 | |
| 18852 | /* Well-known aliases. */ |
| 18853 | REGDEF(wr, 7,RN), REGDEF(sb, 9,RN), REGDEF(sl,10,RN), REGDEF(fp,11,RN), |
| 18854 | REGDEF(ip,12,RN), REGDEF(sp,13,RN), REGDEF(lr,14,RN), REGDEF(pc,15,RN), |
| 18855 | |
| 18856 | REGDEF(WR, 7,RN), REGDEF(SB, 9,RN), REGDEF(SL,10,RN), REGDEF(FP,11,RN), |
| 18857 | REGDEF(IP,12,RN), REGDEF(SP,13,RN), REGDEF(LR,14,RN), REGDEF(PC,15,RN), |
| 18858 | |
| 18859 | /* Coprocessor numbers. */ |
| 18860 | REGSET(p, CP), REGSET(P, CP), |
| 18861 | |
| 18862 | /* Coprocessor register numbers. The "cr" variants are for backward |
| 18863 | compatibility. */ |
| 18864 | REGSET(c, CN), REGSET(C, CN), |
| 18865 | REGSET(cr, CN), REGSET(CR, CN), |
| 18866 | |
| 18867 | /* ARM banked registers. */ |
| 18868 | REGDEF(R8_usr,512|(0<<16),RNB), REGDEF(r8_usr,512|(0<<16),RNB), |
| 18869 | REGDEF(R9_usr,512|(1<<16),RNB), REGDEF(r9_usr,512|(1<<16),RNB), |
| 18870 | REGDEF(R10_usr,512|(2<<16),RNB), REGDEF(r10_usr,512|(2<<16),RNB), |
| 18871 | REGDEF(R11_usr,512|(3<<16),RNB), REGDEF(r11_usr,512|(3<<16),RNB), |
| 18872 | REGDEF(R12_usr,512|(4<<16),RNB), REGDEF(r12_usr,512|(4<<16),RNB), |
| 18873 | REGDEF(SP_usr,512|(5<<16),RNB), REGDEF(sp_usr,512|(5<<16),RNB), |
| 18874 | REGDEF(LR_usr,512|(6<<16),RNB), REGDEF(lr_usr,512|(6<<16),RNB), |
| 18875 | |
| 18876 | REGDEF(R8_fiq,512|(8<<16),RNB), REGDEF(r8_fiq,512|(8<<16),RNB), |
| 18877 | REGDEF(R9_fiq,512|(9<<16),RNB), REGDEF(r9_fiq,512|(9<<16),RNB), |
| 18878 | REGDEF(R10_fiq,512|(10<<16),RNB), REGDEF(r10_fiq,512|(10<<16),RNB), |
| 18879 | REGDEF(R11_fiq,512|(11<<16),RNB), REGDEF(r11_fiq,512|(11<<16),RNB), |
| 18880 | REGDEF(R12_fiq,512|(12<<16),RNB), REGDEF(r12_fiq,512|(12<<16),RNB), |
| 18881 | REGDEF(SP_fiq,512|(13<<16),RNB), REGDEF(sp_fiq,512|(13<<16),RNB), |
| 18882 | REGDEF(LR_fiq,512|(14<<16),RNB), REGDEF(lr_fiq,512|(14<<16),RNB), |
| 18883 | REGDEF(SPSR_fiq,512|(14<<16)|SPSR_BIT,RNB), REGDEF(spsr_fiq,512|(14<<16)|SPSR_BIT,RNB), |
| 18884 | |
| 18885 | SPLRBANK(0,IRQ,RNB), SPLRBANK(0,irq,RNB), |
| 18886 | SPLRBANK(2,SVC,RNB), SPLRBANK(2,svc,RNB), |
| 18887 | SPLRBANK(4,ABT,RNB), SPLRBANK(4,abt,RNB), |
| 18888 | SPLRBANK(6,UND,RNB), SPLRBANK(6,und,RNB), |
| 18889 | SPLRBANK(12,MON,RNB), SPLRBANK(12,mon,RNB), |
| 18890 | REGDEF(elr_hyp,768|(14<<16),RNB), REGDEF(ELR_hyp,768|(14<<16),RNB), |
| 18891 | REGDEF(sp_hyp,768|(15<<16),RNB), REGDEF(SP_hyp,768|(15<<16),RNB), |
| 18892 | REGDEF(spsr_hyp,768|(14<<16)|SPSR_BIT,RNB), |
| 18893 | REGDEF(SPSR_hyp,768|(14<<16)|SPSR_BIT,RNB), |
| 18894 | |
| 18895 | /* FPA registers. */ |
| 18896 | REGNUM(f,0,FN), REGNUM(f,1,FN), REGNUM(f,2,FN), REGNUM(f,3,FN), |
| 18897 | REGNUM(f,4,FN), REGNUM(f,5,FN), REGNUM(f,6,FN), REGNUM(f,7, FN), |
| 18898 | |
| 18899 | REGNUM(F,0,FN), REGNUM(F,1,FN), REGNUM(F,2,FN), REGNUM(F,3,FN), |
| 18900 | REGNUM(F,4,FN), REGNUM(F,5,FN), REGNUM(F,6,FN), REGNUM(F,7, FN), |
| 18901 | |
| 18902 | /* VFP SP registers. */ |
| 18903 | REGSET(s,VFS), REGSET(S,VFS), |
| 18904 | REGSETH(s,VFS), REGSETH(S,VFS), |
| 18905 | |
| 18906 | /* VFP DP Registers. */ |
| 18907 | REGSET(d,VFD), REGSET(D,VFD), |
| 18908 | /* Extra Neon DP registers. */ |
| 18909 | REGSETH(d,VFD), REGSETH(D,VFD), |
| 18910 | |
| 18911 | /* Neon QP registers. */ |
| 18912 | REGSET2(q,NQ), REGSET2(Q,NQ), |
| 18913 | |
| 18914 | /* VFP control registers. */ |
| 18915 | REGDEF(fpsid,0,VFC), REGDEF(fpscr,1,VFC), REGDEF(fpexc,8,VFC), |
| 18916 | REGDEF(FPSID,0,VFC), REGDEF(FPSCR,1,VFC), REGDEF(FPEXC,8,VFC), |
| 18917 | REGDEF(fpinst,9,VFC), REGDEF(fpinst2,10,VFC), |
| 18918 | REGDEF(FPINST,9,VFC), REGDEF(FPINST2,10,VFC), |
| 18919 | REGDEF(mvfr0,7,VFC), REGDEF(mvfr1,6,VFC), |
| 18920 | REGDEF(MVFR0,7,VFC), REGDEF(MVFR1,6,VFC), |
| 18921 | REGDEF(mvfr2,5,VFC), REGDEF(MVFR2,5,VFC), |
| 18922 | |
| 18923 | /* Maverick DSP coprocessor registers. */ |
| 18924 | REGSET(mvf,MVF), REGSET(mvd,MVD), REGSET(mvfx,MVFX), REGSET(mvdx,MVDX), |
| 18925 | REGSET(MVF,MVF), REGSET(MVD,MVD), REGSET(MVFX,MVFX), REGSET(MVDX,MVDX), |
| 18926 | |
| 18927 | REGNUM(mvax,0,MVAX), REGNUM(mvax,1,MVAX), |
| 18928 | REGNUM(mvax,2,MVAX), REGNUM(mvax,3,MVAX), |
| 18929 | REGDEF(dspsc,0,DSPSC), |
| 18930 | |
| 18931 | REGNUM(MVAX,0,MVAX), REGNUM(MVAX,1,MVAX), |
| 18932 | REGNUM(MVAX,2,MVAX), REGNUM(MVAX,3,MVAX), |
| 18933 | REGDEF(DSPSC,0,DSPSC), |
| 18934 | |
| 18935 | /* iWMMXt data registers - p0, c0-15. */ |
| 18936 | REGSET(wr,MMXWR), REGSET(wR,MMXWR), REGSET(WR, MMXWR), |
| 18937 | |
| 18938 | /* iWMMXt control registers - p1, c0-3. */ |
| 18939 | REGDEF(wcid, 0,MMXWC), REGDEF(wCID, 0,MMXWC), REGDEF(WCID, 0,MMXWC), |
| 18940 | REGDEF(wcon, 1,MMXWC), REGDEF(wCon, 1,MMXWC), REGDEF(WCON, 1,MMXWC), |
| 18941 | REGDEF(wcssf, 2,MMXWC), REGDEF(wCSSF, 2,MMXWC), REGDEF(WCSSF, 2,MMXWC), |
| 18942 | REGDEF(wcasf, 3,MMXWC), REGDEF(wCASF, 3,MMXWC), REGDEF(WCASF, 3,MMXWC), |
| 18943 | |
| 18944 | /* iWMMXt scalar (constant/offset) registers - p1, c8-11. */ |
| 18945 | REGDEF(wcgr0, 8,MMXWCG), REGDEF(wCGR0, 8,MMXWCG), REGDEF(WCGR0, 8,MMXWCG), |
| 18946 | REGDEF(wcgr1, 9,MMXWCG), REGDEF(wCGR1, 9,MMXWCG), REGDEF(WCGR1, 9,MMXWCG), |
| 18947 | REGDEF(wcgr2,10,MMXWCG), REGDEF(wCGR2,10,MMXWCG), REGDEF(WCGR2,10,MMXWCG), |
| 18948 | REGDEF(wcgr3,11,MMXWCG), REGDEF(wCGR3,11,MMXWCG), REGDEF(WCGR3,11,MMXWCG), |
| 18949 | |
| 18950 | /* XScale accumulator registers. */ |
| 18951 | REGNUM(acc,0,XSCALE), REGNUM(ACC,0,XSCALE), |
| 18952 | }; |
| 18953 | #undef REGDEF |
| 18954 | #undef REGNUM |
| 18955 | #undef REGSET |
| 18956 | |
| 18957 | /* Table of all PSR suffixes. Bare "CPSR" and "SPSR" are handled |
| 18958 | within psr_required_here. */ |
| 18959 | static const struct asm_psr psrs[] = |
| 18960 | { |
| 18961 | /* Backward compatibility notation. Note that "all" is no longer |
| 18962 | truly all possible PSR bits. */ |
| 18963 | {"all", PSR_c | PSR_f}, |
| 18964 | {"flg", PSR_f}, |
| 18965 | {"ctl", PSR_c}, |
| 18966 | |
| 18967 | /* Individual flags. */ |
| 18968 | {"f", PSR_f}, |
| 18969 | {"c", PSR_c}, |
| 18970 | {"x", PSR_x}, |
| 18971 | {"s", PSR_s}, |
| 18972 | |
| 18973 | /* Combinations of flags. */ |
| 18974 | {"fs", PSR_f | PSR_s}, |
| 18975 | {"fx", PSR_f | PSR_x}, |
| 18976 | {"fc", PSR_f | PSR_c}, |
| 18977 | {"sf", PSR_s | PSR_f}, |
| 18978 | {"sx", PSR_s | PSR_x}, |
| 18979 | {"sc", PSR_s | PSR_c}, |
| 18980 | {"xf", PSR_x | PSR_f}, |
| 18981 | {"xs", PSR_x | PSR_s}, |
| 18982 | {"xc", PSR_x | PSR_c}, |
| 18983 | {"cf", PSR_c | PSR_f}, |
| 18984 | {"cs", PSR_c | PSR_s}, |
| 18985 | {"cx", PSR_c | PSR_x}, |
| 18986 | {"fsx", PSR_f | PSR_s | PSR_x}, |
| 18987 | {"fsc", PSR_f | PSR_s | PSR_c}, |
| 18988 | {"fxs", PSR_f | PSR_x | PSR_s}, |
| 18989 | {"fxc", PSR_f | PSR_x | PSR_c}, |
| 18990 | {"fcs", PSR_f | PSR_c | PSR_s}, |
| 18991 | {"fcx", PSR_f | PSR_c | PSR_x}, |
| 18992 | {"sfx", PSR_s | PSR_f | PSR_x}, |
| 18993 | {"sfc", PSR_s | PSR_f | PSR_c}, |
| 18994 | {"sxf", PSR_s | PSR_x | PSR_f}, |
| 18995 | {"sxc", PSR_s | PSR_x | PSR_c}, |
| 18996 | {"scf", PSR_s | PSR_c | PSR_f}, |
| 18997 | {"scx", PSR_s | PSR_c | PSR_x}, |
| 18998 | {"xfs", PSR_x | PSR_f | PSR_s}, |
| 18999 | {"xfc", PSR_x | PSR_f | PSR_c}, |
| 19000 | {"xsf", PSR_x | PSR_s | PSR_f}, |
| 19001 | {"xsc", PSR_x | PSR_s | PSR_c}, |
| 19002 | {"xcf", PSR_x | PSR_c | PSR_f}, |
| 19003 | {"xcs", PSR_x | PSR_c | PSR_s}, |
| 19004 | {"cfs", PSR_c | PSR_f | PSR_s}, |
| 19005 | {"cfx", PSR_c | PSR_f | PSR_x}, |
| 19006 | {"csf", PSR_c | PSR_s | PSR_f}, |
| 19007 | {"csx", PSR_c | PSR_s | PSR_x}, |
| 19008 | {"cxf", PSR_c | PSR_x | PSR_f}, |
| 19009 | {"cxs", PSR_c | PSR_x | PSR_s}, |
| 19010 | {"fsxc", PSR_f | PSR_s | PSR_x | PSR_c}, |
| 19011 | {"fscx", PSR_f | PSR_s | PSR_c | PSR_x}, |
| 19012 | {"fxsc", PSR_f | PSR_x | PSR_s | PSR_c}, |
| 19013 | {"fxcs", PSR_f | PSR_x | PSR_c | PSR_s}, |
| 19014 | {"fcsx", PSR_f | PSR_c | PSR_s | PSR_x}, |
| 19015 | {"fcxs", PSR_f | PSR_c | PSR_x | PSR_s}, |
| 19016 | {"sfxc", PSR_s | PSR_f | PSR_x | PSR_c}, |
| 19017 | {"sfcx", PSR_s | PSR_f | PSR_c | PSR_x}, |
| 19018 | {"sxfc", PSR_s | PSR_x | PSR_f | PSR_c}, |
| 19019 | {"sxcf", PSR_s | PSR_x | PSR_c | PSR_f}, |
| 19020 | {"scfx", PSR_s | PSR_c | PSR_f | PSR_x}, |
| 19021 | {"scxf", PSR_s | PSR_c | PSR_x | PSR_f}, |
| 19022 | {"xfsc", PSR_x | PSR_f | PSR_s | PSR_c}, |
| 19023 | {"xfcs", PSR_x | PSR_f | PSR_c | PSR_s}, |
| 19024 | {"xsfc", PSR_x | PSR_s | PSR_f | PSR_c}, |
| 19025 | {"xscf", PSR_x | PSR_s | PSR_c | PSR_f}, |
| 19026 | {"xcfs", PSR_x | PSR_c | PSR_f | PSR_s}, |
| 19027 | {"xcsf", PSR_x | PSR_c | PSR_s | PSR_f}, |
| 19028 | {"cfsx", PSR_c | PSR_f | PSR_s | PSR_x}, |
| 19029 | {"cfxs", PSR_c | PSR_f | PSR_x | PSR_s}, |
| 19030 | {"csfx", PSR_c | PSR_s | PSR_f | PSR_x}, |
| 19031 | {"csxf", PSR_c | PSR_s | PSR_x | PSR_f}, |
| 19032 | {"cxfs", PSR_c | PSR_x | PSR_f | PSR_s}, |
| 19033 | {"cxsf", PSR_c | PSR_x | PSR_s | PSR_f}, |
| 19034 | }; |
| 19035 | |
| 19036 | /* Table of V7M psr names. */ |
| 19037 | static const struct asm_psr v7m_psrs[] = |
| 19038 | { |
| 19039 | {"apsr", 0x0 }, {"APSR", 0x0 }, |
| 19040 | {"iapsr", 0x1 }, {"IAPSR", 0x1 }, |
| 19041 | {"eapsr", 0x2 }, {"EAPSR", 0x2 }, |
| 19042 | {"psr", 0x3 }, {"PSR", 0x3 }, |
| 19043 | {"xpsr", 0x3 }, {"XPSR", 0x3 }, {"xPSR", 3 }, |
| 19044 | {"ipsr", 0x5 }, {"IPSR", 0x5 }, |
| 19045 | {"epsr", 0x6 }, {"EPSR", 0x6 }, |
| 19046 | {"iepsr", 0x7 }, {"IEPSR", 0x7 }, |
| 19047 | {"msp", 0x8 }, {"MSP", 0x8 }, |
| 19048 | {"psp", 0x9 }, {"PSP", 0x9 }, |
| 19049 | {"msplim", 0xa }, {"MSPLIM", 0xa }, |
| 19050 | {"psplim", 0xb }, {"PSPLIM", 0xb }, |
| 19051 | {"primask", 0x10}, {"PRIMASK", 0x10}, |
| 19052 | {"basepri", 0x11}, {"BASEPRI", 0x11}, |
| 19053 | {"basepri_max", 0x12}, {"BASEPRI_MAX", 0x12}, |
| 19054 | {"faultmask", 0x13}, {"FAULTMASK", 0x13}, |
| 19055 | {"control", 0x14}, {"CONTROL", 0x14}, |
| 19056 | {"msp_ns", 0x88}, {"MSP_NS", 0x88}, |
| 19057 | {"psp_ns", 0x89}, {"PSP_NS", 0x89}, |
| 19058 | {"msplim_ns", 0x8a}, {"MSPLIM_NS", 0x8a}, |
| 19059 | {"psplim_ns", 0x8b}, {"PSPLIM_NS", 0x8b}, |
| 19060 | {"primask_ns", 0x90}, {"PRIMASK_NS", 0x90}, |
| 19061 | {"basepri_ns", 0x91}, {"BASEPRI_NS", 0x91}, |
| 19062 | {"faultmask_ns", 0x93}, {"FAULTMASK_NS", 0x93}, |
| 19063 | {"control_ns", 0x94}, {"CONTROL_NS", 0x94}, |
| 19064 | {"sp_ns", 0x98}, {"SP_NS", 0x98 } |
| 19065 | }; |
| 19066 | |
| 19067 | /* Table of all shift-in-operand names. */ |
| 19068 | static const struct asm_shift_name shift_names [] = |
| 19069 | { |
| 19070 | { "asl", SHIFT_LSL }, { "ASL", SHIFT_LSL }, |
| 19071 | { "lsl", SHIFT_LSL }, { "LSL", SHIFT_LSL }, |
| 19072 | { "lsr", SHIFT_LSR }, { "LSR", SHIFT_LSR }, |
| 19073 | { "asr", SHIFT_ASR }, { "ASR", SHIFT_ASR }, |
| 19074 | { "ror", SHIFT_ROR }, { "ROR", SHIFT_ROR }, |
| 19075 | { "rrx", SHIFT_RRX }, { "RRX", SHIFT_RRX } |
| 19076 | }; |
| 19077 | |
| 19078 | /* Table of all explicit relocation names. */ |
| 19079 | #ifdef OBJ_ELF |
| 19080 | static struct reloc_entry reloc_names[] = |
| 19081 | { |
| 19082 | { "got", BFD_RELOC_ARM_GOT32 }, { "GOT", BFD_RELOC_ARM_GOT32 }, |
| 19083 | { "gotoff", BFD_RELOC_ARM_GOTOFF }, { "GOTOFF", BFD_RELOC_ARM_GOTOFF }, |
| 19084 | { "plt", BFD_RELOC_ARM_PLT32 }, { "PLT", BFD_RELOC_ARM_PLT32 }, |
| 19085 | { "target1", BFD_RELOC_ARM_TARGET1 }, { "TARGET1", BFD_RELOC_ARM_TARGET1 }, |
| 19086 | { "target2", BFD_RELOC_ARM_TARGET2 }, { "TARGET2", BFD_RELOC_ARM_TARGET2 }, |
| 19087 | { "sbrel", BFD_RELOC_ARM_SBREL32 }, { "SBREL", BFD_RELOC_ARM_SBREL32 }, |
| 19088 | { "tlsgd", BFD_RELOC_ARM_TLS_GD32}, { "TLSGD", BFD_RELOC_ARM_TLS_GD32}, |
| 19089 | { "tlsldm", BFD_RELOC_ARM_TLS_LDM32}, { "TLSLDM", BFD_RELOC_ARM_TLS_LDM32}, |
| 19090 | { "tlsldo", BFD_RELOC_ARM_TLS_LDO32}, { "TLSLDO", BFD_RELOC_ARM_TLS_LDO32}, |
| 19091 | { "gottpoff",BFD_RELOC_ARM_TLS_IE32}, { "GOTTPOFF",BFD_RELOC_ARM_TLS_IE32}, |
| 19092 | { "tpoff", BFD_RELOC_ARM_TLS_LE32}, { "TPOFF", BFD_RELOC_ARM_TLS_LE32}, |
| 19093 | { "got_prel", BFD_RELOC_ARM_GOT_PREL}, { "GOT_PREL", BFD_RELOC_ARM_GOT_PREL}, |
| 19094 | { "tlsdesc", BFD_RELOC_ARM_TLS_GOTDESC}, |
| 19095 | { "TLSDESC", BFD_RELOC_ARM_TLS_GOTDESC}, |
| 19096 | { "tlscall", BFD_RELOC_ARM_TLS_CALL}, |
| 19097 | { "TLSCALL", BFD_RELOC_ARM_TLS_CALL}, |
| 19098 | { "tlsdescseq", BFD_RELOC_ARM_TLS_DESCSEQ}, |
| 19099 | { "TLSDESCSEQ", BFD_RELOC_ARM_TLS_DESCSEQ} |
| 19100 | }; |
| 19101 | #endif |
| 19102 | |
| 19103 | /* Table of all conditional affixes. 0xF is not defined as a condition code. */ |
| 19104 | static const struct asm_cond conds[] = |
| 19105 | { |
| 19106 | {"eq", 0x0}, |
| 19107 | {"ne", 0x1}, |
| 19108 | {"cs", 0x2}, {"hs", 0x2}, |
| 19109 | {"cc", 0x3}, {"ul", 0x3}, {"lo", 0x3}, |
| 19110 | {"mi", 0x4}, |
| 19111 | {"pl", 0x5}, |
| 19112 | {"vs", 0x6}, |
| 19113 | {"vc", 0x7}, |
| 19114 | {"hi", 0x8}, |
| 19115 | {"ls", 0x9}, |
| 19116 | {"ge", 0xa}, |
| 19117 | {"lt", 0xb}, |
| 19118 | {"gt", 0xc}, |
| 19119 | {"le", 0xd}, |
| 19120 | {"al", 0xe} |
| 19121 | }; |
| 19122 | |
| 19123 | #define UL_BARRIER(L,U,CODE,FEAT) \ |
| 19124 | { L, CODE, ARM_FEATURE_CORE_LOW (FEAT) }, \ |
| 19125 | { U, CODE, ARM_FEATURE_CORE_LOW (FEAT) } |
| 19126 | |
| 19127 | static struct asm_barrier_opt barrier_opt_names[] = |
| 19128 | { |
| 19129 | UL_BARRIER ("sy", "SY", 0xf, ARM_EXT_BARRIER), |
| 19130 | UL_BARRIER ("st", "ST", 0xe, ARM_EXT_BARRIER), |
| 19131 | UL_BARRIER ("ld", "LD", 0xd, ARM_EXT_V8), |
| 19132 | UL_BARRIER ("ish", "ISH", 0xb, ARM_EXT_BARRIER), |
| 19133 | UL_BARRIER ("sh", "SH", 0xb, ARM_EXT_BARRIER), |
| 19134 | UL_BARRIER ("ishst", "ISHST", 0xa, ARM_EXT_BARRIER), |
| 19135 | UL_BARRIER ("shst", "SHST", 0xa, ARM_EXT_BARRIER), |
| 19136 | UL_BARRIER ("ishld", "ISHLD", 0x9, ARM_EXT_V8), |
| 19137 | UL_BARRIER ("un", "UN", 0x7, ARM_EXT_BARRIER), |
| 19138 | UL_BARRIER ("nsh", "NSH", 0x7, ARM_EXT_BARRIER), |
| 19139 | UL_BARRIER ("unst", "UNST", 0x6, ARM_EXT_BARRIER), |
| 19140 | UL_BARRIER ("nshst", "NSHST", 0x6, ARM_EXT_BARRIER), |
| 19141 | UL_BARRIER ("nshld", "NSHLD", 0x5, ARM_EXT_V8), |
| 19142 | UL_BARRIER ("osh", "OSH", 0x3, ARM_EXT_BARRIER), |
| 19143 | UL_BARRIER ("oshst", "OSHST", 0x2, ARM_EXT_BARRIER), |
| 19144 | UL_BARRIER ("oshld", "OSHLD", 0x1, ARM_EXT_V8) |
| 19145 | }; |
| 19146 | |
| 19147 | #undef UL_BARRIER |
| 19148 | |
| 19149 | /* Table of ARM-format instructions. */ |
| 19150 | |
| 19151 | /* Macros for gluing together operand strings. N.B. In all cases |
| 19152 | other than OPS0, the trailing OP_stop comes from default |
| 19153 | zero-initialization of the unspecified elements of the array. */ |
| 19154 | #define OPS0() { OP_stop, } |
| 19155 | #define OPS1(a) { OP_##a, } |
| 19156 | #define OPS2(a,b) { OP_##a,OP_##b, } |
| 19157 | #define OPS3(a,b,c) { OP_##a,OP_##b,OP_##c, } |
| 19158 | #define OPS4(a,b,c,d) { OP_##a,OP_##b,OP_##c,OP_##d, } |
| 19159 | #define OPS5(a,b,c,d,e) { OP_##a,OP_##b,OP_##c,OP_##d,OP_##e, } |
| 19160 | #define OPS6(a,b,c,d,e,f) { OP_##a,OP_##b,OP_##c,OP_##d,OP_##e,OP_##f, } |
| 19161 | |
| 19162 | /* These macros are similar to the OPSn, but do not prepend the OP_ prefix. |
| 19163 | This is useful when mixing operands for ARM and THUMB, i.e. using the |
| 19164 | MIX_ARM_THUMB_OPERANDS macro. |
| 19165 | In order to use these macros, prefix the number of operands with _ |
| 19166 | e.g. _3. */ |
| 19167 | #define OPS_1(a) { a, } |
| 19168 | #define OPS_2(a,b) { a,b, } |
| 19169 | #define OPS_3(a,b,c) { a,b,c, } |
| 19170 | #define OPS_4(a,b,c,d) { a,b,c,d, } |
| 19171 | #define OPS_5(a,b,c,d,e) { a,b,c,d,e, } |
| 19172 | #define OPS_6(a,b,c,d,e,f) { a,b,c,d,e,f, } |
| 19173 | |
| 19174 | /* These macros abstract out the exact format of the mnemonic table and |
| 19175 | save some repeated characters. */ |
| 19176 | |
| 19177 | /* The normal sort of mnemonic; has a Thumb variant; takes a conditional suffix. */ |
| 19178 | #define TxCE(mnem, op, top, nops, ops, ae, te) \ |
| 19179 | { mnem, OPS##nops ops, OT_csuffix, 0x##op, top, ARM_VARIANT, \ |
| 19180 | THUMB_VARIANT, do_##ae, do_##te } |
| 19181 | |
| 19182 | /* Two variants of the above - TCE for a numeric Thumb opcode, tCE for |
| 19183 | a T_MNEM_xyz enumerator. */ |
| 19184 | #define TCE(mnem, aop, top, nops, ops, ae, te) \ |
| 19185 | TxCE (mnem, aop, 0x##top, nops, ops, ae, te) |
| 19186 | #define tCE(mnem, aop, top, nops, ops, ae, te) \ |
| 19187 | TxCE (mnem, aop, T_MNEM##top, nops, ops, ae, te) |
| 19188 | |
| 19189 | /* Second most common sort of mnemonic: has a Thumb variant, takes a conditional |
| 19190 | infix after the third character. */ |
| 19191 | #define TxC3(mnem, op, top, nops, ops, ae, te) \ |
| 19192 | { mnem, OPS##nops ops, OT_cinfix3, 0x##op, top, ARM_VARIANT, \ |
| 19193 | THUMB_VARIANT, do_##ae, do_##te } |
| 19194 | #define TxC3w(mnem, op, top, nops, ops, ae, te) \ |
| 19195 | { mnem, OPS##nops ops, OT_cinfix3_deprecated, 0x##op, top, ARM_VARIANT, \ |
| 19196 | THUMB_VARIANT, do_##ae, do_##te } |
| 19197 | #define TC3(mnem, aop, top, nops, ops, ae, te) \ |
| 19198 | TxC3 (mnem, aop, 0x##top, nops, ops, ae, te) |
| 19199 | #define TC3w(mnem, aop, top, nops, ops, ae, te) \ |
| 19200 | TxC3w (mnem, aop, 0x##top, nops, ops, ae, te) |
| 19201 | #define tC3(mnem, aop, top, nops, ops, ae, te) \ |
| 19202 | TxC3 (mnem, aop, T_MNEM##top, nops, ops, ae, te) |
| 19203 | #define tC3w(mnem, aop, top, nops, ops, ae, te) \ |
| 19204 | TxC3w (mnem, aop, T_MNEM##top, nops, ops, ae, te) |
| 19205 | |
| 19206 | /* Mnemonic that cannot be conditionalized. The ARM condition-code |
| 19207 | field is still 0xE. Many of the Thumb variants can be executed |
| 19208 | conditionally, so this is checked separately. */ |
| 19209 | #define TUE(mnem, op, top, nops, ops, ae, te) \ |
| 19210 | { mnem, OPS##nops ops, OT_unconditional, 0x##op, 0x##top, ARM_VARIANT, \ |
| 19211 | THUMB_VARIANT, do_##ae, do_##te } |
| 19212 | |
| 19213 | /* Same as TUE but the encoding function for ARM and Thumb modes is the same. |
| 19214 | Used by mnemonics that have very minimal differences in the encoding for |
| 19215 | ARM and Thumb variants and can be handled in a common function. */ |
| 19216 | #define TUEc(mnem, op, top, nops, ops, en) \ |
| 19217 | { mnem, OPS##nops ops, OT_unconditional, 0x##op, 0x##top, ARM_VARIANT, \ |
| 19218 | THUMB_VARIANT, do_##en, do_##en } |
| 19219 | |
| 19220 | /* Mnemonic that cannot be conditionalized, and bears 0xF in its ARM |
| 19221 | condition code field. */ |
| 19222 | #define TUF(mnem, op, top, nops, ops, ae, te) \ |
| 19223 | { mnem, OPS##nops ops, OT_unconditionalF, 0x##op, 0x##top, ARM_VARIANT, \ |
| 19224 | THUMB_VARIANT, do_##ae, do_##te } |
| 19225 | |
| 19226 | /* ARM-only variants of all the above. */ |
| 19227 | #define CE(mnem, op, nops, ops, ae) \ |
| 19228 | { mnem, OPS##nops ops, OT_csuffix, 0x##op, 0x0, ARM_VARIANT, 0, do_##ae, NULL } |
| 19229 | |
| 19230 | #define C3(mnem, op, nops, ops, ae) \ |
| 19231 | { #mnem, OPS##nops ops, OT_cinfix3, 0x##op, 0x0, ARM_VARIANT, 0, do_##ae, NULL } |
| 19232 | |
| 19233 | /* Legacy mnemonics that always have conditional infix after the third |
| 19234 | character. */ |
| 19235 | #define CL(mnem, op, nops, ops, ae) \ |
| 19236 | { mnem, OPS##nops ops, OT_cinfix3_legacy, \ |
| 19237 | 0x##op, 0x0, ARM_VARIANT, 0, do_##ae, NULL } |
| 19238 | |
| 19239 | /* Coprocessor instructions. Isomorphic between Arm and Thumb-2. */ |
| 19240 | #define cCE(mnem, op, nops, ops, ae) \ |
| 19241 | { mnem, OPS##nops ops, OT_csuffix, 0x##op, 0xe##op, ARM_VARIANT, ARM_VARIANT, do_##ae, do_##ae } |
| 19242 | |
| 19243 | /* Legacy coprocessor instructions where conditional infix and conditional |
| 19244 | suffix are ambiguous. For consistency this includes all FPA instructions, |
| 19245 | not just the potentially ambiguous ones. */ |
| 19246 | #define cCL(mnem, op, nops, ops, ae) \ |
| 19247 | { mnem, OPS##nops ops, OT_cinfix3_legacy, \ |
| 19248 | 0x##op, 0xe##op, ARM_VARIANT, ARM_VARIANT, do_##ae, do_##ae } |
| 19249 | |
| 19250 | /* Coprocessor, takes either a suffix or a position-3 infix |
| 19251 | (for an FPA corner case). */ |
| 19252 | #define C3E(mnem, op, nops, ops, ae) \ |
| 19253 | { mnem, OPS##nops ops, OT_csuf_or_in3, \ |
| 19254 | 0x##op, 0xe##op, ARM_VARIANT, ARM_VARIANT, do_##ae, do_##ae } |
| 19255 | |
| 19256 | #define xCM_(m1, m2, m3, op, nops, ops, ae) \ |
| 19257 | { m1 #m2 m3, OPS##nops ops, \ |
| 19258 | sizeof (#m2) == 1 ? OT_odd_infix_unc : OT_odd_infix_0 + sizeof (m1) - 1, \ |
| 19259 | 0x##op, 0x0, ARM_VARIANT, 0, do_##ae, NULL } |
| 19260 | |
| 19261 | #define CM(m1, m2, op, nops, ops, ae) \ |
| 19262 | xCM_ (m1, , m2, op, nops, ops, ae), \ |
| 19263 | xCM_ (m1, eq, m2, op, nops, ops, ae), \ |
| 19264 | xCM_ (m1, ne, m2, op, nops, ops, ae), \ |
| 19265 | xCM_ (m1, cs, m2, op, nops, ops, ae), \ |
| 19266 | xCM_ (m1, hs, m2, op, nops, ops, ae), \ |
| 19267 | xCM_ (m1, cc, m2, op, nops, ops, ae), \ |
| 19268 | xCM_ (m1, ul, m2, op, nops, ops, ae), \ |
| 19269 | xCM_ (m1, lo, m2, op, nops, ops, ae), \ |
| 19270 | xCM_ (m1, mi, m2, op, nops, ops, ae), \ |
| 19271 | xCM_ (m1, pl, m2, op, nops, ops, ae), \ |
| 19272 | xCM_ (m1, vs, m2, op, nops, ops, ae), \ |
| 19273 | xCM_ (m1, vc, m2, op, nops, ops, ae), \ |
| 19274 | xCM_ (m1, hi, m2, op, nops, ops, ae), \ |
| 19275 | xCM_ (m1, ls, m2, op, nops, ops, ae), \ |
| 19276 | xCM_ (m1, ge, m2, op, nops, ops, ae), \ |
| 19277 | xCM_ (m1, lt, m2, op, nops, ops, ae), \ |
| 19278 | xCM_ (m1, gt, m2, op, nops, ops, ae), \ |
| 19279 | xCM_ (m1, le, m2, op, nops, ops, ae), \ |
| 19280 | xCM_ (m1, al, m2, op, nops, ops, ae) |
| 19281 | |
| 19282 | #define UE(mnem, op, nops, ops, ae) \ |
| 19283 | { #mnem, OPS##nops ops, OT_unconditional, 0x##op, 0, ARM_VARIANT, 0, do_##ae, NULL } |
| 19284 | |
| 19285 | #define UF(mnem, op, nops, ops, ae) \ |
| 19286 | { #mnem, OPS##nops ops, OT_unconditionalF, 0x##op, 0, ARM_VARIANT, 0, do_##ae, NULL } |
| 19287 | |
| 19288 | /* Neon data-processing. ARM versions are unconditional with cond=0xf. |
| 19289 | The Thumb and ARM variants are mostly the same (bits 0-23 and 24/28), so we |
| 19290 | use the same encoding function for each. */ |
| 19291 | #define NUF(mnem, op, nops, ops, enc) \ |
| 19292 | { #mnem, OPS##nops ops, OT_unconditionalF, 0x##op, 0x##op, \ |
| 19293 | ARM_VARIANT, THUMB_VARIANT, do_##enc, do_##enc } |
| 19294 | |
| 19295 | /* Neon data processing, version which indirects through neon_enc_tab for |
| 19296 | the various overloaded versions of opcodes. */ |
| 19297 | #define nUF(mnem, op, nops, ops, enc) \ |
| 19298 | { #mnem, OPS##nops ops, OT_unconditionalF, N_MNEM##op, N_MNEM##op, \ |
| 19299 | ARM_VARIANT, THUMB_VARIANT, do_##enc, do_##enc } |
| 19300 | |
| 19301 | /* Neon insn with conditional suffix for the ARM version, non-overloaded |
| 19302 | version. */ |
| 19303 | #define NCE_tag(mnem, op, nops, ops, enc, tag) \ |
| 19304 | { #mnem, OPS##nops ops, tag, 0x##op, 0x##op, ARM_VARIANT, \ |
| 19305 | THUMB_VARIANT, do_##enc, do_##enc } |
| 19306 | |
| 19307 | #define NCE(mnem, op, nops, ops, enc) \ |
| 19308 | NCE_tag (mnem, op, nops, ops, enc, OT_csuffix) |
| 19309 | |
| 19310 | #define NCEF(mnem, op, nops, ops, enc) \ |
| 19311 | NCE_tag (mnem, op, nops, ops, enc, OT_csuffixF) |
| 19312 | |
| 19313 | /* Neon insn with conditional suffix for the ARM version, overloaded types. */ |
| 19314 | #define nCE_tag(mnem, op, nops, ops, enc, tag) \ |
| 19315 | { #mnem, OPS##nops ops, tag, N_MNEM##op, N_MNEM##op, \ |
| 19316 | ARM_VARIANT, THUMB_VARIANT, do_##enc, do_##enc } |
| 19317 | |
| 19318 | #define nCE(mnem, op, nops, ops, enc) \ |
| 19319 | nCE_tag (mnem, op, nops, ops, enc, OT_csuffix) |
| 19320 | |
| 19321 | #define nCEF(mnem, op, nops, ops, enc) \ |
| 19322 | nCE_tag (mnem, op, nops, ops, enc, OT_csuffixF) |
| 19323 | |
| 19324 | #define do_0 0 |
| 19325 | |
| 19326 | static const struct asm_opcode insns[] = |
| 19327 | { |
| 19328 | #define ARM_VARIANT & arm_ext_v1 /* Core ARM Instructions. */ |
| 19329 | #define THUMB_VARIANT & arm_ext_v4t |
| 19330 | tCE("and", 0000000, _and, 3, (RR, oRR, SH), arit, t_arit3c), |
| 19331 | tC3("ands", 0100000, _ands, 3, (RR, oRR, SH), arit, t_arit3c), |
| 19332 | tCE("eor", 0200000, _eor, 3, (RR, oRR, SH), arit, t_arit3c), |
| 19333 | tC3("eors", 0300000, _eors, 3, (RR, oRR, SH), arit, t_arit3c), |
| 19334 | tCE("sub", 0400000, _sub, 3, (RR, oRR, SH), arit, t_add_sub), |
| 19335 | tC3("subs", 0500000, _subs, 3, (RR, oRR, SH), arit, t_add_sub), |
| 19336 | tCE("add", 0800000, _add, 3, (RR, oRR, SHG), arit, t_add_sub), |
| 19337 | tC3("adds", 0900000, _adds, 3, (RR, oRR, SHG), arit, t_add_sub), |
| 19338 | tCE("adc", 0a00000, _adc, 3, (RR, oRR, SH), arit, t_arit3c), |
| 19339 | tC3("adcs", 0b00000, _adcs, 3, (RR, oRR, SH), arit, t_arit3c), |
| 19340 | tCE("sbc", 0c00000, _sbc, 3, (RR, oRR, SH), arit, t_arit3), |
| 19341 | tC3("sbcs", 0d00000, _sbcs, 3, (RR, oRR, SH), arit, t_arit3), |
| 19342 | tCE("orr", 1800000, _orr, 3, (RR, oRR, SH), arit, t_arit3c), |
| 19343 | tC3("orrs", 1900000, _orrs, 3, (RR, oRR, SH), arit, t_arit3c), |
| 19344 | tCE("bic", 1c00000, _bic, 3, (RR, oRR, SH), arit, t_arit3), |
| 19345 | tC3("bics", 1d00000, _bics, 3, (RR, oRR, SH), arit, t_arit3), |
| 19346 | |
| 19347 | /* The p-variants of tst/cmp/cmn/teq (below) are the pre-V6 mechanism |
| 19348 | for setting PSR flag bits. They are obsolete in V6 and do not |
| 19349 | have Thumb equivalents. */ |
| 19350 | tCE("tst", 1100000, _tst, 2, (RR, SH), cmp, t_mvn_tst), |
| 19351 | tC3w("tsts", 1100000, _tst, 2, (RR, SH), cmp, t_mvn_tst), |
| 19352 | CL("tstp", 110f000, 2, (RR, SH), cmp), |
| 19353 | tCE("cmp", 1500000, _cmp, 2, (RR, SH), cmp, t_mov_cmp), |
| 19354 | tC3w("cmps", 1500000, _cmp, 2, (RR, SH), cmp, t_mov_cmp), |
| 19355 | CL("cmpp", 150f000, 2, (RR, SH), cmp), |
| 19356 | tCE("cmn", 1700000, _cmn, 2, (RR, SH), cmp, t_mvn_tst), |
| 19357 | tC3w("cmns", 1700000, _cmn, 2, (RR, SH), cmp, t_mvn_tst), |
| 19358 | CL("cmnp", 170f000, 2, (RR, SH), cmp), |
| 19359 | |
| 19360 | tCE("mov", 1a00000, _mov, 2, (RR, SH), mov, t_mov_cmp), |
| 19361 | tC3("movs", 1b00000, _movs, 2, (RR, SHG), mov, t_mov_cmp), |
| 19362 | tCE("mvn", 1e00000, _mvn, 2, (RR, SH), mov, t_mvn_tst), |
| 19363 | tC3("mvns", 1f00000, _mvns, 2, (RR, SH), mov, t_mvn_tst), |
| 19364 | |
| 19365 | tCE("ldr", 4100000, _ldr, 2, (RR, ADDRGLDR),ldst, t_ldst), |
| 19366 | tC3("ldrb", 4500000, _ldrb, 2, (RRnpc_npcsp, ADDRGLDR),ldst, t_ldst), |
| 19367 | tCE("str", 4000000, _str, _2, (MIX_ARM_THUMB_OPERANDS (OP_RR, |
| 19368 | OP_RRnpc), |
| 19369 | OP_ADDRGLDR),ldst, t_ldst), |
| 19370 | tC3("strb", 4400000, _strb, 2, (RRnpc_npcsp, ADDRGLDR),ldst, t_ldst), |
| 19371 | |
| 19372 | tCE("stm", 8800000, _stmia, 2, (RRw, REGLST), ldmstm, t_ldmstm), |
| 19373 | tC3("stmia", 8800000, _stmia, 2, (RRw, REGLST), ldmstm, t_ldmstm), |
| 19374 | tC3("stmea", 8800000, _stmia, 2, (RRw, REGLST), ldmstm, t_ldmstm), |
| 19375 | tCE("ldm", 8900000, _ldmia, 2, (RRw, REGLST), ldmstm, t_ldmstm), |
| 19376 | tC3("ldmia", 8900000, _ldmia, 2, (RRw, REGLST), ldmstm, t_ldmstm), |
| 19377 | tC3("ldmfd", 8900000, _ldmia, 2, (RRw, REGLST), ldmstm, t_ldmstm), |
| 19378 | |
| 19379 | tCE("b", a000000, _b, 1, (EXPr), branch, t_branch), |
| 19380 | TCE("bl", b000000, f000f800, 1, (EXPr), bl, t_branch23), |
| 19381 | |
| 19382 | /* Pseudo ops. */ |
| 19383 | tCE("adr", 28f0000, _adr, 2, (RR, EXP), adr, t_adr), |
| 19384 | C3(adrl, 28f0000, 2, (RR, EXP), adrl), |
| 19385 | tCE("nop", 1a00000, _nop, 1, (oI255c), nop, t_nop), |
| 19386 | tCE("udf", 7f000f0, _udf, 1, (oIffffb), bkpt, t_udf), |
| 19387 | |
| 19388 | /* Thumb-compatibility pseudo ops. */ |
| 19389 | tCE("lsl", 1a00000, _lsl, 3, (RR, oRR, SH), shift, t_shift), |
| 19390 | tC3("lsls", 1b00000, _lsls, 3, (RR, oRR, SH), shift, t_shift), |
| 19391 | tCE("lsr", 1a00020, _lsr, 3, (RR, oRR, SH), shift, t_shift), |
| 19392 | tC3("lsrs", 1b00020, _lsrs, 3, (RR, oRR, SH), shift, t_shift), |
| 19393 | tCE("asr", 1a00040, _asr, 3, (RR, oRR, SH), shift, t_shift), |
| 19394 | tC3("asrs", 1b00040, _asrs, 3, (RR, oRR, SH), shift, t_shift), |
| 19395 | tCE("ror", 1a00060, _ror, 3, (RR, oRR, SH), shift, t_shift), |
| 19396 | tC3("rors", 1b00060, _rors, 3, (RR, oRR, SH), shift, t_shift), |
| 19397 | tCE("neg", 2600000, _neg, 2, (RR, RR), rd_rn, t_neg), |
| 19398 | tC3("negs", 2700000, _negs, 2, (RR, RR), rd_rn, t_neg), |
| 19399 | tCE("push", 92d0000, _push, 1, (REGLST), push_pop, t_push_pop), |
| 19400 | tCE("pop", 8bd0000, _pop, 1, (REGLST), push_pop, t_push_pop), |
| 19401 | |
| 19402 | /* These may simplify to neg. */ |
| 19403 | TCE("rsb", 0600000, ebc00000, 3, (RR, oRR, SH), arit, t_rsb), |
| 19404 | TC3("rsbs", 0700000, ebd00000, 3, (RR, oRR, SH), arit, t_rsb), |
| 19405 | |
| 19406 | #undef THUMB_VARIANT |
| 19407 | #define THUMB_VARIANT & arm_ext_os |
| 19408 | |
| 19409 | TCE("swi", f000000, df00, 1, (EXPi), swi, t_swi), |
| 19410 | TCE("svc", f000000, df00, 1, (EXPi), swi, t_swi), |
| 19411 | |
| 19412 | #undef THUMB_VARIANT |
| 19413 | #define THUMB_VARIANT & arm_ext_v6 |
| 19414 | |
| 19415 | TCE("cpy", 1a00000, 4600, 2, (RR, RR), rd_rm, t_cpy), |
| 19416 | |
| 19417 | /* V1 instructions with no Thumb analogue prior to V6T2. */ |
| 19418 | #undef THUMB_VARIANT |
| 19419 | #define THUMB_VARIANT & arm_ext_v6t2 |
| 19420 | |
| 19421 | TCE("teq", 1300000, ea900f00, 2, (RR, SH), cmp, t_mvn_tst), |
| 19422 | TC3w("teqs", 1300000, ea900f00, 2, (RR, SH), cmp, t_mvn_tst), |
| 19423 | CL("teqp", 130f000, 2, (RR, SH), cmp), |
| 19424 | |
| 19425 | TC3("ldrt", 4300000, f8500e00, 2, (RRnpc_npcsp, ADDR),ldstt, t_ldstt), |
| 19426 | TC3("ldrbt", 4700000, f8100e00, 2, (RRnpc_npcsp, ADDR),ldstt, t_ldstt), |
| 19427 | TC3("strt", 4200000, f8400e00, 2, (RR_npcsp, ADDR), ldstt, t_ldstt), |
| 19428 | TC3("strbt", 4600000, f8000e00, 2, (RRnpc_npcsp, ADDR),ldstt, t_ldstt), |
| 19429 | |
| 19430 | TC3("stmdb", 9000000, e9000000, 2, (RRw, REGLST), ldmstm, t_ldmstm), |
| 19431 | TC3("stmfd", 9000000, e9000000, 2, (RRw, REGLST), ldmstm, t_ldmstm), |
| 19432 | |
| 19433 | TC3("ldmdb", 9100000, e9100000, 2, (RRw, REGLST), ldmstm, t_ldmstm), |
| 19434 | TC3("ldmea", 9100000, e9100000, 2, (RRw, REGLST), ldmstm, t_ldmstm), |
| 19435 | |
| 19436 | /* V1 instructions with no Thumb analogue at all. */ |
| 19437 | CE("rsc", 0e00000, 3, (RR, oRR, SH), arit), |
| 19438 | C3(rscs, 0f00000, 3, (RR, oRR, SH), arit), |
| 19439 | |
| 19440 | C3(stmib, 9800000, 2, (RRw, REGLST), ldmstm), |
| 19441 | C3(stmfa, 9800000, 2, (RRw, REGLST), ldmstm), |
| 19442 | C3(stmda, 8000000, 2, (RRw, REGLST), ldmstm), |
| 19443 | C3(stmed, 8000000, 2, (RRw, REGLST), ldmstm), |
| 19444 | C3(ldmib, 9900000, 2, (RRw, REGLST), ldmstm), |
| 19445 | C3(ldmed, 9900000, 2, (RRw, REGLST), ldmstm), |
| 19446 | C3(ldmda, 8100000, 2, (RRw, REGLST), ldmstm), |
| 19447 | C3(ldmfa, 8100000, 2, (RRw, REGLST), ldmstm), |
| 19448 | |
| 19449 | #undef ARM_VARIANT |
| 19450 | #define ARM_VARIANT & arm_ext_v2 /* ARM 2 - multiplies. */ |
| 19451 | #undef THUMB_VARIANT |
| 19452 | #define THUMB_VARIANT & arm_ext_v4t |
| 19453 | |
| 19454 | tCE("mul", 0000090, _mul, 3, (RRnpc, RRnpc, oRR), mul, t_mul), |
| 19455 | tC3("muls", 0100090, _muls, 3, (RRnpc, RRnpc, oRR), mul, t_mul), |
| 19456 | |
| 19457 | #undef THUMB_VARIANT |
| 19458 | #define THUMB_VARIANT & arm_ext_v6t2 |
| 19459 | |
| 19460 | TCE("mla", 0200090, fb000000, 4, (RRnpc, RRnpc, RRnpc, RRnpc), mlas, t_mla), |
| 19461 | C3(mlas, 0300090, 4, (RRnpc, RRnpc, RRnpc, RRnpc), mlas), |
| 19462 | |
| 19463 | /* Generic coprocessor instructions. */ |
| 19464 | TCE("cdp", e000000, ee000000, 6, (RCP, I15b, RCN, RCN, RCN, oI7b), cdp, cdp), |
| 19465 | TCE("ldc", c100000, ec100000, 3, (RCP, RCN, ADDRGLDC), lstc, lstc), |
| 19466 | TC3("ldcl", c500000, ec500000, 3, (RCP, RCN, ADDRGLDC), lstc, lstc), |
| 19467 | TCE("stc", c000000, ec000000, 3, (RCP, RCN, ADDRGLDC), lstc, lstc), |
| 19468 | TC3("stcl", c400000, ec400000, 3, (RCP, RCN, ADDRGLDC), lstc, lstc), |
| 19469 | TCE("mcr", e000010, ee000010, 6, (RCP, I7b, RR, RCN, RCN, oI7b), co_reg, co_reg), |
| 19470 | TCE("mrc", e100010, ee100010, 6, (RCP, I7b, APSR_RR, RCN, RCN, oI7b), co_reg, co_reg), |
| 19471 | |
| 19472 | #undef ARM_VARIANT |
| 19473 | #define ARM_VARIANT & arm_ext_v2s /* ARM 3 - swp instructions. */ |
| 19474 | |
| 19475 | CE("swp", 1000090, 3, (RRnpc, RRnpc, RRnpcb), rd_rm_rn), |
| 19476 | C3(swpb, 1400090, 3, (RRnpc, RRnpc, RRnpcb), rd_rm_rn), |
| 19477 | |
| 19478 | #undef ARM_VARIANT |
| 19479 | #define ARM_VARIANT & arm_ext_v3 /* ARM 6 Status register instructions. */ |
| 19480 | #undef THUMB_VARIANT |
| 19481 | #define THUMB_VARIANT & arm_ext_msr |
| 19482 | |
| 19483 | TCE("mrs", 1000000, f3e08000, 2, (RRnpc, rPSR), mrs, t_mrs), |
| 19484 | TCE("msr", 120f000, f3808000, 2, (wPSR, RR_EXi), msr, t_msr), |
| 19485 | |
| 19486 | #undef ARM_VARIANT |
| 19487 | #define ARM_VARIANT & arm_ext_v3m /* ARM 7M long multiplies. */ |
| 19488 | #undef THUMB_VARIANT |
| 19489 | #define THUMB_VARIANT & arm_ext_v6t2 |
| 19490 | |
| 19491 | TCE("smull", 0c00090, fb800000, 4, (RRnpc, RRnpc, RRnpc, RRnpc), mull, t_mull), |
| 19492 | CM("smull","s", 0d00090, 4, (RRnpc, RRnpc, RRnpc, RRnpc), mull), |
| 19493 | TCE("umull", 0800090, fba00000, 4, (RRnpc, RRnpc, RRnpc, RRnpc), mull, t_mull), |
| 19494 | CM("umull","s", 0900090, 4, (RRnpc, RRnpc, RRnpc, RRnpc), mull), |
| 19495 | TCE("smlal", 0e00090, fbc00000, 4, (RRnpc, RRnpc, RRnpc, RRnpc), mull, t_mull), |
| 19496 | CM("smlal","s", 0f00090, 4, (RRnpc, RRnpc, RRnpc, RRnpc), mull), |
| 19497 | TCE("umlal", 0a00090, fbe00000, 4, (RRnpc, RRnpc, RRnpc, RRnpc), mull, t_mull), |
| 19498 | CM("umlal","s", 0b00090, 4, (RRnpc, RRnpc, RRnpc, RRnpc), mull), |
| 19499 | |
| 19500 | #undef ARM_VARIANT |
| 19501 | #define ARM_VARIANT & arm_ext_v4 /* ARM Architecture 4. */ |
| 19502 | #undef THUMB_VARIANT |
| 19503 | #define THUMB_VARIANT & arm_ext_v4t |
| 19504 | |
| 19505 | tC3("ldrh", 01000b0, _ldrh, 2, (RRnpc_npcsp, ADDRGLDRS), ldstv4, t_ldst), |
| 19506 | tC3("strh", 00000b0, _strh, 2, (RRnpc_npcsp, ADDRGLDRS), ldstv4, t_ldst), |
| 19507 | tC3("ldrsh", 01000f0, _ldrsh, 2, (RRnpc_npcsp, ADDRGLDRS), ldstv4, t_ldst), |
| 19508 | tC3("ldrsb", 01000d0, _ldrsb, 2, (RRnpc_npcsp, ADDRGLDRS), ldstv4, t_ldst), |
| 19509 | tC3("ldsh", 01000f0, _ldrsh, 2, (RRnpc_npcsp, ADDRGLDRS), ldstv4, t_ldst), |
| 19510 | tC3("ldsb", 01000d0, _ldrsb, 2, (RRnpc_npcsp, ADDRGLDRS), ldstv4, t_ldst), |
| 19511 | |
| 19512 | #undef ARM_VARIANT |
| 19513 | #define ARM_VARIANT & arm_ext_v4t_5 |
| 19514 | |
| 19515 | /* ARM Architecture 4T. */ |
| 19516 | /* Note: bx (and blx) are required on V5, even if the processor does |
| 19517 | not support Thumb. */ |
| 19518 | TCE("bx", 12fff10, 4700, 1, (RR), bx, t_bx), |
| 19519 | |
| 19520 | #undef ARM_VARIANT |
| 19521 | #define ARM_VARIANT & arm_ext_v5 /* ARM Architecture 5T. */ |
| 19522 | #undef THUMB_VARIANT |
| 19523 | #define THUMB_VARIANT & arm_ext_v5t |
| 19524 | |
| 19525 | /* Note: blx has 2 variants; the .value coded here is for |
| 19526 | BLX(2). Only this variant has conditional execution. */ |
| 19527 | TCE("blx", 12fff30, 4780, 1, (RR_EXr), blx, t_blx), |
| 19528 | TUE("bkpt", 1200070, be00, 1, (oIffffb), bkpt, t_bkpt), |
| 19529 | |
| 19530 | #undef THUMB_VARIANT |
| 19531 | #define THUMB_VARIANT & arm_ext_v6t2 |
| 19532 | |
| 19533 | TCE("clz", 16f0f10, fab0f080, 2, (RRnpc, RRnpc), rd_rm, t_clz), |
| 19534 | TUF("ldc2", c100000, fc100000, 3, (RCP, RCN, ADDRGLDC), lstc, lstc), |
| 19535 | TUF("ldc2l", c500000, fc500000, 3, (RCP, RCN, ADDRGLDC), lstc, lstc), |
| 19536 | TUF("stc2", c000000, fc000000, 3, (RCP, RCN, ADDRGLDC), lstc, lstc), |
| 19537 | TUF("stc2l", c400000, fc400000, 3, (RCP, RCN, ADDRGLDC), lstc, lstc), |
| 19538 | TUF("cdp2", e000000, fe000000, 6, (RCP, I15b, RCN, RCN, RCN, oI7b), cdp, cdp), |
| 19539 | TUF("mcr2", e000010, fe000010, 6, (RCP, I7b, RR, RCN, RCN, oI7b), co_reg, co_reg), |
| 19540 | TUF("mrc2", e100010, fe100010, 6, (RCP, I7b, RR, RCN, RCN, oI7b), co_reg, co_reg), |
| 19541 | |
| 19542 | #undef ARM_VARIANT |
| 19543 | #define ARM_VARIANT & arm_ext_v5exp /* ARM Architecture 5TExP. */ |
| 19544 | #undef THUMB_VARIANT |
| 19545 | #define THUMB_VARIANT & arm_ext_v5exp |
| 19546 | |
| 19547 | TCE("smlabb", 1000080, fb100000, 4, (RRnpc, RRnpc, RRnpc, RRnpc), smla, t_mla), |
| 19548 | TCE("smlatb", 10000a0, fb100020, 4, (RRnpc, RRnpc, RRnpc, RRnpc), smla, t_mla), |
| 19549 | TCE("smlabt", 10000c0, fb100010, 4, (RRnpc, RRnpc, RRnpc, RRnpc), smla, t_mla), |
| 19550 | TCE("smlatt", 10000e0, fb100030, 4, (RRnpc, RRnpc, RRnpc, RRnpc), smla, t_mla), |
| 19551 | |
| 19552 | TCE("smlawb", 1200080, fb300000, 4, (RRnpc, RRnpc, RRnpc, RRnpc), smla, t_mla), |
| 19553 | TCE("smlawt", 12000c0, fb300010, 4, (RRnpc, RRnpc, RRnpc, RRnpc), smla, t_mla), |
| 19554 | |
| 19555 | TCE("smlalbb", 1400080, fbc00080, 4, (RRnpc, RRnpc, RRnpc, RRnpc), smlal, t_mlal), |
| 19556 | TCE("smlaltb", 14000a0, fbc000a0, 4, (RRnpc, RRnpc, RRnpc, RRnpc), smlal, t_mlal), |
| 19557 | TCE("smlalbt", 14000c0, fbc00090, 4, (RRnpc, RRnpc, RRnpc, RRnpc), smlal, t_mlal), |
| 19558 | TCE("smlaltt", 14000e0, fbc000b0, 4, (RRnpc, RRnpc, RRnpc, RRnpc), smlal, t_mlal), |
| 19559 | |
| 19560 | TCE("smulbb", 1600080, fb10f000, 3, (RRnpc, RRnpc, RRnpc), smul, t_simd), |
| 19561 | TCE("smultb", 16000a0, fb10f020, 3, (RRnpc, RRnpc, RRnpc), smul, t_simd), |
| 19562 | TCE("smulbt", 16000c0, fb10f010, 3, (RRnpc, RRnpc, RRnpc), smul, t_simd), |
| 19563 | TCE("smultt", 16000e0, fb10f030, 3, (RRnpc, RRnpc, RRnpc), smul, t_simd), |
| 19564 | |
| 19565 | TCE("smulwb", 12000a0, fb30f000, 3, (RRnpc, RRnpc, RRnpc), smul, t_simd), |
| 19566 | TCE("smulwt", 12000e0, fb30f010, 3, (RRnpc, RRnpc, RRnpc), smul, t_simd), |
| 19567 | |
| 19568 | TCE("qadd", 1000050, fa80f080, 3, (RRnpc, RRnpc, RRnpc), rd_rm_rn, t_simd2), |
| 19569 | TCE("qdadd", 1400050, fa80f090, 3, (RRnpc, RRnpc, RRnpc), rd_rm_rn, t_simd2), |
| 19570 | TCE("qsub", 1200050, fa80f0a0, 3, (RRnpc, RRnpc, RRnpc), rd_rm_rn, t_simd2), |
| 19571 | TCE("qdsub", 1600050, fa80f0b0, 3, (RRnpc, RRnpc, RRnpc), rd_rm_rn, t_simd2), |
| 19572 | |
| 19573 | #undef ARM_VARIANT |
| 19574 | #define ARM_VARIANT & arm_ext_v5e /* ARM Architecture 5TE. */ |
| 19575 | #undef THUMB_VARIANT |
| 19576 | #define THUMB_VARIANT & arm_ext_v6t2 |
| 19577 | |
| 19578 | TUF("pld", 450f000, f810f000, 1, (ADDR), pld, t_pld), |
| 19579 | TC3("ldrd", 00000d0, e8500000, 3, (RRnpc_npcsp, oRRnpc_npcsp, ADDRGLDRS), |
| 19580 | ldrd, t_ldstd), |
| 19581 | TC3("strd", 00000f0, e8400000, 3, (RRnpc_npcsp, oRRnpc_npcsp, |
| 19582 | ADDRGLDRS), ldrd, t_ldstd), |
| 19583 | |
| 19584 | TCE("mcrr", c400000, ec400000, 5, (RCP, I15b, RRnpc, RRnpc, RCN), co_reg2c, co_reg2c), |
| 19585 | TCE("mrrc", c500000, ec500000, 5, (RCP, I15b, RRnpc, RRnpc, RCN), co_reg2c, co_reg2c), |
| 19586 | |
| 19587 | #undef ARM_VARIANT |
| 19588 | #define ARM_VARIANT & arm_ext_v5j /* ARM Architecture 5TEJ. */ |
| 19589 | |
| 19590 | TCE("bxj", 12fff20, f3c08f00, 1, (RR), bxj, t_bxj), |
| 19591 | |
| 19592 | #undef ARM_VARIANT |
| 19593 | #define ARM_VARIANT & arm_ext_v6 /* ARM V6. */ |
| 19594 | #undef THUMB_VARIANT |
| 19595 | #define THUMB_VARIANT & arm_ext_v6 |
| 19596 | |
| 19597 | TUF("cpsie", 1080000, b660, 2, (CPSF, oI31b), cpsi, t_cpsi), |
| 19598 | TUF("cpsid", 10c0000, b670, 2, (CPSF, oI31b), cpsi, t_cpsi), |
| 19599 | tCE("rev", 6bf0f30, _rev, 2, (RRnpc, RRnpc), rd_rm, t_rev), |
| 19600 | tCE("rev16", 6bf0fb0, _rev16, 2, (RRnpc, RRnpc), rd_rm, t_rev), |
| 19601 | tCE("revsh", 6ff0fb0, _revsh, 2, (RRnpc, RRnpc), rd_rm, t_rev), |
| 19602 | tCE("sxth", 6bf0070, _sxth, 3, (RRnpc, RRnpc, oROR), sxth, t_sxth), |
| 19603 | tCE("uxth", 6ff0070, _uxth, 3, (RRnpc, RRnpc, oROR), sxth, t_sxth), |
| 19604 | tCE("sxtb", 6af0070, _sxtb, 3, (RRnpc, RRnpc, oROR), sxth, t_sxth), |
| 19605 | tCE("uxtb", 6ef0070, _uxtb, 3, (RRnpc, RRnpc, oROR), sxth, t_sxth), |
| 19606 | TUF("setend", 1010000, b650, 1, (ENDI), setend, t_setend), |
| 19607 | |
| 19608 | #undef THUMB_VARIANT |
| 19609 | #define THUMB_VARIANT & arm_ext_v6t2_v8m |
| 19610 | |
| 19611 | TCE("ldrex", 1900f9f, e8500f00, 2, (RRnpc_npcsp, ADDR), ldrex, t_ldrex), |
| 19612 | TCE("strex", 1800f90, e8400000, 3, (RRnpc_npcsp, RRnpc_npcsp, ADDR), |
| 19613 | strex, t_strex), |
| 19614 | #undef THUMB_VARIANT |
| 19615 | #define THUMB_VARIANT & arm_ext_v6t2 |
| 19616 | |
| 19617 | TUF("mcrr2", c400000, fc400000, 5, (RCP, I15b, RRnpc, RRnpc, RCN), co_reg2c, co_reg2c), |
| 19618 | TUF("mrrc2", c500000, fc500000, 5, (RCP, I15b, RRnpc, RRnpc, RCN), co_reg2c, co_reg2c), |
| 19619 | |
| 19620 | TCE("ssat", 6a00010, f3000000, 4, (RRnpc, I32, RRnpc, oSHllar),ssat, t_ssat), |
| 19621 | TCE("usat", 6e00010, f3800000, 4, (RRnpc, I31, RRnpc, oSHllar),usat, t_usat), |
| 19622 | |
| 19623 | /* ARM V6 not included in V7M. */ |
| 19624 | #undef THUMB_VARIANT |
| 19625 | #define THUMB_VARIANT & arm_ext_v6_notm |
| 19626 | TUF("rfeia", 8900a00, e990c000, 1, (RRw), rfe, rfe), |
| 19627 | TUF("rfe", 8900a00, e990c000, 1, (RRw), rfe, rfe), |
| 19628 | UF(rfeib, 9900a00, 1, (RRw), rfe), |
| 19629 | UF(rfeda, 8100a00, 1, (RRw), rfe), |
| 19630 | TUF("rfedb", 9100a00, e810c000, 1, (RRw), rfe, rfe), |
| 19631 | TUF("rfefd", 8900a00, e990c000, 1, (RRw), rfe, rfe), |
| 19632 | UF(rfefa, 8100a00, 1, (RRw), rfe), |
| 19633 | TUF("rfeea", 9100a00, e810c000, 1, (RRw), rfe, rfe), |
| 19634 | UF(rfeed, 9900a00, 1, (RRw), rfe), |
| 19635 | TUF("srsia", 8c00500, e980c000, 2, (oRRw, I31w), srs, srs), |
| 19636 | TUF("srs", 8c00500, e980c000, 2, (oRRw, I31w), srs, srs), |
| 19637 | TUF("srsea", 8c00500, e980c000, 2, (oRRw, I31w), srs, srs), |
| 19638 | UF(srsib, 9c00500, 2, (oRRw, I31w), srs), |
| 19639 | UF(srsfa, 9c00500, 2, (oRRw, I31w), srs), |
| 19640 | UF(srsda, 8400500, 2, (oRRw, I31w), srs), |
| 19641 | UF(srsed, 8400500, 2, (oRRw, I31w), srs), |
| 19642 | TUF("srsdb", 9400500, e800c000, 2, (oRRw, I31w), srs, srs), |
| 19643 | TUF("srsfd", 9400500, e800c000, 2, (oRRw, I31w), srs, srs), |
| 19644 | TUF("cps", 1020000, f3af8100, 1, (I31b), imm0, t_cps), |
| 19645 | |
| 19646 | /* ARM V6 not included in V7M (eg. integer SIMD). */ |
| 19647 | #undef THUMB_VARIANT |
| 19648 | #define THUMB_VARIANT & arm_ext_v6_dsp |
| 19649 | TCE("pkhbt", 6800010, eac00000, 4, (RRnpc, RRnpc, RRnpc, oSHll), pkhbt, t_pkhbt), |
| 19650 | TCE("pkhtb", 6800050, eac00020, 4, (RRnpc, RRnpc, RRnpc, oSHar), pkhtb, t_pkhtb), |
| 19651 | TCE("qadd16", 6200f10, fa90f010, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19652 | TCE("qadd8", 6200f90, fa80f010, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19653 | TCE("qasx", 6200f30, faa0f010, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19654 | /* Old name for QASX. */ |
| 19655 | TCE("qaddsubx",6200f30, faa0f010, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19656 | TCE("qsax", 6200f50, fae0f010, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19657 | /* Old name for QSAX. */ |
| 19658 | TCE("qsubaddx",6200f50, fae0f010, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19659 | TCE("qsub16", 6200f70, fad0f010, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19660 | TCE("qsub8", 6200ff0, fac0f010, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19661 | TCE("sadd16", 6100f10, fa90f000, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19662 | TCE("sadd8", 6100f90, fa80f000, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19663 | TCE("sasx", 6100f30, faa0f000, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19664 | /* Old name for SASX. */ |
| 19665 | TCE("saddsubx",6100f30, faa0f000, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19666 | TCE("shadd16", 6300f10, fa90f020, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19667 | TCE("shadd8", 6300f90, fa80f020, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19668 | TCE("shasx", 6300f30, faa0f020, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19669 | /* Old name for SHASX. */ |
| 19670 | TCE("shaddsubx", 6300f30, faa0f020, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19671 | TCE("shsax", 6300f50, fae0f020, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19672 | /* Old name for SHSAX. */ |
| 19673 | TCE("shsubaddx", 6300f50, fae0f020, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19674 | TCE("shsub16", 6300f70, fad0f020, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19675 | TCE("shsub8", 6300ff0, fac0f020, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19676 | TCE("ssax", 6100f50, fae0f000, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19677 | /* Old name for SSAX. */ |
| 19678 | TCE("ssubaddx",6100f50, fae0f000, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19679 | TCE("ssub16", 6100f70, fad0f000, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19680 | TCE("ssub8", 6100ff0, fac0f000, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19681 | TCE("uadd16", 6500f10, fa90f040, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19682 | TCE("uadd8", 6500f90, fa80f040, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19683 | TCE("uasx", 6500f30, faa0f040, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19684 | /* Old name for UASX. */ |
| 19685 | TCE("uaddsubx",6500f30, faa0f040, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19686 | TCE("uhadd16", 6700f10, fa90f060, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19687 | TCE("uhadd8", 6700f90, fa80f060, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19688 | TCE("uhasx", 6700f30, faa0f060, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19689 | /* Old name for UHASX. */ |
| 19690 | TCE("uhaddsubx", 6700f30, faa0f060, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19691 | TCE("uhsax", 6700f50, fae0f060, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19692 | /* Old name for UHSAX. */ |
| 19693 | TCE("uhsubaddx", 6700f50, fae0f060, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19694 | TCE("uhsub16", 6700f70, fad0f060, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19695 | TCE("uhsub8", 6700ff0, fac0f060, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19696 | TCE("uqadd16", 6600f10, fa90f050, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19697 | TCE("uqadd8", 6600f90, fa80f050, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19698 | TCE("uqasx", 6600f30, faa0f050, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19699 | /* Old name for UQASX. */ |
| 19700 | TCE("uqaddsubx", 6600f30, faa0f050, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19701 | TCE("uqsax", 6600f50, fae0f050, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19702 | /* Old name for UQSAX. */ |
| 19703 | TCE("uqsubaddx", 6600f50, fae0f050, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19704 | TCE("uqsub16", 6600f70, fad0f050, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19705 | TCE("uqsub8", 6600ff0, fac0f050, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19706 | TCE("usub16", 6500f70, fad0f040, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19707 | TCE("usax", 6500f50, fae0f040, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19708 | /* Old name for USAX. */ |
| 19709 | TCE("usubaddx",6500f50, fae0f040, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19710 | TCE("usub8", 6500ff0, fac0f040, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19711 | TCE("sxtah", 6b00070, fa00f080, 4, (RRnpc, RRnpc, RRnpc, oROR), sxtah, t_sxtah), |
| 19712 | TCE("sxtab16", 6800070, fa20f080, 4, (RRnpc, RRnpc, RRnpc, oROR), sxtah, t_sxtah), |
| 19713 | TCE("sxtab", 6a00070, fa40f080, 4, (RRnpc, RRnpc, RRnpc, oROR), sxtah, t_sxtah), |
| 19714 | TCE("sxtb16", 68f0070, fa2ff080, 3, (RRnpc, RRnpc, oROR), sxth, t_sxth), |
| 19715 | TCE("uxtah", 6f00070, fa10f080, 4, (RRnpc, RRnpc, RRnpc, oROR), sxtah, t_sxtah), |
| 19716 | TCE("uxtab16", 6c00070, fa30f080, 4, (RRnpc, RRnpc, RRnpc, oROR), sxtah, t_sxtah), |
| 19717 | TCE("uxtab", 6e00070, fa50f080, 4, (RRnpc, RRnpc, RRnpc, oROR), sxtah, t_sxtah), |
| 19718 | TCE("uxtb16", 6cf0070, fa3ff080, 3, (RRnpc, RRnpc, oROR), sxth, t_sxth), |
| 19719 | TCE("sel", 6800fb0, faa0f080, 3, (RRnpc, RRnpc, RRnpc), rd_rn_rm, t_simd), |
| 19720 | TCE("smlad", 7000010, fb200000, 4, (RRnpc, RRnpc, RRnpc, RRnpc),smla, t_mla), |
| 19721 | TCE("smladx", 7000030, fb200010, 4, (RRnpc, RRnpc, RRnpc, RRnpc),smla, t_mla), |
| 19722 | TCE("smlald", 7400010, fbc000c0, 4, (RRnpc, RRnpc, RRnpc, RRnpc),smlal,t_mlal), |
| 19723 | TCE("smlaldx", 7400030, fbc000d0, 4, (RRnpc, RRnpc, RRnpc, RRnpc),smlal,t_mlal), |
| 19724 | TCE("smlsd", 7000050, fb400000, 4, (RRnpc, RRnpc, RRnpc, RRnpc),smla, t_mla), |
| 19725 | TCE("smlsdx", 7000070, fb400010, 4, (RRnpc, RRnpc, RRnpc, RRnpc),smla, t_mla), |
| 19726 | TCE("smlsld", 7400050, fbd000c0, 4, (RRnpc, RRnpc, RRnpc, RRnpc),smlal,t_mlal), |
| 19727 | TCE("smlsldx", 7400070, fbd000d0, 4, (RRnpc, RRnpc, RRnpc, RRnpc),smlal,t_mlal), |
| 19728 | TCE("smmla", 7500010, fb500000, 4, (RRnpc, RRnpc, RRnpc, RRnpc),smla, t_mla), |
| 19729 | TCE("smmlar", 7500030, fb500010, 4, (RRnpc, RRnpc, RRnpc, RRnpc),smla, t_mla), |
| 19730 | TCE("smmls", 75000d0, fb600000, 4, (RRnpc, RRnpc, RRnpc, RRnpc),smla, t_mla), |
| 19731 | TCE("smmlsr", 75000f0, fb600010, 4, (RRnpc, RRnpc, RRnpc, RRnpc),smla, t_mla), |
| 19732 | TCE("smmul", 750f010, fb50f000, 3, (RRnpc, RRnpc, RRnpc), smul, t_simd), |
| 19733 | TCE("smmulr", 750f030, fb50f010, 3, (RRnpc, RRnpc, RRnpc), smul, t_simd), |
| 19734 | TCE("smuad", 700f010, fb20f000, 3, (RRnpc, RRnpc, RRnpc), smul, t_simd), |
| 19735 | TCE("smuadx", 700f030, fb20f010, 3, (RRnpc, RRnpc, RRnpc), smul, t_simd), |
| 19736 | TCE("smusd", 700f050, fb40f000, 3, (RRnpc, RRnpc, RRnpc), smul, t_simd), |
| 19737 | TCE("smusdx", 700f070, fb40f010, 3, (RRnpc, RRnpc, RRnpc), smul, t_simd), |
| 19738 | TCE("ssat16", 6a00f30, f3200000, 3, (RRnpc, I16, RRnpc), ssat16, t_ssat16), |
| 19739 | TCE("umaal", 0400090, fbe00060, 4, (RRnpc, RRnpc, RRnpc, RRnpc),smlal, t_mlal), |
| 19740 | TCE("usad8", 780f010, fb70f000, 3, (RRnpc, RRnpc, RRnpc), smul, t_simd), |
| 19741 | TCE("usada8", 7800010, fb700000, 4, (RRnpc, RRnpc, RRnpc, RRnpc),smla, t_mla), |
| 19742 | TCE("usat16", 6e00f30, f3a00000, 3, (RRnpc, I15, RRnpc), usat16, t_usat16), |
| 19743 | |
| 19744 | #undef ARM_VARIANT |
| 19745 | #define ARM_VARIANT & arm_ext_v6k |
| 19746 | #undef THUMB_VARIANT |
| 19747 | #define THUMB_VARIANT & arm_ext_v6k |
| 19748 | |
| 19749 | tCE("yield", 320f001, _yield, 0, (), noargs, t_hint), |
| 19750 | tCE("wfe", 320f002, _wfe, 0, (), noargs, t_hint), |
| 19751 | tCE("wfi", 320f003, _wfi, 0, (), noargs, t_hint), |
| 19752 | tCE("sev", 320f004, _sev, 0, (), noargs, t_hint), |
| 19753 | |
| 19754 | #undef THUMB_VARIANT |
| 19755 | #define THUMB_VARIANT & arm_ext_v6_notm |
| 19756 | TCE("ldrexd", 1b00f9f, e8d0007f, 3, (RRnpc_npcsp, oRRnpc_npcsp, RRnpcb), |
| 19757 | ldrexd, t_ldrexd), |
| 19758 | TCE("strexd", 1a00f90, e8c00070, 4, (RRnpc_npcsp, RRnpc_npcsp, oRRnpc_npcsp, |
| 19759 | RRnpcb), strexd, t_strexd), |
| 19760 | |
| 19761 | #undef THUMB_VARIANT |
| 19762 | #define THUMB_VARIANT & arm_ext_v6t2_v8m |
| 19763 | TCE("ldrexb", 1d00f9f, e8d00f4f, 2, (RRnpc_npcsp,RRnpcb), |
| 19764 | rd_rn, rd_rn), |
| 19765 | TCE("ldrexh", 1f00f9f, e8d00f5f, 2, (RRnpc_npcsp, RRnpcb), |
| 19766 | rd_rn, rd_rn), |
| 19767 | TCE("strexb", 1c00f90, e8c00f40, 3, (RRnpc_npcsp, RRnpc_npcsp, ADDR), |
| 19768 | strex, t_strexbh), |
| 19769 | TCE("strexh", 1e00f90, e8c00f50, 3, (RRnpc_npcsp, RRnpc_npcsp, ADDR), |
| 19770 | strex, t_strexbh), |
| 19771 | TUF("clrex", 57ff01f, f3bf8f2f, 0, (), noargs, noargs), |
| 19772 | |
| 19773 | #undef ARM_VARIANT |
| 19774 | #define ARM_VARIANT & arm_ext_sec |
| 19775 | #undef THUMB_VARIANT |
| 19776 | #define THUMB_VARIANT & arm_ext_sec |
| 19777 | |
| 19778 | TCE("smc", 1600070, f7f08000, 1, (EXPi), smc, t_smc), |
| 19779 | |
| 19780 | #undef ARM_VARIANT |
| 19781 | #define ARM_VARIANT & arm_ext_virt |
| 19782 | #undef THUMB_VARIANT |
| 19783 | #define THUMB_VARIANT & arm_ext_virt |
| 19784 | |
| 19785 | TCE("hvc", 1400070, f7e08000, 1, (EXPi), hvc, t_hvc), |
| 19786 | TCE("eret", 160006e, f3de8f00, 0, (), noargs, noargs), |
| 19787 | |
| 19788 | #undef ARM_VARIANT |
| 19789 | #define ARM_VARIANT & arm_ext_pan |
| 19790 | #undef THUMB_VARIANT |
| 19791 | #define THUMB_VARIANT & arm_ext_pan |
| 19792 | |
| 19793 | TUF("setpan", 1100000, b610, 1, (I7), setpan, t_setpan), |
| 19794 | |
| 19795 | #undef ARM_VARIANT |
| 19796 | #define ARM_VARIANT & arm_ext_v6t2 |
| 19797 | #undef THUMB_VARIANT |
| 19798 | #define THUMB_VARIANT & arm_ext_v6t2 |
| 19799 | |
| 19800 | TCE("bfc", 7c0001f, f36f0000, 3, (RRnpc, I31, I32), bfc, t_bfc), |
| 19801 | TCE("bfi", 7c00010, f3600000, 4, (RRnpc, RRnpc_I0, I31, I32), bfi, t_bfi), |
| 19802 | TCE("sbfx", 7a00050, f3400000, 4, (RR, RR, I31, I32), bfx, t_bfx), |
| 19803 | TCE("ubfx", 7e00050, f3c00000, 4, (RR, RR, I31, I32), bfx, t_bfx), |
| 19804 | |
| 19805 | TCE("mls", 0600090, fb000010, 4, (RRnpc, RRnpc, RRnpc, RRnpc), mlas, t_mla), |
| 19806 | TCE("rbit", 6ff0f30, fa90f0a0, 2, (RR, RR), rd_rm, t_rbit), |
| 19807 | |
| 19808 | TC3("ldrht", 03000b0, f8300e00, 2, (RRnpc_npcsp, ADDR), ldsttv4, t_ldstt), |
| 19809 | TC3("ldrsht", 03000f0, f9300e00, 2, (RRnpc_npcsp, ADDR), ldsttv4, t_ldstt), |
| 19810 | TC3("ldrsbt", 03000d0, f9100e00, 2, (RRnpc_npcsp, ADDR), ldsttv4, t_ldstt), |
| 19811 | TC3("strht", 02000b0, f8200e00, 2, (RRnpc_npcsp, ADDR), ldsttv4, t_ldstt), |
| 19812 | |
| 19813 | #undef THUMB_VARIANT |
| 19814 | #define THUMB_VARIANT & arm_ext_v6t2_v8m |
| 19815 | TCE("movw", 3000000, f2400000, 2, (RRnpc, HALF), mov16, t_mov16), |
| 19816 | TCE("movt", 3400000, f2c00000, 2, (RRnpc, HALF), mov16, t_mov16), |
| 19817 | |
| 19818 | /* Thumb-only instructions. */ |
| 19819 | #undef ARM_VARIANT |
| 19820 | #define ARM_VARIANT NULL |
| 19821 | TUE("cbnz", 0, b900, 2, (RR, EXP), 0, t_cbz), |
| 19822 | TUE("cbz", 0, b100, 2, (RR, EXP), 0, t_cbz), |
| 19823 | |
| 19824 | /* ARM does not really have an IT instruction, so always allow it. |
| 19825 | The opcode is copied from Thumb in order to allow warnings in |
| 19826 | -mimplicit-it=[never | arm] modes. */ |
| 19827 | #undef ARM_VARIANT |
| 19828 | #define ARM_VARIANT & arm_ext_v1 |
| 19829 | #undef THUMB_VARIANT |
| 19830 | #define THUMB_VARIANT & arm_ext_v6t2 |
| 19831 | |
| 19832 | TUE("it", bf08, bf08, 1, (COND), it, t_it), |
| 19833 | TUE("itt", bf0c, bf0c, 1, (COND), it, t_it), |
| 19834 | TUE("ite", bf04, bf04, 1, (COND), it, t_it), |
| 19835 | TUE("ittt", bf0e, bf0e, 1, (COND), it, t_it), |
| 19836 | TUE("itet", bf06, bf06, 1, (COND), it, t_it), |
| 19837 | TUE("itte", bf0a, bf0a, 1, (COND), it, t_it), |
| 19838 | TUE("itee", bf02, bf02, 1, (COND), it, t_it), |
| 19839 | TUE("itttt", bf0f, bf0f, 1, (COND), it, t_it), |
| 19840 | TUE("itett", bf07, bf07, 1, (COND), it, t_it), |
| 19841 | TUE("ittet", bf0b, bf0b, 1, (COND), it, t_it), |
| 19842 | TUE("iteet", bf03, bf03, 1, (COND), it, t_it), |
| 19843 | TUE("ittte", bf0d, bf0d, 1, (COND), it, t_it), |
| 19844 | TUE("itete", bf05, bf05, 1, (COND), it, t_it), |
| 19845 | TUE("ittee", bf09, bf09, 1, (COND), it, t_it), |
| 19846 | TUE("iteee", bf01, bf01, 1, (COND), it, t_it), |
| 19847 | /* ARM/Thumb-2 instructions with no Thumb-1 equivalent. */ |
| 19848 | TC3("rrx", 01a00060, ea4f0030, 2, (RR, RR), rd_rm, t_rrx), |
| 19849 | TC3("rrxs", 01b00060, ea5f0030, 2, (RR, RR), rd_rm, t_rrx), |
| 19850 | |
| 19851 | /* Thumb2 only instructions. */ |
| 19852 | #undef ARM_VARIANT |
| 19853 | #define ARM_VARIANT NULL |
| 19854 | |
| 19855 | TCE("addw", 0, f2000000, 3, (RR, RR, EXPi), 0, t_add_sub_w), |
| 19856 | TCE("subw", 0, f2a00000, 3, (RR, RR, EXPi), 0, t_add_sub_w), |
| 19857 | TCE("orn", 0, ea600000, 3, (RR, oRR, SH), 0, t_orn), |
| 19858 | TCE("orns", 0, ea700000, 3, (RR, oRR, SH), 0, t_orn), |
| 19859 | TCE("tbb", 0, e8d0f000, 1, (TB), 0, t_tb), |
| 19860 | TCE("tbh", 0, e8d0f010, 1, (TB), 0, t_tb), |
| 19861 | |
| 19862 | /* Hardware division instructions. */ |
| 19863 | #undef ARM_VARIANT |
| 19864 | #define ARM_VARIANT & arm_ext_adiv |
| 19865 | #undef THUMB_VARIANT |
| 19866 | #define THUMB_VARIANT & arm_ext_div |
| 19867 | |
| 19868 | TCE("sdiv", 710f010, fb90f0f0, 3, (RR, oRR, RR), div, t_div), |
| 19869 | TCE("udiv", 730f010, fbb0f0f0, 3, (RR, oRR, RR), div, t_div), |
| 19870 | |
| 19871 | /* ARM V6M/V7 instructions. */ |
| 19872 | #undef ARM_VARIANT |
| 19873 | #define ARM_VARIANT & arm_ext_barrier |
| 19874 | #undef THUMB_VARIANT |
| 19875 | #define THUMB_VARIANT & arm_ext_barrier |
| 19876 | |
| 19877 | TUF("dmb", 57ff050, f3bf8f50, 1, (oBARRIER_I15), barrier, barrier), |
| 19878 | TUF("dsb", 57ff040, f3bf8f40, 1, (oBARRIER_I15), barrier, barrier), |
| 19879 | TUF("isb", 57ff060, f3bf8f60, 1, (oBARRIER_I15), barrier, barrier), |
| 19880 | |
| 19881 | /* ARM V7 instructions. */ |
| 19882 | #undef ARM_VARIANT |
| 19883 | #define ARM_VARIANT & arm_ext_v7 |
| 19884 | #undef THUMB_VARIANT |
| 19885 | #define THUMB_VARIANT & arm_ext_v7 |
| 19886 | |
| 19887 | TUF("pli", 450f000, f910f000, 1, (ADDR), pli, t_pld), |
| 19888 | TCE("dbg", 320f0f0, f3af80f0, 1, (I15), dbg, t_dbg), |
| 19889 | |
| 19890 | #undef ARM_VARIANT |
| 19891 | #define ARM_VARIANT & arm_ext_mp |
| 19892 | #undef THUMB_VARIANT |
| 19893 | #define THUMB_VARIANT & arm_ext_mp |
| 19894 | |
| 19895 | TUF("pldw", 410f000, f830f000, 1, (ADDR), pld, t_pld), |
| 19896 | |
| 19897 | /* AArchv8 instructions. */ |
| 19898 | #undef ARM_VARIANT |
| 19899 | #define ARM_VARIANT & arm_ext_v8 |
| 19900 | |
| 19901 | /* Instructions shared between armv8-a and armv8-m. */ |
| 19902 | #undef THUMB_VARIANT |
| 19903 | #define THUMB_VARIANT & arm_ext_atomics |
| 19904 | |
| 19905 | TCE("lda", 1900c9f, e8d00faf, 2, (RRnpc, RRnpcb), rd_rn, rd_rn), |
| 19906 | TCE("ldab", 1d00c9f, e8d00f8f, 2, (RRnpc, RRnpcb), rd_rn, rd_rn), |
| 19907 | TCE("ldah", 1f00c9f, e8d00f9f, 2, (RRnpc, RRnpcb), rd_rn, rd_rn), |
| 19908 | TCE("stl", 180fc90, e8c00faf, 2, (RRnpc, RRnpcb), rm_rn, rd_rn), |
| 19909 | TCE("stlb", 1c0fc90, e8c00f8f, 2, (RRnpc, RRnpcb), rm_rn, rd_rn), |
| 19910 | TCE("stlh", 1e0fc90, e8c00f9f, 2, (RRnpc, RRnpcb), rm_rn, rd_rn), |
| 19911 | TCE("ldaex", 1900e9f, e8d00fef, 2, (RRnpc, RRnpcb), rd_rn, rd_rn), |
| 19912 | TCE("ldaexb", 1d00e9f, e8d00fcf, 2, (RRnpc,RRnpcb), rd_rn, rd_rn), |
| 19913 | TCE("ldaexh", 1f00e9f, e8d00fdf, 2, (RRnpc, RRnpcb), rd_rn, rd_rn), |
| 19914 | TCE("stlex", 1800e90, e8c00fe0, 3, (RRnpc, RRnpc, RRnpcb), |
| 19915 | stlex, t_stlex), |
| 19916 | TCE("stlexb", 1c00e90, e8c00fc0, 3, (RRnpc, RRnpc, RRnpcb), |
| 19917 | stlex, t_stlex), |
| 19918 | TCE("stlexh", 1e00e90, e8c00fd0, 3, (RRnpc, RRnpc, RRnpcb), |
| 19919 | stlex, t_stlex), |
| 19920 | #undef THUMB_VARIANT |
| 19921 | #define THUMB_VARIANT & arm_ext_v8 |
| 19922 | |
| 19923 | tCE("sevl", 320f005, _sevl, 0, (), noargs, t_hint), |
| 19924 | TUE("hlt", 1000070, ba80, 1, (oIffffb), bkpt, t_hlt), |
| 19925 | TCE("ldaexd", 1b00e9f, e8d000ff, 3, (RRnpc, oRRnpc, RRnpcb), |
| 19926 | ldrexd, t_ldrexd), |
| 19927 | TCE("stlexd", 1a00e90, e8c000f0, 4, (RRnpc, RRnpc, oRRnpc, RRnpcb), |
| 19928 | strexd, t_strexd), |
| 19929 | /* ARMv8 T32 only. */ |
| 19930 | #undef ARM_VARIANT |
| 19931 | #define ARM_VARIANT NULL |
| 19932 | TUF("dcps1", 0, f78f8001, 0, (), noargs, noargs), |
| 19933 | TUF("dcps2", 0, f78f8002, 0, (), noargs, noargs), |
| 19934 | TUF("dcps3", 0, f78f8003, 0, (), noargs, noargs), |
| 19935 | |
| 19936 | /* FP for ARMv8. */ |
| 19937 | #undef ARM_VARIANT |
| 19938 | #define ARM_VARIANT & fpu_vfp_ext_armv8xd |
| 19939 | #undef THUMB_VARIANT |
| 19940 | #define THUMB_VARIANT & fpu_vfp_ext_armv8xd |
| 19941 | |
| 19942 | nUF(vseleq, _vseleq, 3, (RVSD, RVSD, RVSD), vsel), |
| 19943 | nUF(vselvs, _vselvs, 3, (RVSD, RVSD, RVSD), vsel), |
| 19944 | nUF(vselge, _vselge, 3, (RVSD, RVSD, RVSD), vsel), |
| 19945 | nUF(vselgt, _vselgt, 3, (RVSD, RVSD, RVSD), vsel), |
| 19946 | nUF(vmaxnm, _vmaxnm, 3, (RNSDQ, oRNSDQ, RNSDQ), vmaxnm), |
| 19947 | nUF(vminnm, _vminnm, 3, (RNSDQ, oRNSDQ, RNSDQ), vmaxnm), |
| 19948 | nUF(vcvta, _vcvta, 2, (RNSDQ, oRNSDQ), neon_cvta), |
| 19949 | nUF(vcvtn, _vcvta, 2, (RNSDQ, oRNSDQ), neon_cvtn), |
| 19950 | nUF(vcvtp, _vcvta, 2, (RNSDQ, oRNSDQ), neon_cvtp), |
| 19951 | nUF(vcvtm, _vcvta, 2, (RNSDQ, oRNSDQ), neon_cvtm), |
| 19952 | nCE(vrintr, _vrintr, 2, (RNSDQ, oRNSDQ), vrintr), |
| 19953 | nCE(vrintz, _vrintr, 2, (RNSDQ, oRNSDQ), vrintz), |
| 19954 | nCE(vrintx, _vrintr, 2, (RNSDQ, oRNSDQ), vrintx), |
| 19955 | nUF(vrinta, _vrinta, 2, (RNSDQ, oRNSDQ), vrinta), |
| 19956 | nUF(vrintn, _vrinta, 2, (RNSDQ, oRNSDQ), vrintn), |
| 19957 | nUF(vrintp, _vrinta, 2, (RNSDQ, oRNSDQ), vrintp), |
| 19958 | nUF(vrintm, _vrinta, 2, (RNSDQ, oRNSDQ), vrintm), |
| 19959 | |
| 19960 | /* Crypto v1 extensions. */ |
| 19961 | #undef ARM_VARIANT |
| 19962 | #define ARM_VARIANT & fpu_crypto_ext_armv8 |
| 19963 | #undef THUMB_VARIANT |
| 19964 | #define THUMB_VARIANT & fpu_crypto_ext_armv8 |
| 19965 | |
| 19966 | nUF(aese, _aes, 2, (RNQ, RNQ), aese), |
| 19967 | nUF(aesd, _aes, 2, (RNQ, RNQ), aesd), |
| 19968 | nUF(aesmc, _aes, 2, (RNQ, RNQ), aesmc), |
| 19969 | nUF(aesimc, _aes, 2, (RNQ, RNQ), aesimc), |
| 19970 | nUF(sha1c, _sha3op, 3, (RNQ, RNQ, RNQ), sha1c), |
| 19971 | nUF(sha1p, _sha3op, 3, (RNQ, RNQ, RNQ), sha1p), |
| 19972 | nUF(sha1m, _sha3op, 3, (RNQ, RNQ, RNQ), sha1m), |
| 19973 | nUF(sha1su0, _sha3op, 3, (RNQ, RNQ, RNQ), sha1su0), |
| 19974 | nUF(sha256h, _sha3op, 3, (RNQ, RNQ, RNQ), sha256h), |
| 19975 | nUF(sha256h2, _sha3op, 3, (RNQ, RNQ, RNQ), sha256h2), |
| 19976 | nUF(sha256su1, _sha3op, 3, (RNQ, RNQ, RNQ), sha256su1), |
| 19977 | nUF(sha1h, _sha1h, 2, (RNQ, RNQ), sha1h), |
| 19978 | nUF(sha1su1, _sha2op, 2, (RNQ, RNQ), sha1su1), |
| 19979 | nUF(sha256su0, _sha2op, 2, (RNQ, RNQ), sha256su0), |
| 19980 | |
| 19981 | #undef ARM_VARIANT |
| 19982 | #define ARM_VARIANT & crc_ext_armv8 |
| 19983 | #undef THUMB_VARIANT |
| 19984 | #define THUMB_VARIANT & crc_ext_armv8 |
| 19985 | TUEc("crc32b", 1000040, fac0f080, 3, (RR, oRR, RR), crc32b), |
| 19986 | TUEc("crc32h", 1200040, fac0f090, 3, (RR, oRR, RR), crc32h), |
| 19987 | TUEc("crc32w", 1400040, fac0f0a0, 3, (RR, oRR, RR), crc32w), |
| 19988 | TUEc("crc32cb",1000240, fad0f080, 3, (RR, oRR, RR), crc32cb), |
| 19989 | TUEc("crc32ch",1200240, fad0f090, 3, (RR, oRR, RR), crc32ch), |
| 19990 | TUEc("crc32cw",1400240, fad0f0a0, 3, (RR, oRR, RR), crc32cw), |
| 19991 | |
| 19992 | /* ARMv8.2 RAS extension. */ |
| 19993 | #undef ARM_VARIANT |
| 19994 | #define ARM_VARIANT & arm_ext_ras |
| 19995 | #undef THUMB_VARIANT |
| 19996 | #define THUMB_VARIANT & arm_ext_ras |
| 19997 | TUE ("esb", 320f010, f3af8010, 0, (), noargs, noargs), |
| 19998 | |
| 19999 | #undef ARM_VARIANT |
| 20000 | #define ARM_VARIANT & arm_ext_v8_3 |
| 20001 | #undef THUMB_VARIANT |
| 20002 | #define THUMB_VARIANT & arm_ext_v8_3 |
| 20003 | NCE (vjcvt, eb90bc0, 2, (RVS, RVD), vjcvt), |
| 20004 | NUF (vcmla, 0, 4, (RNDQ, RNDQ, RNDQ_RNSC, EXPi), vcmla), |
| 20005 | NUF (vcadd, 0, 4, (RNDQ, RNDQ, RNDQ, EXPi), vcadd), |
| 20006 | |
| 20007 | #undef ARM_VARIANT |
| 20008 | #define ARM_VARIANT & fpu_neon_ext_dotprod |
| 20009 | #undef THUMB_VARIANT |
| 20010 | #define THUMB_VARIANT & fpu_neon_ext_dotprod |
| 20011 | NUF (vsdot, d00, 3, (RNDQ, RNDQ, RNDQ_RNSC), neon_dotproduct_s), |
| 20012 | NUF (vudot, d00, 3, (RNDQ, RNDQ, RNDQ_RNSC), neon_dotproduct_u), |
| 20013 | |
| 20014 | #undef ARM_VARIANT |
| 20015 | #define ARM_VARIANT & fpu_fpa_ext_v1 /* Core FPA instruction set (V1). */ |
| 20016 | #undef THUMB_VARIANT |
| 20017 | #define THUMB_VARIANT NULL |
| 20018 | |
| 20019 | cCE("wfs", e200110, 1, (RR), rd), |
| 20020 | cCE("rfs", e300110, 1, (RR), rd), |
| 20021 | cCE("wfc", e400110, 1, (RR), rd), |
| 20022 | cCE("rfc", e500110, 1, (RR), rd), |
| 20023 | |
| 20024 | cCL("ldfs", c100100, 2, (RF, ADDRGLDC), rd_cpaddr), |
| 20025 | cCL("ldfd", c108100, 2, (RF, ADDRGLDC), rd_cpaddr), |
| 20026 | cCL("ldfe", c500100, 2, (RF, ADDRGLDC), rd_cpaddr), |
| 20027 | cCL("ldfp", c508100, 2, (RF, ADDRGLDC), rd_cpaddr), |
| 20028 | |
| 20029 | cCL("stfs", c000100, 2, (RF, ADDRGLDC), rd_cpaddr), |
| 20030 | cCL("stfd", c008100, 2, (RF, ADDRGLDC), rd_cpaddr), |
| 20031 | cCL("stfe", c400100, 2, (RF, ADDRGLDC), rd_cpaddr), |
| 20032 | cCL("stfp", c408100, 2, (RF, ADDRGLDC), rd_cpaddr), |
| 20033 | |
| 20034 | cCL("mvfs", e008100, 2, (RF, RF_IF), rd_rm), |
| 20035 | cCL("mvfsp", e008120, 2, (RF, RF_IF), rd_rm), |
| 20036 | cCL("mvfsm", e008140, 2, (RF, RF_IF), rd_rm), |
| 20037 | cCL("mvfsz", e008160, 2, (RF, RF_IF), rd_rm), |
| 20038 | cCL("mvfd", e008180, 2, (RF, RF_IF), rd_rm), |
| 20039 | cCL("mvfdp", e0081a0, 2, (RF, RF_IF), rd_rm), |
| 20040 | cCL("mvfdm", e0081c0, 2, (RF, RF_IF), rd_rm), |
| 20041 | cCL("mvfdz", e0081e0, 2, (RF, RF_IF), rd_rm), |
| 20042 | cCL("mvfe", e088100, 2, (RF, RF_IF), rd_rm), |
| 20043 | cCL("mvfep", e088120, 2, (RF, RF_IF), rd_rm), |
| 20044 | cCL("mvfem", e088140, 2, (RF, RF_IF), rd_rm), |
| 20045 | cCL("mvfez", e088160, 2, (RF, RF_IF), rd_rm), |
| 20046 | |
| 20047 | cCL("mnfs", e108100, 2, (RF, RF_IF), rd_rm), |
| 20048 | cCL("mnfsp", e108120, 2, (RF, RF_IF), rd_rm), |
| 20049 | cCL("mnfsm", e108140, 2, (RF, RF_IF), rd_rm), |
| 20050 | cCL("mnfsz", e108160, 2, (RF, RF_IF), rd_rm), |
| 20051 | cCL("mnfd", e108180, 2, (RF, RF_IF), rd_rm), |
| 20052 | cCL("mnfdp", e1081a0, 2, (RF, RF_IF), rd_rm), |
| 20053 | cCL("mnfdm", e1081c0, 2, (RF, RF_IF), rd_rm), |
| 20054 | cCL("mnfdz", e1081e0, 2, (RF, RF_IF), rd_rm), |
| 20055 | cCL("mnfe", e188100, 2, (RF, RF_IF), rd_rm), |
| 20056 | cCL("mnfep", e188120, 2, (RF, RF_IF), rd_rm), |
| 20057 | cCL("mnfem", e188140, 2, (RF, RF_IF), rd_rm), |
| 20058 | cCL("mnfez", e188160, 2, (RF, RF_IF), rd_rm), |
| 20059 | |
| 20060 | cCL("abss", e208100, 2, (RF, RF_IF), rd_rm), |
| 20061 | cCL("abssp", e208120, 2, (RF, RF_IF), rd_rm), |
| 20062 | cCL("abssm", e208140, 2, (RF, RF_IF), rd_rm), |
| 20063 | cCL("abssz", e208160, 2, (RF, RF_IF), rd_rm), |
| 20064 | cCL("absd", e208180, 2, (RF, RF_IF), rd_rm), |
| 20065 | cCL("absdp", e2081a0, 2, (RF, RF_IF), rd_rm), |
| 20066 | cCL("absdm", e2081c0, 2, (RF, RF_IF), rd_rm), |
| 20067 | cCL("absdz", e2081e0, 2, (RF, RF_IF), rd_rm), |
| 20068 | cCL("abse", e288100, 2, (RF, RF_IF), rd_rm), |
| 20069 | cCL("absep", e288120, 2, (RF, RF_IF), rd_rm), |
| 20070 | cCL("absem", e288140, 2, (RF, RF_IF), rd_rm), |
| 20071 | cCL("absez", e288160, 2, (RF, RF_IF), rd_rm), |
| 20072 | |
| 20073 | cCL("rnds", e308100, 2, (RF, RF_IF), rd_rm), |
| 20074 | cCL("rndsp", e308120, 2, (RF, RF_IF), rd_rm), |
| 20075 | cCL("rndsm", e308140, 2, (RF, RF_IF), rd_rm), |
| 20076 | cCL("rndsz", e308160, 2, (RF, RF_IF), rd_rm), |
| 20077 | cCL("rndd", e308180, 2, (RF, RF_IF), rd_rm), |
| 20078 | cCL("rnddp", e3081a0, 2, (RF, RF_IF), rd_rm), |
| 20079 | cCL("rnddm", e3081c0, 2, (RF, RF_IF), rd_rm), |
| 20080 | cCL("rnddz", e3081e0, 2, (RF, RF_IF), rd_rm), |
| 20081 | cCL("rnde", e388100, 2, (RF, RF_IF), rd_rm), |
| 20082 | cCL("rndep", e388120, 2, (RF, RF_IF), rd_rm), |
| 20083 | cCL("rndem", e388140, 2, (RF, RF_IF), rd_rm), |
| 20084 | cCL("rndez", e388160, 2, (RF, RF_IF), rd_rm), |
| 20085 | |
| 20086 | cCL("sqts", e408100, 2, (RF, RF_IF), rd_rm), |
| 20087 | cCL("sqtsp", e408120, 2, (RF, RF_IF), rd_rm), |
| 20088 | cCL("sqtsm", e408140, 2, (RF, RF_IF), rd_rm), |
| 20089 | cCL("sqtsz", e408160, 2, (RF, RF_IF), rd_rm), |
| 20090 | cCL("sqtd", e408180, 2, (RF, RF_IF), rd_rm), |
| 20091 | cCL("sqtdp", e4081a0, 2, (RF, RF_IF), rd_rm), |
| 20092 | cCL("sqtdm", e4081c0, 2, (RF, RF_IF), rd_rm), |
| 20093 | cCL("sqtdz", e4081e0, 2, (RF, RF_IF), rd_rm), |
| 20094 | cCL("sqte", e488100, 2, (RF, RF_IF), rd_rm), |
| 20095 | cCL("sqtep", e488120, 2, (RF, RF_IF), rd_rm), |
| 20096 | cCL("sqtem", e488140, 2, (RF, RF_IF), rd_rm), |
| 20097 | cCL("sqtez", e488160, 2, (RF, RF_IF), rd_rm), |
| 20098 | |
| 20099 | cCL("logs", e508100, 2, (RF, RF_IF), rd_rm), |
| 20100 | cCL("logsp", e508120, 2, (RF, RF_IF), rd_rm), |
| 20101 | cCL("logsm", e508140, 2, (RF, RF_IF), rd_rm), |
| 20102 | cCL("logsz", e508160, 2, (RF, RF_IF), rd_rm), |
| 20103 | cCL("logd", e508180, 2, (RF, RF_IF), rd_rm), |
| 20104 | cCL("logdp", e5081a0, 2, (RF, RF_IF), rd_rm), |
| 20105 | cCL("logdm", e5081c0, 2, (RF, RF_IF), rd_rm), |
| 20106 | cCL("logdz", e5081e0, 2, (RF, RF_IF), rd_rm), |
| 20107 | cCL("loge", e588100, 2, (RF, RF_IF), rd_rm), |
| 20108 | cCL("logep", e588120, 2, (RF, RF_IF), rd_rm), |
| 20109 | cCL("logem", e588140, 2, (RF, RF_IF), rd_rm), |
| 20110 | cCL("logez", e588160, 2, (RF, RF_IF), rd_rm), |
| 20111 | |
| 20112 | cCL("lgns", e608100, 2, (RF, RF_IF), rd_rm), |
| 20113 | cCL("lgnsp", e608120, 2, (RF, RF_IF), rd_rm), |
| 20114 | cCL("lgnsm", e608140, 2, (RF, RF_IF), rd_rm), |
| 20115 | cCL("lgnsz", e608160, 2, (RF, RF_IF), rd_rm), |
| 20116 | cCL("lgnd", e608180, 2, (RF, RF_IF), rd_rm), |
| 20117 | cCL("lgndp", e6081a0, 2, (RF, RF_IF), rd_rm), |
| 20118 | cCL("lgndm", e6081c0, 2, (RF, RF_IF), rd_rm), |
| 20119 | cCL("lgndz", e6081e0, 2, (RF, RF_IF), rd_rm), |
| 20120 | cCL("lgne", e688100, 2, (RF, RF_IF), rd_rm), |
| 20121 | cCL("lgnep", e688120, 2, (RF, RF_IF), rd_rm), |
| 20122 | cCL("lgnem", e688140, 2, (RF, RF_IF), rd_rm), |
| 20123 | cCL("lgnez", e688160, 2, (RF, RF_IF), rd_rm), |
| 20124 | |
| 20125 | cCL("exps", e708100, 2, (RF, RF_IF), rd_rm), |
| 20126 | cCL("expsp", e708120, 2, (RF, RF_IF), rd_rm), |
| 20127 | cCL("expsm", e708140, 2, (RF, RF_IF), rd_rm), |
| 20128 | cCL("expsz", e708160, 2, (RF, RF_IF), rd_rm), |
| 20129 | cCL("expd", e708180, 2, (RF, RF_IF), rd_rm), |
| 20130 | cCL("expdp", e7081a0, 2, (RF, RF_IF), rd_rm), |
| 20131 | cCL("expdm", e7081c0, 2, (RF, RF_IF), rd_rm), |
| 20132 | cCL("expdz", e7081e0, 2, (RF, RF_IF), rd_rm), |
| 20133 | cCL("expe", e788100, 2, (RF, RF_IF), rd_rm), |
| 20134 | cCL("expep", e788120, 2, (RF, RF_IF), rd_rm), |
| 20135 | cCL("expem", e788140, 2, (RF, RF_IF), rd_rm), |
| 20136 | cCL("expdz", e788160, 2, (RF, RF_IF), rd_rm), |
| 20137 | |
| 20138 | cCL("sins", e808100, 2, (RF, RF_IF), rd_rm), |
| 20139 | cCL("sinsp", e808120, 2, (RF, RF_IF), rd_rm), |
| 20140 | cCL("sinsm", e808140, 2, (RF, RF_IF), rd_rm), |
| 20141 | cCL("sinsz", e808160, 2, (RF, RF_IF), rd_rm), |
| 20142 | cCL("sind", e808180, 2, (RF, RF_IF), rd_rm), |
| 20143 | cCL("sindp", e8081a0, 2, (RF, RF_IF), rd_rm), |
| 20144 | cCL("sindm", e8081c0, 2, (RF, RF_IF), rd_rm), |
| 20145 | cCL("sindz", e8081e0, 2, (RF, RF_IF), rd_rm), |
| 20146 | cCL("sine", e888100, 2, (RF, RF_IF), rd_rm), |
| 20147 | cCL("sinep", e888120, 2, (RF, RF_IF), rd_rm), |
| 20148 | cCL("sinem", e888140, 2, (RF, RF_IF), rd_rm), |
| 20149 | cCL("sinez", e888160, 2, (RF, RF_IF), rd_rm), |
| 20150 | |
| 20151 | cCL("coss", e908100, 2, (RF, RF_IF), rd_rm), |
| 20152 | cCL("cossp", e908120, 2, (RF, RF_IF), rd_rm), |
| 20153 | cCL("cossm", e908140, 2, (RF, RF_IF), rd_rm), |
| 20154 | cCL("cossz", e908160, 2, (RF, RF_IF), rd_rm), |
| 20155 | cCL("cosd", e908180, 2, (RF, RF_IF), rd_rm), |
| 20156 | cCL("cosdp", e9081a0, 2, (RF, RF_IF), rd_rm), |
| 20157 | cCL("cosdm", e9081c0, 2, (RF, RF_IF), rd_rm), |
| 20158 | cCL("cosdz", e9081e0, 2, (RF, RF_IF), rd_rm), |
| 20159 | cCL("cose", e988100, 2, (RF, RF_IF), rd_rm), |
| 20160 | cCL("cosep", e988120, 2, (RF, RF_IF), rd_rm), |
| 20161 | cCL("cosem", e988140, 2, (RF, RF_IF), rd_rm), |
| 20162 | cCL("cosez", e988160, 2, (RF, RF_IF), rd_rm), |
| 20163 | |
| 20164 | cCL("tans", ea08100, 2, (RF, RF_IF), rd_rm), |
| 20165 | cCL("tansp", ea08120, 2, (RF, RF_IF), rd_rm), |
| 20166 | cCL("tansm", ea08140, 2, (RF, RF_IF), rd_rm), |
| 20167 | cCL("tansz", ea08160, 2, (RF, RF_IF), rd_rm), |
| 20168 | cCL("tand", ea08180, 2, (RF, RF_IF), rd_rm), |
| 20169 | cCL("tandp", ea081a0, 2, (RF, RF_IF), rd_rm), |
| 20170 | cCL("tandm", ea081c0, 2, (RF, RF_IF), rd_rm), |
| 20171 | cCL("tandz", ea081e0, 2, (RF, RF_IF), rd_rm), |
| 20172 | cCL("tane", ea88100, 2, (RF, RF_IF), rd_rm), |
| 20173 | cCL("tanep", ea88120, 2, (RF, RF_IF), rd_rm), |
| 20174 | cCL("tanem", ea88140, 2, (RF, RF_IF), rd_rm), |
| 20175 | cCL("tanez", ea88160, 2, (RF, RF_IF), rd_rm), |
| 20176 | |
| 20177 | cCL("asns", eb08100, 2, (RF, RF_IF), rd_rm), |
| 20178 | cCL("asnsp", eb08120, 2, (RF, RF_IF), rd_rm), |
| 20179 | cCL("asnsm", eb08140, 2, (RF, RF_IF), rd_rm), |
| 20180 | cCL("asnsz", eb08160, 2, (RF, RF_IF), rd_rm), |
| 20181 | cCL("asnd", eb08180, 2, (RF, RF_IF), rd_rm), |
| 20182 | cCL("asndp", eb081a0, 2, (RF, RF_IF), rd_rm), |
| 20183 | cCL("asndm", eb081c0, 2, (RF, RF_IF), rd_rm), |
| 20184 | cCL("asndz", eb081e0, 2, (RF, RF_IF), rd_rm), |
| 20185 | cCL("asne", eb88100, 2, (RF, RF_IF), rd_rm), |
| 20186 | cCL("asnep", eb88120, 2, (RF, RF_IF), rd_rm), |
| 20187 | cCL("asnem", eb88140, 2, (RF, RF_IF), rd_rm), |
| 20188 | cCL("asnez", eb88160, 2, (RF, RF_IF), rd_rm), |
| 20189 | |
| 20190 | cCL("acss", ec08100, 2, (RF, RF_IF), rd_rm), |
| 20191 | cCL("acssp", ec08120, 2, (RF, RF_IF), rd_rm), |
| 20192 | cCL("acssm", ec08140, 2, (RF, RF_IF), rd_rm), |
| 20193 | cCL("acssz", ec08160, 2, (RF, RF_IF), rd_rm), |
| 20194 | cCL("acsd", ec08180, 2, (RF, RF_IF), rd_rm), |
| 20195 | cCL("acsdp", ec081a0, 2, (RF, RF_IF), rd_rm), |
| 20196 | cCL("acsdm", ec081c0, 2, (RF, RF_IF), rd_rm), |
| 20197 | cCL("acsdz", ec081e0, 2, (RF, RF_IF), rd_rm), |
| 20198 | cCL("acse", ec88100, 2, (RF, RF_IF), rd_rm), |
| 20199 | cCL("acsep", ec88120, 2, (RF, RF_IF), rd_rm), |
| 20200 | cCL("acsem", ec88140, 2, (RF, RF_IF), rd_rm), |
| 20201 | cCL("acsez", ec88160, 2, (RF, RF_IF), rd_rm), |
| 20202 | |
| 20203 | cCL("atns", ed08100, 2, (RF, RF_IF), rd_rm), |
| 20204 | cCL("atnsp", ed08120, 2, (RF, RF_IF), rd_rm), |
| 20205 | cCL("atnsm", ed08140, 2, (RF, RF_IF), rd_rm), |
| 20206 | cCL("atnsz", ed08160, 2, (RF, RF_IF), rd_rm), |
| 20207 | cCL("atnd", ed08180, 2, (RF, RF_IF), rd_rm), |
| 20208 | cCL("atndp", ed081a0, 2, (RF, RF_IF), rd_rm), |
| 20209 | cCL("atndm", ed081c0, 2, (RF, RF_IF), rd_rm), |
| 20210 | cCL("atndz", ed081e0, 2, (RF, RF_IF), rd_rm), |
| 20211 | cCL("atne", ed88100, 2, (RF, RF_IF), rd_rm), |
| 20212 | cCL("atnep", ed88120, 2, (RF, RF_IF), rd_rm), |
| 20213 | cCL("atnem", ed88140, 2, (RF, RF_IF), rd_rm), |
| 20214 | cCL("atnez", ed88160, 2, (RF, RF_IF), rd_rm), |
| 20215 | |
| 20216 | cCL("urds", ee08100, 2, (RF, RF_IF), rd_rm), |
| 20217 | cCL("urdsp", ee08120, 2, (RF, RF_IF), rd_rm), |
| 20218 | cCL("urdsm", ee08140, 2, (RF, RF_IF), rd_rm), |
| 20219 | cCL("urdsz", ee08160, 2, (RF, RF_IF), rd_rm), |
| 20220 | cCL("urdd", ee08180, 2, (RF, RF_IF), rd_rm), |
| 20221 | cCL("urddp", ee081a0, 2, (RF, RF_IF), rd_rm), |
| 20222 | cCL("urddm", ee081c0, 2, (RF, RF_IF), rd_rm), |
| 20223 | cCL("urddz", ee081e0, 2, (RF, RF_IF), rd_rm), |
| 20224 | cCL("urde", ee88100, 2, (RF, RF_IF), rd_rm), |
| 20225 | cCL("urdep", ee88120, 2, (RF, RF_IF), rd_rm), |
| 20226 | cCL("urdem", ee88140, 2, (RF, RF_IF), rd_rm), |
| 20227 | cCL("urdez", ee88160, 2, (RF, RF_IF), rd_rm), |
| 20228 | |
| 20229 | cCL("nrms", ef08100, 2, (RF, RF_IF), rd_rm), |
| 20230 | cCL("nrmsp", ef08120, 2, (RF, RF_IF), rd_rm), |
| 20231 | cCL("nrmsm", ef08140, 2, (RF, RF_IF), rd_rm), |
| 20232 | cCL("nrmsz", ef08160, 2, (RF, RF_IF), rd_rm), |
| 20233 | cCL("nrmd", ef08180, 2, (RF, RF_IF), rd_rm), |
| 20234 | cCL("nrmdp", ef081a0, 2, (RF, RF_IF), rd_rm), |
| 20235 | cCL("nrmdm", ef081c0, 2, (RF, RF_IF), rd_rm), |
| 20236 | cCL("nrmdz", ef081e0, 2, (RF, RF_IF), rd_rm), |
| 20237 | cCL("nrme", ef88100, 2, (RF, RF_IF), rd_rm), |
| 20238 | cCL("nrmep", ef88120, 2, (RF, RF_IF), rd_rm), |
| 20239 | cCL("nrmem", ef88140, 2, (RF, RF_IF), rd_rm), |
| 20240 | cCL("nrmez", ef88160, 2, (RF, RF_IF), rd_rm), |
| 20241 | |
| 20242 | cCL("adfs", e000100, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20243 | cCL("adfsp", e000120, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20244 | cCL("adfsm", e000140, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20245 | cCL("adfsz", e000160, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20246 | cCL("adfd", e000180, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20247 | cCL("adfdp", e0001a0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20248 | cCL("adfdm", e0001c0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20249 | cCL("adfdz", e0001e0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20250 | cCL("adfe", e080100, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20251 | cCL("adfep", e080120, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20252 | cCL("adfem", e080140, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20253 | cCL("adfez", e080160, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20254 | |
| 20255 | cCL("sufs", e200100, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20256 | cCL("sufsp", e200120, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20257 | cCL("sufsm", e200140, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20258 | cCL("sufsz", e200160, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20259 | cCL("sufd", e200180, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20260 | cCL("sufdp", e2001a0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20261 | cCL("sufdm", e2001c0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20262 | cCL("sufdz", e2001e0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20263 | cCL("sufe", e280100, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20264 | cCL("sufep", e280120, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20265 | cCL("sufem", e280140, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20266 | cCL("sufez", e280160, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20267 | |
| 20268 | cCL("rsfs", e300100, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20269 | cCL("rsfsp", e300120, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20270 | cCL("rsfsm", e300140, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20271 | cCL("rsfsz", e300160, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20272 | cCL("rsfd", e300180, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20273 | cCL("rsfdp", e3001a0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20274 | cCL("rsfdm", e3001c0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20275 | cCL("rsfdz", e3001e0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20276 | cCL("rsfe", e380100, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20277 | cCL("rsfep", e380120, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20278 | cCL("rsfem", e380140, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20279 | cCL("rsfez", e380160, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20280 | |
| 20281 | cCL("mufs", e100100, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20282 | cCL("mufsp", e100120, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20283 | cCL("mufsm", e100140, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20284 | cCL("mufsz", e100160, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20285 | cCL("mufd", e100180, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20286 | cCL("mufdp", e1001a0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20287 | cCL("mufdm", e1001c0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20288 | cCL("mufdz", e1001e0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20289 | cCL("mufe", e180100, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20290 | cCL("mufep", e180120, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20291 | cCL("mufem", e180140, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20292 | cCL("mufez", e180160, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20293 | |
| 20294 | cCL("dvfs", e400100, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20295 | cCL("dvfsp", e400120, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20296 | cCL("dvfsm", e400140, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20297 | cCL("dvfsz", e400160, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20298 | cCL("dvfd", e400180, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20299 | cCL("dvfdp", e4001a0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20300 | cCL("dvfdm", e4001c0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20301 | cCL("dvfdz", e4001e0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20302 | cCL("dvfe", e480100, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20303 | cCL("dvfep", e480120, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20304 | cCL("dvfem", e480140, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20305 | cCL("dvfez", e480160, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20306 | |
| 20307 | cCL("rdfs", e500100, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20308 | cCL("rdfsp", e500120, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20309 | cCL("rdfsm", e500140, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20310 | cCL("rdfsz", e500160, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20311 | cCL("rdfd", e500180, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20312 | cCL("rdfdp", e5001a0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20313 | cCL("rdfdm", e5001c0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20314 | cCL("rdfdz", e5001e0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20315 | cCL("rdfe", e580100, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20316 | cCL("rdfep", e580120, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20317 | cCL("rdfem", e580140, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20318 | cCL("rdfez", e580160, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20319 | |
| 20320 | cCL("pows", e600100, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20321 | cCL("powsp", e600120, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20322 | cCL("powsm", e600140, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20323 | cCL("powsz", e600160, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20324 | cCL("powd", e600180, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20325 | cCL("powdp", e6001a0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20326 | cCL("powdm", e6001c0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20327 | cCL("powdz", e6001e0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20328 | cCL("powe", e680100, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20329 | cCL("powep", e680120, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20330 | cCL("powem", e680140, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20331 | cCL("powez", e680160, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20332 | |
| 20333 | cCL("rpws", e700100, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20334 | cCL("rpwsp", e700120, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20335 | cCL("rpwsm", e700140, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20336 | cCL("rpwsz", e700160, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20337 | cCL("rpwd", e700180, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20338 | cCL("rpwdp", e7001a0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20339 | cCL("rpwdm", e7001c0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20340 | cCL("rpwdz", e7001e0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20341 | cCL("rpwe", e780100, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20342 | cCL("rpwep", e780120, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20343 | cCL("rpwem", e780140, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20344 | cCL("rpwez", e780160, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20345 | |
| 20346 | cCL("rmfs", e800100, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20347 | cCL("rmfsp", e800120, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20348 | cCL("rmfsm", e800140, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20349 | cCL("rmfsz", e800160, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20350 | cCL("rmfd", e800180, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20351 | cCL("rmfdp", e8001a0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20352 | cCL("rmfdm", e8001c0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20353 | cCL("rmfdz", e8001e0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20354 | cCL("rmfe", e880100, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20355 | cCL("rmfep", e880120, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20356 | cCL("rmfem", e880140, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20357 | cCL("rmfez", e880160, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20358 | |
| 20359 | cCL("fmls", e900100, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20360 | cCL("fmlsp", e900120, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20361 | cCL("fmlsm", e900140, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20362 | cCL("fmlsz", e900160, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20363 | cCL("fmld", e900180, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20364 | cCL("fmldp", e9001a0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20365 | cCL("fmldm", e9001c0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20366 | cCL("fmldz", e9001e0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20367 | cCL("fmle", e980100, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20368 | cCL("fmlep", e980120, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20369 | cCL("fmlem", e980140, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20370 | cCL("fmlez", e980160, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20371 | |
| 20372 | cCL("fdvs", ea00100, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20373 | cCL("fdvsp", ea00120, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20374 | cCL("fdvsm", ea00140, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20375 | cCL("fdvsz", ea00160, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20376 | cCL("fdvd", ea00180, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20377 | cCL("fdvdp", ea001a0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20378 | cCL("fdvdm", ea001c0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20379 | cCL("fdvdz", ea001e0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20380 | cCL("fdve", ea80100, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20381 | cCL("fdvep", ea80120, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20382 | cCL("fdvem", ea80140, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20383 | cCL("fdvez", ea80160, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20384 | |
| 20385 | cCL("frds", eb00100, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20386 | cCL("frdsp", eb00120, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20387 | cCL("frdsm", eb00140, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20388 | cCL("frdsz", eb00160, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20389 | cCL("frdd", eb00180, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20390 | cCL("frddp", eb001a0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20391 | cCL("frddm", eb001c0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20392 | cCL("frddz", eb001e0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20393 | cCL("frde", eb80100, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20394 | cCL("frdep", eb80120, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20395 | cCL("frdem", eb80140, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20396 | cCL("frdez", eb80160, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20397 | |
| 20398 | cCL("pols", ec00100, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20399 | cCL("polsp", ec00120, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20400 | cCL("polsm", ec00140, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20401 | cCL("polsz", ec00160, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20402 | cCL("pold", ec00180, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20403 | cCL("poldp", ec001a0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20404 | cCL("poldm", ec001c0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20405 | cCL("poldz", ec001e0, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20406 | cCL("pole", ec80100, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20407 | cCL("polep", ec80120, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20408 | cCL("polem", ec80140, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20409 | cCL("polez", ec80160, 3, (RF, RF, RF_IF), rd_rn_rm), |
| 20410 | |
| 20411 | cCE("cmf", e90f110, 2, (RF, RF_IF), fpa_cmp), |
| 20412 | C3E("cmfe", ed0f110, 2, (RF, RF_IF), fpa_cmp), |
| 20413 | cCE("cnf", eb0f110, 2, (RF, RF_IF), fpa_cmp), |
| 20414 | C3E("cnfe", ef0f110, 2, (RF, RF_IF), fpa_cmp), |
| 20415 | |
| 20416 | cCL("flts", e000110, 2, (RF, RR), rn_rd), |
| 20417 | cCL("fltsp", e000130, 2, (RF, RR), rn_rd), |
| 20418 | cCL("fltsm", e000150, 2, (RF, RR), rn_rd), |
| 20419 | cCL("fltsz", e000170, 2, (RF, RR), rn_rd), |
| 20420 | cCL("fltd", e000190, 2, (RF, RR), rn_rd), |
| 20421 | cCL("fltdp", e0001b0, 2, (RF, RR), rn_rd), |
| 20422 | cCL("fltdm", e0001d0, 2, (RF, RR), rn_rd), |
| 20423 | cCL("fltdz", e0001f0, 2, (RF, RR), rn_rd), |
| 20424 | cCL("flte", e080110, 2, (RF, RR), rn_rd), |
| 20425 | cCL("fltep", e080130, 2, (RF, RR), rn_rd), |
| 20426 | cCL("fltem", e080150, 2, (RF, RR), rn_rd), |
| 20427 | cCL("fltez", e080170, 2, (RF, RR), rn_rd), |
| 20428 | |
| 20429 | /* The implementation of the FIX instruction is broken on some |
| 20430 | assemblers, in that it accepts a precision specifier as well as a |
| 20431 | rounding specifier, despite the fact that this is meaningless. |
| 20432 | To be more compatible, we accept it as well, though of course it |
| 20433 | does not set any bits. */ |
| 20434 | cCE("fix", e100110, 2, (RR, RF), rd_rm), |
| 20435 | cCL("fixp", e100130, 2, (RR, RF), rd_rm), |
| 20436 | cCL("fixm", e100150, 2, (RR, RF), rd_rm), |
| 20437 | cCL("fixz", e100170, 2, (RR, RF), rd_rm), |
| 20438 | cCL("fixsp", e100130, 2, (RR, RF), rd_rm), |
| 20439 | cCL("fixsm", e100150, 2, (RR, RF), rd_rm), |
| 20440 | cCL("fixsz", e100170, 2, (RR, RF), rd_rm), |
| 20441 | cCL("fixdp", e100130, 2, (RR, RF), rd_rm), |
| 20442 | cCL("fixdm", e100150, 2, (RR, RF), rd_rm), |
| 20443 | cCL("fixdz", e100170, 2, (RR, RF), rd_rm), |
| 20444 | cCL("fixep", e100130, 2, (RR, RF), rd_rm), |
| 20445 | cCL("fixem", e100150, 2, (RR, RF), rd_rm), |
| 20446 | cCL("fixez", e100170, 2, (RR, RF), rd_rm), |
| 20447 | |
| 20448 | /* Instructions that were new with the real FPA, call them V2. */ |
| 20449 | #undef ARM_VARIANT |
| 20450 | #define ARM_VARIANT & fpu_fpa_ext_v2 |
| 20451 | |
| 20452 | cCE("lfm", c100200, 3, (RF, I4b, ADDR), fpa_ldmstm), |
| 20453 | cCL("lfmfd", c900200, 3, (RF, I4b, ADDR), fpa_ldmstm), |
| 20454 | cCL("lfmea", d100200, 3, (RF, I4b, ADDR), fpa_ldmstm), |
| 20455 | cCE("sfm", c000200, 3, (RF, I4b, ADDR), fpa_ldmstm), |
| 20456 | cCL("sfmfd", d000200, 3, (RF, I4b, ADDR), fpa_ldmstm), |
| 20457 | cCL("sfmea", c800200, 3, (RF, I4b, ADDR), fpa_ldmstm), |
| 20458 | |
| 20459 | #undef ARM_VARIANT |
| 20460 | #define ARM_VARIANT & fpu_vfp_ext_v1xd /* VFP V1xD (single precision). */ |
| 20461 | |
| 20462 | /* Moves and type conversions. */ |
| 20463 | cCE("fcpys", eb00a40, 2, (RVS, RVS), vfp_sp_monadic), |
| 20464 | cCE("fmrs", e100a10, 2, (RR, RVS), vfp_reg_from_sp), |
| 20465 | cCE("fmsr", e000a10, 2, (RVS, RR), vfp_sp_from_reg), |
| 20466 | cCE("fmstat", ef1fa10, 0, (), noargs), |
| 20467 | cCE("vmrs", ef00a10, 2, (APSR_RR, RVC), vmrs), |
| 20468 | cCE("vmsr", ee00a10, 2, (RVC, RR), vmsr), |
| 20469 | cCE("fsitos", eb80ac0, 2, (RVS, RVS), vfp_sp_monadic), |
| 20470 | cCE("fuitos", eb80a40, 2, (RVS, RVS), vfp_sp_monadic), |
| 20471 | cCE("ftosis", ebd0a40, 2, (RVS, RVS), vfp_sp_monadic), |
| 20472 | cCE("ftosizs", ebd0ac0, 2, (RVS, RVS), vfp_sp_monadic), |
| 20473 | cCE("ftouis", ebc0a40, 2, (RVS, RVS), vfp_sp_monadic), |
| 20474 | cCE("ftouizs", ebc0ac0, 2, (RVS, RVS), vfp_sp_monadic), |
| 20475 | cCE("fmrx", ef00a10, 2, (RR, RVC), rd_rn), |
| 20476 | cCE("fmxr", ee00a10, 2, (RVC, RR), rn_rd), |
| 20477 | |
| 20478 | /* Memory operations. */ |
| 20479 | cCE("flds", d100a00, 2, (RVS, ADDRGLDC), vfp_sp_ldst), |
| 20480 | cCE("fsts", d000a00, 2, (RVS, ADDRGLDC), vfp_sp_ldst), |
| 20481 | cCE("fldmias", c900a00, 2, (RRnpctw, VRSLST), vfp_sp_ldstmia), |
| 20482 | cCE("fldmfds", c900a00, 2, (RRnpctw, VRSLST), vfp_sp_ldstmia), |
| 20483 | cCE("fldmdbs", d300a00, 2, (RRnpctw, VRSLST), vfp_sp_ldstmdb), |
| 20484 | cCE("fldmeas", d300a00, 2, (RRnpctw, VRSLST), vfp_sp_ldstmdb), |
| 20485 | cCE("fldmiax", c900b00, 2, (RRnpctw, VRDLST), vfp_xp_ldstmia), |
| 20486 | cCE("fldmfdx", c900b00, 2, (RRnpctw, VRDLST), vfp_xp_ldstmia), |
| 20487 | cCE("fldmdbx", d300b00, 2, (RRnpctw, VRDLST), vfp_xp_ldstmdb), |
| 20488 | cCE("fldmeax", d300b00, 2, (RRnpctw, VRDLST), vfp_xp_ldstmdb), |
| 20489 | cCE("fstmias", c800a00, 2, (RRnpctw, VRSLST), vfp_sp_ldstmia), |
| 20490 | cCE("fstmeas", c800a00, 2, (RRnpctw, VRSLST), vfp_sp_ldstmia), |
| 20491 | cCE("fstmdbs", d200a00, 2, (RRnpctw, VRSLST), vfp_sp_ldstmdb), |
| 20492 | cCE("fstmfds", d200a00, 2, (RRnpctw, VRSLST), vfp_sp_ldstmdb), |
| 20493 | cCE("fstmiax", c800b00, 2, (RRnpctw, VRDLST), vfp_xp_ldstmia), |
| 20494 | cCE("fstmeax", c800b00, 2, (RRnpctw, VRDLST), vfp_xp_ldstmia), |
| 20495 | cCE("fstmdbx", d200b00, 2, (RRnpctw, VRDLST), vfp_xp_ldstmdb), |
| 20496 | cCE("fstmfdx", d200b00, 2, (RRnpctw, VRDLST), vfp_xp_ldstmdb), |
| 20497 | |
| 20498 | /* Monadic operations. */ |
| 20499 | cCE("fabss", eb00ac0, 2, (RVS, RVS), vfp_sp_monadic), |
| 20500 | cCE("fnegs", eb10a40, 2, (RVS, RVS), vfp_sp_monadic), |
| 20501 | cCE("fsqrts", eb10ac0, 2, (RVS, RVS), vfp_sp_monadic), |
| 20502 | |
| 20503 | /* Dyadic operations. */ |
| 20504 | cCE("fadds", e300a00, 3, (RVS, RVS, RVS), vfp_sp_dyadic), |
| 20505 | cCE("fsubs", e300a40, 3, (RVS, RVS, RVS), vfp_sp_dyadic), |
| 20506 | cCE("fmuls", e200a00, 3, (RVS, RVS, RVS), vfp_sp_dyadic), |
| 20507 | cCE("fdivs", e800a00, 3, (RVS, RVS, RVS), vfp_sp_dyadic), |
| 20508 | cCE("fmacs", e000a00, 3, (RVS, RVS, RVS), vfp_sp_dyadic), |
| 20509 | cCE("fmscs", e100a00, 3, (RVS, RVS, RVS), vfp_sp_dyadic), |
| 20510 | cCE("fnmuls", e200a40, 3, (RVS, RVS, RVS), vfp_sp_dyadic), |
| 20511 | cCE("fnmacs", e000a40, 3, (RVS, RVS, RVS), vfp_sp_dyadic), |
| 20512 | cCE("fnmscs", e100a40, 3, (RVS, RVS, RVS), vfp_sp_dyadic), |
| 20513 | |
| 20514 | /* Comparisons. */ |
| 20515 | cCE("fcmps", eb40a40, 2, (RVS, RVS), vfp_sp_monadic), |
| 20516 | cCE("fcmpzs", eb50a40, 1, (RVS), vfp_sp_compare_z), |
| 20517 | cCE("fcmpes", eb40ac0, 2, (RVS, RVS), vfp_sp_monadic), |
| 20518 | cCE("fcmpezs", eb50ac0, 1, (RVS), vfp_sp_compare_z), |
| 20519 | |
| 20520 | /* Double precision load/store are still present on single precision |
| 20521 | implementations. */ |
| 20522 | cCE("fldd", d100b00, 2, (RVD, ADDRGLDC), vfp_dp_ldst), |
| 20523 | cCE("fstd", d000b00, 2, (RVD, ADDRGLDC), vfp_dp_ldst), |
| 20524 | cCE("fldmiad", c900b00, 2, (RRnpctw, VRDLST), vfp_dp_ldstmia), |
| 20525 | cCE("fldmfdd", c900b00, 2, (RRnpctw, VRDLST), vfp_dp_ldstmia), |
| 20526 | cCE("fldmdbd", d300b00, 2, (RRnpctw, VRDLST), vfp_dp_ldstmdb), |
| 20527 | cCE("fldmead", d300b00, 2, (RRnpctw, VRDLST), vfp_dp_ldstmdb), |
| 20528 | cCE("fstmiad", c800b00, 2, (RRnpctw, VRDLST), vfp_dp_ldstmia), |
| 20529 | cCE("fstmead", c800b00, 2, (RRnpctw, VRDLST), vfp_dp_ldstmia), |
| 20530 | cCE("fstmdbd", d200b00, 2, (RRnpctw, VRDLST), vfp_dp_ldstmdb), |
| 20531 | cCE("fstmfdd", d200b00, 2, (RRnpctw, VRDLST), vfp_dp_ldstmdb), |
| 20532 | |
| 20533 | #undef ARM_VARIANT |
| 20534 | #define ARM_VARIANT & fpu_vfp_ext_v1 /* VFP V1 (Double precision). */ |
| 20535 | |
| 20536 | /* Moves and type conversions. */ |
| 20537 | cCE("fcpyd", eb00b40, 2, (RVD, RVD), vfp_dp_rd_rm), |
| 20538 | cCE("fcvtds", eb70ac0, 2, (RVD, RVS), vfp_dp_sp_cvt), |
| 20539 | cCE("fcvtsd", eb70bc0, 2, (RVS, RVD), vfp_sp_dp_cvt), |
| 20540 | cCE("fmdhr", e200b10, 2, (RVD, RR), vfp_dp_rn_rd), |
| 20541 | cCE("fmdlr", e000b10, 2, (RVD, RR), vfp_dp_rn_rd), |
| 20542 | cCE("fmrdh", e300b10, 2, (RR, RVD), vfp_dp_rd_rn), |
| 20543 | cCE("fmrdl", e100b10, 2, (RR, RVD), vfp_dp_rd_rn), |
| 20544 | cCE("fsitod", eb80bc0, 2, (RVD, RVS), vfp_dp_sp_cvt), |
| 20545 | cCE("fuitod", eb80b40, 2, (RVD, RVS), vfp_dp_sp_cvt), |
| 20546 | cCE("ftosid", ebd0b40, 2, (RVS, RVD), vfp_sp_dp_cvt), |
| 20547 | cCE("ftosizd", ebd0bc0, 2, (RVS, RVD), vfp_sp_dp_cvt), |
| 20548 | cCE("ftouid", ebc0b40, 2, (RVS, RVD), vfp_sp_dp_cvt), |
| 20549 | cCE("ftouizd", ebc0bc0, 2, (RVS, RVD), vfp_sp_dp_cvt), |
| 20550 | |
| 20551 | /* Monadic operations. */ |
| 20552 | cCE("fabsd", eb00bc0, 2, (RVD, RVD), vfp_dp_rd_rm), |
| 20553 | cCE("fnegd", eb10b40, 2, (RVD, RVD), vfp_dp_rd_rm), |
| 20554 | cCE("fsqrtd", eb10bc0, 2, (RVD, RVD), vfp_dp_rd_rm), |
| 20555 | |
| 20556 | /* Dyadic operations. */ |
| 20557 | cCE("faddd", e300b00, 3, (RVD, RVD, RVD), vfp_dp_rd_rn_rm), |
| 20558 | cCE("fsubd", e300b40, 3, (RVD, RVD, RVD), vfp_dp_rd_rn_rm), |
| 20559 | cCE("fmuld", e200b00, 3, (RVD, RVD, RVD), vfp_dp_rd_rn_rm), |
| 20560 | cCE("fdivd", e800b00, 3, (RVD, RVD, RVD), vfp_dp_rd_rn_rm), |
| 20561 | cCE("fmacd", e000b00, 3, (RVD, RVD, RVD), vfp_dp_rd_rn_rm), |
| 20562 | cCE("fmscd", e100b00, 3, (RVD, RVD, RVD), vfp_dp_rd_rn_rm), |
| 20563 | cCE("fnmuld", e200b40, 3, (RVD, RVD, RVD), vfp_dp_rd_rn_rm), |
| 20564 | cCE("fnmacd", e000b40, 3, (RVD, RVD, RVD), vfp_dp_rd_rn_rm), |
| 20565 | cCE("fnmscd", e100b40, 3, (RVD, RVD, RVD), vfp_dp_rd_rn_rm), |
| 20566 | |
| 20567 | /* Comparisons. */ |
| 20568 | cCE("fcmpd", eb40b40, 2, (RVD, RVD), vfp_dp_rd_rm), |
| 20569 | cCE("fcmpzd", eb50b40, 1, (RVD), vfp_dp_rd), |
| 20570 | cCE("fcmped", eb40bc0, 2, (RVD, RVD), vfp_dp_rd_rm), |
| 20571 | cCE("fcmpezd", eb50bc0, 1, (RVD), vfp_dp_rd), |
| 20572 | |
| 20573 | #undef ARM_VARIANT |
| 20574 | #define ARM_VARIANT & fpu_vfp_ext_v2 |
| 20575 | |
| 20576 | cCE("fmsrr", c400a10, 3, (VRSLST, RR, RR), vfp_sp2_from_reg2), |
| 20577 | cCE("fmrrs", c500a10, 3, (RR, RR, VRSLST), vfp_reg2_from_sp2), |
| 20578 | cCE("fmdrr", c400b10, 3, (RVD, RR, RR), vfp_dp_rm_rd_rn), |
| 20579 | cCE("fmrrd", c500b10, 3, (RR, RR, RVD), vfp_dp_rd_rn_rm), |
| 20580 | |
| 20581 | /* Instructions which may belong to either the Neon or VFP instruction sets. |
| 20582 | Individual encoder functions perform additional architecture checks. */ |
| 20583 | #undef ARM_VARIANT |
| 20584 | #define ARM_VARIANT & fpu_vfp_ext_v1xd |
| 20585 | #undef THUMB_VARIANT |
| 20586 | #define THUMB_VARIANT & fpu_vfp_ext_v1xd |
| 20587 | |
| 20588 | /* These mnemonics are unique to VFP. */ |
| 20589 | NCE(vsqrt, 0, 2, (RVSD, RVSD), vfp_nsyn_sqrt), |
| 20590 | NCE(vdiv, 0, 3, (RVSD, RVSD, RVSD), vfp_nsyn_div), |
| 20591 | nCE(vnmul, _vnmul, 3, (RVSD, RVSD, RVSD), vfp_nsyn_nmul), |
| 20592 | nCE(vnmla, _vnmla, 3, (RVSD, RVSD, RVSD), vfp_nsyn_nmul), |
| 20593 | nCE(vnmls, _vnmls, 3, (RVSD, RVSD, RVSD), vfp_nsyn_nmul), |
| 20594 | nCE(vcmp, _vcmp, 2, (RVSD, RSVD_FI0), vfp_nsyn_cmp), |
| 20595 | nCE(vcmpe, _vcmpe, 2, (RVSD, RSVD_FI0), vfp_nsyn_cmp), |
| 20596 | NCE(vpush, 0, 1, (VRSDLST), vfp_nsyn_push), |
| 20597 | NCE(vpop, 0, 1, (VRSDLST), vfp_nsyn_pop), |
| 20598 | NCE(vcvtz, 0, 2, (RVSD, RVSD), vfp_nsyn_cvtz), |
| 20599 | |
| 20600 | /* Mnemonics shared by Neon and VFP. */ |
| 20601 | nCEF(vmul, _vmul, 3, (RNSDQ, oRNSDQ, RNSDQ_RNSC), neon_mul), |
| 20602 | nCEF(vmla, _vmla, 3, (RNSDQ, oRNSDQ, RNSDQ_RNSC), neon_mac_maybe_scalar), |
| 20603 | nCEF(vmls, _vmls, 3, (RNSDQ, oRNSDQ, RNSDQ_RNSC), neon_mac_maybe_scalar), |
| 20604 | |
| 20605 | nCEF(vadd, _vadd, 3, (RNSDQ, oRNSDQ, RNSDQ), neon_addsub_if_i), |
| 20606 | nCEF(vsub, _vsub, 3, (RNSDQ, oRNSDQ, RNSDQ), neon_addsub_if_i), |
| 20607 | |
| 20608 | NCEF(vabs, 1b10300, 2, (RNSDQ, RNSDQ), neon_abs_neg), |
| 20609 | NCEF(vneg, 1b10380, 2, (RNSDQ, RNSDQ), neon_abs_neg), |
| 20610 | |
| 20611 | NCE(vldm, c900b00, 2, (RRnpctw, VRSDLST), neon_ldm_stm), |
| 20612 | NCE(vldmia, c900b00, 2, (RRnpctw, VRSDLST), neon_ldm_stm), |
| 20613 | NCE(vldmdb, d100b00, 2, (RRnpctw, VRSDLST), neon_ldm_stm), |
| 20614 | NCE(vstm, c800b00, 2, (RRnpctw, VRSDLST), neon_ldm_stm), |
| 20615 | NCE(vstmia, c800b00, 2, (RRnpctw, VRSDLST), neon_ldm_stm), |
| 20616 | NCE(vstmdb, d000b00, 2, (RRnpctw, VRSDLST), neon_ldm_stm), |
| 20617 | NCE(vldr, d100b00, 2, (RVSD, ADDRGLDC), neon_ldr_str), |
| 20618 | NCE(vstr, d000b00, 2, (RVSD, ADDRGLDC), neon_ldr_str), |
| 20619 | |
| 20620 | nCEF(vcvt, _vcvt, 3, (RNSDQ, RNSDQ, oI32z), neon_cvt), |
| 20621 | nCEF(vcvtr, _vcvt, 2, (RNSDQ, RNSDQ), neon_cvtr), |
| 20622 | NCEF(vcvtb, eb20a40, 2, (RVSD, RVSD), neon_cvtb), |
| 20623 | NCEF(vcvtt, eb20a40, 2, (RVSD, RVSD), neon_cvtt), |
| 20624 | |
| 20625 | |
| 20626 | /* NOTE: All VMOV encoding is special-cased! */ |
| 20627 | NCE(vmov, 0, 1, (VMOV), neon_mov), |
| 20628 | NCE(vmovq, 0, 1, (VMOV), neon_mov), |
| 20629 | |
| 20630 | #undef ARM_VARIANT |
| 20631 | #define ARM_VARIANT & arm_ext_fp16 |
| 20632 | #undef THUMB_VARIANT |
| 20633 | #define THUMB_VARIANT & arm_ext_fp16 |
| 20634 | /* New instructions added from v8.2, allowing the extraction and insertion of |
| 20635 | the upper 16 bits of a 32-bit vector register. */ |
| 20636 | NCE (vmovx, eb00a40, 2, (RVS, RVS), neon_movhf), |
| 20637 | NCE (vins, eb00ac0, 2, (RVS, RVS), neon_movhf), |
| 20638 | |
| 20639 | #undef THUMB_VARIANT |
| 20640 | #define THUMB_VARIANT & fpu_neon_ext_v1 |
| 20641 | #undef ARM_VARIANT |
| 20642 | #define ARM_VARIANT & fpu_neon_ext_v1 |
| 20643 | |
| 20644 | /* Data processing with three registers of the same length. */ |
| 20645 | /* integer ops, valid types S8 S16 S32 U8 U16 U32. */ |
| 20646 | NUF(vaba, 0000710, 3, (RNDQ, RNDQ, RNDQ), neon_dyadic_i_su), |
| 20647 | NUF(vabaq, 0000710, 3, (RNQ, RNQ, RNQ), neon_dyadic_i_su), |
| 20648 | NUF(vhadd, 0000000, 3, (RNDQ, oRNDQ, RNDQ), neon_dyadic_i_su), |
| 20649 | NUF(vhaddq, 0000000, 3, (RNQ, oRNQ, RNQ), neon_dyadic_i_su), |
| 20650 | NUF(vrhadd, 0000100, 3, (RNDQ, oRNDQ, RNDQ), neon_dyadic_i_su), |
| 20651 | NUF(vrhaddq, 0000100, 3, (RNQ, oRNQ, RNQ), neon_dyadic_i_su), |
| 20652 | NUF(vhsub, 0000200, 3, (RNDQ, oRNDQ, RNDQ), neon_dyadic_i_su), |
| 20653 | NUF(vhsubq, 0000200, 3, (RNQ, oRNQ, RNQ), neon_dyadic_i_su), |
| 20654 | /* integer ops, valid types S8 S16 S32 S64 U8 U16 U32 U64. */ |
| 20655 | NUF(vqadd, 0000010, 3, (RNDQ, oRNDQ, RNDQ), neon_dyadic_i64_su), |
| 20656 | NUF(vqaddq, 0000010, 3, (RNQ, oRNQ, RNQ), neon_dyadic_i64_su), |
| 20657 | NUF(vqsub, 0000210, 3, (RNDQ, oRNDQ, RNDQ), neon_dyadic_i64_su), |
| 20658 | NUF(vqsubq, 0000210, 3, (RNQ, oRNQ, RNQ), neon_dyadic_i64_su), |
| 20659 | NUF(vrshl, 0000500, 3, (RNDQ, oRNDQ, RNDQ), neon_rshl), |
| 20660 | NUF(vrshlq, 0000500, 3, (RNQ, oRNQ, RNQ), neon_rshl), |
| 20661 | NUF(vqrshl, 0000510, 3, (RNDQ, oRNDQ, RNDQ), neon_rshl), |
| 20662 | NUF(vqrshlq, 0000510, 3, (RNQ, oRNQ, RNQ), neon_rshl), |
| 20663 | /* If not immediate, fall back to neon_dyadic_i64_su. |
| 20664 | shl_imm should accept I8 I16 I32 I64, |
| 20665 | qshl_imm should accept S8 S16 S32 S64 U8 U16 U32 U64. */ |
| 20666 | nUF(vshl, _vshl, 3, (RNDQ, oRNDQ, RNDQ_I63b), neon_shl_imm), |
| 20667 | nUF(vshlq, _vshl, 3, (RNQ, oRNQ, RNDQ_I63b), neon_shl_imm), |
| 20668 | nUF(vqshl, _vqshl, 3, (RNDQ, oRNDQ, RNDQ_I63b), neon_qshl_imm), |
| 20669 | nUF(vqshlq, _vqshl, 3, (RNQ, oRNQ, RNDQ_I63b), neon_qshl_imm), |
| 20670 | /* Logic ops, types optional & ignored. */ |
| 20671 | nUF(vand, _vand, 3, (RNDQ, oRNDQ, RNDQ_Ibig), neon_logic), |
| 20672 | nUF(vandq, _vand, 3, (RNQ, oRNQ, RNDQ_Ibig), neon_logic), |
| 20673 | nUF(vbic, _vbic, 3, (RNDQ, oRNDQ, RNDQ_Ibig), neon_logic), |
| 20674 | nUF(vbicq, _vbic, 3, (RNQ, oRNQ, RNDQ_Ibig), neon_logic), |
| 20675 | nUF(vorr, _vorr, 3, (RNDQ, oRNDQ, RNDQ_Ibig), neon_logic), |
| 20676 | nUF(vorrq, _vorr, 3, (RNQ, oRNQ, RNDQ_Ibig), neon_logic), |
| 20677 | nUF(vorn, _vorn, 3, (RNDQ, oRNDQ, RNDQ_Ibig), neon_logic), |
| 20678 | nUF(vornq, _vorn, 3, (RNQ, oRNQ, RNDQ_Ibig), neon_logic), |
| 20679 | nUF(veor, _veor, 3, (RNDQ, oRNDQ, RNDQ), neon_logic), |
| 20680 | nUF(veorq, _veor, 3, (RNQ, oRNQ, RNQ), neon_logic), |
| 20681 | /* Bitfield ops, untyped. */ |
| 20682 | NUF(vbsl, 1100110, 3, (RNDQ, RNDQ, RNDQ), neon_bitfield), |
| 20683 | NUF(vbslq, 1100110, 3, (RNQ, RNQ, RNQ), neon_bitfield), |
| 20684 | NUF(vbit, 1200110, 3, (RNDQ, RNDQ, RNDQ), neon_bitfield), |
| 20685 | NUF(vbitq, 1200110, 3, (RNQ, RNQ, RNQ), neon_bitfield), |
| 20686 | NUF(vbif, 1300110, 3, (RNDQ, RNDQ, RNDQ), neon_bitfield), |
| 20687 | NUF(vbifq, 1300110, 3, (RNQ, RNQ, RNQ), neon_bitfield), |
| 20688 | /* Int and float variants, types S8 S16 S32 U8 U16 U32 F16 F32. */ |
| 20689 | nUF(vabd, _vabd, 3, (RNDQ, oRNDQ, RNDQ), neon_dyadic_if_su), |
| 20690 | nUF(vabdq, _vabd, 3, (RNQ, oRNQ, RNQ), neon_dyadic_if_su), |
| 20691 | nUF(vmax, _vmax, 3, (RNDQ, oRNDQ, RNDQ), neon_dyadic_if_su), |
| 20692 | nUF(vmaxq, _vmax, 3, (RNQ, oRNQ, RNQ), neon_dyadic_if_su), |
| 20693 | nUF(vmin, _vmin, 3, (RNDQ, oRNDQ, RNDQ), neon_dyadic_if_su), |
| 20694 | nUF(vminq, _vmin, 3, (RNQ, oRNQ, RNQ), neon_dyadic_if_su), |
| 20695 | /* Comparisons. Types S8 S16 S32 U8 U16 U32 F32. Non-immediate versions fall |
| 20696 | back to neon_dyadic_if_su. */ |
| 20697 | nUF(vcge, _vcge, 3, (RNDQ, oRNDQ, RNDQ_I0), neon_cmp), |
| 20698 | nUF(vcgeq, _vcge, 3, (RNQ, oRNQ, RNDQ_I0), neon_cmp), |
| 20699 | nUF(vcgt, _vcgt, 3, (RNDQ, oRNDQ, RNDQ_I0), neon_cmp), |
| 20700 | nUF(vcgtq, _vcgt, 3, (RNQ, oRNQ, RNDQ_I0), neon_cmp), |
| 20701 | nUF(vclt, _vclt, 3, (RNDQ, oRNDQ, RNDQ_I0), neon_cmp_inv), |
| 20702 | nUF(vcltq, _vclt, 3, (RNQ, oRNQ, RNDQ_I0), neon_cmp_inv), |
| 20703 | nUF(vcle, _vcle, 3, (RNDQ, oRNDQ, RNDQ_I0), neon_cmp_inv), |
| 20704 | nUF(vcleq, _vcle, 3, (RNQ, oRNQ, RNDQ_I0), neon_cmp_inv), |
| 20705 | /* Comparison. Type I8 I16 I32 F32. */ |
| 20706 | nUF(vceq, _vceq, 3, (RNDQ, oRNDQ, RNDQ_I0), neon_ceq), |
| 20707 | nUF(vceqq, _vceq, 3, (RNQ, oRNQ, RNDQ_I0), neon_ceq), |
| 20708 | /* As above, D registers only. */ |
| 20709 | nUF(vpmax, _vpmax, 3, (RND, oRND, RND), neon_dyadic_if_su_d), |
| 20710 | nUF(vpmin, _vpmin, 3, (RND, oRND, RND), neon_dyadic_if_su_d), |
| 20711 | /* Int and float variants, signedness unimportant. */ |
| 20712 | nUF(vmlaq, _vmla, 3, (RNQ, oRNQ, RNDQ_RNSC), neon_mac_maybe_scalar), |
| 20713 | nUF(vmlsq, _vmls, 3, (RNQ, oRNQ, RNDQ_RNSC), neon_mac_maybe_scalar), |
| 20714 | nUF(vpadd, _vpadd, 3, (RND, oRND, RND), neon_dyadic_if_i_d), |
| 20715 | /* Add/sub take types I8 I16 I32 I64 F32. */ |
| 20716 | nUF(vaddq, _vadd, 3, (RNQ, oRNQ, RNQ), neon_addsub_if_i), |
| 20717 | nUF(vsubq, _vsub, 3, (RNQ, oRNQ, RNQ), neon_addsub_if_i), |
| 20718 | /* vtst takes sizes 8, 16, 32. */ |
| 20719 | NUF(vtst, 0000810, 3, (RNDQ, oRNDQ, RNDQ), neon_tst), |
| 20720 | NUF(vtstq, 0000810, 3, (RNQ, oRNQ, RNQ), neon_tst), |
| 20721 | /* VMUL takes I8 I16 I32 F32 P8. */ |
| 20722 | nUF(vmulq, _vmul, 3, (RNQ, oRNQ, RNDQ_RNSC), neon_mul), |
| 20723 | /* VQD{R}MULH takes S16 S32. */ |
| 20724 | nUF(vqdmulh, _vqdmulh, 3, (RNDQ, oRNDQ, RNDQ_RNSC), neon_qdmulh), |
| 20725 | nUF(vqdmulhq, _vqdmulh, 3, (RNQ, oRNQ, RNDQ_RNSC), neon_qdmulh), |
| 20726 | nUF(vqrdmulh, _vqrdmulh, 3, (RNDQ, oRNDQ, RNDQ_RNSC), neon_qdmulh), |
| 20727 | nUF(vqrdmulhq, _vqrdmulh, 3, (RNQ, oRNQ, RNDQ_RNSC), neon_qdmulh), |
| 20728 | NUF(vacge, 0000e10, 3, (RNDQ, oRNDQ, RNDQ), neon_fcmp_absolute), |
| 20729 | NUF(vacgeq, 0000e10, 3, (RNQ, oRNQ, RNQ), neon_fcmp_absolute), |
| 20730 | NUF(vacgt, 0200e10, 3, (RNDQ, oRNDQ, RNDQ), neon_fcmp_absolute), |
| 20731 | NUF(vacgtq, 0200e10, 3, (RNQ, oRNQ, RNQ), neon_fcmp_absolute), |
| 20732 | NUF(vaclt, 0200e10, 3, (RNDQ, oRNDQ, RNDQ), neon_fcmp_absolute_inv), |
| 20733 | NUF(vacltq, 0200e10, 3, (RNQ, oRNQ, RNQ), neon_fcmp_absolute_inv), |
| 20734 | NUF(vacle, 0000e10, 3, (RNDQ, oRNDQ, RNDQ), neon_fcmp_absolute_inv), |
| 20735 | NUF(vacleq, 0000e10, 3, (RNQ, oRNQ, RNQ), neon_fcmp_absolute_inv), |
| 20736 | NUF(vrecps, 0000f10, 3, (RNDQ, oRNDQ, RNDQ), neon_step), |
| 20737 | NUF(vrecpsq, 0000f10, 3, (RNQ, oRNQ, RNQ), neon_step), |
| 20738 | NUF(vrsqrts, 0200f10, 3, (RNDQ, oRNDQ, RNDQ), neon_step), |
| 20739 | NUF(vrsqrtsq, 0200f10, 3, (RNQ, oRNQ, RNQ), neon_step), |
| 20740 | /* ARM v8.1 extension. */ |
| 20741 | nUF (vqrdmlah, _vqrdmlah, 3, (RNDQ, oRNDQ, RNDQ_RNSC), neon_qrdmlah), |
| 20742 | nUF (vqrdmlahq, _vqrdmlah, 3, (RNQ, oRNQ, RNDQ_RNSC), neon_qrdmlah), |
| 20743 | nUF (vqrdmlsh, _vqrdmlsh, 3, (RNDQ, oRNDQ, RNDQ_RNSC), neon_qrdmlah), |
| 20744 | nUF (vqrdmlshq, _vqrdmlsh, 3, (RNQ, oRNQ, RNDQ_RNSC), neon_qrdmlah), |
| 20745 | |
| 20746 | /* Two address, int/float. Types S8 S16 S32 F32. */ |
| 20747 | NUF(vabsq, 1b10300, 2, (RNQ, RNQ), neon_abs_neg), |
| 20748 | NUF(vnegq, 1b10380, 2, (RNQ, RNQ), neon_abs_neg), |
| 20749 | |
| 20750 | /* Data processing with two registers and a shift amount. */ |
| 20751 | /* Right shifts, and variants with rounding. |
| 20752 | Types accepted S8 S16 S32 S64 U8 U16 U32 U64. */ |
| 20753 | NUF(vshr, 0800010, 3, (RNDQ, oRNDQ, I64z), neon_rshift_round_imm), |
| 20754 | NUF(vshrq, 0800010, 3, (RNQ, oRNQ, I64z), neon_rshift_round_imm), |
| 20755 | NUF(vrshr, 0800210, 3, (RNDQ, oRNDQ, I64z), neon_rshift_round_imm), |
| 20756 | NUF(vrshrq, 0800210, 3, (RNQ, oRNQ, I64z), neon_rshift_round_imm), |
| 20757 | NUF(vsra, 0800110, 3, (RNDQ, oRNDQ, I64), neon_rshift_round_imm), |
| 20758 | NUF(vsraq, 0800110, 3, (RNQ, oRNQ, I64), neon_rshift_round_imm), |
| 20759 | NUF(vrsra, 0800310, 3, (RNDQ, oRNDQ, I64), neon_rshift_round_imm), |
| 20760 | NUF(vrsraq, 0800310, 3, (RNQ, oRNQ, I64), neon_rshift_round_imm), |
| 20761 | /* Shift and insert. Sizes accepted 8 16 32 64. */ |
| 20762 | NUF(vsli, 1800510, 3, (RNDQ, oRNDQ, I63), neon_sli), |
| 20763 | NUF(vsliq, 1800510, 3, (RNQ, oRNQ, I63), neon_sli), |
| 20764 | NUF(vsri, 1800410, 3, (RNDQ, oRNDQ, I64), neon_sri), |
| 20765 | NUF(vsriq, 1800410, 3, (RNQ, oRNQ, I64), neon_sri), |
| 20766 | /* QSHL{U} immediate accepts S8 S16 S32 S64 U8 U16 U32 U64. */ |
| 20767 | NUF(vqshlu, 1800610, 3, (RNDQ, oRNDQ, I63), neon_qshlu_imm), |
| 20768 | NUF(vqshluq, 1800610, 3, (RNQ, oRNQ, I63), neon_qshlu_imm), |
| 20769 | /* Right shift immediate, saturating & narrowing, with rounding variants. |
| 20770 | Types accepted S16 S32 S64 U16 U32 U64. */ |
| 20771 | NUF(vqshrn, 0800910, 3, (RND, RNQ, I32z), neon_rshift_sat_narrow), |
| 20772 | NUF(vqrshrn, 0800950, 3, (RND, RNQ, I32z), neon_rshift_sat_narrow), |
| 20773 | /* As above, unsigned. Types accepted S16 S32 S64. */ |
| 20774 | NUF(vqshrun, 0800810, 3, (RND, RNQ, I32z), neon_rshift_sat_narrow_u), |
| 20775 | NUF(vqrshrun, 0800850, 3, (RND, RNQ, I32z), neon_rshift_sat_narrow_u), |
| 20776 | /* Right shift narrowing. Types accepted I16 I32 I64. */ |
| 20777 | NUF(vshrn, 0800810, 3, (RND, RNQ, I32z), neon_rshift_narrow), |
| 20778 | NUF(vrshrn, 0800850, 3, (RND, RNQ, I32z), neon_rshift_narrow), |
| 20779 | /* Special case. Types S8 S16 S32 U8 U16 U32. Handles max shift variant. */ |
| 20780 | nUF(vshll, _vshll, 3, (RNQ, RND, I32), neon_shll), |
| 20781 | /* CVT with optional immediate for fixed-point variant. */ |
| 20782 | nUF(vcvtq, _vcvt, 3, (RNQ, RNQ, oI32b), neon_cvt), |
| 20783 | |
| 20784 | nUF(vmvn, _vmvn, 2, (RNDQ, RNDQ_Ibig), neon_mvn), |
| 20785 | nUF(vmvnq, _vmvn, 2, (RNQ, RNDQ_Ibig), neon_mvn), |
| 20786 | |
| 20787 | /* Data processing, three registers of different lengths. */ |
| 20788 | /* Dyadic, long insns. Types S8 S16 S32 U8 U16 U32. */ |
| 20789 | NUF(vabal, 0800500, 3, (RNQ, RND, RND), neon_abal), |
| 20790 | NUF(vabdl, 0800700, 3, (RNQ, RND, RND), neon_dyadic_long), |
| 20791 | NUF(vaddl, 0800000, 3, (RNQ, RND, RND), neon_dyadic_long), |
| 20792 | NUF(vsubl, 0800200, 3, (RNQ, RND, RND), neon_dyadic_long), |
| 20793 | /* If not scalar, fall back to neon_dyadic_long. |
| 20794 | Vector types as above, scalar types S16 S32 U16 U32. */ |
| 20795 | nUF(vmlal, _vmlal, 3, (RNQ, RND, RND_RNSC), neon_mac_maybe_scalar_long), |
| 20796 | nUF(vmlsl, _vmlsl, 3, (RNQ, RND, RND_RNSC), neon_mac_maybe_scalar_long), |
| 20797 | /* Dyadic, widening insns. Types S8 S16 S32 U8 U16 U32. */ |
| 20798 | NUF(vaddw, 0800100, 3, (RNQ, oRNQ, RND), neon_dyadic_wide), |
| 20799 | NUF(vsubw, 0800300, 3, (RNQ, oRNQ, RND), neon_dyadic_wide), |
| 20800 | /* Dyadic, narrowing insns. Types I16 I32 I64. */ |
| 20801 | NUF(vaddhn, 0800400, 3, (RND, RNQ, RNQ), neon_dyadic_narrow), |
| 20802 | NUF(vraddhn, 1800400, 3, (RND, RNQ, RNQ), neon_dyadic_narrow), |
| 20803 | NUF(vsubhn, 0800600, 3, (RND, RNQ, RNQ), neon_dyadic_narrow), |
| 20804 | NUF(vrsubhn, 1800600, 3, (RND, RNQ, RNQ), neon_dyadic_narrow), |
| 20805 | /* Saturating doubling multiplies. Types S16 S32. */ |
| 20806 | nUF(vqdmlal, _vqdmlal, 3, (RNQ, RND, RND_RNSC), neon_mul_sat_scalar_long), |
| 20807 | nUF(vqdmlsl, _vqdmlsl, 3, (RNQ, RND, RND_RNSC), neon_mul_sat_scalar_long), |
| 20808 | nUF(vqdmull, _vqdmull, 3, (RNQ, RND, RND_RNSC), neon_mul_sat_scalar_long), |
| 20809 | /* VMULL. Vector types S8 S16 S32 U8 U16 U32 P8, scalar types |
| 20810 | S16 S32 U16 U32. */ |
| 20811 | nUF(vmull, _vmull, 3, (RNQ, RND, RND_RNSC), neon_vmull), |
| 20812 | |
| 20813 | /* Extract. Size 8. */ |
| 20814 | NUF(vext, 0b00000, 4, (RNDQ, oRNDQ, RNDQ, I15), neon_ext), |
| 20815 | NUF(vextq, 0b00000, 4, (RNQ, oRNQ, RNQ, I15), neon_ext), |
| 20816 | |
| 20817 | /* Two registers, miscellaneous. */ |
| 20818 | /* Reverse. Sizes 8 16 32 (must be < size in opcode). */ |
| 20819 | NUF(vrev64, 1b00000, 2, (RNDQ, RNDQ), neon_rev), |
| 20820 | NUF(vrev64q, 1b00000, 2, (RNQ, RNQ), neon_rev), |
| 20821 | NUF(vrev32, 1b00080, 2, (RNDQ, RNDQ), neon_rev), |
| 20822 | NUF(vrev32q, 1b00080, 2, (RNQ, RNQ), neon_rev), |
| 20823 | NUF(vrev16, 1b00100, 2, (RNDQ, RNDQ), neon_rev), |
| 20824 | NUF(vrev16q, 1b00100, 2, (RNQ, RNQ), neon_rev), |
| 20825 | /* Vector replicate. Sizes 8 16 32. */ |
| 20826 | nCE(vdup, _vdup, 2, (RNDQ, RR_RNSC), neon_dup), |
| 20827 | nCE(vdupq, _vdup, 2, (RNQ, RR_RNSC), neon_dup), |
| 20828 | /* VMOVL. Types S8 S16 S32 U8 U16 U32. */ |
| 20829 | NUF(vmovl, 0800a10, 2, (RNQ, RND), neon_movl), |
| 20830 | /* VMOVN. Types I16 I32 I64. */ |
| 20831 | nUF(vmovn, _vmovn, 2, (RND, RNQ), neon_movn), |
| 20832 | /* VQMOVN. Types S16 S32 S64 U16 U32 U64. */ |
| 20833 | nUF(vqmovn, _vqmovn, 2, (RND, RNQ), neon_qmovn), |
| 20834 | /* VQMOVUN. Types S16 S32 S64. */ |
| 20835 | nUF(vqmovun, _vqmovun, 2, (RND, RNQ), neon_qmovun), |
| 20836 | /* VZIP / VUZP. Sizes 8 16 32. */ |
| 20837 | NUF(vzip, 1b20180, 2, (RNDQ, RNDQ), neon_zip_uzp), |
| 20838 | NUF(vzipq, 1b20180, 2, (RNQ, RNQ), neon_zip_uzp), |
| 20839 | NUF(vuzp, 1b20100, 2, (RNDQ, RNDQ), neon_zip_uzp), |
| 20840 | NUF(vuzpq, 1b20100, 2, (RNQ, RNQ), neon_zip_uzp), |
| 20841 | /* VQABS / VQNEG. Types S8 S16 S32. */ |
| 20842 | NUF(vqabs, 1b00700, 2, (RNDQ, RNDQ), neon_sat_abs_neg), |
| 20843 | NUF(vqabsq, 1b00700, 2, (RNQ, RNQ), neon_sat_abs_neg), |
| 20844 | NUF(vqneg, 1b00780, 2, (RNDQ, RNDQ), neon_sat_abs_neg), |
| 20845 | NUF(vqnegq, 1b00780, 2, (RNQ, RNQ), neon_sat_abs_neg), |
| 20846 | /* Pairwise, lengthening. Types S8 S16 S32 U8 U16 U32. */ |
| 20847 | NUF(vpadal, 1b00600, 2, (RNDQ, RNDQ), neon_pair_long), |
| 20848 | NUF(vpadalq, 1b00600, 2, (RNQ, RNQ), neon_pair_long), |
| 20849 | NUF(vpaddl, 1b00200, 2, (RNDQ, RNDQ), neon_pair_long), |
| 20850 | NUF(vpaddlq, 1b00200, 2, (RNQ, RNQ), neon_pair_long), |
| 20851 | /* Reciprocal estimates. Types U32 F16 F32. */ |
| 20852 | NUF(vrecpe, 1b30400, 2, (RNDQ, RNDQ), neon_recip_est), |
| 20853 | NUF(vrecpeq, 1b30400, 2, (RNQ, RNQ), neon_recip_est), |
| 20854 | NUF(vrsqrte, 1b30480, 2, (RNDQ, RNDQ), neon_recip_est), |
| 20855 | NUF(vrsqrteq, 1b30480, 2, (RNQ, RNQ), neon_recip_est), |
| 20856 | /* VCLS. Types S8 S16 S32. */ |
| 20857 | NUF(vcls, 1b00400, 2, (RNDQ, RNDQ), neon_cls), |
| 20858 | NUF(vclsq, 1b00400, 2, (RNQ, RNQ), neon_cls), |
| 20859 | /* VCLZ. Types I8 I16 I32. */ |
| 20860 | NUF(vclz, 1b00480, 2, (RNDQ, RNDQ), neon_clz), |
| 20861 | NUF(vclzq, 1b00480, 2, (RNQ, RNQ), neon_clz), |
| 20862 | /* VCNT. Size 8. */ |
| 20863 | NUF(vcnt, 1b00500, 2, (RNDQ, RNDQ), neon_cnt), |
| 20864 | NUF(vcntq, 1b00500, 2, (RNQ, RNQ), neon_cnt), |
| 20865 | /* Two address, untyped. */ |
| 20866 | NUF(vswp, 1b20000, 2, (RNDQ, RNDQ), neon_swp), |
| 20867 | NUF(vswpq, 1b20000, 2, (RNQ, RNQ), neon_swp), |
| 20868 | /* VTRN. Sizes 8 16 32. */ |
| 20869 | nUF(vtrn, _vtrn, 2, (RNDQ, RNDQ), neon_trn), |
| 20870 | nUF(vtrnq, _vtrn, 2, (RNQ, RNQ), neon_trn), |
| 20871 | |
| 20872 | /* Table lookup. Size 8. */ |
| 20873 | NUF(vtbl, 1b00800, 3, (RND, NRDLST, RND), neon_tbl_tbx), |
| 20874 | NUF(vtbx, 1b00840, 3, (RND, NRDLST, RND), neon_tbl_tbx), |
| 20875 | |
| 20876 | #undef THUMB_VARIANT |
| 20877 | #define THUMB_VARIANT & fpu_vfp_v3_or_neon_ext |
| 20878 | #undef ARM_VARIANT |
| 20879 | #define ARM_VARIANT & fpu_vfp_v3_or_neon_ext |
| 20880 | |
| 20881 | /* Neon element/structure load/store. */ |
| 20882 | nUF(vld1, _vld1, 2, (NSTRLST, ADDR), neon_ldx_stx), |
| 20883 | nUF(vst1, _vst1, 2, (NSTRLST, ADDR), neon_ldx_stx), |
| 20884 | nUF(vld2, _vld2, 2, (NSTRLST, ADDR), neon_ldx_stx), |
| 20885 | nUF(vst2, _vst2, 2, (NSTRLST, ADDR), neon_ldx_stx), |
| 20886 | nUF(vld3, _vld3, 2, (NSTRLST, ADDR), neon_ldx_stx), |
| 20887 | nUF(vst3, _vst3, 2, (NSTRLST, ADDR), neon_ldx_stx), |
| 20888 | nUF(vld4, _vld4, 2, (NSTRLST, ADDR), neon_ldx_stx), |
| 20889 | nUF(vst4, _vst4, 2, (NSTRLST, ADDR), neon_ldx_stx), |
| 20890 | |
| 20891 | #undef THUMB_VARIANT |
| 20892 | #define THUMB_VARIANT & fpu_vfp_ext_v3xd |
| 20893 | #undef ARM_VARIANT |
| 20894 | #define ARM_VARIANT & fpu_vfp_ext_v3xd |
| 20895 | cCE("fconsts", eb00a00, 2, (RVS, I255), vfp_sp_const), |
| 20896 | cCE("fshtos", eba0a40, 2, (RVS, I16z), vfp_sp_conv_16), |
| 20897 | cCE("fsltos", eba0ac0, 2, (RVS, I32), vfp_sp_conv_32), |
| 20898 | cCE("fuhtos", ebb0a40, 2, (RVS, I16z), vfp_sp_conv_16), |
| 20899 | cCE("fultos", ebb0ac0, 2, (RVS, I32), vfp_sp_conv_32), |
| 20900 | cCE("ftoshs", ebe0a40, 2, (RVS, I16z), vfp_sp_conv_16), |
| 20901 | cCE("ftosls", ebe0ac0, 2, (RVS, I32), vfp_sp_conv_32), |
| 20902 | cCE("ftouhs", ebf0a40, 2, (RVS, I16z), vfp_sp_conv_16), |
| 20903 | cCE("ftouls", ebf0ac0, 2, (RVS, I32), vfp_sp_conv_32), |
| 20904 | |
| 20905 | #undef THUMB_VARIANT |
| 20906 | #define THUMB_VARIANT & fpu_vfp_ext_v3 |
| 20907 | #undef ARM_VARIANT |
| 20908 | #define ARM_VARIANT & fpu_vfp_ext_v3 |
| 20909 | |
| 20910 | cCE("fconstd", eb00b00, 2, (RVD, I255), vfp_dp_const), |
| 20911 | cCE("fshtod", eba0b40, 2, (RVD, I16z), vfp_dp_conv_16), |
| 20912 | cCE("fsltod", eba0bc0, 2, (RVD, I32), vfp_dp_conv_32), |
| 20913 | cCE("fuhtod", ebb0b40, 2, (RVD, I16z), vfp_dp_conv_16), |
| 20914 | cCE("fultod", ebb0bc0, 2, (RVD, I32), vfp_dp_conv_32), |
| 20915 | cCE("ftoshd", ebe0b40, 2, (RVD, I16z), vfp_dp_conv_16), |
| 20916 | cCE("ftosld", ebe0bc0, 2, (RVD, I32), vfp_dp_conv_32), |
| 20917 | cCE("ftouhd", ebf0b40, 2, (RVD, I16z), vfp_dp_conv_16), |
| 20918 | cCE("ftould", ebf0bc0, 2, (RVD, I32), vfp_dp_conv_32), |
| 20919 | |
| 20920 | #undef ARM_VARIANT |
| 20921 | #define ARM_VARIANT & fpu_vfp_ext_fma |
| 20922 | #undef THUMB_VARIANT |
| 20923 | #define THUMB_VARIANT & fpu_vfp_ext_fma |
| 20924 | /* Mnemonics shared by Neon and VFP. These are included in the |
| 20925 | VFP FMA variant; NEON and VFP FMA always includes the NEON |
| 20926 | FMA instructions. */ |
| 20927 | nCEF(vfma, _vfma, 3, (RNSDQ, oRNSDQ, RNSDQ), neon_fmac), |
| 20928 | nCEF(vfms, _vfms, 3, (RNSDQ, oRNSDQ, RNSDQ), neon_fmac), |
| 20929 | /* ffmas/ffmad/ffmss/ffmsd are dummy mnemonics to satisfy gas; |
| 20930 | the v form should always be used. */ |
| 20931 | cCE("ffmas", ea00a00, 3, (RVS, RVS, RVS), vfp_sp_dyadic), |
| 20932 | cCE("ffnmas", ea00a40, 3, (RVS, RVS, RVS), vfp_sp_dyadic), |
| 20933 | cCE("ffmad", ea00b00, 3, (RVD, RVD, RVD), vfp_dp_rd_rn_rm), |
| 20934 | cCE("ffnmad", ea00b40, 3, (RVD, RVD, RVD), vfp_dp_rd_rn_rm), |
| 20935 | nCE(vfnma, _vfnma, 3, (RVSD, RVSD, RVSD), vfp_nsyn_nmul), |
| 20936 | nCE(vfnms, _vfnms, 3, (RVSD, RVSD, RVSD), vfp_nsyn_nmul), |
| 20937 | |
| 20938 | #undef THUMB_VARIANT |
| 20939 | #undef ARM_VARIANT |
| 20940 | #define ARM_VARIANT & arm_cext_xscale /* Intel XScale extensions. */ |
| 20941 | |
| 20942 | cCE("mia", e200010, 3, (RXA, RRnpc, RRnpc), xsc_mia), |
| 20943 | cCE("miaph", e280010, 3, (RXA, RRnpc, RRnpc), xsc_mia), |
| 20944 | cCE("miabb", e2c0010, 3, (RXA, RRnpc, RRnpc), xsc_mia), |
| 20945 | cCE("miabt", e2d0010, 3, (RXA, RRnpc, RRnpc), xsc_mia), |
| 20946 | cCE("miatb", e2e0010, 3, (RXA, RRnpc, RRnpc), xsc_mia), |
| 20947 | cCE("miatt", e2f0010, 3, (RXA, RRnpc, RRnpc), xsc_mia), |
| 20948 | cCE("mar", c400000, 3, (RXA, RRnpc, RRnpc), xsc_mar), |
| 20949 | cCE("mra", c500000, 3, (RRnpc, RRnpc, RXA), xsc_mra), |
| 20950 | |
| 20951 | #undef ARM_VARIANT |
| 20952 | #define ARM_VARIANT & arm_cext_iwmmxt /* Intel Wireless MMX technology. */ |
| 20953 | |
| 20954 | cCE("tandcb", e13f130, 1, (RR), iwmmxt_tandorc), |
| 20955 | cCE("tandch", e53f130, 1, (RR), iwmmxt_tandorc), |
| 20956 | cCE("tandcw", e93f130, 1, (RR), iwmmxt_tandorc), |
| 20957 | cCE("tbcstb", e400010, 2, (RIWR, RR), rn_rd), |
| 20958 | cCE("tbcsth", e400050, 2, (RIWR, RR), rn_rd), |
| 20959 | cCE("tbcstw", e400090, 2, (RIWR, RR), rn_rd), |
| 20960 | cCE("textrcb", e130170, 2, (RR, I7), iwmmxt_textrc), |
| 20961 | cCE("textrch", e530170, 2, (RR, I7), iwmmxt_textrc), |
| 20962 | cCE("textrcw", e930170, 2, (RR, I7), iwmmxt_textrc), |
| 20963 | cCE("textrmub",e100070, 3, (RR, RIWR, I7), iwmmxt_textrm), |
| 20964 | cCE("textrmuh",e500070, 3, (RR, RIWR, I7), iwmmxt_textrm), |
| 20965 | cCE("textrmuw",e900070, 3, (RR, RIWR, I7), iwmmxt_textrm), |
| 20966 | cCE("textrmsb",e100078, 3, (RR, RIWR, I7), iwmmxt_textrm), |
| 20967 | cCE("textrmsh",e500078, 3, (RR, RIWR, I7), iwmmxt_textrm), |
| 20968 | cCE("textrmsw",e900078, 3, (RR, RIWR, I7), iwmmxt_textrm), |
| 20969 | cCE("tinsrb", e600010, 3, (RIWR, RR, I7), iwmmxt_tinsr), |
| 20970 | cCE("tinsrh", e600050, 3, (RIWR, RR, I7), iwmmxt_tinsr), |
| 20971 | cCE("tinsrw", e600090, 3, (RIWR, RR, I7), iwmmxt_tinsr), |
| 20972 | cCE("tmcr", e000110, 2, (RIWC_RIWG, RR), rn_rd), |
| 20973 | cCE("tmcrr", c400000, 3, (RIWR, RR, RR), rm_rd_rn), |
| 20974 | cCE("tmia", e200010, 3, (RIWR, RR, RR), iwmmxt_tmia), |
| 20975 | cCE("tmiaph", e280010, 3, (RIWR, RR, RR), iwmmxt_tmia), |
| 20976 | cCE("tmiabb", e2c0010, 3, (RIWR, RR, RR), iwmmxt_tmia), |
| 20977 | cCE("tmiabt", e2d0010, 3, (RIWR, RR, RR), iwmmxt_tmia), |
| 20978 | cCE("tmiatb", e2e0010, 3, (RIWR, RR, RR), iwmmxt_tmia), |
| 20979 | cCE("tmiatt", e2f0010, 3, (RIWR, RR, RR), iwmmxt_tmia), |
| 20980 | cCE("tmovmskb",e100030, 2, (RR, RIWR), rd_rn), |
| 20981 | cCE("tmovmskh",e500030, 2, (RR, RIWR), rd_rn), |
| 20982 | cCE("tmovmskw",e900030, 2, (RR, RIWR), rd_rn), |
| 20983 | cCE("tmrc", e100110, 2, (RR, RIWC_RIWG), rd_rn), |
| 20984 | cCE("tmrrc", c500000, 3, (RR, RR, RIWR), rd_rn_rm), |
| 20985 | cCE("torcb", e13f150, 1, (RR), iwmmxt_tandorc), |
| 20986 | cCE("torch", e53f150, 1, (RR), iwmmxt_tandorc), |
| 20987 | cCE("torcw", e93f150, 1, (RR), iwmmxt_tandorc), |
| 20988 | cCE("waccb", e0001c0, 2, (RIWR, RIWR), rd_rn), |
| 20989 | cCE("wacch", e4001c0, 2, (RIWR, RIWR), rd_rn), |
| 20990 | cCE("waccw", e8001c0, 2, (RIWR, RIWR), rd_rn), |
| 20991 | cCE("waddbss", e300180, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 20992 | cCE("waddb", e000180, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 20993 | cCE("waddbus", e100180, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 20994 | cCE("waddhss", e700180, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 20995 | cCE("waddh", e400180, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 20996 | cCE("waddhus", e500180, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 20997 | cCE("waddwss", eb00180, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 20998 | cCE("waddw", e800180, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 20999 | cCE("waddwus", e900180, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21000 | cCE("waligni", e000020, 4, (RIWR, RIWR, RIWR, I7), iwmmxt_waligni), |
| 21001 | cCE("walignr0",e800020, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21002 | cCE("walignr1",e900020, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21003 | cCE("walignr2",ea00020, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21004 | cCE("walignr3",eb00020, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21005 | cCE("wand", e200000, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21006 | cCE("wandn", e300000, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21007 | cCE("wavg2b", e800000, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21008 | cCE("wavg2br", e900000, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21009 | cCE("wavg2h", ec00000, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21010 | cCE("wavg2hr", ed00000, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21011 | cCE("wcmpeqb", e000060, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21012 | cCE("wcmpeqh", e400060, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21013 | cCE("wcmpeqw", e800060, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21014 | cCE("wcmpgtub",e100060, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21015 | cCE("wcmpgtuh",e500060, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21016 | cCE("wcmpgtuw",e900060, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21017 | cCE("wcmpgtsb",e300060, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21018 | cCE("wcmpgtsh",e700060, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21019 | cCE("wcmpgtsw",eb00060, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21020 | cCE("wldrb", c100000, 2, (RIWR, ADDR), iwmmxt_wldstbh), |
| 21021 | cCE("wldrh", c500000, 2, (RIWR, ADDR), iwmmxt_wldstbh), |
| 21022 | cCE("wldrw", c100100, 2, (RIWR_RIWC, ADDR), iwmmxt_wldstw), |
| 21023 | cCE("wldrd", c500100, 2, (RIWR, ADDR), iwmmxt_wldstd), |
| 21024 | cCE("wmacs", e600100, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21025 | cCE("wmacsz", e700100, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21026 | cCE("wmacu", e400100, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21027 | cCE("wmacuz", e500100, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21028 | cCE("wmadds", ea00100, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21029 | cCE("wmaddu", e800100, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21030 | cCE("wmaxsb", e200160, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21031 | cCE("wmaxsh", e600160, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21032 | cCE("wmaxsw", ea00160, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21033 | cCE("wmaxub", e000160, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21034 | cCE("wmaxuh", e400160, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21035 | cCE("wmaxuw", e800160, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21036 | cCE("wminsb", e300160, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21037 | cCE("wminsh", e700160, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21038 | cCE("wminsw", eb00160, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21039 | cCE("wminub", e100160, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21040 | cCE("wminuh", e500160, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21041 | cCE("wminuw", e900160, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21042 | cCE("wmov", e000000, 2, (RIWR, RIWR), iwmmxt_wmov), |
| 21043 | cCE("wmulsm", e300100, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21044 | cCE("wmulsl", e200100, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21045 | cCE("wmulum", e100100, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21046 | cCE("wmulul", e000100, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21047 | cCE("wor", e000000, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21048 | cCE("wpackhss",e700080, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21049 | cCE("wpackhus",e500080, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21050 | cCE("wpackwss",eb00080, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21051 | cCE("wpackwus",e900080, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21052 | cCE("wpackdss",ef00080, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21053 | cCE("wpackdus",ed00080, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21054 | cCE("wrorh", e700040, 3, (RIWR, RIWR, RIWR_I32z),iwmmxt_wrwrwr_or_imm5), |
| 21055 | cCE("wrorhg", e700148, 3, (RIWR, RIWR, RIWG), rd_rn_rm), |
| 21056 | cCE("wrorw", eb00040, 3, (RIWR, RIWR, RIWR_I32z),iwmmxt_wrwrwr_or_imm5), |
| 21057 | cCE("wrorwg", eb00148, 3, (RIWR, RIWR, RIWG), rd_rn_rm), |
| 21058 | cCE("wrord", ef00040, 3, (RIWR, RIWR, RIWR_I32z),iwmmxt_wrwrwr_or_imm5), |
| 21059 | cCE("wrordg", ef00148, 3, (RIWR, RIWR, RIWG), rd_rn_rm), |
| 21060 | cCE("wsadb", e000120, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21061 | cCE("wsadbz", e100120, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21062 | cCE("wsadh", e400120, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21063 | cCE("wsadhz", e500120, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21064 | cCE("wshufh", e0001e0, 3, (RIWR, RIWR, I255), iwmmxt_wshufh), |
| 21065 | cCE("wsllh", e500040, 3, (RIWR, RIWR, RIWR_I32z),iwmmxt_wrwrwr_or_imm5), |
| 21066 | cCE("wsllhg", e500148, 3, (RIWR, RIWR, RIWG), rd_rn_rm), |
| 21067 | cCE("wsllw", e900040, 3, (RIWR, RIWR, RIWR_I32z),iwmmxt_wrwrwr_or_imm5), |
| 21068 | cCE("wsllwg", e900148, 3, (RIWR, RIWR, RIWG), rd_rn_rm), |
| 21069 | cCE("wslld", ed00040, 3, (RIWR, RIWR, RIWR_I32z),iwmmxt_wrwrwr_or_imm5), |
| 21070 | cCE("wslldg", ed00148, 3, (RIWR, RIWR, RIWG), rd_rn_rm), |
| 21071 | cCE("wsrah", e400040, 3, (RIWR, RIWR, RIWR_I32z),iwmmxt_wrwrwr_or_imm5), |
| 21072 | cCE("wsrahg", e400148, 3, (RIWR, RIWR, RIWG), rd_rn_rm), |
| 21073 | cCE("wsraw", e800040, 3, (RIWR, RIWR, RIWR_I32z),iwmmxt_wrwrwr_or_imm5), |
| 21074 | cCE("wsrawg", e800148, 3, (RIWR, RIWR, RIWG), rd_rn_rm), |
| 21075 | cCE("wsrad", ec00040, 3, (RIWR, RIWR, RIWR_I32z),iwmmxt_wrwrwr_or_imm5), |
| 21076 | cCE("wsradg", ec00148, 3, (RIWR, RIWR, RIWG), rd_rn_rm), |
| 21077 | cCE("wsrlh", e600040, 3, (RIWR, RIWR, RIWR_I32z),iwmmxt_wrwrwr_or_imm5), |
| 21078 | cCE("wsrlhg", e600148, 3, (RIWR, RIWR, RIWG), rd_rn_rm), |
| 21079 | cCE("wsrlw", ea00040, 3, (RIWR, RIWR, RIWR_I32z),iwmmxt_wrwrwr_or_imm5), |
| 21080 | cCE("wsrlwg", ea00148, 3, (RIWR, RIWR, RIWG), rd_rn_rm), |
| 21081 | cCE("wsrld", ee00040, 3, (RIWR, RIWR, RIWR_I32z),iwmmxt_wrwrwr_or_imm5), |
| 21082 | cCE("wsrldg", ee00148, 3, (RIWR, RIWR, RIWG), rd_rn_rm), |
| 21083 | cCE("wstrb", c000000, 2, (RIWR, ADDR), iwmmxt_wldstbh), |
| 21084 | cCE("wstrh", c400000, 2, (RIWR, ADDR), iwmmxt_wldstbh), |
| 21085 | cCE("wstrw", c000100, 2, (RIWR_RIWC, ADDR), iwmmxt_wldstw), |
| 21086 | cCE("wstrd", c400100, 2, (RIWR, ADDR), iwmmxt_wldstd), |
| 21087 | cCE("wsubbss", e3001a0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21088 | cCE("wsubb", e0001a0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21089 | cCE("wsubbus", e1001a0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21090 | cCE("wsubhss", e7001a0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21091 | cCE("wsubh", e4001a0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21092 | cCE("wsubhus", e5001a0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21093 | cCE("wsubwss", eb001a0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21094 | cCE("wsubw", e8001a0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21095 | cCE("wsubwus", e9001a0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21096 | cCE("wunpckehub",e0000c0, 2, (RIWR, RIWR), rd_rn), |
| 21097 | cCE("wunpckehuh",e4000c0, 2, (RIWR, RIWR), rd_rn), |
| 21098 | cCE("wunpckehuw",e8000c0, 2, (RIWR, RIWR), rd_rn), |
| 21099 | cCE("wunpckehsb",e2000c0, 2, (RIWR, RIWR), rd_rn), |
| 21100 | cCE("wunpckehsh",e6000c0, 2, (RIWR, RIWR), rd_rn), |
| 21101 | cCE("wunpckehsw",ea000c0, 2, (RIWR, RIWR), rd_rn), |
| 21102 | cCE("wunpckihb", e1000c0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21103 | cCE("wunpckihh", e5000c0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21104 | cCE("wunpckihw", e9000c0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21105 | cCE("wunpckelub",e0000e0, 2, (RIWR, RIWR), rd_rn), |
| 21106 | cCE("wunpckeluh",e4000e0, 2, (RIWR, RIWR), rd_rn), |
| 21107 | cCE("wunpckeluw",e8000e0, 2, (RIWR, RIWR), rd_rn), |
| 21108 | cCE("wunpckelsb",e2000e0, 2, (RIWR, RIWR), rd_rn), |
| 21109 | cCE("wunpckelsh",e6000e0, 2, (RIWR, RIWR), rd_rn), |
| 21110 | cCE("wunpckelsw",ea000e0, 2, (RIWR, RIWR), rd_rn), |
| 21111 | cCE("wunpckilb", e1000e0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21112 | cCE("wunpckilh", e5000e0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21113 | cCE("wunpckilw", e9000e0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21114 | cCE("wxor", e100000, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21115 | cCE("wzero", e300000, 1, (RIWR), iwmmxt_wzero), |
| 21116 | |
| 21117 | #undef ARM_VARIANT |
| 21118 | #define ARM_VARIANT & arm_cext_iwmmxt2 /* Intel Wireless MMX technology, version 2. */ |
| 21119 | |
| 21120 | cCE("torvscb", e12f190, 1, (RR), iwmmxt_tandorc), |
| 21121 | cCE("torvsch", e52f190, 1, (RR), iwmmxt_tandorc), |
| 21122 | cCE("torvscw", e92f190, 1, (RR), iwmmxt_tandorc), |
| 21123 | cCE("wabsb", e2001c0, 2, (RIWR, RIWR), rd_rn), |
| 21124 | cCE("wabsh", e6001c0, 2, (RIWR, RIWR), rd_rn), |
| 21125 | cCE("wabsw", ea001c0, 2, (RIWR, RIWR), rd_rn), |
| 21126 | cCE("wabsdiffb", e1001c0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21127 | cCE("wabsdiffh", e5001c0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21128 | cCE("wabsdiffw", e9001c0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21129 | cCE("waddbhusl", e2001a0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21130 | cCE("waddbhusm", e6001a0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21131 | cCE("waddhc", e600180, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21132 | cCE("waddwc", ea00180, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21133 | cCE("waddsubhx", ea001a0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21134 | cCE("wavg4", e400000, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21135 | cCE("wavg4r", e500000, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21136 | cCE("wmaddsn", ee00100, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21137 | cCE("wmaddsx", eb00100, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21138 | cCE("wmaddun", ec00100, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21139 | cCE("wmaddux", e900100, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21140 | cCE("wmerge", e000080, 4, (RIWR, RIWR, RIWR, I7), iwmmxt_wmerge), |
| 21141 | cCE("wmiabb", e0000a0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21142 | cCE("wmiabt", e1000a0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21143 | cCE("wmiatb", e2000a0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21144 | cCE("wmiatt", e3000a0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21145 | cCE("wmiabbn", e4000a0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21146 | cCE("wmiabtn", e5000a0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21147 | cCE("wmiatbn", e6000a0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21148 | cCE("wmiattn", e7000a0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21149 | cCE("wmiawbb", e800120, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21150 | cCE("wmiawbt", e900120, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21151 | cCE("wmiawtb", ea00120, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21152 | cCE("wmiawtt", eb00120, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21153 | cCE("wmiawbbn", ec00120, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21154 | cCE("wmiawbtn", ed00120, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21155 | cCE("wmiawtbn", ee00120, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21156 | cCE("wmiawttn", ef00120, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21157 | cCE("wmulsmr", ef00100, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21158 | cCE("wmulumr", ed00100, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21159 | cCE("wmulwumr", ec000c0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21160 | cCE("wmulwsmr", ee000c0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21161 | cCE("wmulwum", ed000c0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21162 | cCE("wmulwsm", ef000c0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21163 | cCE("wmulwl", eb000c0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21164 | cCE("wqmiabb", e8000a0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21165 | cCE("wqmiabt", e9000a0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21166 | cCE("wqmiatb", ea000a0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21167 | cCE("wqmiatt", eb000a0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21168 | cCE("wqmiabbn", ec000a0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21169 | cCE("wqmiabtn", ed000a0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21170 | cCE("wqmiatbn", ee000a0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21171 | cCE("wqmiattn", ef000a0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21172 | cCE("wqmulm", e100080, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21173 | cCE("wqmulmr", e300080, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21174 | cCE("wqmulwm", ec000e0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21175 | cCE("wqmulwmr", ee000e0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21176 | cCE("wsubaddhx", ed001c0, 3, (RIWR, RIWR, RIWR), rd_rn_rm), |
| 21177 | |
| 21178 | #undef ARM_VARIANT |
| 21179 | #define ARM_VARIANT & arm_cext_maverick /* Cirrus Maverick instructions. */ |
| 21180 | |
| 21181 | cCE("cfldrs", c100400, 2, (RMF, ADDRGLDC), rd_cpaddr), |
| 21182 | cCE("cfldrd", c500400, 2, (RMD, ADDRGLDC), rd_cpaddr), |
| 21183 | cCE("cfldr32", c100500, 2, (RMFX, ADDRGLDC), rd_cpaddr), |
| 21184 | cCE("cfldr64", c500500, 2, (RMDX, ADDRGLDC), rd_cpaddr), |
| 21185 | cCE("cfstrs", c000400, 2, (RMF, ADDRGLDC), rd_cpaddr), |
| 21186 | cCE("cfstrd", c400400, 2, (RMD, ADDRGLDC), rd_cpaddr), |
| 21187 | cCE("cfstr32", c000500, 2, (RMFX, ADDRGLDC), rd_cpaddr), |
| 21188 | cCE("cfstr64", c400500, 2, (RMDX, ADDRGLDC), rd_cpaddr), |
| 21189 | cCE("cfmvsr", e000450, 2, (RMF, RR), rn_rd), |
| 21190 | cCE("cfmvrs", e100450, 2, (RR, RMF), rd_rn), |
| 21191 | cCE("cfmvdlr", e000410, 2, (RMD, RR), rn_rd), |
| 21192 | cCE("cfmvrdl", e100410, 2, (RR, RMD), rd_rn), |
| 21193 | cCE("cfmvdhr", e000430, 2, (RMD, RR), rn_rd), |
| 21194 | cCE("cfmvrdh", e100430, 2, (RR, RMD), rd_rn), |
| 21195 | cCE("cfmv64lr",e000510, 2, (RMDX, RR), rn_rd), |
| 21196 | cCE("cfmvr64l",e100510, 2, (RR, RMDX), rd_rn), |
| 21197 | cCE("cfmv64hr",e000530, 2, (RMDX, RR), rn_rd), |
| 21198 | cCE("cfmvr64h",e100530, 2, (RR, RMDX), rd_rn), |
| 21199 | cCE("cfmval32",e200440, 2, (RMAX, RMFX), rd_rn), |
| 21200 | cCE("cfmv32al",e100440, 2, (RMFX, RMAX), rd_rn), |
| 21201 | cCE("cfmvam32",e200460, 2, (RMAX, RMFX), rd_rn), |
| 21202 | cCE("cfmv32am",e100460, 2, (RMFX, RMAX), rd_rn), |
| 21203 | cCE("cfmvah32",e200480, 2, (RMAX, RMFX), rd_rn), |
| 21204 | cCE("cfmv32ah",e100480, 2, (RMFX, RMAX), rd_rn), |
| 21205 | cCE("cfmva32", e2004a0, 2, (RMAX, RMFX), rd_rn), |
| 21206 | cCE("cfmv32a", e1004a0, 2, (RMFX, RMAX), rd_rn), |
| 21207 | cCE("cfmva64", e2004c0, 2, (RMAX, RMDX), rd_rn), |
| 21208 | cCE("cfmv64a", e1004c0, 2, (RMDX, RMAX), rd_rn), |
| 21209 | cCE("cfmvsc32",e2004e0, 2, (RMDS, RMDX), mav_dspsc), |
| 21210 | cCE("cfmv32sc",e1004e0, 2, (RMDX, RMDS), rd), |
| 21211 | cCE("cfcpys", e000400, 2, (RMF, RMF), rd_rn), |
| 21212 | cCE("cfcpyd", e000420, 2, (RMD, RMD), rd_rn), |
| 21213 | cCE("cfcvtsd", e000460, 2, (RMD, RMF), rd_rn), |
| 21214 | cCE("cfcvtds", e000440, 2, (RMF, RMD), rd_rn), |
| 21215 | cCE("cfcvt32s",e000480, 2, (RMF, RMFX), rd_rn), |
| 21216 | cCE("cfcvt32d",e0004a0, 2, (RMD, RMFX), rd_rn), |
| 21217 | cCE("cfcvt64s",e0004c0, 2, (RMF, RMDX), rd_rn), |
| 21218 | cCE("cfcvt64d",e0004e0, 2, (RMD, RMDX), rd_rn), |
| 21219 | cCE("cfcvts32",e100580, 2, (RMFX, RMF), rd_rn), |
| 21220 | cCE("cfcvtd32",e1005a0, 2, (RMFX, RMD), rd_rn), |
| 21221 | cCE("cftruncs32",e1005c0, 2, (RMFX, RMF), rd_rn), |
| 21222 | cCE("cftruncd32",e1005e0, 2, (RMFX, RMD), rd_rn), |
| 21223 | cCE("cfrshl32",e000550, 3, (RMFX, RMFX, RR), mav_triple), |
| 21224 | cCE("cfrshl64",e000570, 3, (RMDX, RMDX, RR), mav_triple), |
| 21225 | cCE("cfsh32", e000500, 3, (RMFX, RMFX, I63s), mav_shift), |
| 21226 | cCE("cfsh64", e200500, 3, (RMDX, RMDX, I63s), mav_shift), |
| 21227 | cCE("cfcmps", e100490, 3, (RR, RMF, RMF), rd_rn_rm), |
| 21228 | cCE("cfcmpd", e1004b0, 3, (RR, RMD, RMD), rd_rn_rm), |
| 21229 | cCE("cfcmp32", e100590, 3, (RR, RMFX, RMFX), rd_rn_rm), |
| 21230 | cCE("cfcmp64", e1005b0, 3, (RR, RMDX, RMDX), rd_rn_rm), |
| 21231 | cCE("cfabss", e300400, 2, (RMF, RMF), rd_rn), |
| 21232 | cCE("cfabsd", e300420, 2, (RMD, RMD), rd_rn), |
| 21233 | cCE("cfnegs", e300440, 2, (RMF, RMF), rd_rn), |
| 21234 | cCE("cfnegd", e300460, 2, (RMD, RMD), rd_rn), |
| 21235 | cCE("cfadds", e300480, 3, (RMF, RMF, RMF), rd_rn_rm), |
| 21236 | cCE("cfaddd", e3004a0, 3, (RMD, RMD, RMD), rd_rn_rm), |
| 21237 | cCE("cfsubs", e3004c0, 3, (RMF, RMF, RMF), rd_rn_rm), |
| 21238 | cCE("cfsubd", e3004e0, 3, (RMD, RMD, RMD), rd_rn_rm), |
| 21239 | cCE("cfmuls", e100400, 3, (RMF, RMF, RMF), rd_rn_rm), |
| 21240 | cCE("cfmuld", e100420, 3, (RMD, RMD, RMD), rd_rn_rm), |
| 21241 | cCE("cfabs32", e300500, 2, (RMFX, RMFX), rd_rn), |
| 21242 | cCE("cfabs64", e300520, 2, (RMDX, RMDX), rd_rn), |
| 21243 | cCE("cfneg32", e300540, 2, (RMFX, RMFX), rd_rn), |
| 21244 | cCE("cfneg64", e300560, 2, (RMDX, RMDX), rd_rn), |
| 21245 | cCE("cfadd32", e300580, 3, (RMFX, RMFX, RMFX), rd_rn_rm), |
| 21246 | cCE("cfadd64", e3005a0, 3, (RMDX, RMDX, RMDX), rd_rn_rm), |
| 21247 | cCE("cfsub32", e3005c0, 3, (RMFX, RMFX, RMFX), rd_rn_rm), |
| 21248 | cCE("cfsub64", e3005e0, 3, (RMDX, RMDX, RMDX), rd_rn_rm), |
| 21249 | cCE("cfmul32", e100500, 3, (RMFX, RMFX, RMFX), rd_rn_rm), |
| 21250 | cCE("cfmul64", e100520, 3, (RMDX, RMDX, RMDX), rd_rn_rm), |
| 21251 | cCE("cfmac32", e100540, 3, (RMFX, RMFX, RMFX), rd_rn_rm), |
| 21252 | cCE("cfmsc32", e100560, 3, (RMFX, RMFX, RMFX), rd_rn_rm), |
| 21253 | cCE("cfmadd32",e000600, 4, (RMAX, RMFX, RMFX, RMFX), mav_quad), |
| 21254 | cCE("cfmsub32",e100600, 4, (RMAX, RMFX, RMFX, RMFX), mav_quad), |
| 21255 | cCE("cfmadda32", e200600, 4, (RMAX, RMAX, RMFX, RMFX), mav_quad), |
| 21256 | cCE("cfmsuba32", e300600, 4, (RMAX, RMAX, RMFX, RMFX), mav_quad), |
| 21257 | |
| 21258 | /* ARMv8-M instructions. */ |
| 21259 | #undef ARM_VARIANT |
| 21260 | #define ARM_VARIANT NULL |
| 21261 | #undef THUMB_VARIANT |
| 21262 | #define THUMB_VARIANT & arm_ext_v8m |
| 21263 | TUE("sg", 0, e97fe97f, 0, (), 0, noargs), |
| 21264 | TUE("blxns", 0, 4784, 1, (RRnpc), 0, t_blx), |
| 21265 | TUE("bxns", 0, 4704, 1, (RRnpc), 0, t_bx), |
| 21266 | TUE("tt", 0, e840f000, 2, (RRnpc, RRnpc), 0, tt), |
| 21267 | TUE("ttt", 0, e840f040, 2, (RRnpc, RRnpc), 0, tt), |
| 21268 | TUE("tta", 0, e840f080, 2, (RRnpc, RRnpc), 0, tt), |
| 21269 | TUE("ttat", 0, e840f0c0, 2, (RRnpc, RRnpc), 0, tt), |
| 21270 | |
| 21271 | /* FP for ARMv8-M Mainline. Enabled for ARMv8-M Mainline because the |
| 21272 | instructions behave as nop if no VFP is present. */ |
| 21273 | #undef THUMB_VARIANT |
| 21274 | #define THUMB_VARIANT & arm_ext_v8m_main |
| 21275 | TUEc("vlldm", 0, ec300a00, 1, (RRnpc), rn), |
| 21276 | TUEc("vlstm", 0, ec200a00, 1, (RRnpc), rn), |
| 21277 | }; |
| 21278 | #undef ARM_VARIANT |
| 21279 | #undef THUMB_VARIANT |
| 21280 | #undef TCE |
| 21281 | #undef TUE |
| 21282 | #undef TUF |
| 21283 | #undef TCC |
| 21284 | #undef cCE |
| 21285 | #undef cCL |
| 21286 | #undef C3E |
| 21287 | #undef CE |
| 21288 | #undef CM |
| 21289 | #undef UE |
| 21290 | #undef UF |
| 21291 | #undef UT |
| 21292 | #undef NUF |
| 21293 | #undef nUF |
| 21294 | #undef NCE |
| 21295 | #undef nCE |
| 21296 | #undef OPS0 |
| 21297 | #undef OPS1 |
| 21298 | #undef OPS2 |
| 21299 | #undef OPS3 |
| 21300 | #undef OPS4 |
| 21301 | #undef OPS5 |
| 21302 | #undef OPS6 |
| 21303 | #undef do_0 |
| 21304 | \f |
| 21305 | /* MD interface: bits in the object file. */ |
| 21306 | |
| 21307 | /* Turn an integer of n bytes (in val) into a stream of bytes appropriate |
| 21308 | for use in the a.out file, and stores them in the array pointed to by buf. |
| 21309 | This knows about the endian-ness of the target machine and does |
| 21310 | THE RIGHT THING, whatever it is. Possible values for n are 1 (byte) |
| 21311 | 2 (short) and 4 (long) Floating numbers are put out as a series of |
| 21312 | LITTLENUMS (shorts, here at least). */ |
| 21313 | |
| 21314 | void |
| 21315 | md_number_to_chars (char * buf, valueT val, int n) |
| 21316 | { |
| 21317 | if (target_big_endian) |
| 21318 | number_to_chars_bigendian (buf, val, n); |
| 21319 | else |
| 21320 | number_to_chars_littleendian (buf, val, n); |
| 21321 | } |
| 21322 | |
| 21323 | static valueT |
| 21324 | md_chars_to_number (char * buf, int n) |
| 21325 | { |
| 21326 | valueT result = 0; |
| 21327 | unsigned char * where = (unsigned char *) buf; |
| 21328 | |
| 21329 | if (target_big_endian) |
| 21330 | { |
| 21331 | while (n--) |
| 21332 | { |
| 21333 | result <<= 8; |
| 21334 | result |= (*where++ & 255); |
| 21335 | } |
| 21336 | } |
| 21337 | else |
| 21338 | { |
| 21339 | while (n--) |
| 21340 | { |
| 21341 | result <<= 8; |
| 21342 | result |= (where[n] & 255); |
| 21343 | } |
| 21344 | } |
| 21345 | |
| 21346 | return result; |
| 21347 | } |
| 21348 | |
| 21349 | /* MD interface: Sections. */ |
| 21350 | |
| 21351 | /* Calculate the maximum variable size (i.e., excluding fr_fix) |
| 21352 | that an rs_machine_dependent frag may reach. */ |
| 21353 | |
| 21354 | unsigned int |
| 21355 | arm_frag_max_var (fragS *fragp) |
| 21356 | { |
| 21357 | /* We only use rs_machine_dependent for variable-size Thumb instructions, |
| 21358 | which are either THUMB_SIZE (2) or INSN_SIZE (4). |
| 21359 | |
| 21360 | Note that we generate relaxable instructions even for cases that don't |
| 21361 | really need it, like an immediate that's a trivial constant. So we're |
| 21362 | overestimating the instruction size for some of those cases. Rather |
| 21363 | than putting more intelligence here, it would probably be better to |
| 21364 | avoid generating a relaxation frag in the first place when it can be |
| 21365 | determined up front that a short instruction will suffice. */ |
| 21366 | |
| 21367 | gas_assert (fragp->fr_type == rs_machine_dependent); |
| 21368 | return INSN_SIZE; |
| 21369 | } |
| 21370 | |
| 21371 | /* Estimate the size of a frag before relaxing. Assume everything fits in |
| 21372 | 2 bytes. */ |
| 21373 | |
| 21374 | int |
| 21375 | md_estimate_size_before_relax (fragS * fragp, |
| 21376 | segT segtype ATTRIBUTE_UNUSED) |
| 21377 | { |
| 21378 | fragp->fr_var = 2; |
| 21379 | return 2; |
| 21380 | } |
| 21381 | |
| 21382 | /* Convert a machine dependent frag. */ |
| 21383 | |
| 21384 | void |
| 21385 | md_convert_frag (bfd *abfd, segT asec ATTRIBUTE_UNUSED, fragS *fragp) |
| 21386 | { |
| 21387 | unsigned long insn; |
| 21388 | unsigned long old_op; |
| 21389 | char *buf; |
| 21390 | expressionS exp; |
| 21391 | fixS *fixp; |
| 21392 | int reloc_type; |
| 21393 | int pc_rel; |
| 21394 | int opcode; |
| 21395 | |
| 21396 | buf = fragp->fr_literal + fragp->fr_fix; |
| 21397 | |
| 21398 | old_op = bfd_get_16(abfd, buf); |
| 21399 | if (fragp->fr_symbol) |
| 21400 | { |
| 21401 | exp.X_op = O_symbol; |
| 21402 | exp.X_add_symbol = fragp->fr_symbol; |
| 21403 | } |
| 21404 | else |
| 21405 | { |
| 21406 | exp.X_op = O_constant; |
| 21407 | } |
| 21408 | exp.X_add_number = fragp->fr_offset; |
| 21409 | opcode = fragp->fr_subtype; |
| 21410 | switch (opcode) |
| 21411 | { |
| 21412 | case T_MNEM_ldr_pc: |
| 21413 | case T_MNEM_ldr_pc2: |
| 21414 | case T_MNEM_ldr_sp: |
| 21415 | case T_MNEM_str_sp: |
| 21416 | case T_MNEM_ldr: |
| 21417 | case T_MNEM_ldrb: |
| 21418 | case T_MNEM_ldrh: |
| 21419 | case T_MNEM_str: |
| 21420 | case T_MNEM_strb: |
| 21421 | case T_MNEM_strh: |
| 21422 | if (fragp->fr_var == 4) |
| 21423 | { |
| 21424 | insn = THUMB_OP32 (opcode); |
| 21425 | if ((old_op >> 12) == 4 || (old_op >> 12) == 9) |
| 21426 | { |
| 21427 | insn |= (old_op & 0x700) << 4; |
| 21428 | } |
| 21429 | else |
| 21430 | { |
| 21431 | insn |= (old_op & 7) << 12; |
| 21432 | insn |= (old_op & 0x38) << 13; |
| 21433 | } |
| 21434 | insn |= 0x00000c00; |
| 21435 | put_thumb32_insn (buf, insn); |
| 21436 | reloc_type = BFD_RELOC_ARM_T32_OFFSET_IMM; |
| 21437 | } |
| 21438 | else |
| 21439 | { |
| 21440 | reloc_type = BFD_RELOC_ARM_THUMB_OFFSET; |
| 21441 | } |
| 21442 | pc_rel = (opcode == T_MNEM_ldr_pc2); |
| 21443 | break; |
| 21444 | case T_MNEM_adr: |
| 21445 | if (fragp->fr_var == 4) |
| 21446 | { |
| 21447 | insn = THUMB_OP32 (opcode); |
| 21448 | insn |= (old_op & 0xf0) << 4; |
| 21449 | put_thumb32_insn (buf, insn); |
| 21450 | reloc_type = BFD_RELOC_ARM_T32_ADD_PC12; |
| 21451 | } |
| 21452 | else |
| 21453 | { |
| 21454 | reloc_type = BFD_RELOC_ARM_THUMB_ADD; |
| 21455 | exp.X_add_number -= 4; |
| 21456 | } |
| 21457 | pc_rel = 1; |
| 21458 | break; |
| 21459 | case T_MNEM_mov: |
| 21460 | case T_MNEM_movs: |
| 21461 | case T_MNEM_cmp: |
| 21462 | case T_MNEM_cmn: |
| 21463 | if (fragp->fr_var == 4) |
| 21464 | { |
| 21465 | int r0off = (opcode == T_MNEM_mov |
| 21466 | || opcode == T_MNEM_movs) ? 0 : 8; |
| 21467 | insn = THUMB_OP32 (opcode); |
| 21468 | insn = (insn & 0xe1ffffff) | 0x10000000; |
| 21469 | insn |= (old_op & 0x700) << r0off; |
| 21470 | put_thumb32_insn (buf, insn); |
| 21471 | reloc_type = BFD_RELOC_ARM_T32_IMMEDIATE; |
| 21472 | } |
| 21473 | else |
| 21474 | { |
| 21475 | reloc_type = BFD_RELOC_ARM_THUMB_IMM; |
| 21476 | } |
| 21477 | pc_rel = 0; |
| 21478 | break; |
| 21479 | case T_MNEM_b: |
| 21480 | if (fragp->fr_var == 4) |
| 21481 | { |
| 21482 | insn = THUMB_OP32(opcode); |
| 21483 | put_thumb32_insn (buf, insn); |
| 21484 | reloc_type = BFD_RELOC_THUMB_PCREL_BRANCH25; |
| 21485 | } |
| 21486 | else |
| 21487 | reloc_type = BFD_RELOC_THUMB_PCREL_BRANCH12; |
| 21488 | pc_rel = 1; |
| 21489 | break; |
| 21490 | case T_MNEM_bcond: |
| 21491 | if (fragp->fr_var == 4) |
| 21492 | { |
| 21493 | insn = THUMB_OP32(opcode); |
| 21494 | insn |= (old_op & 0xf00) << 14; |
| 21495 | put_thumb32_insn (buf, insn); |
| 21496 | reloc_type = BFD_RELOC_THUMB_PCREL_BRANCH20; |
| 21497 | } |
| 21498 | else |
| 21499 | reloc_type = BFD_RELOC_THUMB_PCREL_BRANCH9; |
| 21500 | pc_rel = 1; |
| 21501 | break; |
| 21502 | case T_MNEM_add_sp: |
| 21503 | case T_MNEM_add_pc: |
| 21504 | case T_MNEM_inc_sp: |
| 21505 | case T_MNEM_dec_sp: |
| 21506 | if (fragp->fr_var == 4) |
| 21507 | { |
| 21508 | /* ??? Choose between add and addw. */ |
| 21509 | insn = THUMB_OP32 (opcode); |
| 21510 | insn |= (old_op & 0xf0) << 4; |
| 21511 | put_thumb32_insn (buf, insn); |
| 21512 | if (opcode == T_MNEM_add_pc) |
| 21513 | reloc_type = BFD_RELOC_ARM_T32_IMM12; |
| 21514 | else |
| 21515 | reloc_type = BFD_RELOC_ARM_T32_ADD_IMM; |
| 21516 | } |
| 21517 | else |
| 21518 | reloc_type = BFD_RELOC_ARM_THUMB_ADD; |
| 21519 | pc_rel = 0; |
| 21520 | break; |
| 21521 | |
| 21522 | case T_MNEM_addi: |
| 21523 | case T_MNEM_addis: |
| 21524 | case T_MNEM_subi: |
| 21525 | case T_MNEM_subis: |
| 21526 | if (fragp->fr_var == 4) |
| 21527 | { |
| 21528 | insn = THUMB_OP32 (opcode); |
| 21529 | insn |= (old_op & 0xf0) << 4; |
| 21530 | insn |= (old_op & 0xf) << 16; |
| 21531 | put_thumb32_insn (buf, insn); |
| 21532 | if (insn & (1 << 20)) |
| 21533 | reloc_type = BFD_RELOC_ARM_T32_ADD_IMM; |
| 21534 | else |
| 21535 | reloc_type = BFD_RELOC_ARM_T32_IMMEDIATE; |
| 21536 | } |
| 21537 | else |
| 21538 | reloc_type = BFD_RELOC_ARM_THUMB_ADD; |
| 21539 | pc_rel = 0; |
| 21540 | break; |
| 21541 | default: |
| 21542 | abort (); |
| 21543 | } |
| 21544 | fixp = fix_new_exp (fragp, fragp->fr_fix, fragp->fr_var, &exp, pc_rel, |
| 21545 | (enum bfd_reloc_code_real) reloc_type); |
| 21546 | fixp->fx_file = fragp->fr_file; |
| 21547 | fixp->fx_line = fragp->fr_line; |
| 21548 | fragp->fr_fix += fragp->fr_var; |
| 21549 | |
| 21550 | /* Set whether we use thumb-2 ISA based on final relaxation results. */ |
| 21551 | if (thumb_mode && fragp->fr_var == 4 && no_cpu_selected () |
| 21552 | && !ARM_CPU_HAS_FEATURE (thumb_arch_used, arm_arch_t2)) |
| 21553 | ARM_MERGE_FEATURE_SETS (arm_arch_used, thumb_arch_used, arm_ext_v6t2); |
| 21554 | } |
| 21555 | |
| 21556 | /* Return the size of a relaxable immediate operand instruction. |
| 21557 | SHIFT and SIZE specify the form of the allowable immediate. */ |
| 21558 | static int |
| 21559 | relax_immediate (fragS *fragp, int size, int shift) |
| 21560 | { |
| 21561 | offsetT offset; |
| 21562 | offsetT mask; |
| 21563 | offsetT low; |
| 21564 | |
| 21565 | /* ??? Should be able to do better than this. */ |
| 21566 | if (fragp->fr_symbol) |
| 21567 | return 4; |
| 21568 | |
| 21569 | low = (1 << shift) - 1; |
| 21570 | mask = (1 << (shift + size)) - (1 << shift); |
| 21571 | offset = fragp->fr_offset; |
| 21572 | /* Force misaligned offsets to 32-bit variant. */ |
| 21573 | if (offset & low) |
| 21574 | return 4; |
| 21575 | if (offset & ~mask) |
| 21576 | return 4; |
| 21577 | return 2; |
| 21578 | } |
| 21579 | |
| 21580 | /* Get the address of a symbol during relaxation. */ |
| 21581 | static addressT |
| 21582 | relaxed_symbol_addr (fragS *fragp, long stretch) |
| 21583 | { |
| 21584 | fragS *sym_frag; |
| 21585 | addressT addr; |
| 21586 | symbolS *sym; |
| 21587 | |
| 21588 | sym = fragp->fr_symbol; |
| 21589 | sym_frag = symbol_get_frag (sym); |
| 21590 | know (S_GET_SEGMENT (sym) != absolute_section |
| 21591 | || sym_frag == &zero_address_frag); |
| 21592 | addr = S_GET_VALUE (sym) + fragp->fr_offset; |
| 21593 | |
| 21594 | /* If frag has yet to be reached on this pass, assume it will |
| 21595 | move by STRETCH just as we did. If this is not so, it will |
| 21596 | be because some frag between grows, and that will force |
| 21597 | another pass. */ |
| 21598 | |
| 21599 | if (stretch != 0 |
| 21600 | && sym_frag->relax_marker != fragp->relax_marker) |
| 21601 | { |
| 21602 | fragS *f; |
| 21603 | |
| 21604 | /* Adjust stretch for any alignment frag. Note that if have |
| 21605 | been expanding the earlier code, the symbol may be |
| 21606 | defined in what appears to be an earlier frag. FIXME: |
| 21607 | This doesn't handle the fr_subtype field, which specifies |
| 21608 | a maximum number of bytes to skip when doing an |
| 21609 | alignment. */ |
| 21610 | for (f = fragp; f != NULL && f != sym_frag; f = f->fr_next) |
| 21611 | { |
| 21612 | if (f->fr_type == rs_align || f->fr_type == rs_align_code) |
| 21613 | { |
| 21614 | if (stretch < 0) |
| 21615 | stretch = - ((- stretch) |
| 21616 | & ~ ((1 << (int) f->fr_offset) - 1)); |
| 21617 | else |
| 21618 | stretch &= ~ ((1 << (int) f->fr_offset) - 1); |
| 21619 | if (stretch == 0) |
| 21620 | break; |
| 21621 | } |
| 21622 | } |
| 21623 | if (f != NULL) |
| 21624 | addr += stretch; |
| 21625 | } |
| 21626 | |
| 21627 | return addr; |
| 21628 | } |
| 21629 | |
| 21630 | /* Return the size of a relaxable adr pseudo-instruction or PC-relative |
| 21631 | load. */ |
| 21632 | static int |
| 21633 | relax_adr (fragS *fragp, asection *sec, long stretch) |
| 21634 | { |
| 21635 | addressT addr; |
| 21636 | offsetT val; |
| 21637 | |
| 21638 | /* Assume worst case for symbols not known to be in the same section. */ |
| 21639 | if (fragp->fr_symbol == NULL |
| 21640 | || !S_IS_DEFINED (fragp->fr_symbol) |
| 21641 | || sec != S_GET_SEGMENT (fragp->fr_symbol) |
| 21642 | || S_IS_WEAK (fragp->fr_symbol)) |
| 21643 | return 4; |
| 21644 | |
| 21645 | val = relaxed_symbol_addr (fragp, stretch); |
| 21646 | addr = fragp->fr_address + fragp->fr_fix; |
| 21647 | addr = (addr + 4) & ~3; |
| 21648 | /* Force misaligned targets to 32-bit variant. */ |
| 21649 | if (val & 3) |
| 21650 | return 4; |
| 21651 | val -= addr; |
| 21652 | if (val < 0 || val > 1020) |
| 21653 | return 4; |
| 21654 | return 2; |
| 21655 | } |
| 21656 | |
| 21657 | /* Return the size of a relaxable add/sub immediate instruction. */ |
| 21658 | static int |
| 21659 | relax_addsub (fragS *fragp, asection *sec) |
| 21660 | { |
| 21661 | char *buf; |
| 21662 | int op; |
| 21663 | |
| 21664 | buf = fragp->fr_literal + fragp->fr_fix; |
| 21665 | op = bfd_get_16(sec->owner, buf); |
| 21666 | if ((op & 0xf) == ((op >> 4) & 0xf)) |
| 21667 | return relax_immediate (fragp, 8, 0); |
| 21668 | else |
| 21669 | return relax_immediate (fragp, 3, 0); |
| 21670 | } |
| 21671 | |
| 21672 | /* Return TRUE iff the definition of symbol S could be pre-empted |
| 21673 | (overridden) at link or load time. */ |
| 21674 | static bfd_boolean |
| 21675 | symbol_preemptible (symbolS *s) |
| 21676 | { |
| 21677 | /* Weak symbols can always be pre-empted. */ |
| 21678 | if (S_IS_WEAK (s)) |
| 21679 | return TRUE; |
| 21680 | |
| 21681 | /* Non-global symbols cannot be pre-empted. */ |
| 21682 | if (! S_IS_EXTERNAL (s)) |
| 21683 | return FALSE; |
| 21684 | |
| 21685 | #ifdef OBJ_ELF |
| 21686 | /* In ELF, a global symbol can be marked protected, or private. In that |
| 21687 | case it can't be pre-empted (other definitions in the same link unit |
| 21688 | would violate the ODR). */ |
| 21689 | if (ELF_ST_VISIBILITY (S_GET_OTHER (s)) > STV_DEFAULT) |
| 21690 | return FALSE; |
| 21691 | #endif |
| 21692 | |
| 21693 | /* Other global symbols might be pre-empted. */ |
| 21694 | return TRUE; |
| 21695 | } |
| 21696 | |
| 21697 | /* Return the size of a relaxable branch instruction. BITS is the |
| 21698 | size of the offset field in the narrow instruction. */ |
| 21699 | |
| 21700 | static int |
| 21701 | relax_branch (fragS *fragp, asection *sec, int bits, long stretch) |
| 21702 | { |
| 21703 | addressT addr; |
| 21704 | offsetT val; |
| 21705 | offsetT limit; |
| 21706 | |
| 21707 | /* Assume worst case for symbols not known to be in the same section. */ |
| 21708 | if (!S_IS_DEFINED (fragp->fr_symbol) |
| 21709 | || sec != S_GET_SEGMENT (fragp->fr_symbol) |
| 21710 | || S_IS_WEAK (fragp->fr_symbol)) |
| 21711 | return 4; |
| 21712 | |
| 21713 | #ifdef OBJ_ELF |
| 21714 | /* A branch to a function in ARM state will require interworking. */ |
| 21715 | if (S_IS_DEFINED (fragp->fr_symbol) |
| 21716 | && ARM_IS_FUNC (fragp->fr_symbol)) |
| 21717 | return 4; |
| 21718 | #endif |
| 21719 | |
| 21720 | if (symbol_preemptible (fragp->fr_symbol)) |
| 21721 | return 4; |
| 21722 | |
| 21723 | val = relaxed_symbol_addr (fragp, stretch); |
| 21724 | addr = fragp->fr_address + fragp->fr_fix + 4; |
| 21725 | val -= addr; |
| 21726 | |
| 21727 | /* Offset is a signed value *2 */ |
| 21728 | limit = 1 << bits; |
| 21729 | if (val >= limit || val < -limit) |
| 21730 | return 4; |
| 21731 | return 2; |
| 21732 | } |
| 21733 | |
| 21734 | |
| 21735 | /* Relax a machine dependent frag. This returns the amount by which |
| 21736 | the current size of the frag should change. */ |
| 21737 | |
| 21738 | int |
| 21739 | arm_relax_frag (asection *sec, fragS *fragp, long stretch) |
| 21740 | { |
| 21741 | int oldsize; |
| 21742 | int newsize; |
| 21743 | |
| 21744 | oldsize = fragp->fr_var; |
| 21745 | switch (fragp->fr_subtype) |
| 21746 | { |
| 21747 | case T_MNEM_ldr_pc2: |
| 21748 | newsize = relax_adr (fragp, sec, stretch); |
| 21749 | break; |
| 21750 | case T_MNEM_ldr_pc: |
| 21751 | case T_MNEM_ldr_sp: |
| 21752 | case T_MNEM_str_sp: |
| 21753 | newsize = relax_immediate (fragp, 8, 2); |
| 21754 | break; |
| 21755 | case T_MNEM_ldr: |
| 21756 | case T_MNEM_str: |
| 21757 | newsize = relax_immediate (fragp, 5, 2); |
| 21758 | break; |
| 21759 | case T_MNEM_ldrh: |
| 21760 | case T_MNEM_strh: |
| 21761 | newsize = relax_immediate (fragp, 5, 1); |
| 21762 | break; |
| 21763 | case T_MNEM_ldrb: |
| 21764 | case T_MNEM_strb: |
| 21765 | newsize = relax_immediate (fragp, 5, 0); |
| 21766 | break; |
| 21767 | case T_MNEM_adr: |
| 21768 | newsize = relax_adr (fragp, sec, stretch); |
| 21769 | break; |
| 21770 | case T_MNEM_mov: |
| 21771 | case T_MNEM_movs: |
| 21772 | case T_MNEM_cmp: |
| 21773 | case T_MNEM_cmn: |
| 21774 | newsize = relax_immediate (fragp, 8, 0); |
| 21775 | break; |
| 21776 | case T_MNEM_b: |
| 21777 | newsize = relax_branch (fragp, sec, 11, stretch); |
| 21778 | break; |
| 21779 | case T_MNEM_bcond: |
| 21780 | newsize = relax_branch (fragp, sec, 8, stretch); |
| 21781 | break; |
| 21782 | case T_MNEM_add_sp: |
| 21783 | case T_MNEM_add_pc: |
| 21784 | newsize = relax_immediate (fragp, 8, 2); |
| 21785 | break; |
| 21786 | case T_MNEM_inc_sp: |
| 21787 | case T_MNEM_dec_sp: |
| 21788 | newsize = relax_immediate (fragp, 7, 2); |
| 21789 | break; |
| 21790 | case T_MNEM_addi: |
| 21791 | case T_MNEM_addis: |
| 21792 | case T_MNEM_subi: |
| 21793 | case T_MNEM_subis: |
| 21794 | newsize = relax_addsub (fragp, sec); |
| 21795 | break; |
| 21796 | default: |
| 21797 | abort (); |
| 21798 | } |
| 21799 | |
| 21800 | fragp->fr_var = newsize; |
| 21801 | /* Freeze wide instructions that are at or before the same location as |
| 21802 | in the previous pass. This avoids infinite loops. |
| 21803 | Don't freeze them unconditionally because targets may be artificially |
| 21804 | misaligned by the expansion of preceding frags. */ |
| 21805 | if (stretch <= 0 && newsize > 2) |
| 21806 | { |
| 21807 | md_convert_frag (sec->owner, sec, fragp); |
| 21808 | frag_wane (fragp); |
| 21809 | } |
| 21810 | |
| 21811 | return newsize - oldsize; |
| 21812 | } |
| 21813 | |
| 21814 | /* Round up a section size to the appropriate boundary. */ |
| 21815 | |
| 21816 | valueT |
| 21817 | md_section_align (segT segment ATTRIBUTE_UNUSED, |
| 21818 | valueT size) |
| 21819 | { |
| 21820 | #if (defined (OBJ_AOUT) || defined (OBJ_MAYBE_AOUT)) |
| 21821 | if (OUTPUT_FLAVOR == bfd_target_aout_flavour) |
| 21822 | { |
| 21823 | /* For a.out, force the section size to be aligned. If we don't do |
| 21824 | this, BFD will align it for us, but it will not write out the |
| 21825 | final bytes of the section. This may be a bug in BFD, but it is |
| 21826 | easier to fix it here since that is how the other a.out targets |
| 21827 | work. */ |
| 21828 | int align; |
| 21829 | |
| 21830 | align = bfd_get_section_alignment (stdoutput, segment); |
| 21831 | size = ((size + (1 << align) - 1) & (-((valueT) 1 << align))); |
| 21832 | } |
| 21833 | #endif |
| 21834 | |
| 21835 | return size; |
| 21836 | } |
| 21837 | |
| 21838 | /* This is called from HANDLE_ALIGN in write.c. Fill in the contents |
| 21839 | of an rs_align_code fragment. */ |
| 21840 | |
| 21841 | void |
| 21842 | arm_handle_align (fragS * fragP) |
| 21843 | { |
| 21844 | static unsigned char const arm_noop[2][2][4] = |
| 21845 | { |
| 21846 | { /* ARMv1 */ |
| 21847 | {0x00, 0x00, 0xa0, 0xe1}, /* LE */ |
| 21848 | {0xe1, 0xa0, 0x00, 0x00}, /* BE */ |
| 21849 | }, |
| 21850 | { /* ARMv6k */ |
| 21851 | {0x00, 0xf0, 0x20, 0xe3}, /* LE */ |
| 21852 | {0xe3, 0x20, 0xf0, 0x00}, /* BE */ |
| 21853 | }, |
| 21854 | }; |
| 21855 | static unsigned char const thumb_noop[2][2][2] = |
| 21856 | { |
| 21857 | { /* Thumb-1 */ |
| 21858 | {0xc0, 0x46}, /* LE */ |
| 21859 | {0x46, 0xc0}, /* BE */ |
| 21860 | }, |
| 21861 | { /* Thumb-2 */ |
| 21862 | {0x00, 0xbf}, /* LE */ |
| 21863 | {0xbf, 0x00} /* BE */ |
| 21864 | } |
| 21865 | }; |
| 21866 | static unsigned char const wide_thumb_noop[2][4] = |
| 21867 | { /* Wide Thumb-2 */ |
| 21868 | {0xaf, 0xf3, 0x00, 0x80}, /* LE */ |
| 21869 | {0xf3, 0xaf, 0x80, 0x00}, /* BE */ |
| 21870 | }; |
| 21871 | |
| 21872 | unsigned bytes, fix, noop_size; |
| 21873 | char * p; |
| 21874 | const unsigned char * noop; |
| 21875 | const unsigned char *narrow_noop = NULL; |
| 21876 | #ifdef OBJ_ELF |
| 21877 | enum mstate state; |
| 21878 | #endif |
| 21879 | |
| 21880 | if (fragP->fr_type != rs_align_code) |
| 21881 | return; |
| 21882 | |
| 21883 | bytes = fragP->fr_next->fr_address - fragP->fr_address - fragP->fr_fix; |
| 21884 | p = fragP->fr_literal + fragP->fr_fix; |
| 21885 | fix = 0; |
| 21886 | |
| 21887 | if (bytes > MAX_MEM_FOR_RS_ALIGN_CODE) |
| 21888 | bytes &= MAX_MEM_FOR_RS_ALIGN_CODE; |
| 21889 | |
| 21890 | gas_assert ((fragP->tc_frag_data.thumb_mode & MODE_RECORDED) != 0); |
| 21891 | |
| 21892 | if (fragP->tc_frag_data.thumb_mode & (~ MODE_RECORDED)) |
| 21893 | { |
| 21894 | if (ARM_CPU_HAS_FEATURE (selected_cpu_name[0] |
| 21895 | ? selected_cpu : arm_arch_none, arm_ext_v6t2)) |
| 21896 | { |
| 21897 | narrow_noop = thumb_noop[1][target_big_endian]; |
| 21898 | noop = wide_thumb_noop[target_big_endian]; |
| 21899 | } |
| 21900 | else |
| 21901 | noop = thumb_noop[0][target_big_endian]; |
| 21902 | noop_size = 2; |
| 21903 | #ifdef OBJ_ELF |
| 21904 | state = MAP_THUMB; |
| 21905 | #endif |
| 21906 | } |
| 21907 | else |
| 21908 | { |
| 21909 | noop = arm_noop[ARM_CPU_HAS_FEATURE (selected_cpu_name[0] |
| 21910 | ? selected_cpu : arm_arch_none, |
| 21911 | arm_ext_v6k) != 0] |
| 21912 | [target_big_endian]; |
| 21913 | noop_size = 4; |
| 21914 | #ifdef OBJ_ELF |
| 21915 | state = MAP_ARM; |
| 21916 | #endif |
| 21917 | } |
| 21918 | |
| 21919 | fragP->fr_var = noop_size; |
| 21920 | |
| 21921 | if (bytes & (noop_size - 1)) |
| 21922 | { |
| 21923 | fix = bytes & (noop_size - 1); |
| 21924 | #ifdef OBJ_ELF |
| 21925 | insert_data_mapping_symbol (state, fragP->fr_fix, fragP, fix); |
| 21926 | #endif |
| 21927 | memset (p, 0, fix); |
| 21928 | p += fix; |
| 21929 | bytes -= fix; |
| 21930 | } |
| 21931 | |
| 21932 | if (narrow_noop) |
| 21933 | { |
| 21934 | if (bytes & noop_size) |
| 21935 | { |
| 21936 | /* Insert a narrow noop. */ |
| 21937 | memcpy (p, narrow_noop, noop_size); |
| 21938 | p += noop_size; |
| 21939 | bytes -= noop_size; |
| 21940 | fix += noop_size; |
| 21941 | } |
| 21942 | |
| 21943 | /* Use wide noops for the remainder */ |
| 21944 | noop_size = 4; |
| 21945 | } |
| 21946 | |
| 21947 | while (bytes >= noop_size) |
| 21948 | { |
| 21949 | memcpy (p, noop, noop_size); |
| 21950 | p += noop_size; |
| 21951 | bytes -= noop_size; |
| 21952 | fix += noop_size; |
| 21953 | } |
| 21954 | |
| 21955 | fragP->fr_fix += fix; |
| 21956 | } |
| 21957 | |
| 21958 | /* Called from md_do_align. Used to create an alignment |
| 21959 | frag in a code section. */ |
| 21960 | |
| 21961 | void |
| 21962 | arm_frag_align_code (int n, int max) |
| 21963 | { |
| 21964 | char * p; |
| 21965 | |
| 21966 | /* We assume that there will never be a requirement |
| 21967 | to support alignments greater than MAX_MEM_FOR_RS_ALIGN_CODE bytes. */ |
| 21968 | if (max > MAX_MEM_FOR_RS_ALIGN_CODE) |
| 21969 | { |
| 21970 | char err_msg[128]; |
| 21971 | |
| 21972 | sprintf (err_msg, |
| 21973 | _("alignments greater than %d bytes not supported in .text sections."), |
| 21974 | MAX_MEM_FOR_RS_ALIGN_CODE + 1); |
| 21975 | as_fatal ("%s", err_msg); |
| 21976 | } |
| 21977 | |
| 21978 | p = frag_var (rs_align_code, |
| 21979 | MAX_MEM_FOR_RS_ALIGN_CODE, |
| 21980 | 1, |
| 21981 | (relax_substateT) max, |
| 21982 | (symbolS *) NULL, |
| 21983 | (offsetT) n, |
| 21984 | (char *) NULL); |
| 21985 | *p = 0; |
| 21986 | } |
| 21987 | |
| 21988 | /* Perform target specific initialisation of a frag. |
| 21989 | Note - despite the name this initialisation is not done when the frag |
| 21990 | is created, but only when its type is assigned. A frag can be created |
| 21991 | and used a long time before its type is set, so beware of assuming that |
| 21992 | this initialisation is performed first. */ |
| 21993 | |
| 21994 | #ifndef OBJ_ELF |
| 21995 | void |
| 21996 | arm_init_frag (fragS * fragP, int max_chars ATTRIBUTE_UNUSED) |
| 21997 | { |
| 21998 | /* Record whether this frag is in an ARM or a THUMB area. */ |
| 21999 | fragP->tc_frag_data.thumb_mode = thumb_mode | MODE_RECORDED; |
| 22000 | } |
| 22001 | |
| 22002 | #else /* OBJ_ELF is defined. */ |
| 22003 | void |
| 22004 | arm_init_frag (fragS * fragP, int max_chars) |
| 22005 | { |
| 22006 | bfd_boolean frag_thumb_mode; |
| 22007 | |
| 22008 | /* If the current ARM vs THUMB mode has not already |
| 22009 | been recorded into this frag then do so now. */ |
| 22010 | if ((fragP->tc_frag_data.thumb_mode & MODE_RECORDED) == 0) |
| 22011 | fragP->tc_frag_data.thumb_mode = thumb_mode | MODE_RECORDED; |
| 22012 | |
| 22013 | /* PR 21809: Do not set a mapping state for debug sections |
| 22014 | - it just confuses other tools. */ |
| 22015 | if (bfd_get_section_flags (NULL, now_seg) & SEC_DEBUGGING) |
| 22016 | return; |
| 22017 | |
| 22018 | frag_thumb_mode = fragP->tc_frag_data.thumb_mode ^ MODE_RECORDED; |
| 22019 | |
| 22020 | /* Record a mapping symbol for alignment frags. We will delete this |
| 22021 | later if the alignment ends up empty. */ |
| 22022 | switch (fragP->fr_type) |
| 22023 | { |
| 22024 | case rs_align: |
| 22025 | case rs_align_test: |
| 22026 | case rs_fill: |
| 22027 | mapping_state_2 (MAP_DATA, max_chars); |
| 22028 | break; |
| 22029 | case rs_align_code: |
| 22030 | mapping_state_2 (frag_thumb_mode ? MAP_THUMB : MAP_ARM, max_chars); |
| 22031 | break; |
| 22032 | default: |
| 22033 | break; |
| 22034 | } |
| 22035 | } |
| 22036 | |
| 22037 | /* When we change sections we need to issue a new mapping symbol. */ |
| 22038 | |
| 22039 | void |
| 22040 | arm_elf_change_section (void) |
| 22041 | { |
| 22042 | /* Link an unlinked unwind index table section to the .text section. */ |
| 22043 | if (elf_section_type (now_seg) == SHT_ARM_EXIDX |
| 22044 | && elf_linked_to_section (now_seg) == NULL) |
| 22045 | elf_linked_to_section (now_seg) = text_section; |
| 22046 | } |
| 22047 | |
| 22048 | int |
| 22049 | arm_elf_section_type (const char * str, size_t len) |
| 22050 | { |
| 22051 | if (len == 5 && strncmp (str, "exidx", 5) == 0) |
| 22052 | return SHT_ARM_EXIDX; |
| 22053 | |
| 22054 | return -1; |
| 22055 | } |
| 22056 | \f |
| 22057 | /* Code to deal with unwinding tables. */ |
| 22058 | |
| 22059 | static void add_unwind_adjustsp (offsetT); |
| 22060 | |
| 22061 | /* Generate any deferred unwind frame offset. */ |
| 22062 | |
| 22063 | static void |
| 22064 | flush_pending_unwind (void) |
| 22065 | { |
| 22066 | offsetT offset; |
| 22067 | |
| 22068 | offset = unwind.pending_offset; |
| 22069 | unwind.pending_offset = 0; |
| 22070 | if (offset != 0) |
| 22071 | add_unwind_adjustsp (offset); |
| 22072 | } |
| 22073 | |
| 22074 | /* Add an opcode to this list for this function. Two-byte opcodes should |
| 22075 | be passed as op[0] << 8 | op[1]. The list of opcodes is built in reverse |
| 22076 | order. */ |
| 22077 | |
| 22078 | static void |
| 22079 | add_unwind_opcode (valueT op, int length) |
| 22080 | { |
| 22081 | /* Add any deferred stack adjustment. */ |
| 22082 | if (unwind.pending_offset) |
| 22083 | flush_pending_unwind (); |
| 22084 | |
| 22085 | unwind.sp_restored = 0; |
| 22086 | |
| 22087 | if (unwind.opcode_count + length > unwind.opcode_alloc) |
| 22088 | { |
| 22089 | unwind.opcode_alloc += ARM_OPCODE_CHUNK_SIZE; |
| 22090 | if (unwind.opcodes) |
| 22091 | unwind.opcodes = XRESIZEVEC (unsigned char, unwind.opcodes, |
| 22092 | unwind.opcode_alloc); |
| 22093 | else |
| 22094 | unwind.opcodes = XNEWVEC (unsigned char, unwind.opcode_alloc); |
| 22095 | } |
| 22096 | while (length > 0) |
| 22097 | { |
| 22098 | length--; |
| 22099 | unwind.opcodes[unwind.opcode_count] = op & 0xff; |
| 22100 | op >>= 8; |
| 22101 | unwind.opcode_count++; |
| 22102 | } |
| 22103 | } |
| 22104 | |
| 22105 | /* Add unwind opcodes to adjust the stack pointer. */ |
| 22106 | |
| 22107 | static void |
| 22108 | add_unwind_adjustsp (offsetT offset) |
| 22109 | { |
| 22110 | valueT op; |
| 22111 | |
| 22112 | if (offset > 0x200) |
| 22113 | { |
| 22114 | /* We need at most 5 bytes to hold a 32-bit value in a uleb128. */ |
| 22115 | char bytes[5]; |
| 22116 | int n; |
| 22117 | valueT o; |
| 22118 | |
| 22119 | /* Long form: 0xb2, uleb128. */ |
| 22120 | /* This might not fit in a word so add the individual bytes, |
| 22121 | remembering the list is built in reverse order. */ |
| 22122 | o = (valueT) ((offset - 0x204) >> 2); |
| 22123 | if (o == 0) |
| 22124 | add_unwind_opcode (0, 1); |
| 22125 | |
| 22126 | /* Calculate the uleb128 encoding of the offset. */ |
| 22127 | n = 0; |
| 22128 | while (o) |
| 22129 | { |
| 22130 | bytes[n] = o & 0x7f; |
| 22131 | o >>= 7; |
| 22132 | if (o) |
| 22133 | bytes[n] |= 0x80; |
| 22134 | n++; |
| 22135 | } |
| 22136 | /* Add the insn. */ |
| 22137 | for (; n; n--) |
| 22138 | add_unwind_opcode (bytes[n - 1], 1); |
| 22139 | add_unwind_opcode (0xb2, 1); |
| 22140 | } |
| 22141 | else if (offset > 0x100) |
| 22142 | { |
| 22143 | /* Two short opcodes. */ |
| 22144 | add_unwind_opcode (0x3f, 1); |
| 22145 | op = (offset - 0x104) >> 2; |
| 22146 | add_unwind_opcode (op, 1); |
| 22147 | } |
| 22148 | else if (offset > 0) |
| 22149 | { |
| 22150 | /* Short opcode. */ |
| 22151 | op = (offset - 4) >> 2; |
| 22152 | add_unwind_opcode (op, 1); |
| 22153 | } |
| 22154 | else if (offset < 0) |
| 22155 | { |
| 22156 | offset = -offset; |
| 22157 | while (offset > 0x100) |
| 22158 | { |
| 22159 | add_unwind_opcode (0x7f, 1); |
| 22160 | offset -= 0x100; |
| 22161 | } |
| 22162 | op = ((offset - 4) >> 2) | 0x40; |
| 22163 | add_unwind_opcode (op, 1); |
| 22164 | } |
| 22165 | } |
| 22166 | |
| 22167 | /* Finish the list of unwind opcodes for this function. */ |
| 22168 | static void |
| 22169 | finish_unwind_opcodes (void) |
| 22170 | { |
| 22171 | valueT op; |
| 22172 | |
| 22173 | if (unwind.fp_used) |
| 22174 | { |
| 22175 | /* Adjust sp as necessary. */ |
| 22176 | unwind.pending_offset += unwind.fp_offset - unwind.frame_size; |
| 22177 | flush_pending_unwind (); |
| 22178 | |
| 22179 | /* After restoring sp from the frame pointer. */ |
| 22180 | op = 0x90 | unwind.fp_reg; |
| 22181 | add_unwind_opcode (op, 1); |
| 22182 | } |
| 22183 | else |
| 22184 | flush_pending_unwind (); |
| 22185 | } |
| 22186 | |
| 22187 | |
| 22188 | /* Start an exception table entry. If idx is nonzero this is an index table |
| 22189 | entry. */ |
| 22190 | |
| 22191 | static void |
| 22192 | start_unwind_section (const segT text_seg, int idx) |
| 22193 | { |
| 22194 | const char * text_name; |
| 22195 | const char * prefix; |
| 22196 | const char * prefix_once; |
| 22197 | const char * group_name; |
| 22198 | char * sec_name; |
| 22199 | int type; |
| 22200 | int flags; |
| 22201 | int linkonce; |
| 22202 | |
| 22203 | if (idx) |
| 22204 | { |
| 22205 | prefix = ELF_STRING_ARM_unwind; |
| 22206 | prefix_once = ELF_STRING_ARM_unwind_once; |
| 22207 | type = SHT_ARM_EXIDX; |
| 22208 | } |
| 22209 | else |
| 22210 | { |
| 22211 | prefix = ELF_STRING_ARM_unwind_info; |
| 22212 | prefix_once = ELF_STRING_ARM_unwind_info_once; |
| 22213 | type = SHT_PROGBITS; |
| 22214 | } |
| 22215 | |
| 22216 | text_name = segment_name (text_seg); |
| 22217 | if (streq (text_name, ".text")) |
| 22218 | text_name = ""; |
| 22219 | |
| 22220 | if (strncmp (text_name, ".gnu.linkonce.t.", |
| 22221 | strlen (".gnu.linkonce.t.")) == 0) |
| 22222 | { |
| 22223 | prefix = prefix_once; |
| 22224 | text_name += strlen (".gnu.linkonce.t."); |
| 22225 | } |
| 22226 | |
| 22227 | sec_name = concat (prefix, text_name, (char *) NULL); |
| 22228 | |
| 22229 | flags = SHF_ALLOC; |
| 22230 | linkonce = 0; |
| 22231 | group_name = 0; |
| 22232 | |
| 22233 | /* Handle COMDAT group. */ |
| 22234 | if (prefix != prefix_once && (text_seg->flags & SEC_LINK_ONCE) != 0) |
| 22235 | { |
| 22236 | group_name = elf_group_name (text_seg); |
| 22237 | if (group_name == NULL) |
| 22238 | { |
| 22239 | as_bad (_("Group section `%s' has no group signature"), |
| 22240 | segment_name (text_seg)); |
| 22241 | ignore_rest_of_line (); |
| 22242 | return; |
| 22243 | } |
| 22244 | flags |= SHF_GROUP; |
| 22245 | linkonce = 1; |
| 22246 | } |
| 22247 | |
| 22248 | obj_elf_change_section (sec_name, type, 0, flags, 0, group_name, |
| 22249 | linkonce, 0); |
| 22250 | |
| 22251 | /* Set the section link for index tables. */ |
| 22252 | if (idx) |
| 22253 | elf_linked_to_section (now_seg) = text_seg; |
| 22254 | } |
| 22255 | |
| 22256 | |
| 22257 | /* Start an unwind table entry. HAVE_DATA is nonzero if we have additional |
| 22258 | personality routine data. Returns zero, or the index table value for |
| 22259 | an inline entry. */ |
| 22260 | |
| 22261 | static valueT |
| 22262 | create_unwind_entry (int have_data) |
| 22263 | { |
| 22264 | int size; |
| 22265 | addressT where; |
| 22266 | char *ptr; |
| 22267 | /* The current word of data. */ |
| 22268 | valueT data; |
| 22269 | /* The number of bytes left in this word. */ |
| 22270 | int n; |
| 22271 | |
| 22272 | finish_unwind_opcodes (); |
| 22273 | |
| 22274 | /* Remember the current text section. */ |
| 22275 | unwind.saved_seg = now_seg; |
| 22276 | unwind.saved_subseg = now_subseg; |
| 22277 | |
| 22278 | start_unwind_section (now_seg, 0); |
| 22279 | |
| 22280 | if (unwind.personality_routine == NULL) |
| 22281 | { |
| 22282 | if (unwind.personality_index == -2) |
| 22283 | { |
| 22284 | if (have_data) |
| 22285 | as_bad (_("handlerdata in cantunwind frame")); |
| 22286 | return 1; /* EXIDX_CANTUNWIND. */ |
| 22287 | } |
| 22288 | |
| 22289 | /* Use a default personality routine if none is specified. */ |
| 22290 | if (unwind.personality_index == -1) |
| 22291 | { |
| 22292 | if (unwind.opcode_count > 3) |
| 22293 | unwind.personality_index = 1; |
| 22294 | else |
| 22295 | unwind.personality_index = 0; |
| 22296 | } |
| 22297 | |
| 22298 | /* Space for the personality routine entry. */ |
| 22299 | if (unwind.personality_index == 0) |
| 22300 | { |
| 22301 | if (unwind.opcode_count > 3) |
| 22302 | as_bad (_("too many unwind opcodes for personality routine 0")); |
| 22303 | |
| 22304 | if (!have_data) |
| 22305 | { |
| 22306 | /* All the data is inline in the index table. */ |
| 22307 | data = 0x80; |
| 22308 | n = 3; |
| 22309 | while (unwind.opcode_count > 0) |
| 22310 | { |
| 22311 | unwind.opcode_count--; |
| 22312 | data = (data << 8) | unwind.opcodes[unwind.opcode_count]; |
| 22313 | n--; |
| 22314 | } |
| 22315 | |
| 22316 | /* Pad with "finish" opcodes. */ |
| 22317 | while (n--) |
| 22318 | data = (data << 8) | 0xb0; |
| 22319 | |
| 22320 | return data; |
| 22321 | } |
| 22322 | size = 0; |
| 22323 | } |
| 22324 | else |
| 22325 | /* We get two opcodes "free" in the first word. */ |
| 22326 | size = unwind.opcode_count - 2; |
| 22327 | } |
| 22328 | else |
| 22329 | { |
| 22330 | /* PR 16765: Missing or misplaced unwind directives can trigger this. */ |
| 22331 | if (unwind.personality_index != -1) |
| 22332 | { |
| 22333 | as_bad (_("attempt to recreate an unwind entry")); |
| 22334 | return 1; |
| 22335 | } |
| 22336 | |
| 22337 | /* An extra byte is required for the opcode count. */ |
| 22338 | size = unwind.opcode_count + 1; |
| 22339 | } |
| 22340 | |
| 22341 | size = (size + 3) >> 2; |
| 22342 | if (size > 0xff) |
| 22343 | as_bad (_("too many unwind opcodes")); |
| 22344 | |
| 22345 | frag_align (2, 0, 0); |
| 22346 | record_alignment (now_seg, 2); |
| 22347 | unwind.table_entry = expr_build_dot (); |
| 22348 | |
| 22349 | /* Allocate the table entry. */ |
| 22350 | ptr = frag_more ((size << 2) + 4); |
| 22351 | /* PR 13449: Zero the table entries in case some of them are not used. */ |
| 22352 | memset (ptr, 0, (size << 2) + 4); |
| 22353 | where = frag_now_fix () - ((size << 2) + 4); |
| 22354 | |
| 22355 | switch (unwind.personality_index) |
| 22356 | { |
| 22357 | case -1: |
| 22358 | /* ??? Should this be a PLT generating relocation? */ |
| 22359 | /* Custom personality routine. */ |
| 22360 | fix_new (frag_now, where, 4, unwind.personality_routine, 0, 1, |
| 22361 | BFD_RELOC_ARM_PREL31); |
| 22362 | |
| 22363 | where += 4; |
| 22364 | ptr += 4; |
| 22365 | |
| 22366 | /* Set the first byte to the number of additional words. */ |
| 22367 | data = size > 0 ? size - 1 : 0; |
| 22368 | n = 3; |
| 22369 | break; |
| 22370 | |
| 22371 | /* ABI defined personality routines. */ |
| 22372 | case 0: |
| 22373 | /* Three opcodes bytes are packed into the first word. */ |
| 22374 | data = 0x80; |
| 22375 | n = 3; |
| 22376 | break; |
| 22377 | |
| 22378 | case 1: |
| 22379 | case 2: |
| 22380 | /* The size and first two opcode bytes go in the first word. */ |
| 22381 | data = ((0x80 + unwind.personality_index) << 8) | size; |
| 22382 | n = 2; |
| 22383 | break; |
| 22384 | |
| 22385 | default: |
| 22386 | /* Should never happen. */ |
| 22387 | abort (); |
| 22388 | } |
| 22389 | |
| 22390 | /* Pack the opcodes into words (MSB first), reversing the list at the same |
| 22391 | time. */ |
| 22392 | while (unwind.opcode_count > 0) |
| 22393 | { |
| 22394 | if (n == 0) |
| 22395 | { |
| 22396 | md_number_to_chars (ptr, data, 4); |
| 22397 | ptr += 4; |
| 22398 | n = 4; |
| 22399 | data = 0; |
| 22400 | } |
| 22401 | unwind.opcode_count--; |
| 22402 | n--; |
| 22403 | data = (data << 8) | unwind.opcodes[unwind.opcode_count]; |
| 22404 | } |
| 22405 | |
| 22406 | /* Finish off the last word. */ |
| 22407 | if (n < 4) |
| 22408 | { |
| 22409 | /* Pad with "finish" opcodes. */ |
| 22410 | while (n--) |
| 22411 | data = (data << 8) | 0xb0; |
| 22412 | |
| 22413 | md_number_to_chars (ptr, data, 4); |
| 22414 | } |
| 22415 | |
| 22416 | if (!have_data) |
| 22417 | { |
| 22418 | /* Add an empty descriptor if there is no user-specified data. */ |
| 22419 | ptr = frag_more (4); |
| 22420 | md_number_to_chars (ptr, 0, 4); |
| 22421 | } |
| 22422 | |
| 22423 | return 0; |
| 22424 | } |
| 22425 | |
| 22426 | |
| 22427 | /* Initialize the DWARF-2 unwind information for this procedure. */ |
| 22428 | |
| 22429 | void |
| 22430 | tc_arm_frame_initial_instructions (void) |
| 22431 | { |
| 22432 | cfi_add_CFA_def_cfa (REG_SP, 0); |
| 22433 | } |
| 22434 | #endif /* OBJ_ELF */ |
| 22435 | |
| 22436 | /* Convert REGNAME to a DWARF-2 register number. */ |
| 22437 | |
| 22438 | int |
| 22439 | tc_arm_regname_to_dw2regnum (char *regname) |
| 22440 | { |
| 22441 | int reg = arm_reg_parse (®name, REG_TYPE_RN); |
| 22442 | if (reg != FAIL) |
| 22443 | return reg; |
| 22444 | |
| 22445 | /* PR 16694: Allow VFP registers as well. */ |
| 22446 | reg = arm_reg_parse (®name, REG_TYPE_VFS); |
| 22447 | if (reg != FAIL) |
| 22448 | return 64 + reg; |
| 22449 | |
| 22450 | reg = arm_reg_parse (®name, REG_TYPE_VFD); |
| 22451 | if (reg != FAIL) |
| 22452 | return reg + 256; |
| 22453 | |
| 22454 | return -1; |
| 22455 | } |
| 22456 | |
| 22457 | #ifdef TE_PE |
| 22458 | void |
| 22459 | tc_pe_dwarf2_emit_offset (symbolS *symbol, unsigned int size) |
| 22460 | { |
| 22461 | expressionS exp; |
| 22462 | |
| 22463 | exp.X_op = O_secrel; |
| 22464 | exp.X_add_symbol = symbol; |
| 22465 | exp.X_add_number = 0; |
| 22466 | emit_expr (&exp, size); |
| 22467 | } |
| 22468 | #endif |
| 22469 | |
| 22470 | /* MD interface: Symbol and relocation handling. */ |
| 22471 | |
| 22472 | /* Return the address within the segment that a PC-relative fixup is |
| 22473 | relative to. For ARM, PC-relative fixups applied to instructions |
| 22474 | are generally relative to the location of the fixup plus 8 bytes. |
| 22475 | Thumb branches are offset by 4, and Thumb loads relative to PC |
| 22476 | require special handling. */ |
| 22477 | |
| 22478 | long |
| 22479 | md_pcrel_from_section (fixS * fixP, segT seg) |
| 22480 | { |
| 22481 | offsetT base = fixP->fx_where + fixP->fx_frag->fr_address; |
| 22482 | |
| 22483 | /* If this is pc-relative and we are going to emit a relocation |
| 22484 | then we just want to put out any pipeline compensation that the linker |
| 22485 | will need. Otherwise we want to use the calculated base. |
| 22486 | For WinCE we skip the bias for externals as well, since this |
| 22487 | is how the MS ARM-CE assembler behaves and we want to be compatible. */ |
| 22488 | if (fixP->fx_pcrel |
| 22489 | && ((fixP->fx_addsy && S_GET_SEGMENT (fixP->fx_addsy) != seg) |
| 22490 | || (arm_force_relocation (fixP) |
| 22491 | #ifdef TE_WINCE |
| 22492 | && !S_IS_EXTERNAL (fixP->fx_addsy) |
| 22493 | #endif |
| 22494 | ))) |
| 22495 | base = 0; |
| 22496 | |
| 22497 | |
| 22498 | switch (fixP->fx_r_type) |
| 22499 | { |
| 22500 | /* PC relative addressing on the Thumb is slightly odd as the |
| 22501 | bottom two bits of the PC are forced to zero for the |
| 22502 | calculation. This happens *after* application of the |
| 22503 | pipeline offset. However, Thumb adrl already adjusts for |
| 22504 | this, so we need not do it again. */ |
| 22505 | case BFD_RELOC_ARM_THUMB_ADD: |
| 22506 | return base & ~3; |
| 22507 | |
| 22508 | case BFD_RELOC_ARM_THUMB_OFFSET: |
| 22509 | case BFD_RELOC_ARM_T32_OFFSET_IMM: |
| 22510 | case BFD_RELOC_ARM_T32_ADD_PC12: |
| 22511 | case BFD_RELOC_ARM_T32_CP_OFF_IMM: |
| 22512 | return (base + 4) & ~3; |
| 22513 | |
| 22514 | /* Thumb branches are simply offset by +4. */ |
| 22515 | case BFD_RELOC_THUMB_PCREL_BRANCH7: |
| 22516 | case BFD_RELOC_THUMB_PCREL_BRANCH9: |
| 22517 | case BFD_RELOC_THUMB_PCREL_BRANCH12: |
| 22518 | case BFD_RELOC_THUMB_PCREL_BRANCH20: |
| 22519 | case BFD_RELOC_THUMB_PCREL_BRANCH25: |
| 22520 | return base + 4; |
| 22521 | |
| 22522 | case BFD_RELOC_THUMB_PCREL_BRANCH23: |
| 22523 | if (fixP->fx_addsy |
| 22524 | && (S_GET_SEGMENT (fixP->fx_addsy) == seg) |
| 22525 | && !S_FORCE_RELOC (fixP->fx_addsy, TRUE) |
| 22526 | && ARM_IS_FUNC (fixP->fx_addsy) |
| 22527 | && ARM_CPU_HAS_FEATURE (selected_cpu, arm_ext_v5t)) |
| 22528 | base = fixP->fx_where + fixP->fx_frag->fr_address; |
| 22529 | return base + 4; |
| 22530 | |
| 22531 | /* BLX is like branches above, but forces the low two bits of PC to |
| 22532 | zero. */ |
| 22533 | case BFD_RELOC_THUMB_PCREL_BLX: |
| 22534 | if (fixP->fx_addsy |
| 22535 | && (S_GET_SEGMENT (fixP->fx_addsy) == seg) |
| 22536 | && !S_FORCE_RELOC (fixP->fx_addsy, TRUE) |
| 22537 | && THUMB_IS_FUNC (fixP->fx_addsy) |
| 22538 | && ARM_CPU_HAS_FEATURE (selected_cpu, arm_ext_v5t)) |
| 22539 | base = fixP->fx_where + fixP->fx_frag->fr_address; |
| 22540 | return (base + 4) & ~3; |
| 22541 | |
| 22542 | /* ARM mode branches are offset by +8. However, the Windows CE |
| 22543 | loader expects the relocation not to take this into account. */ |
| 22544 | case BFD_RELOC_ARM_PCREL_BLX: |
| 22545 | if (fixP->fx_addsy |
| 22546 | && (S_GET_SEGMENT (fixP->fx_addsy) == seg) |
| 22547 | && !S_FORCE_RELOC (fixP->fx_addsy, TRUE) |
| 22548 | && ARM_IS_FUNC (fixP->fx_addsy) |
| 22549 | && ARM_CPU_HAS_FEATURE (selected_cpu, arm_ext_v5t)) |
| 22550 | base = fixP->fx_where + fixP->fx_frag->fr_address; |
| 22551 | return base + 8; |
| 22552 | |
| 22553 | case BFD_RELOC_ARM_PCREL_CALL: |
| 22554 | if (fixP->fx_addsy |
| 22555 | && (S_GET_SEGMENT (fixP->fx_addsy) == seg) |
| 22556 | && !S_FORCE_RELOC (fixP->fx_addsy, TRUE) |
| 22557 | && THUMB_IS_FUNC (fixP->fx_addsy) |
| 22558 | && ARM_CPU_HAS_FEATURE (selected_cpu, arm_ext_v5t)) |
| 22559 | base = fixP->fx_where + fixP->fx_frag->fr_address; |
| 22560 | return base + 8; |
| 22561 | |
| 22562 | case BFD_RELOC_ARM_PCREL_BRANCH: |
| 22563 | case BFD_RELOC_ARM_PCREL_JUMP: |
| 22564 | case BFD_RELOC_ARM_PLT32: |
| 22565 | #ifdef TE_WINCE |
| 22566 | /* When handling fixups immediately, because we have already |
| 22567 | discovered the value of a symbol, or the address of the frag involved |
| 22568 | we must account for the offset by +8, as the OS loader will never see the reloc. |
| 22569 | see fixup_segment() in write.c |
| 22570 | The S_IS_EXTERNAL test handles the case of global symbols. |
| 22571 | Those need the calculated base, not just the pipe compensation the linker will need. */ |
| 22572 | if (fixP->fx_pcrel |
| 22573 | && fixP->fx_addsy != NULL |
| 22574 | && (S_GET_SEGMENT (fixP->fx_addsy) == seg) |
| 22575 | && (S_IS_EXTERNAL (fixP->fx_addsy) || !arm_force_relocation (fixP))) |
| 22576 | return base + 8; |
| 22577 | return base; |
| 22578 | #else |
| 22579 | return base + 8; |
| 22580 | #endif |
| 22581 | |
| 22582 | |
| 22583 | /* ARM mode loads relative to PC are also offset by +8. Unlike |
| 22584 | branches, the Windows CE loader *does* expect the relocation |
| 22585 | to take this into account. */ |
| 22586 | case BFD_RELOC_ARM_OFFSET_IMM: |
| 22587 | case BFD_RELOC_ARM_OFFSET_IMM8: |
| 22588 | case BFD_RELOC_ARM_HWLITERAL: |
| 22589 | case BFD_RELOC_ARM_LITERAL: |
| 22590 | case BFD_RELOC_ARM_CP_OFF_IMM: |
| 22591 | return base + 8; |
| 22592 | |
| 22593 | |
| 22594 | /* Other PC-relative relocations are un-offset. */ |
| 22595 | default: |
| 22596 | return base; |
| 22597 | } |
| 22598 | } |
| 22599 | |
| 22600 | static bfd_boolean flag_warn_syms = TRUE; |
| 22601 | |
| 22602 | bfd_boolean |
| 22603 | arm_tc_equal_in_insn (int c ATTRIBUTE_UNUSED, char * name) |
| 22604 | { |
| 22605 | /* PR 18347 - Warn if the user attempts to create a symbol with the same |
| 22606 | name as an ARM instruction. Whilst strictly speaking it is allowed, it |
| 22607 | does mean that the resulting code might be very confusing to the reader. |
| 22608 | Also this warning can be triggered if the user omits an operand before |
| 22609 | an immediate address, eg: |
| 22610 | |
| 22611 | LDR =foo |
| 22612 | |
| 22613 | GAS treats this as an assignment of the value of the symbol foo to a |
| 22614 | symbol LDR, and so (without this code) it will not issue any kind of |
| 22615 | warning or error message. |
| 22616 | |
| 22617 | Note - ARM instructions are case-insensitive but the strings in the hash |
| 22618 | table are all stored in lower case, so we must first ensure that name is |
| 22619 | lower case too. */ |
| 22620 | if (flag_warn_syms && arm_ops_hsh) |
| 22621 | { |
| 22622 | char * nbuf = strdup (name); |
| 22623 | char * p; |
| 22624 | |
| 22625 | for (p = nbuf; *p; p++) |
| 22626 | *p = TOLOWER (*p); |
| 22627 | if (hash_find (arm_ops_hsh, nbuf) != NULL) |
| 22628 | { |
| 22629 | static struct hash_control * already_warned = NULL; |
| 22630 | |
| 22631 | if (already_warned == NULL) |
| 22632 | already_warned = hash_new (); |
| 22633 | /* Only warn about the symbol once. To keep the code |
| 22634 | simple we let hash_insert do the lookup for us. */ |
| 22635 | if (hash_insert (already_warned, name, NULL) == NULL) |
| 22636 | as_warn (_("[-mwarn-syms]: Assignment makes a symbol match an ARM instruction: %s"), name); |
| 22637 | } |
| 22638 | else |
| 22639 | free (nbuf); |
| 22640 | } |
| 22641 | |
| 22642 | return FALSE; |
| 22643 | } |
| 22644 | |
| 22645 | /* Under ELF we need to default _GLOBAL_OFFSET_TABLE. |
| 22646 | Otherwise we have no need to default values of symbols. */ |
| 22647 | |
| 22648 | symbolS * |
| 22649 | md_undefined_symbol (char * name ATTRIBUTE_UNUSED) |
| 22650 | { |
| 22651 | #ifdef OBJ_ELF |
| 22652 | if (name[0] == '_' && name[1] == 'G' |
| 22653 | && streq (name, GLOBAL_OFFSET_TABLE_NAME)) |
| 22654 | { |
| 22655 | if (!GOT_symbol) |
| 22656 | { |
| 22657 | if (symbol_find (name)) |
| 22658 | as_bad (_("GOT already in the symbol table")); |
| 22659 | |
| 22660 | GOT_symbol = symbol_new (name, undefined_section, |
| 22661 | (valueT) 0, & zero_address_frag); |
| 22662 | } |
| 22663 | |
| 22664 | return GOT_symbol; |
| 22665 | } |
| 22666 | #endif |
| 22667 | |
| 22668 | return NULL; |
| 22669 | } |
| 22670 | |
| 22671 | /* Subroutine of md_apply_fix. Check to see if an immediate can be |
| 22672 | computed as two separate immediate values, added together. We |
| 22673 | already know that this value cannot be computed by just one ARM |
| 22674 | instruction. */ |
| 22675 | |
| 22676 | static unsigned int |
| 22677 | validate_immediate_twopart (unsigned int val, |
| 22678 | unsigned int * highpart) |
| 22679 | { |
| 22680 | unsigned int a; |
| 22681 | unsigned int i; |
| 22682 | |
| 22683 | for (i = 0; i < 32; i += 2) |
| 22684 | if (((a = rotate_left (val, i)) & 0xff) != 0) |
| 22685 | { |
| 22686 | if (a & 0xff00) |
| 22687 | { |
| 22688 | if (a & ~ 0xffff) |
| 22689 | continue; |
| 22690 | * highpart = (a >> 8) | ((i + 24) << 7); |
| 22691 | } |
| 22692 | else if (a & 0xff0000) |
| 22693 | { |
| 22694 | if (a & 0xff000000) |
| 22695 | continue; |
| 22696 | * highpart = (a >> 16) | ((i + 16) << 7); |
| 22697 | } |
| 22698 | else |
| 22699 | { |
| 22700 | gas_assert (a & 0xff000000); |
| 22701 | * highpart = (a >> 24) | ((i + 8) << 7); |
| 22702 | } |
| 22703 | |
| 22704 | return (a & 0xff) | (i << 7); |
| 22705 | } |
| 22706 | |
| 22707 | return FAIL; |
| 22708 | } |
| 22709 | |
| 22710 | static int |
| 22711 | validate_offset_imm (unsigned int val, int hwse) |
| 22712 | { |
| 22713 | if ((hwse && val > 255) || val > 4095) |
| 22714 | return FAIL; |
| 22715 | return val; |
| 22716 | } |
| 22717 | |
| 22718 | /* Subroutine of md_apply_fix. Do those data_ops which can take a |
| 22719 | negative immediate constant by altering the instruction. A bit of |
| 22720 | a hack really. |
| 22721 | MOV <-> MVN |
| 22722 | AND <-> BIC |
| 22723 | ADC <-> SBC |
| 22724 | by inverting the second operand, and |
| 22725 | ADD <-> SUB |
| 22726 | CMP <-> CMN |
| 22727 | by negating the second operand. */ |
| 22728 | |
| 22729 | static int |
| 22730 | negate_data_op (unsigned long * instruction, |
| 22731 | unsigned long value) |
| 22732 | { |
| 22733 | int op, new_inst; |
| 22734 | unsigned long negated, inverted; |
| 22735 | |
| 22736 | negated = encode_arm_immediate (-value); |
| 22737 | inverted = encode_arm_immediate (~value); |
| 22738 | |
| 22739 | op = (*instruction >> DATA_OP_SHIFT) & 0xf; |
| 22740 | switch (op) |
| 22741 | { |
| 22742 | /* First negates. */ |
| 22743 | case OPCODE_SUB: /* ADD <-> SUB */ |
| 22744 | new_inst = OPCODE_ADD; |
| 22745 | value = negated; |
| 22746 | break; |
| 22747 | |
| 22748 | case OPCODE_ADD: |
| 22749 | new_inst = OPCODE_SUB; |
| 22750 | value = negated; |
| 22751 | break; |
| 22752 | |
| 22753 | case OPCODE_CMP: /* CMP <-> CMN */ |
| 22754 | new_inst = OPCODE_CMN; |
| 22755 | value = negated; |
| 22756 | break; |
| 22757 | |
| 22758 | case OPCODE_CMN: |
| 22759 | new_inst = OPCODE_CMP; |
| 22760 | value = negated; |
| 22761 | break; |
| 22762 | |
| 22763 | /* Now Inverted ops. */ |
| 22764 | case OPCODE_MOV: /* MOV <-> MVN */ |
| 22765 | new_inst = OPCODE_MVN; |
| 22766 | value = inverted; |
| 22767 | break; |
| 22768 | |
| 22769 | case OPCODE_MVN: |
| 22770 | new_inst = OPCODE_MOV; |
| 22771 | value = inverted; |
| 22772 | break; |
| 22773 | |
| 22774 | case OPCODE_AND: /* AND <-> BIC */ |
| 22775 | new_inst = OPCODE_BIC; |
| 22776 | value = inverted; |
| 22777 | break; |
| 22778 | |
| 22779 | case OPCODE_BIC: |
| 22780 | new_inst = OPCODE_AND; |
| 22781 | value = inverted; |
| 22782 | break; |
| 22783 | |
| 22784 | case OPCODE_ADC: /* ADC <-> SBC */ |
| 22785 | new_inst = OPCODE_SBC; |
| 22786 | value = inverted; |
| 22787 | break; |
| 22788 | |
| 22789 | case OPCODE_SBC: |
| 22790 | new_inst = OPCODE_ADC; |
| 22791 | value = inverted; |
| 22792 | break; |
| 22793 | |
| 22794 | /* We cannot do anything. */ |
| 22795 | default: |
| 22796 | return FAIL; |
| 22797 | } |
| 22798 | |
| 22799 | if (value == (unsigned) FAIL) |
| 22800 | return FAIL; |
| 22801 | |
| 22802 | *instruction &= OPCODE_MASK; |
| 22803 | *instruction |= new_inst << DATA_OP_SHIFT; |
| 22804 | return value; |
| 22805 | } |
| 22806 | |
| 22807 | /* Like negate_data_op, but for Thumb-2. */ |
| 22808 | |
| 22809 | static unsigned int |
| 22810 | thumb32_negate_data_op (offsetT *instruction, unsigned int value) |
| 22811 | { |
| 22812 | int op, new_inst; |
| 22813 | int rd; |
| 22814 | unsigned int negated, inverted; |
| 22815 | |
| 22816 | negated = encode_thumb32_immediate (-value); |
| 22817 | inverted = encode_thumb32_immediate (~value); |
| 22818 | |
| 22819 | rd = (*instruction >> 8) & 0xf; |
| 22820 | op = (*instruction >> T2_DATA_OP_SHIFT) & 0xf; |
| 22821 | switch (op) |
| 22822 | { |
| 22823 | /* ADD <-> SUB. Includes CMP <-> CMN. */ |
| 22824 | case T2_OPCODE_SUB: |
| 22825 | new_inst = T2_OPCODE_ADD; |
| 22826 | value = negated; |
| 22827 | break; |
| 22828 | |
| 22829 | case T2_OPCODE_ADD: |
| 22830 | new_inst = T2_OPCODE_SUB; |
| 22831 | value = negated; |
| 22832 | break; |
| 22833 | |
| 22834 | /* ORR <-> ORN. Includes MOV <-> MVN. */ |
| 22835 | case T2_OPCODE_ORR: |
| 22836 | new_inst = T2_OPCODE_ORN; |
| 22837 | value = inverted; |
| 22838 | break; |
| 22839 | |
| 22840 | case T2_OPCODE_ORN: |
| 22841 | new_inst = T2_OPCODE_ORR; |
| 22842 | value = inverted; |
| 22843 | break; |
| 22844 | |
| 22845 | /* AND <-> BIC. TST has no inverted equivalent. */ |
| 22846 | case T2_OPCODE_AND: |
| 22847 | new_inst = T2_OPCODE_BIC; |
| 22848 | if (rd == 15) |
| 22849 | value = FAIL; |
| 22850 | else |
| 22851 | value = inverted; |
| 22852 | break; |
| 22853 | |
| 22854 | case T2_OPCODE_BIC: |
| 22855 | new_inst = T2_OPCODE_AND; |
| 22856 | value = inverted; |
| 22857 | break; |
| 22858 | |
| 22859 | /* ADC <-> SBC */ |
| 22860 | case T2_OPCODE_ADC: |
| 22861 | new_inst = T2_OPCODE_SBC; |
| 22862 | value = inverted; |
| 22863 | break; |
| 22864 | |
| 22865 | case T2_OPCODE_SBC: |
| 22866 | new_inst = T2_OPCODE_ADC; |
| 22867 | value = inverted; |
| 22868 | break; |
| 22869 | |
| 22870 | /* We cannot do anything. */ |
| 22871 | default: |
| 22872 | return FAIL; |
| 22873 | } |
| 22874 | |
| 22875 | if (value == (unsigned int)FAIL) |
| 22876 | return FAIL; |
| 22877 | |
| 22878 | *instruction &= T2_OPCODE_MASK; |
| 22879 | *instruction |= new_inst << T2_DATA_OP_SHIFT; |
| 22880 | return value; |
| 22881 | } |
| 22882 | |
| 22883 | /* Read a 32-bit thumb instruction from buf. */ |
| 22884 | static unsigned long |
| 22885 | get_thumb32_insn (char * buf) |
| 22886 | { |
| 22887 | unsigned long insn; |
| 22888 | insn = md_chars_to_number (buf, THUMB_SIZE) << 16; |
| 22889 | insn |= md_chars_to_number (buf + THUMB_SIZE, THUMB_SIZE); |
| 22890 | |
| 22891 | return insn; |
| 22892 | } |
| 22893 | |
| 22894 | |
| 22895 | /* We usually want to set the low bit on the address of thumb function |
| 22896 | symbols. In particular .word foo - . should have the low bit set. |
| 22897 | Generic code tries to fold the difference of two symbols to |
| 22898 | a constant. Prevent this and force a relocation when the first symbols |
| 22899 | is a thumb function. */ |
| 22900 | |
| 22901 | bfd_boolean |
| 22902 | arm_optimize_expr (expressionS *l, operatorT op, expressionS *r) |
| 22903 | { |
| 22904 | if (op == O_subtract |
| 22905 | && l->X_op == O_symbol |
| 22906 | && r->X_op == O_symbol |
| 22907 | && THUMB_IS_FUNC (l->X_add_symbol)) |
| 22908 | { |
| 22909 | l->X_op = O_subtract; |
| 22910 | l->X_op_symbol = r->X_add_symbol; |
| 22911 | l->X_add_number -= r->X_add_number; |
| 22912 | return TRUE; |
| 22913 | } |
| 22914 | |
| 22915 | /* Process as normal. */ |
| 22916 | return FALSE; |
| 22917 | } |
| 22918 | |
| 22919 | /* Encode Thumb2 unconditional branches and calls. The encoding |
| 22920 | for the 2 are identical for the immediate values. */ |
| 22921 | |
| 22922 | static void |
| 22923 | encode_thumb2_b_bl_offset (char * buf, offsetT value) |
| 22924 | { |
| 22925 | #define T2I1I2MASK ((1 << 13) | (1 << 11)) |
| 22926 | offsetT newval; |
| 22927 | offsetT newval2; |
| 22928 | addressT S, I1, I2, lo, hi; |
| 22929 | |
| 22930 | S = (value >> 24) & 0x01; |
| 22931 | I1 = (value >> 23) & 0x01; |
| 22932 | I2 = (value >> 22) & 0x01; |
| 22933 | hi = (value >> 12) & 0x3ff; |
| 22934 | lo = (value >> 1) & 0x7ff; |
| 22935 | newval = md_chars_to_number (buf, THUMB_SIZE); |
| 22936 | newval2 = md_chars_to_number (buf + THUMB_SIZE, THUMB_SIZE); |
| 22937 | newval |= (S << 10) | hi; |
| 22938 | newval2 &= ~T2I1I2MASK; |
| 22939 | newval2 |= (((I1 ^ S) << 13) | ((I2 ^ S) << 11) | lo) ^ T2I1I2MASK; |
| 22940 | md_number_to_chars (buf, newval, THUMB_SIZE); |
| 22941 | md_number_to_chars (buf + THUMB_SIZE, newval2, THUMB_SIZE); |
| 22942 | } |
| 22943 | |
| 22944 | void |
| 22945 | md_apply_fix (fixS * fixP, |
| 22946 | valueT * valP, |
| 22947 | segT seg) |
| 22948 | { |
| 22949 | offsetT value = * valP; |
| 22950 | offsetT newval; |
| 22951 | unsigned int newimm; |
| 22952 | unsigned long temp; |
| 22953 | int sign; |
| 22954 | char * buf = fixP->fx_where + fixP->fx_frag->fr_literal; |
| 22955 | |
| 22956 | gas_assert (fixP->fx_r_type <= BFD_RELOC_UNUSED); |
| 22957 | |
| 22958 | /* Note whether this will delete the relocation. */ |
| 22959 | |
| 22960 | if (fixP->fx_addsy == 0 && !fixP->fx_pcrel) |
| 22961 | fixP->fx_done = 1; |
| 22962 | |
| 22963 | /* On a 64-bit host, silently truncate 'value' to 32 bits for |
| 22964 | consistency with the behaviour on 32-bit hosts. Remember value |
| 22965 | for emit_reloc. */ |
| 22966 | value &= 0xffffffff; |
| 22967 | value ^= 0x80000000; |
| 22968 | value -= 0x80000000; |
| 22969 | |
| 22970 | *valP = value; |
| 22971 | fixP->fx_addnumber = value; |
| 22972 | |
| 22973 | /* Same treatment for fixP->fx_offset. */ |
| 22974 | fixP->fx_offset &= 0xffffffff; |
| 22975 | fixP->fx_offset ^= 0x80000000; |
| 22976 | fixP->fx_offset -= 0x80000000; |
| 22977 | |
| 22978 | switch (fixP->fx_r_type) |
| 22979 | { |
| 22980 | case BFD_RELOC_NONE: |
| 22981 | /* This will need to go in the object file. */ |
| 22982 | fixP->fx_done = 0; |
| 22983 | break; |
| 22984 | |
| 22985 | case BFD_RELOC_ARM_IMMEDIATE: |
| 22986 | /* We claim that this fixup has been processed here, |
| 22987 | even if in fact we generate an error because we do |
| 22988 | not have a reloc for it, so tc_gen_reloc will reject it. */ |
| 22989 | fixP->fx_done = 1; |
| 22990 | |
| 22991 | if (fixP->fx_addsy) |
| 22992 | { |
| 22993 | const char *msg = 0; |
| 22994 | |
| 22995 | if (! S_IS_DEFINED (fixP->fx_addsy)) |
| 22996 | msg = _("undefined symbol %s used as an immediate value"); |
| 22997 | else if (S_GET_SEGMENT (fixP->fx_addsy) != seg) |
| 22998 | msg = _("symbol %s is in a different section"); |
| 22999 | else if (S_IS_WEAK (fixP->fx_addsy)) |
| 23000 | msg = _("symbol %s is weak and may be overridden later"); |
| 23001 | |
| 23002 | if (msg) |
| 23003 | { |
| 23004 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23005 | msg, S_GET_NAME (fixP->fx_addsy)); |
| 23006 | break; |
| 23007 | } |
| 23008 | } |
| 23009 | |
| 23010 | temp = md_chars_to_number (buf, INSN_SIZE); |
| 23011 | |
| 23012 | /* If the offset is negative, we should use encoding A2 for ADR. */ |
| 23013 | if ((temp & 0xfff0000) == 0x28f0000 && value < 0) |
| 23014 | newimm = negate_data_op (&temp, value); |
| 23015 | else |
| 23016 | { |
| 23017 | newimm = encode_arm_immediate (value); |
| 23018 | |
| 23019 | /* If the instruction will fail, see if we can fix things up by |
| 23020 | changing the opcode. */ |
| 23021 | if (newimm == (unsigned int) FAIL) |
| 23022 | newimm = negate_data_op (&temp, value); |
| 23023 | /* MOV accepts both ARM modified immediate (A1 encoding) and |
| 23024 | UINT16 (A2 encoding) when possible, MOVW only accepts UINT16. |
| 23025 | When disassembling, MOV is preferred when there is no encoding |
| 23026 | overlap. */ |
| 23027 | if (newimm == (unsigned int) FAIL |
| 23028 | && ((temp >> DATA_OP_SHIFT) & 0xf) == OPCODE_MOV |
| 23029 | && ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v6t2) |
| 23030 | && !((temp >> SBIT_SHIFT) & 0x1) |
| 23031 | && value >= 0 && value <= 0xffff) |
| 23032 | { |
| 23033 | /* Clear bits[23:20] to change encoding from A1 to A2. */ |
| 23034 | temp &= 0xff0fffff; |
| 23035 | /* Encoding high 4bits imm. Code below will encode the remaining |
| 23036 | low 12bits. */ |
| 23037 | temp |= (value & 0x0000f000) << 4; |
| 23038 | newimm = value & 0x00000fff; |
| 23039 | } |
| 23040 | } |
| 23041 | |
| 23042 | if (newimm == (unsigned int) FAIL) |
| 23043 | { |
| 23044 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23045 | _("invalid constant (%lx) after fixup"), |
| 23046 | (unsigned long) value); |
| 23047 | break; |
| 23048 | } |
| 23049 | |
| 23050 | newimm |= (temp & 0xfffff000); |
| 23051 | md_number_to_chars (buf, (valueT) newimm, INSN_SIZE); |
| 23052 | break; |
| 23053 | |
| 23054 | case BFD_RELOC_ARM_ADRL_IMMEDIATE: |
| 23055 | { |
| 23056 | unsigned int highpart = 0; |
| 23057 | unsigned int newinsn = 0xe1a00000; /* nop. */ |
| 23058 | |
| 23059 | if (fixP->fx_addsy) |
| 23060 | { |
| 23061 | const char *msg = 0; |
| 23062 | |
| 23063 | if (! S_IS_DEFINED (fixP->fx_addsy)) |
| 23064 | msg = _("undefined symbol %s used as an immediate value"); |
| 23065 | else if (S_GET_SEGMENT (fixP->fx_addsy) != seg) |
| 23066 | msg = _("symbol %s is in a different section"); |
| 23067 | else if (S_IS_WEAK (fixP->fx_addsy)) |
| 23068 | msg = _("symbol %s is weak and may be overridden later"); |
| 23069 | |
| 23070 | if (msg) |
| 23071 | { |
| 23072 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23073 | msg, S_GET_NAME (fixP->fx_addsy)); |
| 23074 | break; |
| 23075 | } |
| 23076 | } |
| 23077 | |
| 23078 | newimm = encode_arm_immediate (value); |
| 23079 | temp = md_chars_to_number (buf, INSN_SIZE); |
| 23080 | |
| 23081 | /* If the instruction will fail, see if we can fix things up by |
| 23082 | changing the opcode. */ |
| 23083 | if (newimm == (unsigned int) FAIL |
| 23084 | && (newimm = negate_data_op (& temp, value)) == (unsigned int) FAIL) |
| 23085 | { |
| 23086 | /* No ? OK - try using two ADD instructions to generate |
| 23087 | the value. */ |
| 23088 | newimm = validate_immediate_twopart (value, & highpart); |
| 23089 | |
| 23090 | /* Yes - then make sure that the second instruction is |
| 23091 | also an add. */ |
| 23092 | if (newimm != (unsigned int) FAIL) |
| 23093 | newinsn = temp; |
| 23094 | /* Still No ? Try using a negated value. */ |
| 23095 | else if ((newimm = validate_immediate_twopart (- value, & highpart)) != (unsigned int) FAIL) |
| 23096 | temp = newinsn = (temp & OPCODE_MASK) | OPCODE_SUB << DATA_OP_SHIFT; |
| 23097 | /* Otherwise - give up. */ |
| 23098 | else |
| 23099 | { |
| 23100 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23101 | _("unable to compute ADRL instructions for PC offset of 0x%lx"), |
| 23102 | (long) value); |
| 23103 | break; |
| 23104 | } |
| 23105 | |
| 23106 | /* Replace the first operand in the 2nd instruction (which |
| 23107 | is the PC) with the destination register. We have |
| 23108 | already added in the PC in the first instruction and we |
| 23109 | do not want to do it again. */ |
| 23110 | newinsn &= ~ 0xf0000; |
| 23111 | newinsn |= ((newinsn & 0x0f000) << 4); |
| 23112 | } |
| 23113 | |
| 23114 | newimm |= (temp & 0xfffff000); |
| 23115 | md_number_to_chars (buf, (valueT) newimm, INSN_SIZE); |
| 23116 | |
| 23117 | highpart |= (newinsn & 0xfffff000); |
| 23118 | md_number_to_chars (buf + INSN_SIZE, (valueT) highpart, INSN_SIZE); |
| 23119 | } |
| 23120 | break; |
| 23121 | |
| 23122 | case BFD_RELOC_ARM_OFFSET_IMM: |
| 23123 | if (!fixP->fx_done && seg->use_rela_p) |
| 23124 | value = 0; |
| 23125 | /* Fall through. */ |
| 23126 | |
| 23127 | case BFD_RELOC_ARM_LITERAL: |
| 23128 | sign = value > 0; |
| 23129 | |
| 23130 | if (value < 0) |
| 23131 | value = - value; |
| 23132 | |
| 23133 | if (validate_offset_imm (value, 0) == FAIL) |
| 23134 | { |
| 23135 | if (fixP->fx_r_type == BFD_RELOC_ARM_LITERAL) |
| 23136 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23137 | _("invalid literal constant: pool needs to be closer")); |
| 23138 | else |
| 23139 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23140 | _("bad immediate value for offset (%ld)"), |
| 23141 | (long) value); |
| 23142 | break; |
| 23143 | } |
| 23144 | |
| 23145 | newval = md_chars_to_number (buf, INSN_SIZE); |
| 23146 | if (value == 0) |
| 23147 | newval &= 0xfffff000; |
| 23148 | else |
| 23149 | { |
| 23150 | newval &= 0xff7ff000; |
| 23151 | newval |= value | (sign ? INDEX_UP : 0); |
| 23152 | } |
| 23153 | md_number_to_chars (buf, newval, INSN_SIZE); |
| 23154 | break; |
| 23155 | |
| 23156 | case BFD_RELOC_ARM_OFFSET_IMM8: |
| 23157 | case BFD_RELOC_ARM_HWLITERAL: |
| 23158 | sign = value > 0; |
| 23159 | |
| 23160 | if (value < 0) |
| 23161 | value = - value; |
| 23162 | |
| 23163 | if (validate_offset_imm (value, 1) == FAIL) |
| 23164 | { |
| 23165 | if (fixP->fx_r_type == BFD_RELOC_ARM_HWLITERAL) |
| 23166 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23167 | _("invalid literal constant: pool needs to be closer")); |
| 23168 | else |
| 23169 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23170 | _("bad immediate value for 8-bit offset (%ld)"), |
| 23171 | (long) value); |
| 23172 | break; |
| 23173 | } |
| 23174 | |
| 23175 | newval = md_chars_to_number (buf, INSN_SIZE); |
| 23176 | if (value == 0) |
| 23177 | newval &= 0xfffff0f0; |
| 23178 | else |
| 23179 | { |
| 23180 | newval &= 0xff7ff0f0; |
| 23181 | newval |= ((value >> 4) << 8) | (value & 0xf) | (sign ? INDEX_UP : 0); |
| 23182 | } |
| 23183 | md_number_to_chars (buf, newval, INSN_SIZE); |
| 23184 | break; |
| 23185 | |
| 23186 | case BFD_RELOC_ARM_T32_OFFSET_U8: |
| 23187 | if (value < 0 || value > 1020 || value % 4 != 0) |
| 23188 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23189 | _("bad immediate value for offset (%ld)"), (long) value); |
| 23190 | value /= 4; |
| 23191 | |
| 23192 | newval = md_chars_to_number (buf+2, THUMB_SIZE); |
| 23193 | newval |= value; |
| 23194 | md_number_to_chars (buf+2, newval, THUMB_SIZE); |
| 23195 | break; |
| 23196 | |
| 23197 | case BFD_RELOC_ARM_T32_OFFSET_IMM: |
| 23198 | /* This is a complicated relocation used for all varieties of Thumb32 |
| 23199 | load/store instruction with immediate offset: |
| 23200 | |
| 23201 | 1110 100P u1WL NNNN XXXX YYYY iiii iiii - +/-(U) pre/post(P) 8-bit, |
| 23202 | *4, optional writeback(W) |
| 23203 | (doubleword load/store) |
| 23204 | |
| 23205 | 1111 100S uTTL 1111 XXXX iiii iiii iiii - +/-(U) 12-bit PC-rel |
| 23206 | 1111 100S 0TTL NNNN XXXX 1Pu1 iiii iiii - +/-(U) pre/post(P) 8-bit |
| 23207 | 1111 100S 0TTL NNNN XXXX 1110 iiii iiii - positive 8-bit (T instruction) |
| 23208 | 1111 100S 1TTL NNNN XXXX iiii iiii iiii - positive 12-bit |
| 23209 | 1111 100S 0TTL NNNN XXXX 1100 iiii iiii - negative 8-bit |
| 23210 | |
| 23211 | Uppercase letters indicate bits that are already encoded at |
| 23212 | this point. Lowercase letters are our problem. For the |
| 23213 | second block of instructions, the secondary opcode nybble |
| 23214 | (bits 8..11) is present, and bit 23 is zero, even if this is |
| 23215 | a PC-relative operation. */ |
| 23216 | newval = md_chars_to_number (buf, THUMB_SIZE); |
| 23217 | newval <<= 16; |
| 23218 | newval |= md_chars_to_number (buf+THUMB_SIZE, THUMB_SIZE); |
| 23219 | |
| 23220 | if ((newval & 0xf0000000) == 0xe0000000) |
| 23221 | { |
| 23222 | /* Doubleword load/store: 8-bit offset, scaled by 4. */ |
| 23223 | if (value >= 0) |
| 23224 | newval |= (1 << 23); |
| 23225 | else |
| 23226 | value = -value; |
| 23227 | if (value % 4 != 0) |
| 23228 | { |
| 23229 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23230 | _("offset not a multiple of 4")); |
| 23231 | break; |
| 23232 | } |
| 23233 | value /= 4; |
| 23234 | if (value > 0xff) |
| 23235 | { |
| 23236 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23237 | _("offset out of range")); |
| 23238 | break; |
| 23239 | } |
| 23240 | newval &= ~0xff; |
| 23241 | } |
| 23242 | else if ((newval & 0x000f0000) == 0x000f0000) |
| 23243 | { |
| 23244 | /* PC-relative, 12-bit offset. */ |
| 23245 | if (value >= 0) |
| 23246 | newval |= (1 << 23); |
| 23247 | else |
| 23248 | value = -value; |
| 23249 | if (value > 0xfff) |
| 23250 | { |
| 23251 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23252 | _("offset out of range")); |
| 23253 | break; |
| 23254 | } |
| 23255 | newval &= ~0xfff; |
| 23256 | } |
| 23257 | else if ((newval & 0x00000100) == 0x00000100) |
| 23258 | { |
| 23259 | /* Writeback: 8-bit, +/- offset. */ |
| 23260 | if (value >= 0) |
| 23261 | newval |= (1 << 9); |
| 23262 | else |
| 23263 | value = -value; |
| 23264 | if (value > 0xff) |
| 23265 | { |
| 23266 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23267 | _("offset out of range")); |
| 23268 | break; |
| 23269 | } |
| 23270 | newval &= ~0xff; |
| 23271 | } |
| 23272 | else if ((newval & 0x00000f00) == 0x00000e00) |
| 23273 | { |
| 23274 | /* T-instruction: positive 8-bit offset. */ |
| 23275 | if (value < 0 || value > 0xff) |
| 23276 | { |
| 23277 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23278 | _("offset out of range")); |
| 23279 | break; |
| 23280 | } |
| 23281 | newval &= ~0xff; |
| 23282 | newval |= value; |
| 23283 | } |
| 23284 | else |
| 23285 | { |
| 23286 | /* Positive 12-bit or negative 8-bit offset. */ |
| 23287 | int limit; |
| 23288 | if (value >= 0) |
| 23289 | { |
| 23290 | newval |= (1 << 23); |
| 23291 | limit = 0xfff; |
| 23292 | } |
| 23293 | else |
| 23294 | { |
| 23295 | value = -value; |
| 23296 | limit = 0xff; |
| 23297 | } |
| 23298 | if (value > limit) |
| 23299 | { |
| 23300 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23301 | _("offset out of range")); |
| 23302 | break; |
| 23303 | } |
| 23304 | newval &= ~limit; |
| 23305 | } |
| 23306 | |
| 23307 | newval |= value; |
| 23308 | md_number_to_chars (buf, (newval >> 16) & 0xffff, THUMB_SIZE); |
| 23309 | md_number_to_chars (buf + THUMB_SIZE, newval & 0xffff, THUMB_SIZE); |
| 23310 | break; |
| 23311 | |
| 23312 | case BFD_RELOC_ARM_SHIFT_IMM: |
| 23313 | newval = md_chars_to_number (buf, INSN_SIZE); |
| 23314 | if (((unsigned long) value) > 32 |
| 23315 | || (value == 32 |
| 23316 | && (((newval & 0x60) == 0) || (newval & 0x60) == 0x60))) |
| 23317 | { |
| 23318 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23319 | _("shift expression is too large")); |
| 23320 | break; |
| 23321 | } |
| 23322 | |
| 23323 | if (value == 0) |
| 23324 | /* Shifts of zero must be done as lsl. */ |
| 23325 | newval &= ~0x60; |
| 23326 | else if (value == 32) |
| 23327 | value = 0; |
| 23328 | newval &= 0xfffff07f; |
| 23329 | newval |= (value & 0x1f) << 7; |
| 23330 | md_number_to_chars (buf, newval, INSN_SIZE); |
| 23331 | break; |
| 23332 | |
| 23333 | case BFD_RELOC_ARM_T32_IMMEDIATE: |
| 23334 | case BFD_RELOC_ARM_T32_ADD_IMM: |
| 23335 | case BFD_RELOC_ARM_T32_IMM12: |
| 23336 | case BFD_RELOC_ARM_T32_ADD_PC12: |
| 23337 | /* We claim that this fixup has been processed here, |
| 23338 | even if in fact we generate an error because we do |
| 23339 | not have a reloc for it, so tc_gen_reloc will reject it. */ |
| 23340 | fixP->fx_done = 1; |
| 23341 | |
| 23342 | if (fixP->fx_addsy |
| 23343 | && ! S_IS_DEFINED (fixP->fx_addsy)) |
| 23344 | { |
| 23345 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23346 | _("undefined symbol %s used as an immediate value"), |
| 23347 | S_GET_NAME (fixP->fx_addsy)); |
| 23348 | break; |
| 23349 | } |
| 23350 | |
| 23351 | newval = md_chars_to_number (buf, THUMB_SIZE); |
| 23352 | newval <<= 16; |
| 23353 | newval |= md_chars_to_number (buf+2, THUMB_SIZE); |
| 23354 | |
| 23355 | newimm = FAIL; |
| 23356 | if ((fixP->fx_r_type == BFD_RELOC_ARM_T32_IMMEDIATE |
| 23357 | /* ARMv8-M Baseline MOV will reach here, but it doesn't support |
| 23358 | Thumb2 modified immediate encoding (T2). */ |
| 23359 | && ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v6t2)) |
| 23360 | || fixP->fx_r_type == BFD_RELOC_ARM_T32_ADD_IMM) |
| 23361 | { |
| 23362 | newimm = encode_thumb32_immediate (value); |
| 23363 | if (newimm == (unsigned int) FAIL) |
| 23364 | newimm = thumb32_negate_data_op (&newval, value); |
| 23365 | } |
| 23366 | if (newimm == (unsigned int) FAIL) |
| 23367 | { |
| 23368 | if (fixP->fx_r_type != BFD_RELOC_ARM_T32_IMMEDIATE) |
| 23369 | { |
| 23370 | /* Turn add/sum into addw/subw. */ |
| 23371 | if (fixP->fx_r_type == BFD_RELOC_ARM_T32_ADD_IMM) |
| 23372 | newval = (newval & 0xfeffffff) | 0x02000000; |
| 23373 | /* No flat 12-bit imm encoding for addsw/subsw. */ |
| 23374 | if ((newval & 0x00100000) == 0) |
| 23375 | { |
| 23376 | /* 12 bit immediate for addw/subw. */ |
| 23377 | if (value < 0) |
| 23378 | { |
| 23379 | value = -value; |
| 23380 | newval ^= 0x00a00000; |
| 23381 | } |
| 23382 | if (value > 0xfff) |
| 23383 | newimm = (unsigned int) FAIL; |
| 23384 | else |
| 23385 | newimm = value; |
| 23386 | } |
| 23387 | } |
| 23388 | else |
| 23389 | { |
| 23390 | /* MOV accepts both Thumb2 modified immediate (T2 encoding) and |
| 23391 | UINT16 (T3 encoding), MOVW only accepts UINT16. When |
| 23392 | disassembling, MOV is preferred when there is no encoding |
| 23393 | overlap. |
| 23394 | NOTE: MOV is using ORR opcode under Thumb 2 mode. */ |
| 23395 | if (((newval >> T2_DATA_OP_SHIFT) & 0xf) == T2_OPCODE_ORR |
| 23396 | && ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v6t2_v8m) |
| 23397 | && !((newval >> T2_SBIT_SHIFT) & 0x1) |
| 23398 | && value >= 0 && value <=0xffff) |
| 23399 | { |
| 23400 | /* Toggle bit[25] to change encoding from T2 to T3. */ |
| 23401 | newval ^= 1 << 25; |
| 23402 | /* Clear bits[19:16]. */ |
| 23403 | newval &= 0xfff0ffff; |
| 23404 | /* Encoding high 4bits imm. Code below will encode the |
| 23405 | remaining low 12bits. */ |
| 23406 | newval |= (value & 0x0000f000) << 4; |
| 23407 | newimm = value & 0x00000fff; |
| 23408 | } |
| 23409 | } |
| 23410 | } |
| 23411 | |
| 23412 | if (newimm == (unsigned int)FAIL) |
| 23413 | { |
| 23414 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23415 | _("invalid constant (%lx) after fixup"), |
| 23416 | (unsigned long) value); |
| 23417 | break; |
| 23418 | } |
| 23419 | |
| 23420 | newval |= (newimm & 0x800) << 15; |
| 23421 | newval |= (newimm & 0x700) << 4; |
| 23422 | newval |= (newimm & 0x0ff); |
| 23423 | |
| 23424 | md_number_to_chars (buf, (valueT) ((newval >> 16) & 0xffff), THUMB_SIZE); |
| 23425 | md_number_to_chars (buf+2, (valueT) (newval & 0xffff), THUMB_SIZE); |
| 23426 | break; |
| 23427 | |
| 23428 | case BFD_RELOC_ARM_SMC: |
| 23429 | if (((unsigned long) value) > 0xffff) |
| 23430 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23431 | _("invalid smc expression")); |
| 23432 | newval = md_chars_to_number (buf, INSN_SIZE); |
| 23433 | newval |= (value & 0xf) | ((value & 0xfff0) << 4); |
| 23434 | md_number_to_chars (buf, newval, INSN_SIZE); |
| 23435 | break; |
| 23436 | |
| 23437 | case BFD_RELOC_ARM_HVC: |
| 23438 | if (((unsigned long) value) > 0xffff) |
| 23439 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23440 | _("invalid hvc expression")); |
| 23441 | newval = md_chars_to_number (buf, INSN_SIZE); |
| 23442 | newval |= (value & 0xf) | ((value & 0xfff0) << 4); |
| 23443 | md_number_to_chars (buf, newval, INSN_SIZE); |
| 23444 | break; |
| 23445 | |
| 23446 | case BFD_RELOC_ARM_SWI: |
| 23447 | if (fixP->tc_fix_data != 0) |
| 23448 | { |
| 23449 | if (((unsigned long) value) > 0xff) |
| 23450 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23451 | _("invalid swi expression")); |
| 23452 | newval = md_chars_to_number (buf, THUMB_SIZE); |
| 23453 | newval |= value; |
| 23454 | md_number_to_chars (buf, newval, THUMB_SIZE); |
| 23455 | } |
| 23456 | else |
| 23457 | { |
| 23458 | if (((unsigned long) value) > 0x00ffffff) |
| 23459 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23460 | _("invalid swi expression")); |
| 23461 | newval = md_chars_to_number (buf, INSN_SIZE); |
| 23462 | newval |= value; |
| 23463 | md_number_to_chars (buf, newval, INSN_SIZE); |
| 23464 | } |
| 23465 | break; |
| 23466 | |
| 23467 | case BFD_RELOC_ARM_MULTI: |
| 23468 | if (((unsigned long) value) > 0xffff) |
| 23469 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23470 | _("invalid expression in load/store multiple")); |
| 23471 | newval = value | md_chars_to_number (buf, INSN_SIZE); |
| 23472 | md_number_to_chars (buf, newval, INSN_SIZE); |
| 23473 | break; |
| 23474 | |
| 23475 | #ifdef OBJ_ELF |
| 23476 | case BFD_RELOC_ARM_PCREL_CALL: |
| 23477 | |
| 23478 | if (ARM_CPU_HAS_FEATURE (selected_cpu, arm_ext_v5t) |
| 23479 | && fixP->fx_addsy |
| 23480 | && !S_FORCE_RELOC (fixP->fx_addsy, TRUE) |
| 23481 | && (S_GET_SEGMENT (fixP->fx_addsy) == seg) |
| 23482 | && THUMB_IS_FUNC (fixP->fx_addsy)) |
| 23483 | /* Flip the bl to blx. This is a simple flip |
| 23484 | bit here because we generate PCREL_CALL for |
| 23485 | unconditional bls. */ |
| 23486 | { |
| 23487 | newval = md_chars_to_number (buf, INSN_SIZE); |
| 23488 | newval = newval | 0x10000000; |
| 23489 | md_number_to_chars (buf, newval, INSN_SIZE); |
| 23490 | temp = 1; |
| 23491 | fixP->fx_done = 1; |
| 23492 | } |
| 23493 | else |
| 23494 | temp = 3; |
| 23495 | goto arm_branch_common; |
| 23496 | |
| 23497 | case BFD_RELOC_ARM_PCREL_JUMP: |
| 23498 | if (ARM_CPU_HAS_FEATURE (selected_cpu, arm_ext_v5t) |
| 23499 | && fixP->fx_addsy |
| 23500 | && !S_FORCE_RELOC (fixP->fx_addsy, TRUE) |
| 23501 | && (S_GET_SEGMENT (fixP->fx_addsy) == seg) |
| 23502 | && THUMB_IS_FUNC (fixP->fx_addsy)) |
| 23503 | { |
| 23504 | /* This would map to a bl<cond>, b<cond>, |
| 23505 | b<always> to a Thumb function. We |
| 23506 | need to force a relocation for this particular |
| 23507 | case. */ |
| 23508 | newval = md_chars_to_number (buf, INSN_SIZE); |
| 23509 | fixP->fx_done = 0; |
| 23510 | } |
| 23511 | /* Fall through. */ |
| 23512 | |
| 23513 | case BFD_RELOC_ARM_PLT32: |
| 23514 | #endif |
| 23515 | case BFD_RELOC_ARM_PCREL_BRANCH: |
| 23516 | temp = 3; |
| 23517 | goto arm_branch_common; |
| 23518 | |
| 23519 | case BFD_RELOC_ARM_PCREL_BLX: |
| 23520 | |
| 23521 | temp = 1; |
| 23522 | if (ARM_CPU_HAS_FEATURE (selected_cpu, arm_ext_v5t) |
| 23523 | && fixP->fx_addsy |
| 23524 | && !S_FORCE_RELOC (fixP->fx_addsy, TRUE) |
| 23525 | && (S_GET_SEGMENT (fixP->fx_addsy) == seg) |
| 23526 | && ARM_IS_FUNC (fixP->fx_addsy)) |
| 23527 | { |
| 23528 | /* Flip the blx to a bl and warn. */ |
| 23529 | const char *name = S_GET_NAME (fixP->fx_addsy); |
| 23530 | newval = 0xeb000000; |
| 23531 | as_warn_where (fixP->fx_file, fixP->fx_line, |
| 23532 | _("blx to '%s' an ARM ISA state function changed to bl"), |
| 23533 | name); |
| 23534 | md_number_to_chars (buf, newval, INSN_SIZE); |
| 23535 | temp = 3; |
| 23536 | fixP->fx_done = 1; |
| 23537 | } |
| 23538 | |
| 23539 | #ifdef OBJ_ELF |
| 23540 | if (EF_ARM_EABI_VERSION (meabi_flags) >= EF_ARM_EABI_VER4) |
| 23541 | fixP->fx_r_type = BFD_RELOC_ARM_PCREL_CALL; |
| 23542 | #endif |
| 23543 | |
| 23544 | arm_branch_common: |
| 23545 | /* We are going to store value (shifted right by two) in the |
| 23546 | instruction, in a 24 bit, signed field. Bits 26 through 32 either |
| 23547 | all clear or all set and bit 0 must be clear. For B/BL bit 1 must |
| 23548 | also be clear. */ |
| 23549 | if (value & temp) |
| 23550 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23551 | _("misaligned branch destination")); |
| 23552 | if ((value & (offsetT)0xfe000000) != (offsetT)0 |
| 23553 | && (value & (offsetT)0xfe000000) != (offsetT)0xfe000000) |
| 23554 | as_bad_where (fixP->fx_file, fixP->fx_line, BAD_RANGE); |
| 23555 | |
| 23556 | if (fixP->fx_done || !seg->use_rela_p) |
| 23557 | { |
| 23558 | newval = md_chars_to_number (buf, INSN_SIZE); |
| 23559 | newval |= (value >> 2) & 0x00ffffff; |
| 23560 | /* Set the H bit on BLX instructions. */ |
| 23561 | if (temp == 1) |
| 23562 | { |
| 23563 | if (value & 2) |
| 23564 | newval |= 0x01000000; |
| 23565 | else |
| 23566 | newval &= ~0x01000000; |
| 23567 | } |
| 23568 | md_number_to_chars (buf, newval, INSN_SIZE); |
| 23569 | } |
| 23570 | break; |
| 23571 | |
| 23572 | case BFD_RELOC_THUMB_PCREL_BRANCH7: /* CBZ */ |
| 23573 | /* CBZ can only branch forward. */ |
| 23574 | |
| 23575 | /* Attempts to use CBZ to branch to the next instruction |
| 23576 | (which, strictly speaking, are prohibited) will be turned into |
| 23577 | no-ops. |
| 23578 | |
| 23579 | FIXME: It may be better to remove the instruction completely and |
| 23580 | perform relaxation. */ |
| 23581 | if (value == -2) |
| 23582 | { |
| 23583 | newval = md_chars_to_number (buf, THUMB_SIZE); |
| 23584 | newval = 0xbf00; /* NOP encoding T1 */ |
| 23585 | md_number_to_chars (buf, newval, THUMB_SIZE); |
| 23586 | } |
| 23587 | else |
| 23588 | { |
| 23589 | if (value & ~0x7e) |
| 23590 | as_bad_where (fixP->fx_file, fixP->fx_line, BAD_RANGE); |
| 23591 | |
| 23592 | if (fixP->fx_done || !seg->use_rela_p) |
| 23593 | { |
| 23594 | newval = md_chars_to_number (buf, THUMB_SIZE); |
| 23595 | newval |= ((value & 0x3e) << 2) | ((value & 0x40) << 3); |
| 23596 | md_number_to_chars (buf, newval, THUMB_SIZE); |
| 23597 | } |
| 23598 | } |
| 23599 | break; |
| 23600 | |
| 23601 | case BFD_RELOC_THUMB_PCREL_BRANCH9: /* Conditional branch. */ |
| 23602 | if ((value & ~0xff) && ((value & ~0xff) != ~0xff)) |
| 23603 | as_bad_where (fixP->fx_file, fixP->fx_line, BAD_RANGE); |
| 23604 | |
| 23605 | if (fixP->fx_done || !seg->use_rela_p) |
| 23606 | { |
| 23607 | newval = md_chars_to_number (buf, THUMB_SIZE); |
| 23608 | newval |= (value & 0x1ff) >> 1; |
| 23609 | md_number_to_chars (buf, newval, THUMB_SIZE); |
| 23610 | } |
| 23611 | break; |
| 23612 | |
| 23613 | case BFD_RELOC_THUMB_PCREL_BRANCH12: /* Unconditional branch. */ |
| 23614 | if ((value & ~0x7ff) && ((value & ~0x7ff) != ~0x7ff)) |
| 23615 | as_bad_where (fixP->fx_file, fixP->fx_line, BAD_RANGE); |
| 23616 | |
| 23617 | if (fixP->fx_done || !seg->use_rela_p) |
| 23618 | { |
| 23619 | newval = md_chars_to_number (buf, THUMB_SIZE); |
| 23620 | newval |= (value & 0xfff) >> 1; |
| 23621 | md_number_to_chars (buf, newval, THUMB_SIZE); |
| 23622 | } |
| 23623 | break; |
| 23624 | |
| 23625 | case BFD_RELOC_THUMB_PCREL_BRANCH20: |
| 23626 | if (fixP->fx_addsy |
| 23627 | && (S_GET_SEGMENT (fixP->fx_addsy) == seg) |
| 23628 | && !S_FORCE_RELOC (fixP->fx_addsy, TRUE) |
| 23629 | && ARM_IS_FUNC (fixP->fx_addsy) |
| 23630 | && ARM_CPU_HAS_FEATURE (selected_cpu, arm_ext_v5t)) |
| 23631 | { |
| 23632 | /* Force a relocation for a branch 20 bits wide. */ |
| 23633 | fixP->fx_done = 0; |
| 23634 | } |
| 23635 | if ((value & ~0x1fffff) && ((value & ~0x0fffff) != ~0x0fffff)) |
| 23636 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23637 | _("conditional branch out of range")); |
| 23638 | |
| 23639 | if (fixP->fx_done || !seg->use_rela_p) |
| 23640 | { |
| 23641 | offsetT newval2; |
| 23642 | addressT S, J1, J2, lo, hi; |
| 23643 | |
| 23644 | S = (value & 0x00100000) >> 20; |
| 23645 | J2 = (value & 0x00080000) >> 19; |
| 23646 | J1 = (value & 0x00040000) >> 18; |
| 23647 | hi = (value & 0x0003f000) >> 12; |
| 23648 | lo = (value & 0x00000ffe) >> 1; |
| 23649 | |
| 23650 | newval = md_chars_to_number (buf, THUMB_SIZE); |
| 23651 | newval2 = md_chars_to_number (buf + THUMB_SIZE, THUMB_SIZE); |
| 23652 | newval |= (S << 10) | hi; |
| 23653 | newval2 |= (J1 << 13) | (J2 << 11) | lo; |
| 23654 | md_number_to_chars (buf, newval, THUMB_SIZE); |
| 23655 | md_number_to_chars (buf + THUMB_SIZE, newval2, THUMB_SIZE); |
| 23656 | } |
| 23657 | break; |
| 23658 | |
| 23659 | case BFD_RELOC_THUMB_PCREL_BLX: |
| 23660 | /* If there is a blx from a thumb state function to |
| 23661 | another thumb function flip this to a bl and warn |
| 23662 | about it. */ |
| 23663 | |
| 23664 | if (fixP->fx_addsy |
| 23665 | && !S_FORCE_RELOC (fixP->fx_addsy, TRUE) |
| 23666 | && (S_GET_SEGMENT (fixP->fx_addsy) == seg) |
| 23667 | && THUMB_IS_FUNC (fixP->fx_addsy)) |
| 23668 | { |
| 23669 | const char *name = S_GET_NAME (fixP->fx_addsy); |
| 23670 | as_warn_where (fixP->fx_file, fixP->fx_line, |
| 23671 | _("blx to Thumb func '%s' from Thumb ISA state changed to bl"), |
| 23672 | name); |
| 23673 | newval = md_chars_to_number (buf + THUMB_SIZE, THUMB_SIZE); |
| 23674 | newval = newval | 0x1000; |
| 23675 | md_number_to_chars (buf+THUMB_SIZE, newval, THUMB_SIZE); |
| 23676 | fixP->fx_r_type = BFD_RELOC_THUMB_PCREL_BRANCH23; |
| 23677 | fixP->fx_done = 1; |
| 23678 | } |
| 23679 | |
| 23680 | |
| 23681 | goto thumb_bl_common; |
| 23682 | |
| 23683 | case BFD_RELOC_THUMB_PCREL_BRANCH23: |
| 23684 | /* A bl from Thumb state ISA to an internal ARM state function |
| 23685 | is converted to a blx. */ |
| 23686 | if (fixP->fx_addsy |
| 23687 | && (S_GET_SEGMENT (fixP->fx_addsy) == seg) |
| 23688 | && !S_FORCE_RELOC (fixP->fx_addsy, TRUE) |
| 23689 | && ARM_IS_FUNC (fixP->fx_addsy) |
| 23690 | && ARM_CPU_HAS_FEATURE (selected_cpu, arm_ext_v5t)) |
| 23691 | { |
| 23692 | newval = md_chars_to_number (buf + THUMB_SIZE, THUMB_SIZE); |
| 23693 | newval = newval & ~0x1000; |
| 23694 | md_number_to_chars (buf+THUMB_SIZE, newval, THUMB_SIZE); |
| 23695 | fixP->fx_r_type = BFD_RELOC_THUMB_PCREL_BLX; |
| 23696 | fixP->fx_done = 1; |
| 23697 | } |
| 23698 | |
| 23699 | thumb_bl_common: |
| 23700 | |
| 23701 | if (fixP->fx_r_type == BFD_RELOC_THUMB_PCREL_BLX) |
| 23702 | /* For a BLX instruction, make sure that the relocation is rounded up |
| 23703 | to a word boundary. This follows the semantics of the instruction |
| 23704 | which specifies that bit 1 of the target address will come from bit |
| 23705 | 1 of the base address. */ |
| 23706 | value = (value + 3) & ~ 3; |
| 23707 | |
| 23708 | #ifdef OBJ_ELF |
| 23709 | if (EF_ARM_EABI_VERSION (meabi_flags) >= EF_ARM_EABI_VER4 |
| 23710 | && fixP->fx_r_type == BFD_RELOC_THUMB_PCREL_BLX) |
| 23711 | fixP->fx_r_type = BFD_RELOC_THUMB_PCREL_BRANCH23; |
| 23712 | #endif |
| 23713 | |
| 23714 | if ((value & ~0x3fffff) && ((value & ~0x3fffff) != ~0x3fffff)) |
| 23715 | { |
| 23716 | if (!(ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v6t2))) |
| 23717 | as_bad_where (fixP->fx_file, fixP->fx_line, BAD_RANGE); |
| 23718 | else if ((value & ~0x1ffffff) |
| 23719 | && ((value & ~0x1ffffff) != ~0x1ffffff)) |
| 23720 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23721 | _("Thumb2 branch out of range")); |
| 23722 | } |
| 23723 | |
| 23724 | if (fixP->fx_done || !seg->use_rela_p) |
| 23725 | encode_thumb2_b_bl_offset (buf, value); |
| 23726 | |
| 23727 | break; |
| 23728 | |
| 23729 | case BFD_RELOC_THUMB_PCREL_BRANCH25: |
| 23730 | if ((value & ~0x0ffffff) && ((value & ~0x0ffffff) != ~0x0ffffff)) |
| 23731 | as_bad_where (fixP->fx_file, fixP->fx_line, BAD_RANGE); |
| 23732 | |
| 23733 | if (fixP->fx_done || !seg->use_rela_p) |
| 23734 | encode_thumb2_b_bl_offset (buf, value); |
| 23735 | |
| 23736 | break; |
| 23737 | |
| 23738 | case BFD_RELOC_8: |
| 23739 | if (fixP->fx_done || !seg->use_rela_p) |
| 23740 | *buf = value; |
| 23741 | break; |
| 23742 | |
| 23743 | case BFD_RELOC_16: |
| 23744 | if (fixP->fx_done || !seg->use_rela_p) |
| 23745 | md_number_to_chars (buf, value, 2); |
| 23746 | break; |
| 23747 | |
| 23748 | #ifdef OBJ_ELF |
| 23749 | case BFD_RELOC_ARM_TLS_CALL: |
| 23750 | case BFD_RELOC_ARM_THM_TLS_CALL: |
| 23751 | case BFD_RELOC_ARM_TLS_DESCSEQ: |
| 23752 | case BFD_RELOC_ARM_THM_TLS_DESCSEQ: |
| 23753 | case BFD_RELOC_ARM_TLS_GOTDESC: |
| 23754 | case BFD_RELOC_ARM_TLS_GD32: |
| 23755 | case BFD_RELOC_ARM_TLS_LE32: |
| 23756 | case BFD_RELOC_ARM_TLS_IE32: |
| 23757 | case BFD_RELOC_ARM_TLS_LDM32: |
| 23758 | case BFD_RELOC_ARM_TLS_LDO32: |
| 23759 | S_SET_THREAD_LOCAL (fixP->fx_addsy); |
| 23760 | break; |
| 23761 | |
| 23762 | case BFD_RELOC_ARM_GOT32: |
| 23763 | case BFD_RELOC_ARM_GOTOFF: |
| 23764 | break; |
| 23765 | |
| 23766 | case BFD_RELOC_ARM_GOT_PREL: |
| 23767 | if (fixP->fx_done || !seg->use_rela_p) |
| 23768 | md_number_to_chars (buf, value, 4); |
| 23769 | break; |
| 23770 | |
| 23771 | case BFD_RELOC_ARM_TARGET2: |
| 23772 | /* TARGET2 is not partial-inplace, so we need to write the |
| 23773 | addend here for REL targets, because it won't be written out |
| 23774 | during reloc processing later. */ |
| 23775 | if (fixP->fx_done || !seg->use_rela_p) |
| 23776 | md_number_to_chars (buf, fixP->fx_offset, 4); |
| 23777 | break; |
| 23778 | #endif |
| 23779 | |
| 23780 | case BFD_RELOC_RVA: |
| 23781 | case BFD_RELOC_32: |
| 23782 | case BFD_RELOC_ARM_TARGET1: |
| 23783 | case BFD_RELOC_ARM_ROSEGREL32: |
| 23784 | case BFD_RELOC_ARM_SBREL32: |
| 23785 | case BFD_RELOC_32_PCREL: |
| 23786 | #ifdef TE_PE |
| 23787 | case BFD_RELOC_32_SECREL: |
| 23788 | #endif |
| 23789 | if (fixP->fx_done || !seg->use_rela_p) |
| 23790 | #ifdef TE_WINCE |
| 23791 | /* For WinCE we only do this for pcrel fixups. */ |
| 23792 | if (fixP->fx_done || fixP->fx_pcrel) |
| 23793 | #endif |
| 23794 | md_number_to_chars (buf, value, 4); |
| 23795 | break; |
| 23796 | |
| 23797 | #ifdef OBJ_ELF |
| 23798 | case BFD_RELOC_ARM_PREL31: |
| 23799 | if (fixP->fx_done || !seg->use_rela_p) |
| 23800 | { |
| 23801 | newval = md_chars_to_number (buf, 4) & 0x80000000; |
| 23802 | if ((value ^ (value >> 1)) & 0x40000000) |
| 23803 | { |
| 23804 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23805 | _("rel31 relocation overflow")); |
| 23806 | } |
| 23807 | newval |= value & 0x7fffffff; |
| 23808 | md_number_to_chars (buf, newval, 4); |
| 23809 | } |
| 23810 | break; |
| 23811 | #endif |
| 23812 | |
| 23813 | case BFD_RELOC_ARM_CP_OFF_IMM: |
| 23814 | case BFD_RELOC_ARM_T32_CP_OFF_IMM: |
| 23815 | if (fixP->fx_r_type == BFD_RELOC_ARM_CP_OFF_IMM) |
| 23816 | newval = md_chars_to_number (buf, INSN_SIZE); |
| 23817 | else |
| 23818 | newval = get_thumb32_insn (buf); |
| 23819 | if ((newval & 0x0f200f00) == 0x0d000900) |
| 23820 | { |
| 23821 | /* This is a fp16 vstr/vldr. The immediate offset in the mnemonic |
| 23822 | has permitted values that are multiples of 2, in the range 0 |
| 23823 | to 510. */ |
| 23824 | if (value < -510 || value > 510 || (value & 1)) |
| 23825 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23826 | _("co-processor offset out of range")); |
| 23827 | } |
| 23828 | else if (value < -1023 || value > 1023 || (value & 3)) |
| 23829 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23830 | _("co-processor offset out of range")); |
| 23831 | cp_off_common: |
| 23832 | sign = value > 0; |
| 23833 | if (value < 0) |
| 23834 | value = -value; |
| 23835 | if (fixP->fx_r_type == BFD_RELOC_ARM_CP_OFF_IMM |
| 23836 | || fixP->fx_r_type == BFD_RELOC_ARM_CP_OFF_IMM_S2) |
| 23837 | newval = md_chars_to_number (buf, INSN_SIZE); |
| 23838 | else |
| 23839 | newval = get_thumb32_insn (buf); |
| 23840 | if (value == 0) |
| 23841 | newval &= 0xffffff00; |
| 23842 | else |
| 23843 | { |
| 23844 | newval &= 0xff7fff00; |
| 23845 | if ((newval & 0x0f200f00) == 0x0d000900) |
| 23846 | { |
| 23847 | /* This is a fp16 vstr/vldr. |
| 23848 | |
| 23849 | It requires the immediate offset in the instruction is shifted |
| 23850 | left by 1 to be a half-word offset. |
| 23851 | |
| 23852 | Here, left shift by 1 first, and later right shift by 2 |
| 23853 | should get the right offset. */ |
| 23854 | value <<= 1; |
| 23855 | } |
| 23856 | newval |= (value >> 2) | (sign ? INDEX_UP : 0); |
| 23857 | } |
| 23858 | if (fixP->fx_r_type == BFD_RELOC_ARM_CP_OFF_IMM |
| 23859 | || fixP->fx_r_type == BFD_RELOC_ARM_CP_OFF_IMM_S2) |
| 23860 | md_number_to_chars (buf, newval, INSN_SIZE); |
| 23861 | else |
| 23862 | put_thumb32_insn (buf, newval); |
| 23863 | break; |
| 23864 | |
| 23865 | case BFD_RELOC_ARM_CP_OFF_IMM_S2: |
| 23866 | case BFD_RELOC_ARM_T32_CP_OFF_IMM_S2: |
| 23867 | if (value < -255 || value > 255) |
| 23868 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23869 | _("co-processor offset out of range")); |
| 23870 | value *= 4; |
| 23871 | goto cp_off_common; |
| 23872 | |
| 23873 | case BFD_RELOC_ARM_THUMB_OFFSET: |
| 23874 | newval = md_chars_to_number (buf, THUMB_SIZE); |
| 23875 | /* Exactly what ranges, and where the offset is inserted depends |
| 23876 | on the type of instruction, we can establish this from the |
| 23877 | top 4 bits. */ |
| 23878 | switch (newval >> 12) |
| 23879 | { |
| 23880 | case 4: /* PC load. */ |
| 23881 | /* Thumb PC loads are somewhat odd, bit 1 of the PC is |
| 23882 | forced to zero for these loads; md_pcrel_from has already |
| 23883 | compensated for this. */ |
| 23884 | if (value & 3) |
| 23885 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23886 | _("invalid offset, target not word aligned (0x%08lX)"), |
| 23887 | (((unsigned long) fixP->fx_frag->fr_address |
| 23888 | + (unsigned long) fixP->fx_where) & ~3) |
| 23889 | + (unsigned long) value); |
| 23890 | |
| 23891 | if (value & ~0x3fc) |
| 23892 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23893 | _("invalid offset, value too big (0x%08lX)"), |
| 23894 | (long) value); |
| 23895 | |
| 23896 | newval |= value >> 2; |
| 23897 | break; |
| 23898 | |
| 23899 | case 9: /* SP load/store. */ |
| 23900 | if (value & ~0x3fc) |
| 23901 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23902 | _("invalid offset, value too big (0x%08lX)"), |
| 23903 | (long) value); |
| 23904 | newval |= value >> 2; |
| 23905 | break; |
| 23906 | |
| 23907 | case 6: /* Word load/store. */ |
| 23908 | if (value & ~0x7c) |
| 23909 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23910 | _("invalid offset, value too big (0x%08lX)"), |
| 23911 | (long) value); |
| 23912 | newval |= value << 4; /* 6 - 2. */ |
| 23913 | break; |
| 23914 | |
| 23915 | case 7: /* Byte load/store. */ |
| 23916 | if (value & ~0x1f) |
| 23917 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23918 | _("invalid offset, value too big (0x%08lX)"), |
| 23919 | (long) value); |
| 23920 | newval |= value << 6; |
| 23921 | break; |
| 23922 | |
| 23923 | case 8: /* Halfword load/store. */ |
| 23924 | if (value & ~0x3e) |
| 23925 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23926 | _("invalid offset, value too big (0x%08lX)"), |
| 23927 | (long) value); |
| 23928 | newval |= value << 5; /* 6 - 1. */ |
| 23929 | break; |
| 23930 | |
| 23931 | default: |
| 23932 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23933 | "Unable to process relocation for thumb opcode: %lx", |
| 23934 | (unsigned long) newval); |
| 23935 | break; |
| 23936 | } |
| 23937 | md_number_to_chars (buf, newval, THUMB_SIZE); |
| 23938 | break; |
| 23939 | |
| 23940 | case BFD_RELOC_ARM_THUMB_ADD: |
| 23941 | /* This is a complicated relocation, since we use it for all of |
| 23942 | the following immediate relocations: |
| 23943 | |
| 23944 | 3bit ADD/SUB |
| 23945 | 8bit ADD/SUB |
| 23946 | 9bit ADD/SUB SP word-aligned |
| 23947 | 10bit ADD PC/SP word-aligned |
| 23948 | |
| 23949 | The type of instruction being processed is encoded in the |
| 23950 | instruction field: |
| 23951 | |
| 23952 | 0x8000 SUB |
| 23953 | 0x00F0 Rd |
| 23954 | 0x000F Rs |
| 23955 | */ |
| 23956 | newval = md_chars_to_number (buf, THUMB_SIZE); |
| 23957 | { |
| 23958 | int rd = (newval >> 4) & 0xf; |
| 23959 | int rs = newval & 0xf; |
| 23960 | int subtract = !!(newval & 0x8000); |
| 23961 | |
| 23962 | /* Check for HI regs, only very restricted cases allowed: |
| 23963 | Adjusting SP, and using PC or SP to get an address. */ |
| 23964 | if ((rd > 7 && (rd != REG_SP || rs != REG_SP)) |
| 23965 | || (rs > 7 && rs != REG_SP && rs != REG_PC)) |
| 23966 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23967 | _("invalid Hi register with immediate")); |
| 23968 | |
| 23969 | /* If value is negative, choose the opposite instruction. */ |
| 23970 | if (value < 0) |
| 23971 | { |
| 23972 | value = -value; |
| 23973 | subtract = !subtract; |
| 23974 | if (value < 0) |
| 23975 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23976 | _("immediate value out of range")); |
| 23977 | } |
| 23978 | |
| 23979 | if (rd == REG_SP) |
| 23980 | { |
| 23981 | if (value & ~0x1fc) |
| 23982 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 23983 | _("invalid immediate for stack address calculation")); |
| 23984 | newval = subtract ? T_OPCODE_SUB_ST : T_OPCODE_ADD_ST; |
| 23985 | newval |= value >> 2; |
| 23986 | } |
| 23987 | else if (rs == REG_PC || rs == REG_SP) |
| 23988 | { |
| 23989 | /* PR gas/18541. If the addition is for a defined symbol |
| 23990 | within range of an ADR instruction then accept it. */ |
| 23991 | if (subtract |
| 23992 | && value == 4 |
| 23993 | && fixP->fx_addsy != NULL) |
| 23994 | { |
| 23995 | subtract = 0; |
| 23996 | |
| 23997 | if (! S_IS_DEFINED (fixP->fx_addsy) |
| 23998 | || S_GET_SEGMENT (fixP->fx_addsy) != seg |
| 23999 | || S_IS_WEAK (fixP->fx_addsy)) |
| 24000 | { |
| 24001 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 24002 | _("address calculation needs a strongly defined nearby symbol")); |
| 24003 | } |
| 24004 | else |
| 24005 | { |
| 24006 | offsetT v = fixP->fx_where + fixP->fx_frag->fr_address; |
| 24007 | |
| 24008 | /* Round up to the next 4-byte boundary. */ |
| 24009 | if (v & 3) |
| 24010 | v = (v + 3) & ~ 3; |
| 24011 | else |
| 24012 | v += 4; |
| 24013 | v = S_GET_VALUE (fixP->fx_addsy) - v; |
| 24014 | |
| 24015 | if (v & ~0x3fc) |
| 24016 | { |
| 24017 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 24018 | _("symbol too far away")); |
| 24019 | } |
| 24020 | else |
| 24021 | { |
| 24022 | fixP->fx_done = 1; |
| 24023 | value = v; |
| 24024 | } |
| 24025 | } |
| 24026 | } |
| 24027 | |
| 24028 | if (subtract || value & ~0x3fc) |
| 24029 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 24030 | _("invalid immediate for address calculation (value = 0x%08lX)"), |
| 24031 | (unsigned long) (subtract ? - value : value)); |
| 24032 | newval = (rs == REG_PC ? T_OPCODE_ADD_PC : T_OPCODE_ADD_SP); |
| 24033 | newval |= rd << 8; |
| 24034 | newval |= value >> 2; |
| 24035 | } |
| 24036 | else if (rs == rd) |
| 24037 | { |
| 24038 | if (value & ~0xff) |
| 24039 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 24040 | _("immediate value out of range")); |
| 24041 | newval = subtract ? T_OPCODE_SUB_I8 : T_OPCODE_ADD_I8; |
| 24042 | newval |= (rd << 8) | value; |
| 24043 | } |
| 24044 | else |
| 24045 | { |
| 24046 | if (value & ~0x7) |
| 24047 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 24048 | _("immediate value out of range")); |
| 24049 | newval = subtract ? T_OPCODE_SUB_I3 : T_OPCODE_ADD_I3; |
| 24050 | newval |= rd | (rs << 3) | (value << 6); |
| 24051 | } |
| 24052 | } |
| 24053 | md_number_to_chars (buf, newval, THUMB_SIZE); |
| 24054 | break; |
| 24055 | |
| 24056 | case BFD_RELOC_ARM_THUMB_IMM: |
| 24057 | newval = md_chars_to_number (buf, THUMB_SIZE); |
| 24058 | if (value < 0 || value > 255) |
| 24059 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 24060 | _("invalid immediate: %ld is out of range"), |
| 24061 | (long) value); |
| 24062 | newval |= value; |
| 24063 | md_number_to_chars (buf, newval, THUMB_SIZE); |
| 24064 | break; |
| 24065 | |
| 24066 | case BFD_RELOC_ARM_THUMB_SHIFT: |
| 24067 | /* 5bit shift value (0..32). LSL cannot take 32. */ |
| 24068 | newval = md_chars_to_number (buf, THUMB_SIZE) & 0xf83f; |
| 24069 | temp = newval & 0xf800; |
| 24070 | if (value < 0 || value > 32 || (value == 32 && temp == T_OPCODE_LSL_I)) |
| 24071 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 24072 | _("invalid shift value: %ld"), (long) value); |
| 24073 | /* Shifts of zero must be encoded as LSL. */ |
| 24074 | if (value == 0) |
| 24075 | newval = (newval & 0x003f) | T_OPCODE_LSL_I; |
| 24076 | /* Shifts of 32 are encoded as zero. */ |
| 24077 | else if (value == 32) |
| 24078 | value = 0; |
| 24079 | newval |= value << 6; |
| 24080 | md_number_to_chars (buf, newval, THUMB_SIZE); |
| 24081 | break; |
| 24082 | |
| 24083 | case BFD_RELOC_VTABLE_INHERIT: |
| 24084 | case BFD_RELOC_VTABLE_ENTRY: |
| 24085 | fixP->fx_done = 0; |
| 24086 | return; |
| 24087 | |
| 24088 | case BFD_RELOC_ARM_MOVW: |
| 24089 | case BFD_RELOC_ARM_MOVT: |
| 24090 | case BFD_RELOC_ARM_THUMB_MOVW: |
| 24091 | case BFD_RELOC_ARM_THUMB_MOVT: |
| 24092 | if (fixP->fx_done || !seg->use_rela_p) |
| 24093 | { |
| 24094 | /* REL format relocations are limited to a 16-bit addend. */ |
| 24095 | if (!fixP->fx_done) |
| 24096 | { |
| 24097 | if (value < -0x8000 || value > 0x7fff) |
| 24098 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 24099 | _("offset out of range")); |
| 24100 | } |
| 24101 | else if (fixP->fx_r_type == BFD_RELOC_ARM_MOVT |
| 24102 | || fixP->fx_r_type == BFD_RELOC_ARM_THUMB_MOVT) |
| 24103 | { |
| 24104 | value >>= 16; |
| 24105 | } |
| 24106 | |
| 24107 | if (fixP->fx_r_type == BFD_RELOC_ARM_THUMB_MOVW |
| 24108 | || fixP->fx_r_type == BFD_RELOC_ARM_THUMB_MOVT) |
| 24109 | { |
| 24110 | newval = get_thumb32_insn (buf); |
| 24111 | newval &= 0xfbf08f00; |
| 24112 | newval |= (value & 0xf000) << 4; |
| 24113 | newval |= (value & 0x0800) << 15; |
| 24114 | newval |= (value & 0x0700) << 4; |
| 24115 | newval |= (value & 0x00ff); |
| 24116 | put_thumb32_insn (buf, newval); |
| 24117 | } |
| 24118 | else |
| 24119 | { |
| 24120 | newval = md_chars_to_number (buf, 4); |
| 24121 | newval &= 0xfff0f000; |
| 24122 | newval |= value & 0x0fff; |
| 24123 | newval |= (value & 0xf000) << 4; |
| 24124 | md_number_to_chars (buf, newval, 4); |
| 24125 | } |
| 24126 | } |
| 24127 | return; |
| 24128 | |
| 24129 | case BFD_RELOC_ARM_THUMB_ALU_ABS_G0_NC: |
| 24130 | case BFD_RELOC_ARM_THUMB_ALU_ABS_G1_NC: |
| 24131 | case BFD_RELOC_ARM_THUMB_ALU_ABS_G2_NC: |
| 24132 | case BFD_RELOC_ARM_THUMB_ALU_ABS_G3_NC: |
| 24133 | gas_assert (!fixP->fx_done); |
| 24134 | { |
| 24135 | bfd_vma insn; |
| 24136 | bfd_boolean is_mov; |
| 24137 | bfd_vma encoded_addend = value; |
| 24138 | |
| 24139 | /* Check that addend can be encoded in instruction. */ |
| 24140 | if (!seg->use_rela_p && (value < 0 || value > 255)) |
| 24141 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 24142 | _("the offset 0x%08lX is not representable"), |
| 24143 | (unsigned long) encoded_addend); |
| 24144 | |
| 24145 | /* Extract the instruction. */ |
| 24146 | insn = md_chars_to_number (buf, THUMB_SIZE); |
| 24147 | is_mov = (insn & 0xf800) == 0x2000; |
| 24148 | |
| 24149 | /* Encode insn. */ |
| 24150 | if (is_mov) |
| 24151 | { |
| 24152 | if (!seg->use_rela_p) |
| 24153 | insn |= encoded_addend; |
| 24154 | } |
| 24155 | else |
| 24156 | { |
| 24157 | int rd, rs; |
| 24158 | |
| 24159 | /* Extract the instruction. */ |
| 24160 | /* Encoding is the following |
| 24161 | 0x8000 SUB |
| 24162 | 0x00F0 Rd |
| 24163 | 0x000F Rs |
| 24164 | */ |
| 24165 | /* The following conditions must be true : |
| 24166 | - ADD |
| 24167 | - Rd == Rs |
| 24168 | - Rd <= 7 |
| 24169 | */ |
| 24170 | rd = (insn >> 4) & 0xf; |
| 24171 | rs = insn & 0xf; |
| 24172 | if ((insn & 0x8000) || (rd != rs) || rd > 7) |
| 24173 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 24174 | _("Unable to process relocation for thumb opcode: %lx"), |
| 24175 | (unsigned long) insn); |
| 24176 | |
| 24177 | /* Encode as ADD immediate8 thumb 1 code. */ |
| 24178 | insn = 0x3000 | (rd << 8); |
| 24179 | |
| 24180 | /* Place the encoded addend into the first 8 bits of the |
| 24181 | instruction. */ |
| 24182 | if (!seg->use_rela_p) |
| 24183 | insn |= encoded_addend; |
| 24184 | } |
| 24185 | |
| 24186 | /* Update the instruction. */ |
| 24187 | md_number_to_chars (buf, insn, THUMB_SIZE); |
| 24188 | } |
| 24189 | break; |
| 24190 | |
| 24191 | case BFD_RELOC_ARM_ALU_PC_G0_NC: |
| 24192 | case BFD_RELOC_ARM_ALU_PC_G0: |
| 24193 | case BFD_RELOC_ARM_ALU_PC_G1_NC: |
| 24194 | case BFD_RELOC_ARM_ALU_PC_G1: |
| 24195 | case BFD_RELOC_ARM_ALU_PC_G2: |
| 24196 | case BFD_RELOC_ARM_ALU_SB_G0_NC: |
| 24197 | case BFD_RELOC_ARM_ALU_SB_G0: |
| 24198 | case BFD_RELOC_ARM_ALU_SB_G1_NC: |
| 24199 | case BFD_RELOC_ARM_ALU_SB_G1: |
| 24200 | case BFD_RELOC_ARM_ALU_SB_G2: |
| 24201 | gas_assert (!fixP->fx_done); |
| 24202 | if (!seg->use_rela_p) |
| 24203 | { |
| 24204 | bfd_vma insn; |
| 24205 | bfd_vma encoded_addend; |
| 24206 | bfd_vma addend_abs = abs (value); |
| 24207 | |
| 24208 | /* Check that the absolute value of the addend can be |
| 24209 | expressed as an 8-bit constant plus a rotation. */ |
| 24210 | encoded_addend = encode_arm_immediate (addend_abs); |
| 24211 | if (encoded_addend == (unsigned int) FAIL) |
| 24212 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 24213 | _("the offset 0x%08lX is not representable"), |
| 24214 | (unsigned long) addend_abs); |
| 24215 | |
| 24216 | /* Extract the instruction. */ |
| 24217 | insn = md_chars_to_number (buf, INSN_SIZE); |
| 24218 | |
| 24219 | /* If the addend is positive, use an ADD instruction. |
| 24220 | Otherwise use a SUB. Take care not to destroy the S bit. */ |
| 24221 | insn &= 0xff1fffff; |
| 24222 | if (value < 0) |
| 24223 | insn |= 1 << 22; |
| 24224 | else |
| 24225 | insn |= 1 << 23; |
| 24226 | |
| 24227 | /* Place the encoded addend into the first 12 bits of the |
| 24228 | instruction. */ |
| 24229 | insn &= 0xfffff000; |
| 24230 | insn |= encoded_addend; |
| 24231 | |
| 24232 | /* Update the instruction. */ |
| 24233 | md_number_to_chars (buf, insn, INSN_SIZE); |
| 24234 | } |
| 24235 | break; |
| 24236 | |
| 24237 | case BFD_RELOC_ARM_LDR_PC_G0: |
| 24238 | case BFD_RELOC_ARM_LDR_PC_G1: |
| 24239 | case BFD_RELOC_ARM_LDR_PC_G2: |
| 24240 | case BFD_RELOC_ARM_LDR_SB_G0: |
| 24241 | case BFD_RELOC_ARM_LDR_SB_G1: |
| 24242 | case BFD_RELOC_ARM_LDR_SB_G2: |
| 24243 | gas_assert (!fixP->fx_done); |
| 24244 | if (!seg->use_rela_p) |
| 24245 | { |
| 24246 | bfd_vma insn; |
| 24247 | bfd_vma addend_abs = abs (value); |
| 24248 | |
| 24249 | /* Check that the absolute value of the addend can be |
| 24250 | encoded in 12 bits. */ |
| 24251 | if (addend_abs >= 0x1000) |
| 24252 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 24253 | _("bad offset 0x%08lX (only 12 bits available for the magnitude)"), |
| 24254 | (unsigned long) addend_abs); |
| 24255 | |
| 24256 | /* Extract the instruction. */ |
| 24257 | insn = md_chars_to_number (buf, INSN_SIZE); |
| 24258 | |
| 24259 | /* If the addend is negative, clear bit 23 of the instruction. |
| 24260 | Otherwise set it. */ |
| 24261 | if (value < 0) |
| 24262 | insn &= ~(1 << 23); |
| 24263 | else |
| 24264 | insn |= 1 << 23; |
| 24265 | |
| 24266 | /* Place the absolute value of the addend into the first 12 bits |
| 24267 | of the instruction. */ |
| 24268 | insn &= 0xfffff000; |
| 24269 | insn |= addend_abs; |
| 24270 | |
| 24271 | /* Update the instruction. */ |
| 24272 | md_number_to_chars (buf, insn, INSN_SIZE); |
| 24273 | } |
| 24274 | break; |
| 24275 | |
| 24276 | case BFD_RELOC_ARM_LDRS_PC_G0: |
| 24277 | case BFD_RELOC_ARM_LDRS_PC_G1: |
| 24278 | case BFD_RELOC_ARM_LDRS_PC_G2: |
| 24279 | case BFD_RELOC_ARM_LDRS_SB_G0: |
| 24280 | case BFD_RELOC_ARM_LDRS_SB_G1: |
| 24281 | case BFD_RELOC_ARM_LDRS_SB_G2: |
| 24282 | gas_assert (!fixP->fx_done); |
| 24283 | if (!seg->use_rela_p) |
| 24284 | { |
| 24285 | bfd_vma insn; |
| 24286 | bfd_vma addend_abs = abs (value); |
| 24287 | |
| 24288 | /* Check that the absolute value of the addend can be |
| 24289 | encoded in 8 bits. */ |
| 24290 | if (addend_abs >= 0x100) |
| 24291 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 24292 | _("bad offset 0x%08lX (only 8 bits available for the magnitude)"), |
| 24293 | (unsigned long) addend_abs); |
| 24294 | |
| 24295 | /* Extract the instruction. */ |
| 24296 | insn = md_chars_to_number (buf, INSN_SIZE); |
| 24297 | |
| 24298 | /* If the addend is negative, clear bit 23 of the instruction. |
| 24299 | Otherwise set it. */ |
| 24300 | if (value < 0) |
| 24301 | insn &= ~(1 << 23); |
| 24302 | else |
| 24303 | insn |= 1 << 23; |
| 24304 | |
| 24305 | /* Place the first four bits of the absolute value of the addend |
| 24306 | into the first 4 bits of the instruction, and the remaining |
| 24307 | four into bits 8 .. 11. */ |
| 24308 | insn &= 0xfffff0f0; |
| 24309 | insn |= (addend_abs & 0xf) | ((addend_abs & 0xf0) << 4); |
| 24310 | |
| 24311 | /* Update the instruction. */ |
| 24312 | md_number_to_chars (buf, insn, INSN_SIZE); |
| 24313 | } |
| 24314 | break; |
| 24315 | |
| 24316 | case BFD_RELOC_ARM_LDC_PC_G0: |
| 24317 | case BFD_RELOC_ARM_LDC_PC_G1: |
| 24318 | case BFD_RELOC_ARM_LDC_PC_G2: |
| 24319 | case BFD_RELOC_ARM_LDC_SB_G0: |
| 24320 | case BFD_RELOC_ARM_LDC_SB_G1: |
| 24321 | case BFD_RELOC_ARM_LDC_SB_G2: |
| 24322 | gas_assert (!fixP->fx_done); |
| 24323 | if (!seg->use_rela_p) |
| 24324 | { |
| 24325 | bfd_vma insn; |
| 24326 | bfd_vma addend_abs = abs (value); |
| 24327 | |
| 24328 | /* Check that the absolute value of the addend is a multiple of |
| 24329 | four and, when divided by four, fits in 8 bits. */ |
| 24330 | if (addend_abs & 0x3) |
| 24331 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 24332 | _("bad offset 0x%08lX (must be word-aligned)"), |
| 24333 | (unsigned long) addend_abs); |
| 24334 | |
| 24335 | if ((addend_abs >> 2) > 0xff) |
| 24336 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 24337 | _("bad offset 0x%08lX (must be an 8-bit number of words)"), |
| 24338 | (unsigned long) addend_abs); |
| 24339 | |
| 24340 | /* Extract the instruction. */ |
| 24341 | insn = md_chars_to_number (buf, INSN_SIZE); |
| 24342 | |
| 24343 | /* If the addend is negative, clear bit 23 of the instruction. |
| 24344 | Otherwise set it. */ |
| 24345 | if (value < 0) |
| 24346 | insn &= ~(1 << 23); |
| 24347 | else |
| 24348 | insn |= 1 << 23; |
| 24349 | |
| 24350 | /* Place the addend (divided by four) into the first eight |
| 24351 | bits of the instruction. */ |
| 24352 | insn &= 0xfffffff0; |
| 24353 | insn |= addend_abs >> 2; |
| 24354 | |
| 24355 | /* Update the instruction. */ |
| 24356 | md_number_to_chars (buf, insn, INSN_SIZE); |
| 24357 | } |
| 24358 | break; |
| 24359 | |
| 24360 | case BFD_RELOC_ARM_V4BX: |
| 24361 | /* This will need to go in the object file. */ |
| 24362 | fixP->fx_done = 0; |
| 24363 | break; |
| 24364 | |
| 24365 | case BFD_RELOC_UNUSED: |
| 24366 | default: |
| 24367 | as_bad_where (fixP->fx_file, fixP->fx_line, |
| 24368 | _("bad relocation fixup type (%d)"), fixP->fx_r_type); |
| 24369 | } |
| 24370 | } |
| 24371 | |
| 24372 | /* Translate internal representation of relocation info to BFD target |
| 24373 | format. */ |
| 24374 | |
| 24375 | arelent * |
| 24376 | tc_gen_reloc (asection *section, fixS *fixp) |
| 24377 | { |
| 24378 | arelent * reloc; |
| 24379 | bfd_reloc_code_real_type code; |
| 24380 | |
| 24381 | reloc = XNEW (arelent); |
| 24382 | |
| 24383 | reloc->sym_ptr_ptr = XNEW (asymbol *); |
| 24384 | *reloc->sym_ptr_ptr = symbol_get_bfdsym (fixp->fx_addsy); |
| 24385 | reloc->address = fixp->fx_frag->fr_address + fixp->fx_where; |
| 24386 | |
| 24387 | if (fixp->fx_pcrel) |
| 24388 | { |
| 24389 | if (section->use_rela_p) |
| 24390 | fixp->fx_offset -= md_pcrel_from_section (fixp, section); |
| 24391 | else |
| 24392 | fixp->fx_offset = reloc->address; |
| 24393 | } |
| 24394 | reloc->addend = fixp->fx_offset; |
| 24395 | |
| 24396 | switch (fixp->fx_r_type) |
| 24397 | { |
| 24398 | case BFD_RELOC_8: |
| 24399 | if (fixp->fx_pcrel) |
| 24400 | { |
| 24401 | code = BFD_RELOC_8_PCREL; |
| 24402 | break; |
| 24403 | } |
| 24404 | /* Fall through. */ |
| 24405 | |
| 24406 | case BFD_RELOC_16: |
| 24407 | if (fixp->fx_pcrel) |
| 24408 | { |
| 24409 | code = BFD_RELOC_16_PCREL; |
| 24410 | break; |
| 24411 | } |
| 24412 | /* Fall through. */ |
| 24413 | |
| 24414 | case BFD_RELOC_32: |
| 24415 | if (fixp->fx_pcrel) |
| 24416 | { |
| 24417 | code = BFD_RELOC_32_PCREL; |
| 24418 | break; |
| 24419 | } |
| 24420 | /* Fall through. */ |
| 24421 | |
| 24422 | case BFD_RELOC_ARM_MOVW: |
| 24423 | if (fixp->fx_pcrel) |
| 24424 | { |
| 24425 | code = BFD_RELOC_ARM_MOVW_PCREL; |
| 24426 | break; |
| 24427 | } |
| 24428 | /* Fall through. */ |
| 24429 | |
| 24430 | case BFD_RELOC_ARM_MOVT: |
| 24431 | if (fixp->fx_pcrel) |
| 24432 | { |
| 24433 | code = BFD_RELOC_ARM_MOVT_PCREL; |
| 24434 | break; |
| 24435 | } |
| 24436 | /* Fall through. */ |
| 24437 | |
| 24438 | case BFD_RELOC_ARM_THUMB_MOVW: |
| 24439 | if (fixp->fx_pcrel) |
| 24440 | { |
| 24441 | code = BFD_RELOC_ARM_THUMB_MOVW_PCREL; |
| 24442 | break; |
| 24443 | } |
| 24444 | /* Fall through. */ |
| 24445 | |
| 24446 | case BFD_RELOC_ARM_THUMB_MOVT: |
| 24447 | if (fixp->fx_pcrel) |
| 24448 | { |
| 24449 | code = BFD_RELOC_ARM_THUMB_MOVT_PCREL; |
| 24450 | break; |
| 24451 | } |
| 24452 | /* Fall through. */ |
| 24453 | |
| 24454 | case BFD_RELOC_NONE: |
| 24455 | case BFD_RELOC_ARM_PCREL_BRANCH: |
| 24456 | case BFD_RELOC_ARM_PCREL_BLX: |
| 24457 | case BFD_RELOC_RVA: |
| 24458 | case BFD_RELOC_THUMB_PCREL_BRANCH7: |
| 24459 | case BFD_RELOC_THUMB_PCREL_BRANCH9: |
| 24460 | case BFD_RELOC_THUMB_PCREL_BRANCH12: |
| 24461 | case BFD_RELOC_THUMB_PCREL_BRANCH20: |
| 24462 | case BFD_RELOC_THUMB_PCREL_BRANCH23: |
| 24463 | case BFD_RELOC_THUMB_PCREL_BRANCH25: |
| 24464 | case BFD_RELOC_VTABLE_ENTRY: |
| 24465 | case BFD_RELOC_VTABLE_INHERIT: |
| 24466 | #ifdef TE_PE |
| 24467 | case BFD_RELOC_32_SECREL: |
| 24468 | #endif |
| 24469 | code = fixp->fx_r_type; |
| 24470 | break; |
| 24471 | |
| 24472 | case BFD_RELOC_THUMB_PCREL_BLX: |
| 24473 | #ifdef OBJ_ELF |
| 24474 | if (EF_ARM_EABI_VERSION (meabi_flags) >= EF_ARM_EABI_VER4) |
| 24475 | code = BFD_RELOC_THUMB_PCREL_BRANCH23; |
| 24476 | else |
| 24477 | #endif |
| 24478 | code = BFD_RELOC_THUMB_PCREL_BLX; |
| 24479 | break; |
| 24480 | |
| 24481 | case BFD_RELOC_ARM_LITERAL: |
| 24482 | case BFD_RELOC_ARM_HWLITERAL: |
| 24483 | /* If this is called then the a literal has |
| 24484 | been referenced across a section boundary. */ |
| 24485 | as_bad_where (fixp->fx_file, fixp->fx_line, |
| 24486 | _("literal referenced across section boundary")); |
| 24487 | return NULL; |
| 24488 | |
| 24489 | #ifdef OBJ_ELF |
| 24490 | case BFD_RELOC_ARM_TLS_CALL: |
| 24491 | case BFD_RELOC_ARM_THM_TLS_CALL: |
| 24492 | case BFD_RELOC_ARM_TLS_DESCSEQ: |
| 24493 | case BFD_RELOC_ARM_THM_TLS_DESCSEQ: |
| 24494 | case BFD_RELOC_ARM_GOT32: |
| 24495 | case BFD_RELOC_ARM_GOTOFF: |
| 24496 | case BFD_RELOC_ARM_GOT_PREL: |
| 24497 | case BFD_RELOC_ARM_PLT32: |
| 24498 | case BFD_RELOC_ARM_TARGET1: |
| 24499 | case BFD_RELOC_ARM_ROSEGREL32: |
| 24500 | case BFD_RELOC_ARM_SBREL32: |
| 24501 | case BFD_RELOC_ARM_PREL31: |
| 24502 | case BFD_RELOC_ARM_TARGET2: |
| 24503 | case BFD_RELOC_ARM_TLS_LDO32: |
| 24504 | case BFD_RELOC_ARM_PCREL_CALL: |
| 24505 | case BFD_RELOC_ARM_PCREL_JUMP: |
| 24506 | case BFD_RELOC_ARM_ALU_PC_G0_NC: |
| 24507 | case BFD_RELOC_ARM_ALU_PC_G0: |
| 24508 | case BFD_RELOC_ARM_ALU_PC_G1_NC: |
| 24509 | case BFD_RELOC_ARM_ALU_PC_G1: |
| 24510 | case BFD_RELOC_ARM_ALU_PC_G2: |
| 24511 | case BFD_RELOC_ARM_LDR_PC_G0: |
| 24512 | case BFD_RELOC_ARM_LDR_PC_G1: |
| 24513 | case BFD_RELOC_ARM_LDR_PC_G2: |
| 24514 | case BFD_RELOC_ARM_LDRS_PC_G0: |
| 24515 | case BFD_RELOC_ARM_LDRS_PC_G1: |
| 24516 | case BFD_RELOC_ARM_LDRS_PC_G2: |
| 24517 | case BFD_RELOC_ARM_LDC_PC_G0: |
| 24518 | case BFD_RELOC_ARM_LDC_PC_G1: |
| 24519 | case BFD_RELOC_ARM_LDC_PC_G2: |
| 24520 | case BFD_RELOC_ARM_ALU_SB_G0_NC: |
| 24521 | case BFD_RELOC_ARM_ALU_SB_G0: |
| 24522 | case BFD_RELOC_ARM_ALU_SB_G1_NC: |
| 24523 | case BFD_RELOC_ARM_ALU_SB_G1: |
| 24524 | case BFD_RELOC_ARM_ALU_SB_G2: |
| 24525 | case BFD_RELOC_ARM_LDR_SB_G0: |
| 24526 | case BFD_RELOC_ARM_LDR_SB_G1: |
| 24527 | case BFD_RELOC_ARM_LDR_SB_G2: |
| 24528 | case BFD_RELOC_ARM_LDRS_SB_G0: |
| 24529 | case BFD_RELOC_ARM_LDRS_SB_G1: |
| 24530 | case BFD_RELOC_ARM_LDRS_SB_G2: |
| 24531 | case BFD_RELOC_ARM_LDC_SB_G0: |
| 24532 | case BFD_RELOC_ARM_LDC_SB_G1: |
| 24533 | case BFD_RELOC_ARM_LDC_SB_G2: |
| 24534 | case BFD_RELOC_ARM_V4BX: |
| 24535 | case BFD_RELOC_ARM_THUMB_ALU_ABS_G0_NC: |
| 24536 | case BFD_RELOC_ARM_THUMB_ALU_ABS_G1_NC: |
| 24537 | case BFD_RELOC_ARM_THUMB_ALU_ABS_G2_NC: |
| 24538 | case BFD_RELOC_ARM_THUMB_ALU_ABS_G3_NC: |
| 24539 | code = fixp->fx_r_type; |
| 24540 | break; |
| 24541 | |
| 24542 | case BFD_RELOC_ARM_TLS_GOTDESC: |
| 24543 | case BFD_RELOC_ARM_TLS_GD32: |
| 24544 | case BFD_RELOC_ARM_TLS_LE32: |
| 24545 | case BFD_RELOC_ARM_TLS_IE32: |
| 24546 | case BFD_RELOC_ARM_TLS_LDM32: |
| 24547 | /* BFD will include the symbol's address in the addend. |
| 24548 | But we don't want that, so subtract it out again here. */ |
| 24549 | if (!S_IS_COMMON (fixp->fx_addsy)) |
| 24550 | reloc->addend -= (*reloc->sym_ptr_ptr)->value; |
| 24551 | code = fixp->fx_r_type; |
| 24552 | break; |
| 24553 | #endif |
| 24554 | |
| 24555 | case BFD_RELOC_ARM_IMMEDIATE: |
| 24556 | as_bad_where (fixp->fx_file, fixp->fx_line, |
| 24557 | _("internal relocation (type: IMMEDIATE) not fixed up")); |
| 24558 | return NULL; |
| 24559 | |
| 24560 | case BFD_RELOC_ARM_ADRL_IMMEDIATE: |
| 24561 | as_bad_where (fixp->fx_file, fixp->fx_line, |
| 24562 | _("ADRL used for a symbol not defined in the same file")); |
| 24563 | return NULL; |
| 24564 | |
| 24565 | case BFD_RELOC_ARM_OFFSET_IMM: |
| 24566 | if (section->use_rela_p) |
| 24567 | { |
| 24568 | code = fixp->fx_r_type; |
| 24569 | break; |
| 24570 | } |
| 24571 | |
| 24572 | if (fixp->fx_addsy != NULL |
| 24573 | && !S_IS_DEFINED (fixp->fx_addsy) |
| 24574 | && S_IS_LOCAL (fixp->fx_addsy)) |
| 24575 | { |
| 24576 | as_bad_where (fixp->fx_file, fixp->fx_line, |
| 24577 | _("undefined local label `%s'"), |
| 24578 | S_GET_NAME (fixp->fx_addsy)); |
| 24579 | return NULL; |
| 24580 | } |
| 24581 | |
| 24582 | as_bad_where (fixp->fx_file, fixp->fx_line, |
| 24583 | _("internal_relocation (type: OFFSET_IMM) not fixed up")); |
| 24584 | return NULL; |
| 24585 | |
| 24586 | default: |
| 24587 | { |
| 24588 | const char * type; |
| 24589 | |
| 24590 | switch (fixp->fx_r_type) |
| 24591 | { |
| 24592 | case BFD_RELOC_NONE: type = "NONE"; break; |
| 24593 | case BFD_RELOC_ARM_OFFSET_IMM8: type = "OFFSET_IMM8"; break; |
| 24594 | case BFD_RELOC_ARM_SHIFT_IMM: type = "SHIFT_IMM"; break; |
| 24595 | case BFD_RELOC_ARM_SMC: type = "SMC"; break; |
| 24596 | case BFD_RELOC_ARM_SWI: type = "SWI"; break; |
| 24597 | case BFD_RELOC_ARM_MULTI: type = "MULTI"; break; |
| 24598 | case BFD_RELOC_ARM_CP_OFF_IMM: type = "CP_OFF_IMM"; break; |
| 24599 | case BFD_RELOC_ARM_T32_OFFSET_IMM: type = "T32_OFFSET_IMM"; break; |
| 24600 | case BFD_RELOC_ARM_T32_CP_OFF_IMM: type = "T32_CP_OFF_IMM"; break; |
| 24601 | case BFD_RELOC_ARM_THUMB_ADD: type = "THUMB_ADD"; break; |
| 24602 | case BFD_RELOC_ARM_THUMB_SHIFT: type = "THUMB_SHIFT"; break; |
| 24603 | case BFD_RELOC_ARM_THUMB_IMM: type = "THUMB_IMM"; break; |
| 24604 | case BFD_RELOC_ARM_THUMB_OFFSET: type = "THUMB_OFFSET"; break; |
| 24605 | default: type = _("<unknown>"); break; |
| 24606 | } |
| 24607 | as_bad_where (fixp->fx_file, fixp->fx_line, |
| 24608 | _("cannot represent %s relocation in this object file format"), |
| 24609 | type); |
| 24610 | return NULL; |
| 24611 | } |
| 24612 | } |
| 24613 | |
| 24614 | #ifdef OBJ_ELF |
| 24615 | if ((code == BFD_RELOC_32_PCREL || code == BFD_RELOC_32) |
| 24616 | && GOT_symbol |
| 24617 | && fixp->fx_addsy == GOT_symbol) |
| 24618 | { |
| 24619 | code = BFD_RELOC_ARM_GOTPC; |
| 24620 | reloc->addend = fixp->fx_offset = reloc->address; |
| 24621 | } |
| 24622 | #endif |
| 24623 | |
| 24624 | reloc->howto = bfd_reloc_type_lookup (stdoutput, code); |
| 24625 | |
| 24626 | if (reloc->howto == NULL) |
| 24627 | { |
| 24628 | as_bad_where (fixp->fx_file, fixp->fx_line, |
| 24629 | _("cannot represent %s relocation in this object file format"), |
| 24630 | bfd_get_reloc_code_name (code)); |
| 24631 | return NULL; |
| 24632 | } |
| 24633 | |
| 24634 | /* HACK: Since arm ELF uses Rel instead of Rela, encode the |
| 24635 | vtable entry to be used in the relocation's section offset. */ |
| 24636 | if (fixp->fx_r_type == BFD_RELOC_VTABLE_ENTRY) |
| 24637 | reloc->address = fixp->fx_offset; |
| 24638 | |
| 24639 | return reloc; |
| 24640 | } |
| 24641 | |
| 24642 | /* This fix_new is called by cons via TC_CONS_FIX_NEW. */ |
| 24643 | |
| 24644 | void |
| 24645 | cons_fix_new_arm (fragS * frag, |
| 24646 | int where, |
| 24647 | int size, |
| 24648 | expressionS * exp, |
| 24649 | bfd_reloc_code_real_type reloc) |
| 24650 | { |
| 24651 | int pcrel = 0; |
| 24652 | |
| 24653 | /* Pick a reloc. |
| 24654 | FIXME: @@ Should look at CPU word size. */ |
| 24655 | switch (size) |
| 24656 | { |
| 24657 | case 1: |
| 24658 | reloc = BFD_RELOC_8; |
| 24659 | break; |
| 24660 | case 2: |
| 24661 | reloc = BFD_RELOC_16; |
| 24662 | break; |
| 24663 | case 4: |
| 24664 | default: |
| 24665 | reloc = BFD_RELOC_32; |
| 24666 | break; |
| 24667 | case 8: |
| 24668 | reloc = BFD_RELOC_64; |
| 24669 | break; |
| 24670 | } |
| 24671 | |
| 24672 | #ifdef TE_PE |
| 24673 | if (exp->X_op == O_secrel) |
| 24674 | { |
| 24675 | exp->X_op = O_symbol; |
| 24676 | reloc = BFD_RELOC_32_SECREL; |
| 24677 | } |
| 24678 | #endif |
| 24679 | |
| 24680 | fix_new_exp (frag, where, size, exp, pcrel, reloc); |
| 24681 | } |
| 24682 | |
| 24683 | #if defined (OBJ_COFF) |
| 24684 | void |
| 24685 | arm_validate_fix (fixS * fixP) |
| 24686 | { |
| 24687 | /* If the destination of the branch is a defined symbol which does not have |
| 24688 | the THUMB_FUNC attribute, then we must be calling a function which has |
| 24689 | the (interfacearm) attribute. We look for the Thumb entry point to that |
| 24690 | function and change the branch to refer to that function instead. */ |
| 24691 | if (fixP->fx_r_type == BFD_RELOC_THUMB_PCREL_BRANCH23 |
| 24692 | && fixP->fx_addsy != NULL |
| 24693 | && S_IS_DEFINED (fixP->fx_addsy) |
| 24694 | && ! THUMB_IS_FUNC (fixP->fx_addsy)) |
| 24695 | { |
| 24696 | fixP->fx_addsy = find_real_start (fixP->fx_addsy); |
| 24697 | } |
| 24698 | } |
| 24699 | #endif |
| 24700 | |
| 24701 | |
| 24702 | int |
| 24703 | arm_force_relocation (struct fix * fixp) |
| 24704 | { |
| 24705 | #if defined (OBJ_COFF) && defined (TE_PE) |
| 24706 | if (fixp->fx_r_type == BFD_RELOC_RVA) |
| 24707 | return 1; |
| 24708 | #endif |
| 24709 | |
| 24710 | /* In case we have a call or a branch to a function in ARM ISA mode from |
| 24711 | a thumb function or vice-versa force the relocation. These relocations |
| 24712 | are cleared off for some cores that might have blx and simple transformations |
| 24713 | are possible. */ |
| 24714 | |
| 24715 | #ifdef OBJ_ELF |
| 24716 | switch (fixp->fx_r_type) |
| 24717 | { |
| 24718 | case BFD_RELOC_ARM_PCREL_JUMP: |
| 24719 | case BFD_RELOC_ARM_PCREL_CALL: |
| 24720 | case BFD_RELOC_THUMB_PCREL_BLX: |
| 24721 | if (THUMB_IS_FUNC (fixp->fx_addsy)) |
| 24722 | return 1; |
| 24723 | break; |
| 24724 | |
| 24725 | case BFD_RELOC_ARM_PCREL_BLX: |
| 24726 | case BFD_RELOC_THUMB_PCREL_BRANCH25: |
| 24727 | case BFD_RELOC_THUMB_PCREL_BRANCH20: |
| 24728 | case BFD_RELOC_THUMB_PCREL_BRANCH23: |
| 24729 | if (ARM_IS_FUNC (fixp->fx_addsy)) |
| 24730 | return 1; |
| 24731 | break; |
| 24732 | |
| 24733 | default: |
| 24734 | break; |
| 24735 | } |
| 24736 | #endif |
| 24737 | |
| 24738 | /* Resolve these relocations even if the symbol is extern or weak. |
| 24739 | Technically this is probably wrong due to symbol preemption. |
| 24740 | In practice these relocations do not have enough range to be useful |
| 24741 | at dynamic link time, and some code (e.g. in the Linux kernel) |
| 24742 | expects these references to be resolved. */ |
| 24743 | if (fixp->fx_r_type == BFD_RELOC_ARM_IMMEDIATE |
| 24744 | || fixp->fx_r_type == BFD_RELOC_ARM_OFFSET_IMM |
| 24745 | || fixp->fx_r_type == BFD_RELOC_ARM_OFFSET_IMM8 |
| 24746 | || fixp->fx_r_type == BFD_RELOC_ARM_ADRL_IMMEDIATE |
| 24747 | || fixp->fx_r_type == BFD_RELOC_ARM_CP_OFF_IMM |
| 24748 | || fixp->fx_r_type == BFD_RELOC_ARM_CP_OFF_IMM_S2 |
| 24749 | || fixp->fx_r_type == BFD_RELOC_ARM_THUMB_OFFSET |
| 24750 | || fixp->fx_r_type == BFD_RELOC_ARM_T32_ADD_IMM |
| 24751 | || fixp->fx_r_type == BFD_RELOC_ARM_T32_IMMEDIATE |
| 24752 | || fixp->fx_r_type == BFD_RELOC_ARM_T32_IMM12 |
| 24753 | || fixp->fx_r_type == BFD_RELOC_ARM_T32_OFFSET_IMM |
| 24754 | || fixp->fx_r_type == BFD_RELOC_ARM_T32_ADD_PC12 |
| 24755 | || fixp->fx_r_type == BFD_RELOC_ARM_T32_CP_OFF_IMM |
| 24756 | || fixp->fx_r_type == BFD_RELOC_ARM_T32_CP_OFF_IMM_S2) |
| 24757 | return 0; |
| 24758 | |
| 24759 | /* Always leave these relocations for the linker. */ |
| 24760 | if ((fixp->fx_r_type >= BFD_RELOC_ARM_ALU_PC_G0_NC |
| 24761 | && fixp->fx_r_type <= BFD_RELOC_ARM_LDC_SB_G2) |
| 24762 | || fixp->fx_r_type == BFD_RELOC_ARM_LDR_PC_G0) |
| 24763 | return 1; |
| 24764 | |
| 24765 | /* Always generate relocations against function symbols. */ |
| 24766 | if (fixp->fx_r_type == BFD_RELOC_32 |
| 24767 | && fixp->fx_addsy |
| 24768 | && (symbol_get_bfdsym (fixp->fx_addsy)->flags & BSF_FUNCTION)) |
| 24769 | return 1; |
| 24770 | |
| 24771 | return generic_force_reloc (fixp); |
| 24772 | } |
| 24773 | |
| 24774 | #if defined (OBJ_ELF) || defined (OBJ_COFF) |
| 24775 | /* Relocations against function names must be left unadjusted, |
| 24776 | so that the linker can use this information to generate interworking |
| 24777 | stubs. The MIPS version of this function |
| 24778 | also prevents relocations that are mips-16 specific, but I do not |
| 24779 | know why it does this. |
| 24780 | |
| 24781 | FIXME: |
| 24782 | There is one other problem that ought to be addressed here, but |
| 24783 | which currently is not: Taking the address of a label (rather |
| 24784 | than a function) and then later jumping to that address. Such |
| 24785 | addresses also ought to have their bottom bit set (assuming that |
| 24786 | they reside in Thumb code), but at the moment they will not. */ |
| 24787 | |
| 24788 | bfd_boolean |
| 24789 | arm_fix_adjustable (fixS * fixP) |
| 24790 | { |
| 24791 | if (fixP->fx_addsy == NULL) |
| 24792 | return 1; |
| 24793 | |
| 24794 | /* Preserve relocations against symbols with function type. */ |
| 24795 | if (symbol_get_bfdsym (fixP->fx_addsy)->flags & BSF_FUNCTION) |
| 24796 | return FALSE; |
| 24797 | |
| 24798 | if (THUMB_IS_FUNC (fixP->fx_addsy) |
| 24799 | && fixP->fx_subsy == NULL) |
| 24800 | return FALSE; |
| 24801 | |
| 24802 | /* We need the symbol name for the VTABLE entries. */ |
| 24803 | if ( fixP->fx_r_type == BFD_RELOC_VTABLE_INHERIT |
| 24804 | || fixP->fx_r_type == BFD_RELOC_VTABLE_ENTRY) |
| 24805 | return FALSE; |
| 24806 | |
| 24807 | /* Don't allow symbols to be discarded on GOT related relocs. */ |
| 24808 | if (fixP->fx_r_type == BFD_RELOC_ARM_PLT32 |
| 24809 | || fixP->fx_r_type == BFD_RELOC_ARM_GOT32 |
| 24810 | || fixP->fx_r_type == BFD_RELOC_ARM_GOTOFF |
| 24811 | || fixP->fx_r_type == BFD_RELOC_ARM_TLS_GD32 |
| 24812 | || fixP->fx_r_type == BFD_RELOC_ARM_TLS_LE32 |
| 24813 | || fixP->fx_r_type == BFD_RELOC_ARM_TLS_IE32 |
| 24814 | || fixP->fx_r_type == BFD_RELOC_ARM_TLS_LDM32 |
| 24815 | || fixP->fx_r_type == BFD_RELOC_ARM_TLS_LDO32 |
| 24816 | || fixP->fx_r_type == BFD_RELOC_ARM_TLS_GOTDESC |
| 24817 | || fixP->fx_r_type == BFD_RELOC_ARM_TLS_CALL |
| 24818 | || fixP->fx_r_type == BFD_RELOC_ARM_THM_TLS_CALL |
| 24819 | || fixP->fx_r_type == BFD_RELOC_ARM_TLS_DESCSEQ |
| 24820 | || fixP->fx_r_type == BFD_RELOC_ARM_THM_TLS_DESCSEQ |
| 24821 | || fixP->fx_r_type == BFD_RELOC_ARM_TARGET2) |
| 24822 | return FALSE; |
| 24823 | |
| 24824 | /* Similarly for group relocations. */ |
| 24825 | if ((fixP->fx_r_type >= BFD_RELOC_ARM_ALU_PC_G0_NC |
| 24826 | && fixP->fx_r_type <= BFD_RELOC_ARM_LDC_SB_G2) |
| 24827 | || fixP->fx_r_type == BFD_RELOC_ARM_LDR_PC_G0) |
| 24828 | return FALSE; |
| 24829 | |
| 24830 | /* MOVW/MOVT REL relocations have limited offsets, so keep the symbols. */ |
| 24831 | if (fixP->fx_r_type == BFD_RELOC_ARM_MOVW |
| 24832 | || fixP->fx_r_type == BFD_RELOC_ARM_MOVT |
| 24833 | || fixP->fx_r_type == BFD_RELOC_ARM_MOVW_PCREL |
| 24834 | || fixP->fx_r_type == BFD_RELOC_ARM_MOVT_PCREL |
| 24835 | || fixP->fx_r_type == BFD_RELOC_ARM_THUMB_MOVW |
| 24836 | || fixP->fx_r_type == BFD_RELOC_ARM_THUMB_MOVT |
| 24837 | || fixP->fx_r_type == BFD_RELOC_ARM_THUMB_MOVW_PCREL |
| 24838 | || fixP->fx_r_type == BFD_RELOC_ARM_THUMB_MOVT_PCREL) |
| 24839 | return FALSE; |
| 24840 | |
| 24841 | /* BFD_RELOC_ARM_THUMB_ALU_ABS_Gx_NC relocations have VERY limited |
| 24842 | offsets, so keep these symbols. */ |
| 24843 | if (fixP->fx_r_type >= BFD_RELOC_ARM_THUMB_ALU_ABS_G0_NC |
| 24844 | && fixP->fx_r_type <= BFD_RELOC_ARM_THUMB_ALU_ABS_G3_NC) |
| 24845 | return FALSE; |
| 24846 | |
| 24847 | return TRUE; |
| 24848 | } |
| 24849 | #endif /* defined (OBJ_ELF) || defined (OBJ_COFF) */ |
| 24850 | |
| 24851 | #ifdef OBJ_ELF |
| 24852 | const char * |
| 24853 | elf32_arm_target_format (void) |
| 24854 | { |
| 24855 | #ifdef TE_SYMBIAN |
| 24856 | return (target_big_endian |
| 24857 | ? "elf32-bigarm-symbian" |
| 24858 | : "elf32-littlearm-symbian"); |
| 24859 | #elif defined (TE_VXWORKS) |
| 24860 | return (target_big_endian |
| 24861 | ? "elf32-bigarm-vxworks" |
| 24862 | : "elf32-littlearm-vxworks"); |
| 24863 | #elif defined (TE_NACL) |
| 24864 | return (target_big_endian |
| 24865 | ? "elf32-bigarm-nacl" |
| 24866 | : "elf32-littlearm-nacl"); |
| 24867 | #else |
| 24868 | if (target_big_endian) |
| 24869 | return "elf32-bigarm"; |
| 24870 | else |
| 24871 | return "elf32-littlearm"; |
| 24872 | #endif |
| 24873 | } |
| 24874 | |
| 24875 | void |
| 24876 | armelf_frob_symbol (symbolS * symp, |
| 24877 | int * puntp) |
| 24878 | { |
| 24879 | elf_frob_symbol (symp, puntp); |
| 24880 | } |
| 24881 | #endif |
| 24882 | |
| 24883 | /* MD interface: Finalization. */ |
| 24884 | |
| 24885 | void |
| 24886 | arm_cleanup (void) |
| 24887 | { |
| 24888 | literal_pool * pool; |
| 24889 | |
| 24890 | /* Ensure that all the IT blocks are properly closed. */ |
| 24891 | check_it_blocks_finished (); |
| 24892 | |
| 24893 | for (pool = list_of_pools; pool; pool = pool->next) |
| 24894 | { |
| 24895 | /* Put it at the end of the relevant section. */ |
| 24896 | subseg_set (pool->section, pool->sub_section); |
| 24897 | #ifdef OBJ_ELF |
| 24898 | arm_elf_change_section (); |
| 24899 | #endif |
| 24900 | s_ltorg (0); |
| 24901 | } |
| 24902 | } |
| 24903 | |
| 24904 | #ifdef OBJ_ELF |
| 24905 | /* Remove any excess mapping symbols generated for alignment frags in |
| 24906 | SEC. We may have created a mapping symbol before a zero byte |
| 24907 | alignment; remove it if there's a mapping symbol after the |
| 24908 | alignment. */ |
| 24909 | static void |
| 24910 | check_mapping_symbols (bfd *abfd ATTRIBUTE_UNUSED, asection *sec, |
| 24911 | void *dummy ATTRIBUTE_UNUSED) |
| 24912 | { |
| 24913 | segment_info_type *seginfo = seg_info (sec); |
| 24914 | fragS *fragp; |
| 24915 | |
| 24916 | if (seginfo == NULL || seginfo->frchainP == NULL) |
| 24917 | return; |
| 24918 | |
| 24919 | for (fragp = seginfo->frchainP->frch_root; |
| 24920 | fragp != NULL; |
| 24921 | fragp = fragp->fr_next) |
| 24922 | { |
| 24923 | symbolS *sym = fragp->tc_frag_data.last_map; |
| 24924 | fragS *next = fragp->fr_next; |
| 24925 | |
| 24926 | /* Variable-sized frags have been converted to fixed size by |
| 24927 | this point. But if this was variable-sized to start with, |
| 24928 | there will be a fixed-size frag after it. So don't handle |
| 24929 | next == NULL. */ |
| 24930 | if (sym == NULL || next == NULL) |
| 24931 | continue; |
| 24932 | |
| 24933 | if (S_GET_VALUE (sym) < next->fr_address) |
| 24934 | /* Not at the end of this frag. */ |
| 24935 | continue; |
| 24936 | know (S_GET_VALUE (sym) == next->fr_address); |
| 24937 | |
| 24938 | do |
| 24939 | { |
| 24940 | if (next->tc_frag_data.first_map != NULL) |
| 24941 | { |
| 24942 | /* Next frag starts with a mapping symbol. Discard this |
| 24943 | one. */ |
| 24944 | symbol_remove (sym, &symbol_rootP, &symbol_lastP); |
| 24945 | break; |
| 24946 | } |
| 24947 | |
| 24948 | if (next->fr_next == NULL) |
| 24949 | { |
| 24950 | /* This mapping symbol is at the end of the section. Discard |
| 24951 | it. */ |
| 24952 | know (next->fr_fix == 0 && next->fr_var == 0); |
| 24953 | symbol_remove (sym, &symbol_rootP, &symbol_lastP); |
| 24954 | break; |
| 24955 | } |
| 24956 | |
| 24957 | /* As long as we have empty frags without any mapping symbols, |
| 24958 | keep looking. */ |
| 24959 | /* If the next frag is non-empty and does not start with a |
| 24960 | mapping symbol, then this mapping symbol is required. */ |
| 24961 | if (next->fr_address != next->fr_next->fr_address) |
| 24962 | break; |
| 24963 | |
| 24964 | next = next->fr_next; |
| 24965 | } |
| 24966 | while (next != NULL); |
| 24967 | } |
| 24968 | } |
| 24969 | #endif |
| 24970 | |
| 24971 | /* Adjust the symbol table. This marks Thumb symbols as distinct from |
| 24972 | ARM ones. */ |
| 24973 | |
| 24974 | void |
| 24975 | arm_adjust_symtab (void) |
| 24976 | { |
| 24977 | #ifdef OBJ_COFF |
| 24978 | symbolS * sym; |
| 24979 | |
| 24980 | for (sym = symbol_rootP; sym != NULL; sym = symbol_next (sym)) |
| 24981 | { |
| 24982 | if (ARM_IS_THUMB (sym)) |
| 24983 | { |
| 24984 | if (THUMB_IS_FUNC (sym)) |
| 24985 | { |
| 24986 | /* Mark the symbol as a Thumb function. */ |
| 24987 | if ( S_GET_STORAGE_CLASS (sym) == C_STAT |
| 24988 | || S_GET_STORAGE_CLASS (sym) == C_LABEL) /* This can happen! */ |
| 24989 | S_SET_STORAGE_CLASS (sym, C_THUMBSTATFUNC); |
| 24990 | |
| 24991 | else if (S_GET_STORAGE_CLASS (sym) == C_EXT) |
| 24992 | S_SET_STORAGE_CLASS (sym, C_THUMBEXTFUNC); |
| 24993 | else |
| 24994 | as_bad (_("%s: unexpected function type: %d"), |
| 24995 | S_GET_NAME (sym), S_GET_STORAGE_CLASS (sym)); |
| 24996 | } |
| 24997 | else switch (S_GET_STORAGE_CLASS (sym)) |
| 24998 | { |
| 24999 | case C_EXT: |
| 25000 | S_SET_STORAGE_CLASS (sym, C_THUMBEXT); |
| 25001 | break; |
| 25002 | case C_STAT: |
| 25003 | S_SET_STORAGE_CLASS (sym, C_THUMBSTAT); |
| 25004 | break; |
| 25005 | case C_LABEL: |
| 25006 | S_SET_STORAGE_CLASS (sym, C_THUMBLABEL); |
| 25007 | break; |
| 25008 | default: |
| 25009 | /* Do nothing. */ |
| 25010 | break; |
| 25011 | } |
| 25012 | } |
| 25013 | |
| 25014 | if (ARM_IS_INTERWORK (sym)) |
| 25015 | coffsymbol (symbol_get_bfdsym (sym))->native->u.syment.n_flags = 0xFF; |
| 25016 | } |
| 25017 | #endif |
| 25018 | #ifdef OBJ_ELF |
| 25019 | symbolS * sym; |
| 25020 | char bind; |
| 25021 | |
| 25022 | for (sym = symbol_rootP; sym != NULL; sym = symbol_next (sym)) |
| 25023 | { |
| 25024 | if (ARM_IS_THUMB (sym)) |
| 25025 | { |
| 25026 | elf_symbol_type * elf_sym; |
| 25027 | |
| 25028 | elf_sym = elf_symbol (symbol_get_bfdsym (sym)); |
| 25029 | bind = ELF_ST_BIND (elf_sym->internal_elf_sym.st_info); |
| 25030 | |
| 25031 | if (! bfd_is_arm_special_symbol_name (elf_sym->symbol.name, |
| 25032 | BFD_ARM_SPECIAL_SYM_TYPE_ANY)) |
| 25033 | { |
| 25034 | /* If it's a .thumb_func, declare it as so, |
| 25035 | otherwise tag label as .code 16. */ |
| 25036 | if (THUMB_IS_FUNC (sym)) |
| 25037 | ARM_SET_SYM_BRANCH_TYPE (elf_sym->internal_elf_sym.st_target_internal, |
| 25038 | ST_BRANCH_TO_THUMB); |
| 25039 | else if (EF_ARM_EABI_VERSION (meabi_flags) < EF_ARM_EABI_VER4) |
| 25040 | elf_sym->internal_elf_sym.st_info = |
| 25041 | ELF_ST_INFO (bind, STT_ARM_16BIT); |
| 25042 | } |
| 25043 | } |
| 25044 | } |
| 25045 | |
| 25046 | /* Remove any overlapping mapping symbols generated by alignment frags. */ |
| 25047 | bfd_map_over_sections (stdoutput, check_mapping_symbols, (char *) 0); |
| 25048 | /* Now do generic ELF adjustments. */ |
| 25049 | elf_adjust_symtab (); |
| 25050 | #endif |
| 25051 | } |
| 25052 | |
| 25053 | /* MD interface: Initialization. */ |
| 25054 | |
| 25055 | static void |
| 25056 | set_constant_flonums (void) |
| 25057 | { |
| 25058 | int i; |
| 25059 | |
| 25060 | for (i = 0; i < NUM_FLOAT_VALS; i++) |
| 25061 | if (atof_ieee ((char *) fp_const[i], 'x', fp_values[i]) == NULL) |
| 25062 | abort (); |
| 25063 | } |
| 25064 | |
| 25065 | /* Auto-select Thumb mode if it's the only available instruction set for the |
| 25066 | given architecture. */ |
| 25067 | |
| 25068 | static void |
| 25069 | autoselect_thumb_from_cpu_variant (void) |
| 25070 | { |
| 25071 | if (!ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v1)) |
| 25072 | opcode_select (16); |
| 25073 | } |
| 25074 | |
| 25075 | void |
| 25076 | md_begin (void) |
| 25077 | { |
| 25078 | unsigned mach; |
| 25079 | unsigned int i; |
| 25080 | |
| 25081 | if ( (arm_ops_hsh = hash_new ()) == NULL |
| 25082 | || (arm_cond_hsh = hash_new ()) == NULL |
| 25083 | || (arm_shift_hsh = hash_new ()) == NULL |
| 25084 | || (arm_psr_hsh = hash_new ()) == NULL |
| 25085 | || (arm_v7m_psr_hsh = hash_new ()) == NULL |
| 25086 | || (arm_reg_hsh = hash_new ()) == NULL |
| 25087 | || (arm_reloc_hsh = hash_new ()) == NULL |
| 25088 | || (arm_barrier_opt_hsh = hash_new ()) == NULL) |
| 25089 | as_fatal (_("virtual memory exhausted")); |
| 25090 | |
| 25091 | for (i = 0; i < sizeof (insns) / sizeof (struct asm_opcode); i++) |
| 25092 | hash_insert (arm_ops_hsh, insns[i].template_name, (void *) (insns + i)); |
| 25093 | for (i = 0; i < sizeof (conds) / sizeof (struct asm_cond); i++) |
| 25094 | hash_insert (arm_cond_hsh, conds[i].template_name, (void *) (conds + i)); |
| 25095 | for (i = 0; i < sizeof (shift_names) / sizeof (struct asm_shift_name); i++) |
| 25096 | hash_insert (arm_shift_hsh, shift_names[i].name, (void *) (shift_names + i)); |
| 25097 | for (i = 0; i < sizeof (psrs) / sizeof (struct asm_psr); i++) |
| 25098 | hash_insert (arm_psr_hsh, psrs[i].template_name, (void *) (psrs + i)); |
| 25099 | for (i = 0; i < sizeof (v7m_psrs) / sizeof (struct asm_psr); i++) |
| 25100 | hash_insert (arm_v7m_psr_hsh, v7m_psrs[i].template_name, |
| 25101 | (void *) (v7m_psrs + i)); |
| 25102 | for (i = 0; i < sizeof (reg_names) / sizeof (struct reg_entry); i++) |
| 25103 | hash_insert (arm_reg_hsh, reg_names[i].name, (void *) (reg_names + i)); |
| 25104 | for (i = 0; |
| 25105 | i < sizeof (barrier_opt_names) / sizeof (struct asm_barrier_opt); |
| 25106 | i++) |
| 25107 | hash_insert (arm_barrier_opt_hsh, barrier_opt_names[i].template_name, |
| 25108 | (void *) (barrier_opt_names + i)); |
| 25109 | #ifdef OBJ_ELF |
| 25110 | for (i = 0; i < ARRAY_SIZE (reloc_names); i++) |
| 25111 | { |
| 25112 | struct reloc_entry * entry = reloc_names + i; |
| 25113 | |
| 25114 | if (arm_is_eabi() && entry->reloc == BFD_RELOC_ARM_PLT32) |
| 25115 | /* This makes encode_branch() use the EABI versions of this relocation. */ |
| 25116 | entry->reloc = BFD_RELOC_UNUSED; |
| 25117 | |
| 25118 | hash_insert (arm_reloc_hsh, entry->name, (void *) entry); |
| 25119 | } |
| 25120 | #endif |
| 25121 | |
| 25122 | set_constant_flonums (); |
| 25123 | |
| 25124 | /* Set the cpu variant based on the command-line options. We prefer |
| 25125 | -mcpu= over -march= if both are set (as for GCC); and we prefer |
| 25126 | -mfpu= over any other way of setting the floating point unit. |
| 25127 | Use of legacy options with new options are faulted. */ |
| 25128 | if (legacy_cpu) |
| 25129 | { |
| 25130 | if (mcpu_cpu_opt || march_cpu_opt) |
| 25131 | as_bad (_("use of old and new-style options to set CPU type")); |
| 25132 | |
| 25133 | mcpu_cpu_opt = legacy_cpu; |
| 25134 | } |
| 25135 | else if (!mcpu_cpu_opt) |
| 25136 | { |
| 25137 | mcpu_cpu_opt = march_cpu_opt; |
| 25138 | dyn_mcpu_ext_opt = dyn_march_ext_opt; |
| 25139 | /* Avoid double free in arm_md_end. */ |
| 25140 | dyn_march_ext_opt = NULL; |
| 25141 | } |
| 25142 | |
| 25143 | if (legacy_fpu) |
| 25144 | { |
| 25145 | if (mfpu_opt) |
| 25146 | as_bad (_("use of old and new-style options to set FPU type")); |
| 25147 | |
| 25148 | mfpu_opt = legacy_fpu; |
| 25149 | } |
| 25150 | else if (!mfpu_opt) |
| 25151 | { |
| 25152 | #if !(defined (EABI_DEFAULT) || defined (TE_LINUX) \ |
| 25153 | || defined (TE_NetBSD) || defined (TE_VXWORKS)) |
| 25154 | /* Some environments specify a default FPU. If they don't, infer it |
| 25155 | from the processor. */ |
| 25156 | if (mcpu_fpu_opt) |
| 25157 | mfpu_opt = mcpu_fpu_opt; |
| 25158 | else |
| 25159 | mfpu_opt = march_fpu_opt; |
| 25160 | #else |
| 25161 | mfpu_opt = &fpu_default; |
| 25162 | #endif |
| 25163 | } |
| 25164 | |
| 25165 | if (!mfpu_opt) |
| 25166 | { |
| 25167 | if (mcpu_cpu_opt != NULL) |
| 25168 | mfpu_opt = &fpu_default; |
| 25169 | else if (mcpu_fpu_opt != NULL && ARM_CPU_HAS_FEATURE (*mcpu_fpu_opt, arm_ext_v5)) |
| 25170 | mfpu_opt = &fpu_arch_vfp_v2; |
| 25171 | else |
| 25172 | mfpu_opt = &fpu_arch_fpa; |
| 25173 | } |
| 25174 | |
| 25175 | #ifdef CPU_DEFAULT |
| 25176 | if (!mcpu_cpu_opt) |
| 25177 | { |
| 25178 | mcpu_cpu_opt = &cpu_default; |
| 25179 | selected_cpu = cpu_default; |
| 25180 | } |
| 25181 | else if (dyn_mcpu_ext_opt) |
| 25182 | ARM_MERGE_FEATURE_SETS (selected_cpu, *mcpu_cpu_opt, *dyn_mcpu_ext_opt); |
| 25183 | else |
| 25184 | selected_cpu = *mcpu_cpu_opt; |
| 25185 | #else |
| 25186 | if (mcpu_cpu_opt && dyn_mcpu_ext_opt) |
| 25187 | ARM_MERGE_FEATURE_SETS (selected_cpu, *mcpu_cpu_opt, *dyn_mcpu_ext_opt); |
| 25188 | else if (mcpu_cpu_opt) |
| 25189 | selected_cpu = *mcpu_cpu_opt; |
| 25190 | else |
| 25191 | mcpu_cpu_opt = &arm_arch_any; |
| 25192 | #endif |
| 25193 | |
| 25194 | ARM_MERGE_FEATURE_SETS (cpu_variant, *mcpu_cpu_opt, *mfpu_opt); |
| 25195 | if (dyn_mcpu_ext_opt) |
| 25196 | ARM_MERGE_FEATURE_SETS (cpu_variant, cpu_variant, *dyn_mcpu_ext_opt); |
| 25197 | |
| 25198 | autoselect_thumb_from_cpu_variant (); |
| 25199 | |
| 25200 | arm_arch_used = thumb_arch_used = arm_arch_none; |
| 25201 | |
| 25202 | #if defined OBJ_COFF || defined OBJ_ELF |
| 25203 | { |
| 25204 | unsigned int flags = 0; |
| 25205 | |
| 25206 | #if defined OBJ_ELF |
| 25207 | flags = meabi_flags; |
| 25208 | |
| 25209 | switch (meabi_flags) |
| 25210 | { |
| 25211 | case EF_ARM_EABI_UNKNOWN: |
| 25212 | #endif |
| 25213 | /* Set the flags in the private structure. */ |
| 25214 | if (uses_apcs_26) flags |= F_APCS26; |
| 25215 | if (support_interwork) flags |= F_INTERWORK; |
| 25216 | if (uses_apcs_float) flags |= F_APCS_FLOAT; |
| 25217 | if (pic_code) flags |= F_PIC; |
| 25218 | if (!ARM_CPU_HAS_FEATURE (cpu_variant, fpu_any_hard)) |
| 25219 | flags |= F_SOFT_FLOAT; |
| 25220 | |
| 25221 | switch (mfloat_abi_opt) |
| 25222 | { |
| 25223 | case ARM_FLOAT_ABI_SOFT: |
| 25224 | case ARM_FLOAT_ABI_SOFTFP: |
| 25225 | flags |= F_SOFT_FLOAT; |
| 25226 | break; |
| 25227 | |
| 25228 | case ARM_FLOAT_ABI_HARD: |
| 25229 | if (flags & F_SOFT_FLOAT) |
| 25230 | as_bad (_("hard-float conflicts with specified fpu")); |
| 25231 | break; |
| 25232 | } |
| 25233 | |
| 25234 | /* Using pure-endian doubles (even if soft-float). */ |
| 25235 | if (ARM_CPU_HAS_FEATURE (cpu_variant, fpu_endian_pure)) |
| 25236 | flags |= F_VFP_FLOAT; |
| 25237 | |
| 25238 | #if defined OBJ_ELF |
| 25239 | if (ARM_CPU_HAS_FEATURE (cpu_variant, fpu_arch_maverick)) |
| 25240 | flags |= EF_ARM_MAVERICK_FLOAT; |
| 25241 | break; |
| 25242 | |
| 25243 | case EF_ARM_EABI_VER4: |
| 25244 | case EF_ARM_EABI_VER5: |
| 25245 | /* No additional flags to set. */ |
| 25246 | break; |
| 25247 | |
| 25248 | default: |
| 25249 | abort (); |
| 25250 | } |
| 25251 | #endif |
| 25252 | bfd_set_private_flags (stdoutput, flags); |
| 25253 | |
| 25254 | /* We have run out flags in the COFF header to encode the |
| 25255 | status of ATPCS support, so instead we create a dummy, |
| 25256 | empty, debug section called .arm.atpcs. */ |
| 25257 | if (atpcs) |
| 25258 | { |
| 25259 | asection * sec; |
| 25260 | |
| 25261 | sec = bfd_make_section (stdoutput, ".arm.atpcs"); |
| 25262 | |
| 25263 | if (sec != NULL) |
| 25264 | { |
| 25265 | bfd_set_section_flags |
| 25266 | (stdoutput, sec, SEC_READONLY | SEC_DEBUGGING /* | SEC_HAS_CONTENTS */); |
| 25267 | bfd_set_section_size (stdoutput, sec, 0); |
| 25268 | bfd_set_section_contents (stdoutput, sec, NULL, 0, 0); |
| 25269 | } |
| 25270 | } |
| 25271 | } |
| 25272 | #endif |
| 25273 | |
| 25274 | /* Record the CPU type as well. */ |
| 25275 | if (ARM_CPU_HAS_FEATURE (cpu_variant, arm_cext_iwmmxt2)) |
| 25276 | mach = bfd_mach_arm_iWMMXt2; |
| 25277 | else if (ARM_CPU_HAS_FEATURE (cpu_variant, arm_cext_iwmmxt)) |
| 25278 | mach = bfd_mach_arm_iWMMXt; |
| 25279 | else if (ARM_CPU_HAS_FEATURE (cpu_variant, arm_cext_xscale)) |
| 25280 | mach = bfd_mach_arm_XScale; |
| 25281 | else if (ARM_CPU_HAS_FEATURE (cpu_variant, arm_cext_maverick)) |
| 25282 | mach = bfd_mach_arm_ep9312; |
| 25283 | else if (ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v5e)) |
| 25284 | mach = bfd_mach_arm_5TE; |
| 25285 | else if (ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v5)) |
| 25286 | { |
| 25287 | if (ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v4t)) |
| 25288 | mach = bfd_mach_arm_5T; |
| 25289 | else |
| 25290 | mach = bfd_mach_arm_5; |
| 25291 | } |
| 25292 | else if (ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v4)) |
| 25293 | { |
| 25294 | if (ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v4t)) |
| 25295 | mach = bfd_mach_arm_4T; |
| 25296 | else |
| 25297 | mach = bfd_mach_arm_4; |
| 25298 | } |
| 25299 | else if (ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v3m)) |
| 25300 | mach = bfd_mach_arm_3M; |
| 25301 | else if (ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v3)) |
| 25302 | mach = bfd_mach_arm_3; |
| 25303 | else if (ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v2s)) |
| 25304 | mach = bfd_mach_arm_2a; |
| 25305 | else if (ARM_CPU_HAS_FEATURE (cpu_variant, arm_ext_v2)) |
| 25306 | mach = bfd_mach_arm_2; |
| 25307 | else |
| 25308 | mach = bfd_mach_arm_unknown; |
| 25309 | |
| 25310 | bfd_set_arch_mach (stdoutput, TARGET_ARCH, mach); |
| 25311 | } |
| 25312 | |
| 25313 | /* Command line processing. */ |
| 25314 | |
| 25315 | /* md_parse_option |
| 25316 | Invocation line includes a switch not recognized by the base assembler. |
| 25317 | See if it's a processor-specific option. |
| 25318 | |
| 25319 | This routine is somewhat complicated by the need for backwards |
| 25320 | compatibility (since older releases of gcc can't be changed). |
| 25321 | The new options try to make the interface as compatible as |
| 25322 | possible with GCC. |
| 25323 | |
| 25324 | New options (supported) are: |
| 25325 | |
| 25326 | -mcpu=<cpu name> Assemble for selected processor |
| 25327 | -march=<architecture name> Assemble for selected architecture |
| 25328 | -mfpu=<fpu architecture> Assemble for selected FPU. |
| 25329 | -EB/-mbig-endian Big-endian |
| 25330 | -EL/-mlittle-endian Little-endian |
| 25331 | -k Generate PIC code |
| 25332 | -mthumb Start in Thumb mode |
| 25333 | -mthumb-interwork Code supports ARM/Thumb interworking |
| 25334 | |
| 25335 | -m[no-]warn-deprecated Warn about deprecated features |
| 25336 | -m[no-]warn-syms Warn when symbols match instructions |
| 25337 | |
| 25338 | For now we will also provide support for: |
| 25339 | |
| 25340 | -mapcs-32 32-bit Program counter |
| 25341 | -mapcs-26 26-bit Program counter |
| 25342 | -macps-float Floats passed in FP registers |
| 25343 | -mapcs-reentrant Reentrant code |
| 25344 | -matpcs |
| 25345 | (sometime these will probably be replaced with -mapcs=<list of options> |
| 25346 | and -matpcs=<list of options>) |
| 25347 | |
| 25348 | The remaining options are only supported for back-wards compatibility. |
| 25349 | Cpu variants, the arm part is optional: |
| 25350 | -m[arm]1 Currently not supported. |
| 25351 | -m[arm]2, -m[arm]250 Arm 2 and Arm 250 processor |
| 25352 | -m[arm]3 Arm 3 processor |
| 25353 | -m[arm]6[xx], Arm 6 processors |
| 25354 | -m[arm]7[xx][t][[d]m] Arm 7 processors |
| 25355 | -m[arm]8[10] Arm 8 processors |
| 25356 | -m[arm]9[20][tdmi] Arm 9 processors |
| 25357 | -mstrongarm[110[0]] StrongARM processors |
| 25358 | -mxscale XScale processors |
| 25359 | -m[arm]v[2345[t[e]]] Arm architectures |
| 25360 | -mall All (except the ARM1) |
| 25361 | FP variants: |
| 25362 | -mfpa10, -mfpa11 FPA10 and 11 co-processor instructions |
| 25363 | -mfpe-old (No float load/store multiples) |
| 25364 | -mvfpxd VFP Single precision |
| 25365 | -mvfp All VFP |
| 25366 | -mno-fpu Disable all floating point instructions |
| 25367 | |
| 25368 | The following CPU names are recognized: |
| 25369 | arm1, arm2, arm250, arm3, arm6, arm600, arm610, arm620, |
| 25370 | arm7, arm7m, arm7d, arm7dm, arm7di, arm7dmi, arm70, arm700, |
| 25371 | arm700i, arm710 arm710t, arm720, arm720t, arm740t, arm710c, |
| 25372 | arm7100, arm7500, arm7500fe, arm7tdmi, arm8, arm810, arm9, |
| 25373 | arm920, arm920t, arm940t, arm946, arm966, arm9tdmi, arm9e, |
| 25374 | arm10t arm10e, arm1020t, arm1020e, arm10200e, |
| 25375 | strongarm, strongarm110, strongarm1100, strongarm1110, xscale. |
| 25376 | |
| 25377 | */ |
| 25378 | |
| 25379 | const char * md_shortopts = "m:k"; |
| 25380 | |
| 25381 | #ifdef ARM_BI_ENDIAN |
| 25382 | #define OPTION_EB (OPTION_MD_BASE + 0) |
| 25383 | #define OPTION_EL (OPTION_MD_BASE + 1) |
| 25384 | #else |
| 25385 | #if TARGET_BYTES_BIG_ENDIAN |
| 25386 | #define OPTION_EB (OPTION_MD_BASE + 0) |
| 25387 | #else |
| 25388 | #define OPTION_EL (OPTION_MD_BASE + 1) |
| 25389 | #endif |
| 25390 | #endif |
| 25391 | #define OPTION_FIX_V4BX (OPTION_MD_BASE + 2) |
| 25392 | |
| 25393 | struct option md_longopts[] = |
| 25394 | { |
| 25395 | #ifdef OPTION_EB |
| 25396 | {"EB", no_argument, NULL, OPTION_EB}, |
| 25397 | #endif |
| 25398 | #ifdef OPTION_EL |
| 25399 | {"EL", no_argument, NULL, OPTION_EL}, |
| 25400 | #endif |
| 25401 | {"fix-v4bx", no_argument, NULL, OPTION_FIX_V4BX}, |
| 25402 | {NULL, no_argument, NULL, 0} |
| 25403 | }; |
| 25404 | |
| 25405 | |
| 25406 | size_t md_longopts_size = sizeof (md_longopts); |
| 25407 | |
| 25408 | struct arm_option_table |
| 25409 | { |
| 25410 | const char *option; /* Option name to match. */ |
| 25411 | const char *help; /* Help information. */ |
| 25412 | int *var; /* Variable to change. */ |
| 25413 | int value; /* What to change it to. */ |
| 25414 | const char *deprecated; /* If non-null, print this message. */ |
| 25415 | }; |
| 25416 | |
| 25417 | struct arm_option_table arm_opts[] = |
| 25418 | { |
| 25419 | {"k", N_("generate PIC code"), &pic_code, 1, NULL}, |
| 25420 | {"mthumb", N_("assemble Thumb code"), &thumb_mode, 1, NULL}, |
| 25421 | {"mthumb-interwork", N_("support ARM/Thumb interworking"), |
| 25422 | &support_interwork, 1, NULL}, |
| 25423 | {"mapcs-32", N_("code uses 32-bit program counter"), &uses_apcs_26, 0, NULL}, |
| 25424 | {"mapcs-26", N_("code uses 26-bit program counter"), &uses_apcs_26, 1, NULL}, |
| 25425 | {"mapcs-float", N_("floating point args are in fp regs"), &uses_apcs_float, |
| 25426 | 1, NULL}, |
| 25427 | {"mapcs-reentrant", N_("re-entrant code"), &pic_code, 1, NULL}, |
| 25428 | {"matpcs", N_("code is ATPCS conformant"), &atpcs, 1, NULL}, |
| 25429 | {"mbig-endian", N_("assemble for big-endian"), &target_big_endian, 1, NULL}, |
| 25430 | {"mlittle-endian", N_("assemble for little-endian"), &target_big_endian, 0, |
| 25431 | NULL}, |
| 25432 | |
| 25433 | /* These are recognized by the assembler, but have no affect on code. */ |
| 25434 | {"mapcs-frame", N_("use frame pointer"), NULL, 0, NULL}, |
| 25435 | {"mapcs-stack-check", N_("use stack size checking"), NULL, 0, NULL}, |
| 25436 | |
| 25437 | {"mwarn-deprecated", NULL, &warn_on_deprecated, 1, NULL}, |
| 25438 | {"mno-warn-deprecated", N_("do not warn on use of deprecated feature"), |
| 25439 | &warn_on_deprecated, 0, NULL}, |
| 25440 | {"mwarn-syms", N_("warn about symbols that match instruction names [default]"), (int *) (& flag_warn_syms), TRUE, NULL}, |
| 25441 | {"mno-warn-syms", N_("disable warnings about symobls that match instructions"), (int *) (& flag_warn_syms), FALSE, NULL}, |
| 25442 | {NULL, NULL, NULL, 0, NULL} |
| 25443 | }; |
| 25444 | |
| 25445 | struct arm_legacy_option_table |
| 25446 | { |
| 25447 | const char *option; /* Option name to match. */ |
| 25448 | const arm_feature_set **var; /* Variable to change. */ |
| 25449 | const arm_feature_set value; /* What to change it to. */ |
| 25450 | const char *deprecated; /* If non-null, print this message. */ |
| 25451 | }; |
| 25452 | |
| 25453 | const struct arm_legacy_option_table arm_legacy_opts[] = |
| 25454 | { |
| 25455 | /* DON'T add any new processors to this list -- we want the whole list |
| 25456 | to go away... Add them to the processors table instead. */ |
| 25457 | {"marm1", &legacy_cpu, ARM_ARCH_V1, N_("use -mcpu=arm1")}, |
| 25458 | {"m1", &legacy_cpu, ARM_ARCH_V1, N_("use -mcpu=arm1")}, |
| 25459 | {"marm2", &legacy_cpu, ARM_ARCH_V2, N_("use -mcpu=arm2")}, |
| 25460 | {"m2", &legacy_cpu, ARM_ARCH_V2, N_("use -mcpu=arm2")}, |
| 25461 | {"marm250", &legacy_cpu, ARM_ARCH_V2S, N_("use -mcpu=arm250")}, |
| 25462 | {"m250", &legacy_cpu, ARM_ARCH_V2S, N_("use -mcpu=arm250")}, |
| 25463 | {"marm3", &legacy_cpu, ARM_ARCH_V2S, N_("use -mcpu=arm3")}, |
| 25464 | {"m3", &legacy_cpu, ARM_ARCH_V2S, N_("use -mcpu=arm3")}, |
| 25465 | {"marm6", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm6")}, |
| 25466 | {"m6", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm6")}, |
| 25467 | {"marm600", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm600")}, |
| 25468 | {"m600", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm600")}, |
| 25469 | {"marm610", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm610")}, |
| 25470 | {"m610", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm610")}, |
| 25471 | {"marm620", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm620")}, |
| 25472 | {"m620", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm620")}, |
| 25473 | {"marm7", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm7")}, |
| 25474 | {"m7", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm7")}, |
| 25475 | {"marm70", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm70")}, |
| 25476 | {"m70", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm70")}, |
| 25477 | {"marm700", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm700")}, |
| 25478 | {"m700", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm700")}, |
| 25479 | {"marm700i", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm700i")}, |
| 25480 | {"m700i", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm700i")}, |
| 25481 | {"marm710", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm710")}, |
| 25482 | {"m710", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm710")}, |
| 25483 | {"marm710c", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm710c")}, |
| 25484 | {"m710c", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm710c")}, |
| 25485 | {"marm720", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm720")}, |
| 25486 | {"m720", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm720")}, |
| 25487 | {"marm7d", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm7d")}, |
| 25488 | {"m7d", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm7d")}, |
| 25489 | {"marm7di", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm7di")}, |
| 25490 | {"m7di", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm7di")}, |
| 25491 | {"marm7m", &legacy_cpu, ARM_ARCH_V3M, N_("use -mcpu=arm7m")}, |
| 25492 | {"m7m", &legacy_cpu, ARM_ARCH_V3M, N_("use -mcpu=arm7m")}, |
| 25493 | {"marm7dm", &legacy_cpu, ARM_ARCH_V3M, N_("use -mcpu=arm7dm")}, |
| 25494 | {"m7dm", &legacy_cpu, ARM_ARCH_V3M, N_("use -mcpu=arm7dm")}, |
| 25495 | {"marm7dmi", &legacy_cpu, ARM_ARCH_V3M, N_("use -mcpu=arm7dmi")}, |
| 25496 | {"m7dmi", &legacy_cpu, ARM_ARCH_V3M, N_("use -mcpu=arm7dmi")}, |
| 25497 | {"marm7100", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm7100")}, |
| 25498 | {"m7100", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm7100")}, |
| 25499 | {"marm7500", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm7500")}, |
| 25500 | {"m7500", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm7500")}, |
| 25501 | {"marm7500fe", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm7500fe")}, |
| 25502 | {"m7500fe", &legacy_cpu, ARM_ARCH_V3, N_("use -mcpu=arm7500fe")}, |
| 25503 | {"marm7t", &legacy_cpu, ARM_ARCH_V4T, N_("use -mcpu=arm7tdmi")}, |
| 25504 | {"m7t", &legacy_cpu, ARM_ARCH_V4T, N_("use -mcpu=arm7tdmi")}, |
| 25505 | {"marm7tdmi", &legacy_cpu, ARM_ARCH_V4T, N_("use -mcpu=arm7tdmi")}, |
| 25506 | {"m7tdmi", &legacy_cpu, ARM_ARCH_V4T, N_("use -mcpu=arm7tdmi")}, |
| 25507 | {"marm710t", &legacy_cpu, ARM_ARCH_V4T, N_("use -mcpu=arm710t")}, |
| 25508 | {"m710t", &legacy_cpu, ARM_ARCH_V4T, N_("use -mcpu=arm710t")}, |
| 25509 | {"marm720t", &legacy_cpu, ARM_ARCH_V4T, N_("use -mcpu=arm720t")}, |
| 25510 | {"m720t", &legacy_cpu, ARM_ARCH_V4T, N_("use -mcpu=arm720t")}, |
| 25511 | {"marm740t", &legacy_cpu, ARM_ARCH_V4T, N_("use -mcpu=arm740t")}, |
| 25512 | {"m740t", &legacy_cpu, ARM_ARCH_V4T, N_("use -mcpu=arm740t")}, |
| 25513 | {"marm8", &legacy_cpu, ARM_ARCH_V4, N_("use -mcpu=arm8")}, |
| 25514 | {"m8", &legacy_cpu, ARM_ARCH_V4, N_("use -mcpu=arm8")}, |
| 25515 | {"marm810", &legacy_cpu, ARM_ARCH_V4, N_("use -mcpu=arm810")}, |
| 25516 | {"m810", &legacy_cpu, ARM_ARCH_V4, N_("use -mcpu=arm810")}, |
| 25517 | {"marm9", &legacy_cpu, ARM_ARCH_V4T, N_("use -mcpu=arm9")}, |
| 25518 | {"m9", &legacy_cpu, ARM_ARCH_V4T, N_("use -mcpu=arm9")}, |
| 25519 | {"marm9tdmi", &legacy_cpu, ARM_ARCH_V4T, N_("use -mcpu=arm9tdmi")}, |
| 25520 | {"m9tdmi", &legacy_cpu, ARM_ARCH_V4T, N_("use -mcpu=arm9tdmi")}, |
| 25521 | {"marm920", &legacy_cpu, ARM_ARCH_V4T, N_("use -mcpu=arm920")}, |
| 25522 | {"m920", &legacy_cpu, ARM_ARCH_V4T, N_("use -mcpu=arm920")}, |
| 25523 | {"marm940", &legacy_cpu, ARM_ARCH_V4T, N_("use -mcpu=arm940")}, |
| 25524 | {"m940", &legacy_cpu, ARM_ARCH_V4T, N_("use -mcpu=arm940")}, |
| 25525 | {"mstrongarm", &legacy_cpu, ARM_ARCH_V4, N_("use -mcpu=strongarm")}, |
| 25526 | {"mstrongarm110", &legacy_cpu, ARM_ARCH_V4, |
| 25527 | N_("use -mcpu=strongarm110")}, |
| 25528 | {"mstrongarm1100", &legacy_cpu, ARM_ARCH_V4, |
| 25529 | N_("use -mcpu=strongarm1100")}, |
| 25530 | {"mstrongarm1110", &legacy_cpu, ARM_ARCH_V4, |
| 25531 | N_("use -mcpu=strongarm1110")}, |
| 25532 | {"mxscale", &legacy_cpu, ARM_ARCH_XSCALE, N_("use -mcpu=xscale")}, |
| 25533 | {"miwmmxt", &legacy_cpu, ARM_ARCH_IWMMXT, N_("use -mcpu=iwmmxt")}, |
| 25534 | {"mall", &legacy_cpu, ARM_ANY, N_("use -mcpu=all")}, |
| 25535 | |
| 25536 | /* Architecture variants -- don't add any more to this list either. */ |
| 25537 | {"mv2", &legacy_cpu, ARM_ARCH_V2, N_("use -march=armv2")}, |
| 25538 | {"marmv2", &legacy_cpu, ARM_ARCH_V2, N_("use -march=armv2")}, |
| 25539 | {"mv2a", &legacy_cpu, ARM_ARCH_V2S, N_("use -march=armv2a")}, |
| 25540 | {"marmv2a", &legacy_cpu, ARM_ARCH_V2S, N_("use -march=armv2a")}, |
| 25541 | {"mv3", &legacy_cpu, ARM_ARCH_V3, N_("use -march=armv3")}, |
| 25542 | {"marmv3", &legacy_cpu, ARM_ARCH_V3, N_("use -march=armv3")}, |
| 25543 | {"mv3m", &legacy_cpu, ARM_ARCH_V3M, N_("use -march=armv3m")}, |
| 25544 | {"marmv3m", &legacy_cpu, ARM_ARCH_V3M, N_("use -march=armv3m")}, |
| 25545 | {"mv4", &legacy_cpu, ARM_ARCH_V4, N_("use -march=armv4")}, |
| 25546 | {"marmv4", &legacy_cpu, ARM_ARCH_V4, N_("use -march=armv4")}, |
| 25547 | {"mv4t", &legacy_cpu, ARM_ARCH_V4T, N_("use -march=armv4t")}, |
| 25548 | {"marmv4t", &legacy_cpu, ARM_ARCH_V4T, N_("use -march=armv4t")}, |
| 25549 | {"mv5", &legacy_cpu, ARM_ARCH_V5, N_("use -march=armv5")}, |
| 25550 | {"marmv5", &legacy_cpu, ARM_ARCH_V5, N_("use -march=armv5")}, |
| 25551 | {"mv5t", &legacy_cpu, ARM_ARCH_V5T, N_("use -march=armv5t")}, |
| 25552 | {"marmv5t", &legacy_cpu, ARM_ARCH_V5T, N_("use -march=armv5t")}, |
| 25553 | {"mv5e", &legacy_cpu, ARM_ARCH_V5TE, N_("use -march=armv5te")}, |
| 25554 | {"marmv5e", &legacy_cpu, ARM_ARCH_V5TE, N_("use -march=armv5te")}, |
| 25555 | |
| 25556 | /* Floating point variants -- don't add any more to this list either. */ |
| 25557 | {"mfpe-old", &legacy_fpu, FPU_ARCH_FPE, N_("use -mfpu=fpe")}, |
| 25558 | {"mfpa10", &legacy_fpu, FPU_ARCH_FPA, N_("use -mfpu=fpa10")}, |
| 25559 | {"mfpa11", &legacy_fpu, FPU_ARCH_FPA, N_("use -mfpu=fpa11")}, |
| 25560 | {"mno-fpu", &legacy_fpu, ARM_ARCH_NONE, |
| 25561 | N_("use either -mfpu=softfpa or -mfpu=softvfp")}, |
| 25562 | |
| 25563 | {NULL, NULL, ARM_ARCH_NONE, NULL} |
| 25564 | }; |
| 25565 | |
| 25566 | struct arm_cpu_option_table |
| 25567 | { |
| 25568 | const char *name; |
| 25569 | size_t name_len; |
| 25570 | const arm_feature_set value; |
| 25571 | const arm_feature_set ext; |
| 25572 | /* For some CPUs we assume an FPU unless the user explicitly sets |
| 25573 | -mfpu=... */ |
| 25574 | const arm_feature_set default_fpu; |
| 25575 | /* The canonical name of the CPU, or NULL to use NAME converted to upper |
| 25576 | case. */ |
| 25577 | const char *canonical_name; |
| 25578 | }; |
| 25579 | |
| 25580 | /* This list should, at a minimum, contain all the cpu names |
| 25581 | recognized by GCC. */ |
| 25582 | #define ARM_CPU_OPT(N, CN, V, E, DF) { N, sizeof (N) - 1, V, E, DF, CN } |
| 25583 | static const struct arm_cpu_option_table arm_cpus[] = |
| 25584 | { |
| 25585 | ARM_CPU_OPT ("all", NULL, ARM_ANY, |
| 25586 | ARM_ARCH_NONE, |
| 25587 | FPU_ARCH_FPA), |
| 25588 | ARM_CPU_OPT ("arm1", NULL, ARM_ARCH_V1, |
| 25589 | ARM_ARCH_NONE, |
| 25590 | FPU_ARCH_FPA), |
| 25591 | ARM_CPU_OPT ("arm2", NULL, ARM_ARCH_V2, |
| 25592 | ARM_ARCH_NONE, |
| 25593 | FPU_ARCH_FPA), |
| 25594 | ARM_CPU_OPT ("arm250", NULL, ARM_ARCH_V2S, |
| 25595 | ARM_ARCH_NONE, |
| 25596 | FPU_ARCH_FPA), |
| 25597 | ARM_CPU_OPT ("arm3", NULL, ARM_ARCH_V2S, |
| 25598 | ARM_ARCH_NONE, |
| 25599 | FPU_ARCH_FPA), |
| 25600 | ARM_CPU_OPT ("arm6", NULL, ARM_ARCH_V3, |
| 25601 | ARM_ARCH_NONE, |
| 25602 | FPU_ARCH_FPA), |
| 25603 | ARM_CPU_OPT ("arm60", NULL, ARM_ARCH_V3, |
| 25604 | ARM_ARCH_NONE, |
| 25605 | FPU_ARCH_FPA), |
| 25606 | ARM_CPU_OPT ("arm600", NULL, ARM_ARCH_V3, |
| 25607 | ARM_ARCH_NONE, |
| 25608 | FPU_ARCH_FPA), |
| 25609 | ARM_CPU_OPT ("arm610", NULL, ARM_ARCH_V3, |
| 25610 | ARM_ARCH_NONE, |
| 25611 | FPU_ARCH_FPA), |
| 25612 | ARM_CPU_OPT ("arm620", NULL, ARM_ARCH_V3, |
| 25613 | ARM_ARCH_NONE, |
| 25614 | FPU_ARCH_FPA), |
| 25615 | ARM_CPU_OPT ("arm7", NULL, ARM_ARCH_V3, |
| 25616 | ARM_ARCH_NONE, |
| 25617 | FPU_ARCH_FPA), |
| 25618 | ARM_CPU_OPT ("arm7m", NULL, ARM_ARCH_V3M, |
| 25619 | ARM_ARCH_NONE, |
| 25620 | FPU_ARCH_FPA), |
| 25621 | ARM_CPU_OPT ("arm7d", NULL, ARM_ARCH_V3, |
| 25622 | ARM_ARCH_NONE, |
| 25623 | FPU_ARCH_FPA), |
| 25624 | ARM_CPU_OPT ("arm7dm", NULL, ARM_ARCH_V3M, |
| 25625 | ARM_ARCH_NONE, |
| 25626 | FPU_ARCH_FPA), |
| 25627 | ARM_CPU_OPT ("arm7di", NULL, ARM_ARCH_V3, |
| 25628 | ARM_ARCH_NONE, |
| 25629 | FPU_ARCH_FPA), |
| 25630 | ARM_CPU_OPT ("arm7dmi", NULL, ARM_ARCH_V3M, |
| 25631 | ARM_ARCH_NONE, |
| 25632 | FPU_ARCH_FPA), |
| 25633 | ARM_CPU_OPT ("arm70", NULL, ARM_ARCH_V3, |
| 25634 | ARM_ARCH_NONE, |
| 25635 | FPU_ARCH_FPA), |
| 25636 | ARM_CPU_OPT ("arm700", NULL, ARM_ARCH_V3, |
| 25637 | ARM_ARCH_NONE, |
| 25638 | FPU_ARCH_FPA), |
| 25639 | ARM_CPU_OPT ("arm700i", NULL, ARM_ARCH_V3, |
| 25640 | ARM_ARCH_NONE, |
| 25641 | FPU_ARCH_FPA), |
| 25642 | ARM_CPU_OPT ("arm710", NULL, ARM_ARCH_V3, |
| 25643 | ARM_ARCH_NONE, |
| 25644 | FPU_ARCH_FPA), |
| 25645 | ARM_CPU_OPT ("arm710t", NULL, ARM_ARCH_V4T, |
| 25646 | ARM_ARCH_NONE, |
| 25647 | FPU_ARCH_FPA), |
| 25648 | ARM_CPU_OPT ("arm720", NULL, ARM_ARCH_V3, |
| 25649 | ARM_ARCH_NONE, |
| 25650 | FPU_ARCH_FPA), |
| 25651 | ARM_CPU_OPT ("arm720t", NULL, ARM_ARCH_V4T, |
| 25652 | ARM_ARCH_NONE, |
| 25653 | FPU_ARCH_FPA), |
| 25654 | ARM_CPU_OPT ("arm740t", NULL, ARM_ARCH_V4T, |
| 25655 | ARM_ARCH_NONE, |
| 25656 | FPU_ARCH_FPA), |
| 25657 | ARM_CPU_OPT ("arm710c", NULL, ARM_ARCH_V3, |
| 25658 | ARM_ARCH_NONE, |
| 25659 | FPU_ARCH_FPA), |
| 25660 | ARM_CPU_OPT ("arm7100", NULL, ARM_ARCH_V3, |
| 25661 | ARM_ARCH_NONE, |
| 25662 | FPU_ARCH_FPA), |
| 25663 | ARM_CPU_OPT ("arm7500", NULL, ARM_ARCH_V3, |
| 25664 | ARM_ARCH_NONE, |
| 25665 | FPU_ARCH_FPA), |
| 25666 | ARM_CPU_OPT ("arm7500fe", NULL, ARM_ARCH_V3, |
| 25667 | ARM_ARCH_NONE, |
| 25668 | FPU_ARCH_FPA), |
| 25669 | ARM_CPU_OPT ("arm7t", NULL, ARM_ARCH_V4T, |
| 25670 | ARM_ARCH_NONE, |
| 25671 | FPU_ARCH_FPA), |
| 25672 | ARM_CPU_OPT ("arm7tdmi", NULL, ARM_ARCH_V4T, |
| 25673 | ARM_ARCH_NONE, |
| 25674 | FPU_ARCH_FPA), |
| 25675 | ARM_CPU_OPT ("arm7tdmi-s", NULL, ARM_ARCH_V4T, |
| 25676 | ARM_ARCH_NONE, |
| 25677 | FPU_ARCH_FPA), |
| 25678 | ARM_CPU_OPT ("arm8", NULL, ARM_ARCH_V4, |
| 25679 | ARM_ARCH_NONE, |
| 25680 | FPU_ARCH_FPA), |
| 25681 | ARM_CPU_OPT ("arm810", NULL, ARM_ARCH_V4, |
| 25682 | ARM_ARCH_NONE, |
| 25683 | FPU_ARCH_FPA), |
| 25684 | ARM_CPU_OPT ("strongarm", NULL, ARM_ARCH_V4, |
| 25685 | ARM_ARCH_NONE, |
| 25686 | FPU_ARCH_FPA), |
| 25687 | ARM_CPU_OPT ("strongarm1", NULL, ARM_ARCH_V4, |
| 25688 | ARM_ARCH_NONE, |
| 25689 | FPU_ARCH_FPA), |
| 25690 | ARM_CPU_OPT ("strongarm110", NULL, ARM_ARCH_V4, |
| 25691 | ARM_ARCH_NONE, |
| 25692 | FPU_ARCH_FPA), |
| 25693 | ARM_CPU_OPT ("strongarm1100", NULL, ARM_ARCH_V4, |
| 25694 | ARM_ARCH_NONE, |
| 25695 | FPU_ARCH_FPA), |
| 25696 | ARM_CPU_OPT ("strongarm1110", NULL, ARM_ARCH_V4, |
| 25697 | ARM_ARCH_NONE, |
| 25698 | FPU_ARCH_FPA), |
| 25699 | ARM_CPU_OPT ("arm9", NULL, ARM_ARCH_V4T, |
| 25700 | ARM_ARCH_NONE, |
| 25701 | FPU_ARCH_FPA), |
| 25702 | ARM_CPU_OPT ("arm920", "ARM920T", ARM_ARCH_V4T, |
| 25703 | ARM_ARCH_NONE, |
| 25704 | FPU_ARCH_FPA), |
| 25705 | ARM_CPU_OPT ("arm920t", NULL, ARM_ARCH_V4T, |
| 25706 | ARM_ARCH_NONE, |
| 25707 | FPU_ARCH_FPA), |
| 25708 | ARM_CPU_OPT ("arm922t", NULL, ARM_ARCH_V4T, |
| 25709 | ARM_ARCH_NONE, |
| 25710 | FPU_ARCH_FPA), |
| 25711 | ARM_CPU_OPT ("arm940t", NULL, ARM_ARCH_V4T, |
| 25712 | ARM_ARCH_NONE, |
| 25713 | FPU_ARCH_FPA), |
| 25714 | ARM_CPU_OPT ("arm9tdmi", NULL, ARM_ARCH_V4T, |
| 25715 | ARM_ARCH_NONE, |
| 25716 | FPU_ARCH_FPA), |
| 25717 | ARM_CPU_OPT ("fa526", NULL, ARM_ARCH_V4, |
| 25718 | ARM_ARCH_NONE, |
| 25719 | FPU_ARCH_FPA), |
| 25720 | ARM_CPU_OPT ("fa626", NULL, ARM_ARCH_V4, |
| 25721 | ARM_ARCH_NONE, |
| 25722 | FPU_ARCH_FPA), |
| 25723 | |
| 25724 | /* For V5 or later processors we default to using VFP; but the user |
| 25725 | should really set the FPU type explicitly. */ |
| 25726 | ARM_CPU_OPT ("arm9e-r0", NULL, ARM_ARCH_V5TExP, |
| 25727 | ARM_ARCH_NONE, |
| 25728 | FPU_ARCH_VFP_V2), |
| 25729 | ARM_CPU_OPT ("arm9e", NULL, ARM_ARCH_V5TE, |
| 25730 | ARM_ARCH_NONE, |
| 25731 | FPU_ARCH_VFP_V2), |
| 25732 | ARM_CPU_OPT ("arm926ej", "ARM926EJ-S", ARM_ARCH_V5TEJ, |
| 25733 | ARM_ARCH_NONE, |
| 25734 | FPU_ARCH_VFP_V2), |
| 25735 | ARM_CPU_OPT ("arm926ejs", "ARM926EJ-S", ARM_ARCH_V5TEJ, |
| 25736 | ARM_ARCH_NONE, |
| 25737 | FPU_ARCH_VFP_V2), |
| 25738 | ARM_CPU_OPT ("arm926ej-s", NULL, ARM_ARCH_V5TEJ, |
| 25739 | ARM_ARCH_NONE, |
| 25740 | FPU_ARCH_VFP_V2), |
| 25741 | ARM_CPU_OPT ("arm946e-r0", NULL, ARM_ARCH_V5TExP, |
| 25742 | ARM_ARCH_NONE, |
| 25743 | FPU_ARCH_VFP_V2), |
| 25744 | ARM_CPU_OPT ("arm946e", "ARM946E-S", ARM_ARCH_V5TE, |
| 25745 | ARM_ARCH_NONE, |
| 25746 | FPU_ARCH_VFP_V2), |
| 25747 | ARM_CPU_OPT ("arm946e-s", NULL, ARM_ARCH_V5TE, |
| 25748 | ARM_ARCH_NONE, |
| 25749 | FPU_ARCH_VFP_V2), |
| 25750 | ARM_CPU_OPT ("arm966e-r0", NULL, ARM_ARCH_V5TExP, |
| 25751 | ARM_ARCH_NONE, |
| 25752 | FPU_ARCH_VFP_V2), |
| 25753 | ARM_CPU_OPT ("arm966e", "ARM966E-S", ARM_ARCH_V5TE, |
| 25754 | ARM_ARCH_NONE, |
| 25755 | FPU_ARCH_VFP_V2), |
| 25756 | ARM_CPU_OPT ("arm966e-s", NULL, ARM_ARCH_V5TE, |
| 25757 | ARM_ARCH_NONE, |
| 25758 | FPU_ARCH_VFP_V2), |
| 25759 | ARM_CPU_OPT ("arm968e-s", NULL, ARM_ARCH_V5TE, |
| 25760 | ARM_ARCH_NONE, |
| 25761 | FPU_ARCH_VFP_V2), |
| 25762 | ARM_CPU_OPT ("arm10t", NULL, ARM_ARCH_V5T, |
| 25763 | ARM_ARCH_NONE, |
| 25764 | FPU_ARCH_VFP_V1), |
| 25765 | ARM_CPU_OPT ("arm10tdmi", NULL, ARM_ARCH_V5T, |
| 25766 | ARM_ARCH_NONE, |
| 25767 | FPU_ARCH_VFP_V1), |
| 25768 | ARM_CPU_OPT ("arm10e", NULL, ARM_ARCH_V5TE, |
| 25769 | ARM_ARCH_NONE, |
| 25770 | FPU_ARCH_VFP_V2), |
| 25771 | ARM_CPU_OPT ("arm1020", "ARM1020E", ARM_ARCH_V5TE, |
| 25772 | ARM_ARCH_NONE, |
| 25773 | FPU_ARCH_VFP_V2), |
| 25774 | ARM_CPU_OPT ("arm1020t", NULL, ARM_ARCH_V5T, |
| 25775 | ARM_ARCH_NONE, |
| 25776 | FPU_ARCH_VFP_V1), |
| 25777 | ARM_CPU_OPT ("arm1020e", NULL, ARM_ARCH_V5TE, |
| 25778 | ARM_ARCH_NONE, |
| 25779 | FPU_ARCH_VFP_V2), |
| 25780 | ARM_CPU_OPT ("arm1022e", NULL, ARM_ARCH_V5TE, |
| 25781 | ARM_ARCH_NONE, |
| 25782 | FPU_ARCH_VFP_V2), |
| 25783 | ARM_CPU_OPT ("arm1026ejs", "ARM1026EJ-S", ARM_ARCH_V5TEJ, |
| 25784 | ARM_ARCH_NONE, |
| 25785 | FPU_ARCH_VFP_V2), |
| 25786 | ARM_CPU_OPT ("arm1026ej-s", NULL, ARM_ARCH_V5TEJ, |
| 25787 | ARM_ARCH_NONE, |
| 25788 | FPU_ARCH_VFP_V2), |
| 25789 | ARM_CPU_OPT ("fa606te", NULL, ARM_ARCH_V5TE, |
| 25790 | ARM_ARCH_NONE, |
| 25791 | FPU_ARCH_VFP_V2), |
| 25792 | ARM_CPU_OPT ("fa616te", NULL, ARM_ARCH_V5TE, |
| 25793 | ARM_ARCH_NONE, |
| 25794 | FPU_ARCH_VFP_V2), |
| 25795 | ARM_CPU_OPT ("fa626te", NULL, ARM_ARCH_V5TE, |
| 25796 | ARM_ARCH_NONE, |
| 25797 | FPU_ARCH_VFP_V2), |
| 25798 | ARM_CPU_OPT ("fmp626", NULL, ARM_ARCH_V5TE, |
| 25799 | ARM_ARCH_NONE, |
| 25800 | FPU_ARCH_VFP_V2), |
| 25801 | ARM_CPU_OPT ("fa726te", NULL, ARM_ARCH_V5TE, |
| 25802 | ARM_ARCH_NONE, |
| 25803 | FPU_ARCH_VFP_V2), |
| 25804 | ARM_CPU_OPT ("arm1136js", "ARM1136J-S", ARM_ARCH_V6, |
| 25805 | ARM_ARCH_NONE, |
| 25806 | FPU_NONE), |
| 25807 | ARM_CPU_OPT ("arm1136j-s", NULL, ARM_ARCH_V6, |
| 25808 | ARM_ARCH_NONE, |
| 25809 | FPU_NONE), |
| 25810 | ARM_CPU_OPT ("arm1136jfs", "ARM1136JF-S", ARM_ARCH_V6, |
| 25811 | ARM_ARCH_NONE, |
| 25812 | FPU_ARCH_VFP_V2), |
| 25813 | ARM_CPU_OPT ("arm1136jf-s", NULL, ARM_ARCH_V6, |
| 25814 | ARM_ARCH_NONE, |
| 25815 | FPU_ARCH_VFP_V2), |
| 25816 | ARM_CPU_OPT ("mpcore", "MPCore", ARM_ARCH_V6K, |
| 25817 | ARM_ARCH_NONE, |
| 25818 | FPU_ARCH_VFP_V2), |
| 25819 | ARM_CPU_OPT ("mpcorenovfp", "MPCore", ARM_ARCH_V6K, |
| 25820 | ARM_ARCH_NONE, |
| 25821 | FPU_NONE), |
| 25822 | ARM_CPU_OPT ("arm1156t2-s", NULL, ARM_ARCH_V6T2, |
| 25823 | ARM_ARCH_NONE, |
| 25824 | FPU_NONE), |
| 25825 | ARM_CPU_OPT ("arm1156t2f-s", NULL, ARM_ARCH_V6T2, |
| 25826 | ARM_ARCH_NONE, |
| 25827 | FPU_ARCH_VFP_V2), |
| 25828 | ARM_CPU_OPT ("arm1176jz-s", NULL, ARM_ARCH_V6KZ, |
| 25829 | ARM_ARCH_NONE, |
| 25830 | FPU_NONE), |
| 25831 | ARM_CPU_OPT ("arm1176jzf-s", NULL, ARM_ARCH_V6KZ, |
| 25832 | ARM_ARCH_NONE, |
| 25833 | FPU_ARCH_VFP_V2), |
| 25834 | ARM_CPU_OPT ("cortex-a5", "Cortex-A5", ARM_ARCH_V7A, |
| 25835 | ARM_FEATURE_CORE_LOW (ARM_EXT_MP | ARM_EXT_SEC), |
| 25836 | FPU_NONE), |
| 25837 | ARM_CPU_OPT ("cortex-a7", "Cortex-A7", ARM_ARCH_V7VE, |
| 25838 | ARM_ARCH_NONE, |
| 25839 | FPU_ARCH_NEON_VFP_V4), |
| 25840 | ARM_CPU_OPT ("cortex-a8", "Cortex-A8", ARM_ARCH_V7A, |
| 25841 | ARM_FEATURE_CORE_LOW (ARM_EXT_SEC), |
| 25842 | ARM_FEATURE_COPROC (FPU_VFP_V3 | FPU_NEON_EXT_V1)), |
| 25843 | ARM_CPU_OPT ("cortex-a9", "Cortex-A9", ARM_ARCH_V7A, |
| 25844 | ARM_FEATURE_CORE_LOW (ARM_EXT_MP | ARM_EXT_SEC), |
| 25845 | ARM_FEATURE_COPROC (FPU_VFP_V3 | FPU_NEON_EXT_V1)), |
| 25846 | ARM_CPU_OPT ("cortex-a12", "Cortex-A12", ARM_ARCH_V7VE, |
| 25847 | ARM_ARCH_NONE, |
| 25848 | FPU_ARCH_NEON_VFP_V4), |
| 25849 | ARM_CPU_OPT ("cortex-a15", "Cortex-A15", ARM_ARCH_V7VE, |
| 25850 | ARM_ARCH_NONE, |
| 25851 | FPU_ARCH_NEON_VFP_V4), |
| 25852 | ARM_CPU_OPT ("cortex-a17", "Cortex-A17", ARM_ARCH_V7VE, |
| 25853 | ARM_ARCH_NONE, |
| 25854 | FPU_ARCH_NEON_VFP_V4), |
| 25855 | ARM_CPU_OPT ("cortex-a32", "Cortex-A32", ARM_ARCH_V8A, |
| 25856 | ARM_FEATURE_COPROC (CRC_EXT_ARMV8), |
| 25857 | FPU_ARCH_CRYPTO_NEON_VFP_ARMV8), |
| 25858 | ARM_CPU_OPT ("cortex-a35", "Cortex-A35", ARM_ARCH_V8A, |
| 25859 | ARM_FEATURE_COPROC (CRC_EXT_ARMV8), |
| 25860 | FPU_ARCH_CRYPTO_NEON_VFP_ARMV8), |
| 25861 | ARM_CPU_OPT ("cortex-a53", "Cortex-A53", ARM_ARCH_V8A, |
| 25862 | ARM_FEATURE_COPROC (CRC_EXT_ARMV8), |
| 25863 | FPU_ARCH_CRYPTO_NEON_VFP_ARMV8), |
| 25864 | ARM_CPU_OPT ("cortex-a55", "Cortex-A55", ARM_ARCH_V8_2A, |
| 25865 | ARM_FEATURE_CORE_HIGH (ARM_EXT2_FP16_INST), |
| 25866 | FPU_ARCH_CRYPTO_NEON_VFP_ARMV8), |
| 25867 | ARM_CPU_OPT ("cortex-a57", "Cortex-A57", ARM_ARCH_V8A, |
| 25868 | ARM_FEATURE_COPROC (CRC_EXT_ARMV8), |
| 25869 | FPU_ARCH_CRYPTO_NEON_VFP_ARMV8), |
| 25870 | ARM_CPU_OPT ("cortex-a72", "Cortex-A72", ARM_ARCH_V8A, |
| 25871 | ARM_FEATURE_COPROC (CRC_EXT_ARMV8), |
| 25872 | FPU_ARCH_CRYPTO_NEON_VFP_ARMV8), |
| 25873 | ARM_CPU_OPT ("cortex-a73", "Cortex-A73", ARM_ARCH_V8A, |
| 25874 | ARM_FEATURE_COPROC (CRC_EXT_ARMV8), |
| 25875 | FPU_ARCH_CRYPTO_NEON_VFP_ARMV8), |
| 25876 | ARM_CPU_OPT ("cortex-a75", "Cortex-A75", ARM_ARCH_V8_2A, |
| 25877 | ARM_FEATURE_CORE_HIGH (ARM_EXT2_FP16_INST), |
| 25878 | FPU_ARCH_CRYPTO_NEON_VFP_ARMV8), |
| 25879 | ARM_CPU_OPT ("cortex-r4", "Cortex-R4", ARM_ARCH_V7R, |
| 25880 | ARM_ARCH_NONE, |
| 25881 | FPU_NONE), |
| 25882 | ARM_CPU_OPT ("cortex-r4f", "Cortex-R4F", ARM_ARCH_V7R, |
| 25883 | ARM_ARCH_NONE, |
| 25884 | FPU_ARCH_VFP_V3D16), |
| 25885 | ARM_CPU_OPT ("cortex-r5", "Cortex-R5", ARM_ARCH_V7R, |
| 25886 | ARM_FEATURE_CORE_LOW (ARM_EXT_ADIV), |
| 25887 | FPU_NONE), |
| 25888 | ARM_CPU_OPT ("cortex-r7", "Cortex-R7", ARM_ARCH_V7R, |
| 25889 | ARM_FEATURE_CORE_LOW (ARM_EXT_ADIV), |
| 25890 | FPU_ARCH_VFP_V3D16), |
| 25891 | ARM_CPU_OPT ("cortex-r8", "Cortex-R8", ARM_ARCH_V7R, |
| 25892 | ARM_FEATURE_CORE_LOW (ARM_EXT_ADIV), |
| 25893 | FPU_ARCH_VFP_V3D16), |
| 25894 | ARM_CPU_OPT ("cortex-r52", "Cortex-R52", ARM_ARCH_V8R, |
| 25895 | ARM_FEATURE_COPROC (CRC_EXT_ARMV8), |
| 25896 | FPU_ARCH_NEON_VFP_ARMV8), |
| 25897 | ARM_CPU_OPT ("cortex-m33", "Cortex-M33", ARM_ARCH_V8M_MAIN, |
| 25898 | ARM_FEATURE_CORE_LOW (ARM_EXT_V5ExP | ARM_EXT_V6_DSP), |
| 25899 | FPU_NONE), |
| 25900 | ARM_CPU_OPT ("cortex-m23", "Cortex-M23", ARM_ARCH_V8M_BASE, |
| 25901 | ARM_ARCH_NONE, |
| 25902 | FPU_NONE), |
| 25903 | ARM_CPU_OPT ("cortex-m7", "Cortex-M7", ARM_ARCH_V7EM, |
| 25904 | ARM_ARCH_NONE, |
| 25905 | FPU_NONE), |
| 25906 | ARM_CPU_OPT ("cortex-m4", "Cortex-M4", ARM_ARCH_V7EM, |
| 25907 | ARM_ARCH_NONE, |
| 25908 | FPU_NONE), |
| 25909 | ARM_CPU_OPT ("cortex-m3", "Cortex-M3", ARM_ARCH_V7M, |
| 25910 | ARM_ARCH_NONE, |
| 25911 | FPU_NONE), |
| 25912 | ARM_CPU_OPT ("cortex-m1", "Cortex-M1", ARM_ARCH_V6SM, |
| 25913 | ARM_ARCH_NONE, |
| 25914 | FPU_NONE), |
| 25915 | ARM_CPU_OPT ("cortex-m0", "Cortex-M0", ARM_ARCH_V6SM, |
| 25916 | ARM_ARCH_NONE, |
| 25917 | FPU_NONE), |
| 25918 | ARM_CPU_OPT ("cortex-m0plus", "Cortex-M0+", ARM_ARCH_V6SM, |
| 25919 | ARM_ARCH_NONE, |
| 25920 | FPU_NONE), |
| 25921 | ARM_CPU_OPT ("exynos-m1", "Samsung Exynos M1", ARM_ARCH_V8A, |
| 25922 | ARM_FEATURE_COPROC (CRC_EXT_ARMV8), |
| 25923 | FPU_ARCH_CRYPTO_NEON_VFP_ARMV8), |
| 25924 | |
| 25925 | /* ??? XSCALE is really an architecture. */ |
| 25926 | ARM_CPU_OPT ("xscale", NULL, ARM_ARCH_XSCALE, |
| 25927 | ARM_ARCH_NONE, |
| 25928 | FPU_ARCH_VFP_V2), |
| 25929 | |
| 25930 | /* ??? iwmmxt is not a processor. */ |
| 25931 | ARM_CPU_OPT ("iwmmxt", NULL, ARM_ARCH_IWMMXT, |
| 25932 | ARM_ARCH_NONE, |
| 25933 | FPU_ARCH_VFP_V2), |
| 25934 | ARM_CPU_OPT ("iwmmxt2", NULL, ARM_ARCH_IWMMXT2, |
| 25935 | ARM_ARCH_NONE, |
| 25936 | FPU_ARCH_VFP_V2), |
| 25937 | ARM_CPU_OPT ("i80200", NULL, ARM_ARCH_XSCALE, |
| 25938 | ARM_ARCH_NONE, |
| 25939 | FPU_ARCH_VFP_V2), |
| 25940 | |
| 25941 | /* Maverick */ |
| 25942 | ARM_CPU_OPT ("ep9312", "ARM920T", |
| 25943 | ARM_FEATURE_LOW (ARM_AEXT_V4T, ARM_CEXT_MAVERICK), |
| 25944 | ARM_ARCH_NONE, FPU_ARCH_MAVERICK), |
| 25945 | |
| 25946 | /* Marvell processors. */ |
| 25947 | ARM_CPU_OPT ("marvell-pj4", NULL, ARM_ARCH_V7A, |
| 25948 | ARM_FEATURE_CORE_LOW (ARM_EXT_MP | ARM_EXT_SEC), |
| 25949 | FPU_ARCH_VFP_V3D16), |
| 25950 | ARM_CPU_OPT ("marvell-whitney", NULL, ARM_ARCH_V7A, |
| 25951 | ARM_FEATURE_CORE_LOW (ARM_EXT_MP | ARM_EXT_SEC), |
| 25952 | FPU_ARCH_NEON_VFP_V4), |
| 25953 | |
| 25954 | /* APM X-Gene family. */ |
| 25955 | ARM_CPU_OPT ("xgene1", "APM X-Gene 1", ARM_ARCH_V8A, |
| 25956 | ARM_ARCH_NONE, |
| 25957 | FPU_ARCH_CRYPTO_NEON_VFP_ARMV8), |
| 25958 | ARM_CPU_OPT ("xgene2", "APM X-Gene 2", ARM_ARCH_V8A, |
| 25959 | ARM_FEATURE_COPROC (CRC_EXT_ARMV8), |
| 25960 | FPU_ARCH_CRYPTO_NEON_VFP_ARMV8), |
| 25961 | |
| 25962 | { NULL, 0, ARM_ARCH_NONE, ARM_ARCH_NONE, ARM_ARCH_NONE, NULL } |
| 25963 | }; |
| 25964 | #undef ARM_CPU_OPT |
| 25965 | |
| 25966 | struct arm_arch_option_table |
| 25967 | { |
| 25968 | const char *name; |
| 25969 | size_t name_len; |
| 25970 | const arm_feature_set value; |
| 25971 | const arm_feature_set default_fpu; |
| 25972 | }; |
| 25973 | |
| 25974 | /* This list should, at a minimum, contain all the architecture names |
| 25975 | recognized by GCC. */ |
| 25976 | #define ARM_ARCH_OPT(N, V, DF) { N, sizeof (N) - 1, V, DF } |
| 25977 | static const struct arm_arch_option_table arm_archs[] = |
| 25978 | { |
| 25979 | ARM_ARCH_OPT ("all", ARM_ANY, FPU_ARCH_FPA), |
| 25980 | ARM_ARCH_OPT ("armv1", ARM_ARCH_V1, FPU_ARCH_FPA), |
| 25981 | ARM_ARCH_OPT ("armv2", ARM_ARCH_V2, FPU_ARCH_FPA), |
| 25982 | ARM_ARCH_OPT ("armv2a", ARM_ARCH_V2S, FPU_ARCH_FPA), |
| 25983 | ARM_ARCH_OPT ("armv2s", ARM_ARCH_V2S, FPU_ARCH_FPA), |
| 25984 | ARM_ARCH_OPT ("armv3", ARM_ARCH_V3, FPU_ARCH_FPA), |
| 25985 | ARM_ARCH_OPT ("armv3m", ARM_ARCH_V3M, FPU_ARCH_FPA), |
| 25986 | ARM_ARCH_OPT ("armv4", ARM_ARCH_V4, FPU_ARCH_FPA), |
| 25987 | ARM_ARCH_OPT ("armv4xm", ARM_ARCH_V4xM, FPU_ARCH_FPA), |
| 25988 | ARM_ARCH_OPT ("armv4t", ARM_ARCH_V4T, FPU_ARCH_FPA), |
| 25989 | ARM_ARCH_OPT ("armv4txm", ARM_ARCH_V4TxM, FPU_ARCH_FPA), |
| 25990 | ARM_ARCH_OPT ("armv5", ARM_ARCH_V5, FPU_ARCH_VFP), |
| 25991 | ARM_ARCH_OPT ("armv5t", ARM_ARCH_V5T, FPU_ARCH_VFP), |
| 25992 | ARM_ARCH_OPT ("armv5txm", ARM_ARCH_V5TxM, FPU_ARCH_VFP), |
| 25993 | ARM_ARCH_OPT ("armv5te", ARM_ARCH_V5TE, FPU_ARCH_VFP), |
| 25994 | ARM_ARCH_OPT ("armv5texp", ARM_ARCH_V5TExP, FPU_ARCH_VFP), |
| 25995 | ARM_ARCH_OPT ("armv5tej", ARM_ARCH_V5TEJ, FPU_ARCH_VFP), |
| 25996 | ARM_ARCH_OPT ("armv6", ARM_ARCH_V6, FPU_ARCH_VFP), |
| 25997 | ARM_ARCH_OPT ("armv6j", ARM_ARCH_V6, FPU_ARCH_VFP), |
| 25998 | ARM_ARCH_OPT ("armv6k", ARM_ARCH_V6K, FPU_ARCH_VFP), |
| 25999 | ARM_ARCH_OPT ("armv6z", ARM_ARCH_V6Z, FPU_ARCH_VFP), |
| 26000 | /* The official spelling of this variant is ARMv6KZ, the name "armv6zk" is |
| 26001 | kept to preserve existing behaviour. */ |
| 26002 | ARM_ARCH_OPT ("armv6kz", ARM_ARCH_V6KZ, FPU_ARCH_VFP), |
| 26003 | ARM_ARCH_OPT ("armv6zk", ARM_ARCH_V6KZ, FPU_ARCH_VFP), |
| 26004 | ARM_ARCH_OPT ("armv6t2", ARM_ARCH_V6T2, FPU_ARCH_VFP), |
| 26005 | ARM_ARCH_OPT ("armv6kt2", ARM_ARCH_V6KT2, FPU_ARCH_VFP), |
| 26006 | ARM_ARCH_OPT ("armv6zt2", ARM_ARCH_V6ZT2, FPU_ARCH_VFP), |
| 26007 | /* The official spelling of this variant is ARMv6KZ, the name "armv6zkt2" is |
| 26008 | kept to preserve existing behaviour. */ |
| 26009 | ARM_ARCH_OPT ("armv6kzt2", ARM_ARCH_V6KZT2, FPU_ARCH_VFP), |
| 26010 | ARM_ARCH_OPT ("armv6zkt2", ARM_ARCH_V6KZT2, FPU_ARCH_VFP), |
| 26011 | ARM_ARCH_OPT ("armv6-m", ARM_ARCH_V6M, FPU_ARCH_VFP), |
| 26012 | ARM_ARCH_OPT ("armv6s-m", ARM_ARCH_V6SM, FPU_ARCH_VFP), |
| 26013 | ARM_ARCH_OPT ("armv7", ARM_ARCH_V7, FPU_ARCH_VFP), |
| 26014 | /* The official spelling of the ARMv7 profile variants is the dashed form. |
| 26015 | Accept the non-dashed form for compatibility with old toolchains. */ |
| 26016 | ARM_ARCH_OPT ("armv7a", ARM_ARCH_V7A, FPU_ARCH_VFP), |
| 26017 | ARM_ARCH_OPT ("armv7ve", ARM_ARCH_V7VE, FPU_ARCH_VFP), |
| 26018 | ARM_ARCH_OPT ("armv7r", ARM_ARCH_V7R, FPU_ARCH_VFP), |
| 26019 | ARM_ARCH_OPT ("armv7m", ARM_ARCH_V7M, FPU_ARCH_VFP), |
| 26020 | ARM_ARCH_OPT ("armv7-a", ARM_ARCH_V7A, FPU_ARCH_VFP), |
| 26021 | ARM_ARCH_OPT ("armv7-r", ARM_ARCH_V7R, FPU_ARCH_VFP), |
| 26022 | ARM_ARCH_OPT ("armv7-m", ARM_ARCH_V7M, FPU_ARCH_VFP), |
| 26023 | ARM_ARCH_OPT ("armv7e-m", ARM_ARCH_V7EM, FPU_ARCH_VFP), |
| 26024 | ARM_ARCH_OPT ("armv8-m.base", ARM_ARCH_V8M_BASE, FPU_ARCH_VFP), |
| 26025 | ARM_ARCH_OPT ("armv8-m.main", ARM_ARCH_V8M_MAIN, FPU_ARCH_VFP), |
| 26026 | ARM_ARCH_OPT ("armv8-a", ARM_ARCH_V8A, FPU_ARCH_VFP), |
| 26027 | ARM_ARCH_OPT ("armv8.1-a", ARM_ARCH_V8_1A, FPU_ARCH_VFP), |
| 26028 | ARM_ARCH_OPT ("armv8.2-a", ARM_ARCH_V8_2A, FPU_ARCH_VFP), |
| 26029 | ARM_ARCH_OPT ("armv8.3-a", ARM_ARCH_V8_3A, FPU_ARCH_VFP), |
| 26030 | ARM_ARCH_OPT ("armv8-r", ARM_ARCH_V8R, FPU_ARCH_VFP), |
| 26031 | ARM_ARCH_OPT ("xscale", ARM_ARCH_XSCALE, FPU_ARCH_VFP), |
| 26032 | ARM_ARCH_OPT ("iwmmxt", ARM_ARCH_IWMMXT, FPU_ARCH_VFP), |
| 26033 | ARM_ARCH_OPT ("iwmmxt2", ARM_ARCH_IWMMXT2,FPU_ARCH_VFP), |
| 26034 | { NULL, 0, ARM_ARCH_NONE, ARM_ARCH_NONE } |
| 26035 | }; |
| 26036 | #undef ARM_ARCH_OPT |
| 26037 | |
| 26038 | /* ISA extensions in the co-processor and main instruction set space. */ |
| 26039 | struct arm_option_extension_value_table |
| 26040 | { |
| 26041 | const char *name; |
| 26042 | size_t name_len; |
| 26043 | const arm_feature_set merge_value; |
| 26044 | const arm_feature_set clear_value; |
| 26045 | /* List of architectures for which an extension is available. ARM_ARCH_NONE |
| 26046 | indicates that an extension is available for all architectures while |
| 26047 | ARM_ANY marks an empty entry. */ |
| 26048 | const arm_feature_set allowed_archs[2]; |
| 26049 | }; |
| 26050 | |
| 26051 | /* The following table must be in alphabetical order with a NULL last entry. |
| 26052 | */ |
| 26053 | #define ARM_EXT_OPT(N, M, C, AA) { N, sizeof (N) - 1, M, C, { AA, ARM_ANY } } |
| 26054 | #define ARM_EXT_OPT2(N, M, C, AA1, AA2) { N, sizeof (N) - 1, M, C, {AA1, AA2} } |
| 26055 | static const struct arm_option_extension_value_table arm_extensions[] = |
| 26056 | { |
| 26057 | ARM_EXT_OPT ("crc", ARCH_CRC_ARMV8, ARM_FEATURE_COPROC (CRC_EXT_ARMV8), |
| 26058 | ARM_FEATURE_CORE_LOW (ARM_EXT_V8)), |
| 26059 | ARM_EXT_OPT ("crypto", FPU_ARCH_CRYPTO_NEON_VFP_ARMV8, |
| 26060 | ARM_FEATURE_COPROC (FPU_CRYPTO_ARMV8), |
| 26061 | ARM_FEATURE_CORE_LOW (ARM_EXT_V8)), |
| 26062 | ARM_EXT_OPT ("dotprod", FPU_ARCH_DOTPROD_NEON_VFP_ARMV8, |
| 26063 | ARM_FEATURE_COPROC (FPU_NEON_EXT_DOTPROD), |
| 26064 | ARM_ARCH_V8_2A), |
| 26065 | ARM_EXT_OPT ("dsp", ARM_FEATURE_CORE_LOW (ARM_EXT_V5ExP | ARM_EXT_V6_DSP), |
| 26066 | ARM_FEATURE_CORE_LOW (ARM_EXT_V5ExP | ARM_EXT_V6_DSP), |
| 26067 | ARM_FEATURE_CORE (ARM_EXT_V7M, ARM_EXT2_V8M)), |
| 26068 | ARM_EXT_OPT ("fp", FPU_ARCH_VFP_ARMV8, ARM_FEATURE_COPROC (FPU_VFP_ARMV8), |
| 26069 | ARM_FEATURE_CORE_LOW (ARM_EXT_V8)), |
| 26070 | ARM_EXT_OPT ("fp16", ARM_FEATURE_CORE_HIGH (ARM_EXT2_FP16_INST), |
| 26071 | ARM_FEATURE_CORE_HIGH (ARM_EXT2_FP16_INST), |
| 26072 | ARM_ARCH_V8_2A), |
| 26073 | ARM_EXT_OPT2 ("idiv", ARM_FEATURE_CORE_LOW (ARM_EXT_ADIV | ARM_EXT_DIV), |
| 26074 | ARM_FEATURE_CORE_LOW (ARM_EXT_ADIV | ARM_EXT_DIV), |
| 26075 | ARM_FEATURE_CORE_LOW (ARM_EXT_V7A), |
| 26076 | ARM_FEATURE_CORE_LOW (ARM_EXT_V7R)), |
| 26077 | /* Duplicate entry for the purpose of allowing ARMv7 to match in presence of |
| 26078 | Thumb divide instruction. Due to this having the same name as the |
| 26079 | previous entry, this will be ignored when doing command-line parsing and |
| 26080 | only considered by build attribute selection code. */ |
| 26081 | ARM_EXT_OPT ("idiv", ARM_FEATURE_CORE_LOW (ARM_EXT_DIV), |
| 26082 | ARM_FEATURE_CORE_LOW (ARM_EXT_DIV), |
| 26083 | ARM_FEATURE_CORE_LOW (ARM_EXT_V7)), |
| 26084 | ARM_EXT_OPT ("iwmmxt",ARM_FEATURE_COPROC (ARM_CEXT_IWMMXT), |
| 26085 | ARM_FEATURE_COPROC (ARM_CEXT_IWMMXT), ARM_ARCH_NONE), |
| 26086 | ARM_EXT_OPT ("iwmmxt2", ARM_FEATURE_COPROC (ARM_CEXT_IWMMXT2), |
| 26087 | ARM_FEATURE_COPROC (ARM_CEXT_IWMMXT2), ARM_ARCH_NONE), |
| 26088 | ARM_EXT_OPT ("maverick", ARM_FEATURE_COPROC (ARM_CEXT_MAVERICK), |
| 26089 | ARM_FEATURE_COPROC (ARM_CEXT_MAVERICK), ARM_ARCH_NONE), |
| 26090 | ARM_EXT_OPT2 ("mp", ARM_FEATURE_CORE_LOW (ARM_EXT_MP), |
| 26091 | ARM_FEATURE_CORE_LOW (ARM_EXT_MP), |
| 26092 | ARM_FEATURE_CORE_LOW (ARM_EXT_V7A), |
| 26093 | ARM_FEATURE_CORE_LOW (ARM_EXT_V7R)), |
| 26094 | ARM_EXT_OPT ("os", ARM_FEATURE_CORE_LOW (ARM_EXT_OS), |
| 26095 | ARM_FEATURE_CORE_LOW (ARM_EXT_OS), |
| 26096 | ARM_FEATURE_CORE_LOW (ARM_EXT_V6M)), |
| 26097 | ARM_EXT_OPT ("pan", ARM_FEATURE_CORE_HIGH (ARM_EXT2_PAN), |
| 26098 | ARM_FEATURE (ARM_EXT_V8, ARM_EXT2_PAN, 0), |
| 26099 | ARM_FEATURE_CORE_HIGH (ARM_EXT2_V8A)), |
| 26100 | ARM_EXT_OPT ("ras", ARM_FEATURE_CORE_HIGH (ARM_EXT2_RAS), |
| 26101 | ARM_FEATURE (ARM_EXT_V8, ARM_EXT2_RAS, 0), |
| 26102 | ARM_FEATURE_CORE_HIGH (ARM_EXT2_V8A)), |
| 26103 | ARM_EXT_OPT ("rdma", FPU_ARCH_NEON_VFP_ARMV8_1, |
| 26104 | ARM_FEATURE_COPROC (FPU_NEON_ARMV8 | FPU_NEON_EXT_RDMA), |
| 26105 | ARM_FEATURE_CORE_HIGH (ARM_EXT2_V8A)), |
| 26106 | ARM_EXT_OPT2 ("sec", ARM_FEATURE_CORE_LOW (ARM_EXT_SEC), |
| 26107 | ARM_FEATURE_CORE_LOW (ARM_EXT_SEC), |
| 26108 | ARM_FEATURE_CORE_LOW (ARM_EXT_V6K), |
| 26109 | ARM_FEATURE_CORE_LOW (ARM_EXT_V7A)), |
| 26110 | ARM_EXT_OPT ("simd", FPU_ARCH_NEON_VFP_ARMV8, |
| 26111 | ARM_FEATURE_COPROC (FPU_NEON_ARMV8), |
| 26112 | ARM_FEATURE_CORE_LOW (ARM_EXT_V8)), |
| 26113 | ARM_EXT_OPT ("virt", ARM_FEATURE_CORE_LOW (ARM_EXT_VIRT | ARM_EXT_ADIV |
| 26114 | | ARM_EXT_DIV), |
| 26115 | ARM_FEATURE_CORE_LOW (ARM_EXT_VIRT), |
| 26116 | ARM_FEATURE_CORE_LOW (ARM_EXT_V7A)), |
| 26117 | ARM_EXT_OPT ("xscale",ARM_FEATURE_COPROC (ARM_CEXT_XSCALE), |
| 26118 | ARM_FEATURE_COPROC (ARM_CEXT_XSCALE), ARM_ARCH_NONE), |
| 26119 | { NULL, 0, ARM_ARCH_NONE, ARM_ARCH_NONE, { ARM_ARCH_NONE, ARM_ARCH_NONE } } |
| 26120 | }; |
| 26121 | #undef ARM_EXT_OPT |
| 26122 | |
| 26123 | /* ISA floating-point and Advanced SIMD extensions. */ |
| 26124 | struct arm_option_fpu_value_table |
| 26125 | { |
| 26126 | const char *name; |
| 26127 | const arm_feature_set value; |
| 26128 | }; |
| 26129 | |
| 26130 | /* This list should, at a minimum, contain all the fpu names |
| 26131 | recognized by GCC. */ |
| 26132 | static const struct arm_option_fpu_value_table arm_fpus[] = |
| 26133 | { |
| 26134 | {"softfpa", FPU_NONE}, |
| 26135 | {"fpe", FPU_ARCH_FPE}, |
| 26136 | {"fpe2", FPU_ARCH_FPE}, |
| 26137 | {"fpe3", FPU_ARCH_FPA}, /* Third release supports LFM/SFM. */ |
| 26138 | {"fpa", FPU_ARCH_FPA}, |
| 26139 | {"fpa10", FPU_ARCH_FPA}, |
| 26140 | {"fpa11", FPU_ARCH_FPA}, |
| 26141 | {"arm7500fe", FPU_ARCH_FPA}, |
| 26142 | {"softvfp", FPU_ARCH_VFP}, |
| 26143 | {"softvfp+vfp", FPU_ARCH_VFP_V2}, |
| 26144 | {"vfp", FPU_ARCH_VFP_V2}, |
| 26145 | {"vfp9", FPU_ARCH_VFP_V2}, |
| 26146 | {"vfp3", FPU_ARCH_VFP_V3}, /* Undocumented, use vfpv3. */ |
| 26147 | {"vfp10", FPU_ARCH_VFP_V2}, |
| 26148 | {"vfp10-r0", FPU_ARCH_VFP_V1}, |
| 26149 | {"vfpxd", FPU_ARCH_VFP_V1xD}, |
| 26150 | {"vfpv2", FPU_ARCH_VFP_V2}, |
| 26151 | {"vfpv3", FPU_ARCH_VFP_V3}, |
| 26152 | {"vfpv3-fp16", FPU_ARCH_VFP_V3_FP16}, |
| 26153 | {"vfpv3-d16", FPU_ARCH_VFP_V3D16}, |
| 26154 | {"vfpv3-d16-fp16", FPU_ARCH_VFP_V3D16_FP16}, |
| 26155 | {"vfpv3xd", FPU_ARCH_VFP_V3xD}, |
| 26156 | {"vfpv3xd-fp16", FPU_ARCH_VFP_V3xD_FP16}, |
| 26157 | {"arm1020t", FPU_ARCH_VFP_V1}, |
| 26158 | {"arm1020e", FPU_ARCH_VFP_V2}, |
| 26159 | {"arm1136jfs", FPU_ARCH_VFP_V2}, /* Undocumented, use arm1136jf-s. */ |
| 26160 | {"arm1136jf-s", FPU_ARCH_VFP_V2}, |
| 26161 | {"maverick", FPU_ARCH_MAVERICK}, |
| 26162 | {"neon", FPU_ARCH_VFP_V3_PLUS_NEON_V1}, |
| 26163 | {"neon-vfpv3", FPU_ARCH_VFP_V3_PLUS_NEON_V1}, |
| 26164 | {"neon-fp16", FPU_ARCH_NEON_FP16}, |
| 26165 | {"vfpv4", FPU_ARCH_VFP_V4}, |
| 26166 | {"vfpv4-d16", FPU_ARCH_VFP_V4D16}, |
| 26167 | {"fpv4-sp-d16", FPU_ARCH_VFP_V4_SP_D16}, |
| 26168 | {"fpv5-d16", FPU_ARCH_VFP_V5D16}, |
| 26169 | {"fpv5-sp-d16", FPU_ARCH_VFP_V5_SP_D16}, |
| 26170 | {"neon-vfpv4", FPU_ARCH_NEON_VFP_V4}, |
| 26171 | {"fp-armv8", FPU_ARCH_VFP_ARMV8}, |
| 26172 | {"neon-fp-armv8", FPU_ARCH_NEON_VFP_ARMV8}, |
| 26173 | {"crypto-neon-fp-armv8", |
| 26174 | FPU_ARCH_CRYPTO_NEON_VFP_ARMV8}, |
| 26175 | {"neon-fp-armv8.1", FPU_ARCH_NEON_VFP_ARMV8_1}, |
| 26176 | {"crypto-neon-fp-armv8.1", |
| 26177 | FPU_ARCH_CRYPTO_NEON_VFP_ARMV8_1}, |
| 26178 | {NULL, ARM_ARCH_NONE} |
| 26179 | }; |
| 26180 | |
| 26181 | struct arm_option_value_table |
| 26182 | { |
| 26183 | const char *name; |
| 26184 | long value; |
| 26185 | }; |
| 26186 | |
| 26187 | static const struct arm_option_value_table arm_float_abis[] = |
| 26188 | { |
| 26189 | {"hard", ARM_FLOAT_ABI_HARD}, |
| 26190 | {"softfp", ARM_FLOAT_ABI_SOFTFP}, |
| 26191 | {"soft", ARM_FLOAT_ABI_SOFT}, |
| 26192 | {NULL, 0} |
| 26193 | }; |
| 26194 | |
| 26195 | #ifdef OBJ_ELF |
| 26196 | /* We only know how to output GNU and ver 4/5 (AAELF) formats. */ |
| 26197 | static const struct arm_option_value_table arm_eabis[] = |
| 26198 | { |
| 26199 | {"gnu", EF_ARM_EABI_UNKNOWN}, |
| 26200 | {"4", EF_ARM_EABI_VER4}, |
| 26201 | {"5", EF_ARM_EABI_VER5}, |
| 26202 | {NULL, 0} |
| 26203 | }; |
| 26204 | #endif |
| 26205 | |
| 26206 | struct arm_long_option_table |
| 26207 | { |
| 26208 | const char * option; /* Substring to match. */ |
| 26209 | const char * help; /* Help information. */ |
| 26210 | int (* func) (const char * subopt); /* Function to decode sub-option. */ |
| 26211 | const char * deprecated; /* If non-null, print this message. */ |
| 26212 | }; |
| 26213 | |
| 26214 | static bfd_boolean |
| 26215 | arm_parse_extension (const char *str, const arm_feature_set *opt_set, |
| 26216 | arm_feature_set **ext_set_p) |
| 26217 | { |
| 26218 | /* We insist on extensions being specified in alphabetical order, and with |
| 26219 | extensions being added before being removed. We achieve this by having |
| 26220 | the global ARM_EXTENSIONS table in alphabetical order, and using the |
| 26221 | ADDING_VALUE variable to indicate whether we are adding an extension (1) |
| 26222 | or removing it (0) and only allowing it to change in the order |
| 26223 | -1 -> 1 -> 0. */ |
| 26224 | const struct arm_option_extension_value_table * opt = NULL; |
| 26225 | const arm_feature_set arm_any = ARM_ANY; |
| 26226 | int adding_value = -1; |
| 26227 | |
| 26228 | if (!*ext_set_p) |
| 26229 | { |
| 26230 | *ext_set_p = XNEW (arm_feature_set); |
| 26231 | **ext_set_p = arm_arch_none; |
| 26232 | } |
| 26233 | |
| 26234 | while (str != NULL && *str != 0) |
| 26235 | { |
| 26236 | const char *ext; |
| 26237 | size_t len; |
| 26238 | |
| 26239 | if (*str != '+') |
| 26240 | { |
| 26241 | as_bad (_("invalid architectural extension")); |
| 26242 | return FALSE; |
| 26243 | } |
| 26244 | |
| 26245 | str++; |
| 26246 | ext = strchr (str, '+'); |
| 26247 | |
| 26248 | if (ext != NULL) |
| 26249 | len = ext - str; |
| 26250 | else |
| 26251 | len = strlen (str); |
| 26252 | |
| 26253 | if (len >= 2 && strncmp (str, "no", 2) == 0) |
| 26254 | { |
| 26255 | if (adding_value != 0) |
| 26256 | { |
| 26257 | adding_value = 0; |
| 26258 | opt = arm_extensions; |
| 26259 | } |
| 26260 | |
| 26261 | len -= 2; |
| 26262 | str += 2; |
| 26263 | } |
| 26264 | else if (len > 0) |
| 26265 | { |
| 26266 | if (adding_value == -1) |
| 26267 | { |
| 26268 | adding_value = 1; |
| 26269 | opt = arm_extensions; |
| 26270 | } |
| 26271 | else if (adding_value != 1) |
| 26272 | { |
| 26273 | as_bad (_("must specify extensions to add before specifying " |
| 26274 | "those to remove")); |
| 26275 | return FALSE; |
| 26276 | } |
| 26277 | } |
| 26278 | |
| 26279 | if (len == 0) |
| 26280 | { |
| 26281 | as_bad (_("missing architectural extension")); |
| 26282 | return FALSE; |
| 26283 | } |
| 26284 | |
| 26285 | gas_assert (adding_value != -1); |
| 26286 | gas_assert (opt != NULL); |
| 26287 | |
| 26288 | /* Scan over the options table trying to find an exact match. */ |
| 26289 | for (; opt->name != NULL; opt++) |
| 26290 | if (opt->name_len == len && strncmp (opt->name, str, len) == 0) |
| 26291 | { |
| 26292 | int i, nb_allowed_archs = |
| 26293 | sizeof (opt->allowed_archs) / sizeof (opt->allowed_archs[0]); |
| 26294 | /* Check we can apply the extension to this architecture. */ |
| 26295 | for (i = 0; i < nb_allowed_archs; i++) |
| 26296 | { |
| 26297 | /* Empty entry. */ |
| 26298 | if (ARM_FEATURE_EQUAL (opt->allowed_archs[i], arm_any)) |
| 26299 | continue; |
| 26300 | if (ARM_FSET_CPU_SUBSET (opt->allowed_archs[i], *opt_set)) |
| 26301 | break; |
| 26302 | } |
| 26303 | if (i == nb_allowed_archs) |
| 26304 | { |
| 26305 | as_bad (_("extension does not apply to the base architecture")); |
| 26306 | return FALSE; |
| 26307 | } |
| 26308 | |
| 26309 | /* Add or remove the extension. */ |
| 26310 | if (adding_value) |
| 26311 | ARM_MERGE_FEATURE_SETS (**ext_set_p, **ext_set_p, |
| 26312 | opt->merge_value); |
| 26313 | else |
| 26314 | ARM_CLEAR_FEATURE (**ext_set_p, **ext_set_p, opt->clear_value); |
| 26315 | |
| 26316 | /* Allowing Thumb division instructions for ARMv7 in autodetection |
| 26317 | rely on this break so that duplicate extensions (extensions |
| 26318 | with the same name as a previous extension in the list) are not |
| 26319 | considered for command-line parsing. */ |
| 26320 | break; |
| 26321 | } |
| 26322 | |
| 26323 | if (opt->name == NULL) |
| 26324 | { |
| 26325 | /* Did we fail to find an extension because it wasn't specified in |
| 26326 | alphabetical order, or because it does not exist? */ |
| 26327 | |
| 26328 | for (opt = arm_extensions; opt->name != NULL; opt++) |
| 26329 | if (opt->name_len == len && strncmp (opt->name, str, len) == 0) |
| 26330 | break; |
| 26331 | |
| 26332 | if (opt->name == NULL) |
| 26333 | as_bad (_("unknown architectural extension `%s'"), str); |
| 26334 | else |
| 26335 | as_bad (_("architectural extensions must be specified in " |
| 26336 | "alphabetical order")); |
| 26337 | |
| 26338 | return FALSE; |
| 26339 | } |
| 26340 | else |
| 26341 | { |
| 26342 | /* We should skip the extension we've just matched the next time |
| 26343 | round. */ |
| 26344 | opt++; |
| 26345 | } |
| 26346 | |
| 26347 | str = ext; |
| 26348 | }; |
| 26349 | |
| 26350 | return TRUE; |
| 26351 | } |
| 26352 | |
| 26353 | static bfd_boolean |
| 26354 | arm_parse_cpu (const char *str) |
| 26355 | { |
| 26356 | const struct arm_cpu_option_table *opt; |
| 26357 | const char *ext = strchr (str, '+'); |
| 26358 | size_t len; |
| 26359 | |
| 26360 | if (ext != NULL) |
| 26361 | len = ext - str; |
| 26362 | else |
| 26363 | len = strlen (str); |
| 26364 | |
| 26365 | if (len == 0) |
| 26366 | { |
| 26367 | as_bad (_("missing cpu name `%s'"), str); |
| 26368 | return FALSE; |
| 26369 | } |
| 26370 | |
| 26371 | for (opt = arm_cpus; opt->name != NULL; opt++) |
| 26372 | if (opt->name_len == len && strncmp (opt->name, str, len) == 0) |
| 26373 | { |
| 26374 | mcpu_cpu_opt = &opt->value; |
| 26375 | if (!dyn_mcpu_ext_opt) |
| 26376 | dyn_mcpu_ext_opt = XNEW (arm_feature_set); |
| 26377 | *dyn_mcpu_ext_opt = opt->ext; |
| 26378 | mcpu_fpu_opt = &opt->default_fpu; |
| 26379 | if (opt->canonical_name) |
| 26380 | { |
| 26381 | gas_assert (sizeof selected_cpu_name > strlen (opt->canonical_name)); |
| 26382 | strcpy (selected_cpu_name, opt->canonical_name); |
| 26383 | } |
| 26384 | else |
| 26385 | { |
| 26386 | size_t i; |
| 26387 | |
| 26388 | if (len >= sizeof selected_cpu_name) |
| 26389 | len = (sizeof selected_cpu_name) - 1; |
| 26390 | |
| 26391 | for (i = 0; i < len; i++) |
| 26392 | selected_cpu_name[i] = TOUPPER (opt->name[i]); |
| 26393 | selected_cpu_name[i] = 0; |
| 26394 | } |
| 26395 | |
| 26396 | if (ext != NULL) |
| 26397 | return arm_parse_extension (ext, mcpu_cpu_opt, &dyn_mcpu_ext_opt); |
| 26398 | |
| 26399 | return TRUE; |
| 26400 | } |
| 26401 | |
| 26402 | as_bad (_("unknown cpu `%s'"), str); |
| 26403 | return FALSE; |
| 26404 | } |
| 26405 | |
| 26406 | static bfd_boolean |
| 26407 | arm_parse_arch (const char *str) |
| 26408 | { |
| 26409 | const struct arm_arch_option_table *opt; |
| 26410 | const char *ext = strchr (str, '+'); |
| 26411 | size_t len; |
| 26412 | |
| 26413 | if (ext != NULL) |
| 26414 | len = ext - str; |
| 26415 | else |
| 26416 | len = strlen (str); |
| 26417 | |
| 26418 | if (len == 0) |
| 26419 | { |
| 26420 | as_bad (_("missing architecture name `%s'"), str); |
| 26421 | return FALSE; |
| 26422 | } |
| 26423 | |
| 26424 | for (opt = arm_archs; opt->name != NULL; opt++) |
| 26425 | if (opt->name_len == len && strncmp (opt->name, str, len) == 0) |
| 26426 | { |
| 26427 | march_cpu_opt = &opt->value; |
| 26428 | march_fpu_opt = &opt->default_fpu; |
| 26429 | strcpy (selected_cpu_name, opt->name); |
| 26430 | |
| 26431 | if (ext != NULL) |
| 26432 | return arm_parse_extension (ext, march_cpu_opt, &dyn_march_ext_opt); |
| 26433 | |
| 26434 | return TRUE; |
| 26435 | } |
| 26436 | |
| 26437 | as_bad (_("unknown architecture `%s'\n"), str); |
| 26438 | return FALSE; |
| 26439 | } |
| 26440 | |
| 26441 | static bfd_boolean |
| 26442 | arm_parse_fpu (const char * str) |
| 26443 | { |
| 26444 | const struct arm_option_fpu_value_table * opt; |
| 26445 | |
| 26446 | for (opt = arm_fpus; opt->name != NULL; opt++) |
| 26447 | if (streq (opt->name, str)) |
| 26448 | { |
| 26449 | mfpu_opt = &opt->value; |
| 26450 | return TRUE; |
| 26451 | } |
| 26452 | |
| 26453 | as_bad (_("unknown floating point format `%s'\n"), str); |
| 26454 | return FALSE; |
| 26455 | } |
| 26456 | |
| 26457 | static bfd_boolean |
| 26458 | arm_parse_float_abi (const char * str) |
| 26459 | { |
| 26460 | const struct arm_option_value_table * opt; |
| 26461 | |
| 26462 | for (opt = arm_float_abis; opt->name != NULL; opt++) |
| 26463 | if (streq (opt->name, str)) |
| 26464 | { |
| 26465 | mfloat_abi_opt = opt->value; |
| 26466 | return TRUE; |
| 26467 | } |
| 26468 | |
| 26469 | as_bad (_("unknown floating point abi `%s'\n"), str); |
| 26470 | return FALSE; |
| 26471 | } |
| 26472 | |
| 26473 | #ifdef OBJ_ELF |
| 26474 | static bfd_boolean |
| 26475 | arm_parse_eabi (const char * str) |
| 26476 | { |
| 26477 | const struct arm_option_value_table *opt; |
| 26478 | |
| 26479 | for (opt = arm_eabis; opt->name != NULL; opt++) |
| 26480 | if (streq (opt->name, str)) |
| 26481 | { |
| 26482 | meabi_flags = opt->value; |
| 26483 | return TRUE; |
| 26484 | } |
| 26485 | as_bad (_("unknown EABI `%s'\n"), str); |
| 26486 | return FALSE; |
| 26487 | } |
| 26488 | #endif |
| 26489 | |
| 26490 | static bfd_boolean |
| 26491 | arm_parse_it_mode (const char * str) |
| 26492 | { |
| 26493 | bfd_boolean ret = TRUE; |
| 26494 | |
| 26495 | if (streq ("arm", str)) |
| 26496 | implicit_it_mode = IMPLICIT_IT_MODE_ARM; |
| 26497 | else if (streq ("thumb", str)) |
| 26498 | implicit_it_mode = IMPLICIT_IT_MODE_THUMB; |
| 26499 | else if (streq ("always", str)) |
| 26500 | implicit_it_mode = IMPLICIT_IT_MODE_ALWAYS; |
| 26501 | else if (streq ("never", str)) |
| 26502 | implicit_it_mode = IMPLICIT_IT_MODE_NEVER; |
| 26503 | else |
| 26504 | { |
| 26505 | as_bad (_("unknown implicit IT mode `%s', should be "\ |
| 26506 | "arm, thumb, always, or never."), str); |
| 26507 | ret = FALSE; |
| 26508 | } |
| 26509 | |
| 26510 | return ret; |
| 26511 | } |
| 26512 | |
| 26513 | static bfd_boolean |
| 26514 | arm_ccs_mode (const char * unused ATTRIBUTE_UNUSED) |
| 26515 | { |
| 26516 | codecomposer_syntax = TRUE; |
| 26517 | arm_comment_chars[0] = ';'; |
| 26518 | arm_line_separator_chars[0] = 0; |
| 26519 | return TRUE; |
| 26520 | } |
| 26521 | |
| 26522 | struct arm_long_option_table arm_long_opts[] = |
| 26523 | { |
| 26524 | {"mcpu=", N_("<cpu name>\t assemble for CPU <cpu name>"), |
| 26525 | arm_parse_cpu, NULL}, |
| 26526 | {"march=", N_("<arch name>\t assemble for architecture <arch name>"), |
| 26527 | arm_parse_arch, NULL}, |
| 26528 | {"mfpu=", N_("<fpu name>\t assemble for FPU architecture <fpu name>"), |
| 26529 | arm_parse_fpu, NULL}, |
| 26530 | {"mfloat-abi=", N_("<abi>\t assemble for floating point ABI <abi>"), |
| 26531 | arm_parse_float_abi, NULL}, |
| 26532 | #ifdef OBJ_ELF |
| 26533 | {"meabi=", N_("<ver>\t\t assemble for eabi version <ver>"), |
| 26534 | arm_parse_eabi, NULL}, |
| 26535 | #endif |
| 26536 | {"mimplicit-it=", N_("<mode>\t controls implicit insertion of IT instructions"), |
| 26537 | arm_parse_it_mode, NULL}, |
| 26538 | {"mccs", N_("\t\t\t TI CodeComposer Studio syntax compatibility mode"), |
| 26539 | arm_ccs_mode, NULL}, |
| 26540 | {NULL, NULL, 0, NULL} |
| 26541 | }; |
| 26542 | |
| 26543 | int |
| 26544 | md_parse_option (int c, const char * arg) |
| 26545 | { |
| 26546 | struct arm_option_table *opt; |
| 26547 | const struct arm_legacy_option_table *fopt; |
| 26548 | struct arm_long_option_table *lopt; |
| 26549 | |
| 26550 | switch (c) |
| 26551 | { |
| 26552 | #ifdef OPTION_EB |
| 26553 | case OPTION_EB: |
| 26554 | target_big_endian = 1; |
| 26555 | break; |
| 26556 | #endif |
| 26557 | |
| 26558 | #ifdef OPTION_EL |
| 26559 | case OPTION_EL: |
| 26560 | target_big_endian = 0; |
| 26561 | break; |
| 26562 | #endif |
| 26563 | |
| 26564 | case OPTION_FIX_V4BX: |
| 26565 | fix_v4bx = TRUE; |
| 26566 | break; |
| 26567 | |
| 26568 | case 'a': |
| 26569 | /* Listing option. Just ignore these, we don't support additional |
| 26570 | ones. */ |
| 26571 | return 0; |
| 26572 | |
| 26573 | default: |
| 26574 | for (opt = arm_opts; opt->option != NULL; opt++) |
| 26575 | { |
| 26576 | if (c == opt->option[0] |
| 26577 | && ((arg == NULL && opt->option[1] == 0) |
| 26578 | || streq (arg, opt->option + 1))) |
| 26579 | { |
| 26580 | /* If the option is deprecated, tell the user. */ |
| 26581 | if (warn_on_deprecated && opt->deprecated != NULL) |
| 26582 | as_tsktsk (_("option `-%c%s' is deprecated: %s"), c, |
| 26583 | arg ? arg : "", _(opt->deprecated)); |
| 26584 | |
| 26585 | if (opt->var != NULL) |
| 26586 | *opt->var = opt->value; |
| 26587 | |
| 26588 | return 1; |
| 26589 | } |
| 26590 | } |
| 26591 | |
| 26592 | for (fopt = arm_legacy_opts; fopt->option != NULL; fopt++) |
| 26593 | { |
| 26594 | if (c == fopt->option[0] |
| 26595 | && ((arg == NULL && fopt->option[1] == 0) |
| 26596 | || streq (arg, fopt->option + 1))) |
| 26597 | { |
| 26598 | /* If the option is deprecated, tell the user. */ |
| 26599 | if (warn_on_deprecated && fopt->deprecated != NULL) |
| 26600 | as_tsktsk (_("option `-%c%s' is deprecated: %s"), c, |
| 26601 | arg ? arg : "", _(fopt->deprecated)); |
| 26602 | |
| 26603 | if (fopt->var != NULL) |
| 26604 | *fopt->var = &fopt->value; |
| 26605 | |
| 26606 | return 1; |
| 26607 | } |
| 26608 | } |
| 26609 | |
| 26610 | for (lopt = arm_long_opts; lopt->option != NULL; lopt++) |
| 26611 | { |
| 26612 | /* These options are expected to have an argument. */ |
| 26613 | if (c == lopt->option[0] |
| 26614 | && arg != NULL |
| 26615 | && strncmp (arg, lopt->option + 1, |
| 26616 | strlen (lopt->option + 1)) == 0) |
| 26617 | { |
| 26618 | /* If the option is deprecated, tell the user. */ |
| 26619 | if (warn_on_deprecated && lopt->deprecated != NULL) |
| 26620 | as_tsktsk (_("option `-%c%s' is deprecated: %s"), c, arg, |
| 26621 | _(lopt->deprecated)); |
| 26622 | |
| 26623 | /* Call the sup-option parser. */ |
| 26624 | return lopt->func (arg + strlen (lopt->option) - 1); |
| 26625 | } |
| 26626 | } |
| 26627 | |
| 26628 | return 0; |
| 26629 | } |
| 26630 | |
| 26631 | return 1; |
| 26632 | } |
| 26633 | |
| 26634 | void |
| 26635 | md_show_usage (FILE * fp) |
| 26636 | { |
| 26637 | struct arm_option_table *opt; |
| 26638 | struct arm_long_option_table *lopt; |
| 26639 | |
| 26640 | fprintf (fp, _(" ARM-specific assembler options:\n")); |
| 26641 | |
| 26642 | for (opt = arm_opts; opt->option != NULL; opt++) |
| 26643 | if (opt->help != NULL) |
| 26644 | fprintf (fp, " -%-23s%s\n", opt->option, _(opt->help)); |
| 26645 | |
| 26646 | for (lopt = arm_long_opts; lopt->option != NULL; lopt++) |
| 26647 | if (lopt->help != NULL) |
| 26648 | fprintf (fp, " -%s%s\n", lopt->option, _(lopt->help)); |
| 26649 | |
| 26650 | #ifdef OPTION_EB |
| 26651 | fprintf (fp, _("\ |
| 26652 | -EB assemble code for a big-endian cpu\n")); |
| 26653 | #endif |
| 26654 | |
| 26655 | #ifdef OPTION_EL |
| 26656 | fprintf (fp, _("\ |
| 26657 | -EL assemble code for a little-endian cpu\n")); |
| 26658 | #endif |
| 26659 | |
| 26660 | fprintf (fp, _("\ |
| 26661 | --fix-v4bx Allow BX in ARMv4 code\n")); |
| 26662 | } |
| 26663 | |
| 26664 | |
| 26665 | #ifdef OBJ_ELF |
| 26666 | typedef struct |
| 26667 | { |
| 26668 | int val; |
| 26669 | arm_feature_set flags; |
| 26670 | } cpu_arch_ver_table; |
| 26671 | |
| 26672 | /* Mapping from CPU features to EABI CPU arch values. Table must be sorted |
| 26673 | chronologically for architectures, with an exception for ARMv6-M and |
| 26674 | ARMv6S-M due to legacy reasons. No new architecture should have a |
| 26675 | special case. This allows for build attribute selection results to be |
| 26676 | stable when new architectures are added. */ |
| 26677 | static const cpu_arch_ver_table cpu_arch_ver[] = |
| 26678 | { |
| 26679 | {0, ARM_ARCH_V1}, |
| 26680 | {0, ARM_ARCH_V2}, |
| 26681 | {0, ARM_ARCH_V2S}, |
| 26682 | {0, ARM_ARCH_V3}, |
| 26683 | {0, ARM_ARCH_V3M}, |
| 26684 | {1, ARM_ARCH_V4xM}, |
| 26685 | {1, ARM_ARCH_V4}, |
| 26686 | {2, ARM_ARCH_V4TxM}, |
| 26687 | {2, ARM_ARCH_V4T}, |
| 26688 | {3, ARM_ARCH_V5xM}, |
| 26689 | {3, ARM_ARCH_V5}, |
| 26690 | {3, ARM_ARCH_V5TxM}, |
| 26691 | {3, ARM_ARCH_V5T}, |
| 26692 | {4, ARM_ARCH_V5TExP}, |
| 26693 | {4, ARM_ARCH_V5TE}, |
| 26694 | {5, ARM_ARCH_V5TEJ}, |
| 26695 | {6, ARM_ARCH_V6}, |
| 26696 | {7, ARM_ARCH_V6Z}, |
| 26697 | {7, ARM_ARCH_V6KZ}, |
| 26698 | {9, ARM_ARCH_V6K}, |
| 26699 | {8, ARM_ARCH_V6T2}, |
| 26700 | {8, ARM_ARCH_V6KT2}, |
| 26701 | {8, ARM_ARCH_V6ZT2}, |
| 26702 | {8, ARM_ARCH_V6KZT2}, |
| 26703 | |
| 26704 | /* When assembling a file with only ARMv6-M or ARMv6S-M instruction, GNU as |
| 26705 | always selected build attributes to match those of ARMv6-M |
| 26706 | (resp. ARMv6S-M). However, due to these architectures being a strict |
| 26707 | subset of ARMv7-M in terms of instructions available, ARMv7-M attributes |
| 26708 | would be selected when fully respecting chronology of architectures. |
| 26709 | It is thus necessary to make a special case of ARMv6-M and ARMv6S-M and |
| 26710 | move them before ARMv7 architectures. */ |
| 26711 | {11, ARM_ARCH_V6M}, |
| 26712 | {12, ARM_ARCH_V6SM}, |
| 26713 | |
| 26714 | {10, ARM_ARCH_V7}, |
| 26715 | {10, ARM_ARCH_V7A}, |
| 26716 | {10, ARM_ARCH_V7R}, |
| 26717 | {10, ARM_ARCH_V7M}, |
| 26718 | {10, ARM_ARCH_V7VE}, |
| 26719 | {13, ARM_ARCH_V7EM}, |
| 26720 | {14, ARM_ARCH_V8A}, |
| 26721 | {14, ARM_ARCH_V8_1A}, |
| 26722 | {14, ARM_ARCH_V8_2A}, |
| 26723 | {14, ARM_ARCH_V8_3A}, |
| 26724 | {16, ARM_ARCH_V8M_BASE}, |
| 26725 | {17, ARM_ARCH_V8M_MAIN}, |
| 26726 | {15, ARM_ARCH_V8R}, |
| 26727 | {-1, ARM_ARCH_NONE} |
| 26728 | }; |
| 26729 | |
| 26730 | /* Set an attribute if it has not already been set by the user. */ |
| 26731 | static void |
| 26732 | aeabi_set_attribute_int (int tag, int value) |
| 26733 | { |
| 26734 | if (tag < 1 |
| 26735 | || tag >= NUM_KNOWN_OBJ_ATTRIBUTES |
| 26736 | || !attributes_set_explicitly[tag]) |
| 26737 | bfd_elf_add_proc_attr_int (stdoutput, tag, value); |
| 26738 | } |
| 26739 | |
| 26740 | static void |
| 26741 | aeabi_set_attribute_string (int tag, const char *value) |
| 26742 | { |
| 26743 | if (tag < 1 |
| 26744 | || tag >= NUM_KNOWN_OBJ_ATTRIBUTES |
| 26745 | || !attributes_set_explicitly[tag]) |
| 26746 | bfd_elf_add_proc_attr_string (stdoutput, tag, value); |
| 26747 | } |
| 26748 | |
| 26749 | /* Return whether features in the *NEEDED feature set are available via |
| 26750 | extensions for the architecture whose feature set is *ARCH_FSET. */ |
| 26751 | static bfd_boolean |
| 26752 | have_ext_for_needed_feat_p (const arm_feature_set *arch_fset, |
| 26753 | const arm_feature_set *needed) |
| 26754 | { |
| 26755 | int i, nb_allowed_archs; |
| 26756 | arm_feature_set ext_fset; |
| 26757 | const struct arm_option_extension_value_table *opt; |
| 26758 | |
| 26759 | ext_fset = arm_arch_none; |
| 26760 | for (opt = arm_extensions; opt->name != NULL; opt++) |
| 26761 | { |
| 26762 | /* Extension does not provide any feature we need. */ |
| 26763 | if (!ARM_CPU_HAS_FEATURE (*needed, opt->merge_value)) |
| 26764 | continue; |
| 26765 | |
| 26766 | nb_allowed_archs = |
| 26767 | sizeof (opt->allowed_archs) / sizeof (opt->allowed_archs[0]); |
| 26768 | for (i = 0; i < nb_allowed_archs; i++) |
| 26769 | { |
| 26770 | /* Empty entry. */ |
| 26771 | if (ARM_FEATURE_EQUAL (opt->allowed_archs[i], arm_arch_any)) |
| 26772 | break; |
| 26773 | |
| 26774 | /* Extension is available, add it. */ |
| 26775 | if (ARM_FSET_CPU_SUBSET (opt->allowed_archs[i], *arch_fset)) |
| 26776 | ARM_MERGE_FEATURE_SETS (ext_fset, ext_fset, opt->merge_value); |
| 26777 | } |
| 26778 | } |
| 26779 | |
| 26780 | /* Can we enable all features in *needed? */ |
| 26781 | return ARM_FSET_CPU_SUBSET (*needed, ext_fset); |
| 26782 | } |
| 26783 | |
| 26784 | /* Select value for Tag_CPU_arch and Tag_CPU_arch_profile build attributes for |
| 26785 | a given architecture feature set *ARCH_EXT_FSET including extension feature |
| 26786 | set *EXT_FSET. Selection logic used depend on EXACT_MATCH: |
| 26787 | - if true, check for an exact match of the architecture modulo extensions; |
| 26788 | - otherwise, select build attribute value of the first superset |
| 26789 | architecture released so that results remains stable when new architectures |
| 26790 | are added. |
| 26791 | For -march/-mcpu=all the build attribute value of the most featureful |
| 26792 | architecture is returned. Tag_CPU_arch_profile result is returned in |
| 26793 | PROFILE. */ |
| 26794 | static int |
| 26795 | get_aeabi_cpu_arch_from_fset (const arm_feature_set *arch_ext_fset, |
| 26796 | const arm_feature_set *ext_fset, |
| 26797 | char *profile, int exact_match) |
| 26798 | { |
| 26799 | arm_feature_set arch_fset; |
| 26800 | const cpu_arch_ver_table *p_ver, *p_ver_ret = NULL; |
| 26801 | |
| 26802 | /* Select most featureful architecture with all its extensions if building |
| 26803 | for -march=all as the feature sets used to set build attributes. */ |
| 26804 | if (ARM_FEATURE_EQUAL (*arch_ext_fset, arm_arch_any)) |
| 26805 | { |
| 26806 | /* Force revisiting of decision for each new architecture. */ |
| 26807 | gas_assert (MAX_TAG_CPU_ARCH <= TAG_CPU_ARCH_V8M_MAIN); |
| 26808 | *profile = 'A'; |
| 26809 | return TAG_CPU_ARCH_V8; |
| 26810 | } |
| 26811 | |
| 26812 | ARM_CLEAR_FEATURE (arch_fset, *arch_ext_fset, *ext_fset); |
| 26813 | |
| 26814 | for (p_ver = cpu_arch_ver; p_ver->val != -1; p_ver++) |
| 26815 | { |
| 26816 | arm_feature_set known_arch_fset; |
| 26817 | |
| 26818 | ARM_CLEAR_FEATURE (known_arch_fset, p_ver->flags, fpu_any); |
| 26819 | if (exact_match) |
| 26820 | { |
| 26821 | /* Base architecture match user-specified architecture and |
| 26822 | extensions, eg. ARMv6S-M matching -march=armv6-m+os. */ |
| 26823 | if (ARM_FEATURE_EQUAL (*arch_ext_fset, known_arch_fset)) |
| 26824 | { |
| 26825 | p_ver_ret = p_ver; |
| 26826 | goto found; |
| 26827 | } |
| 26828 | /* Base architecture match user-specified architecture only |
| 26829 | (eg. ARMv6-M in the same case as above). Record it in case we |
| 26830 | find a match with above condition. */ |
| 26831 | else if (p_ver_ret == NULL |
| 26832 | && ARM_FEATURE_EQUAL (arch_fset, known_arch_fset)) |
| 26833 | p_ver_ret = p_ver; |
| 26834 | } |
| 26835 | else |
| 26836 | { |
| 26837 | |
| 26838 | /* Architecture has all features wanted. */ |
| 26839 | if (ARM_FSET_CPU_SUBSET (arch_fset, known_arch_fset)) |
| 26840 | { |
| 26841 | arm_feature_set added_fset; |
| 26842 | |
| 26843 | /* Compute features added by this architecture over the one |
| 26844 | recorded in p_ver_ret. */ |
| 26845 | if (p_ver_ret != NULL) |
| 26846 | ARM_CLEAR_FEATURE (added_fset, known_arch_fset, |
| 26847 | p_ver_ret->flags); |
| 26848 | /* First architecture that match incl. with extensions, or the |
| 26849 | only difference in features over the recorded match is |
| 26850 | features that were optional and are now mandatory. */ |
| 26851 | if (p_ver_ret == NULL |
| 26852 | || ARM_FSET_CPU_SUBSET (added_fset, arch_fset)) |
| 26853 | { |
| 26854 | p_ver_ret = p_ver; |
| 26855 | goto found; |
| 26856 | } |
| 26857 | } |
| 26858 | else if (p_ver_ret == NULL) |
| 26859 | { |
| 26860 | arm_feature_set needed_ext_fset; |
| 26861 | |
| 26862 | ARM_CLEAR_FEATURE (needed_ext_fset, arch_fset, known_arch_fset); |
| 26863 | |
| 26864 | /* Architecture has all features needed when using some |
| 26865 | extensions. Record it and continue searching in case there |
| 26866 | exist an architecture providing all needed features without |
| 26867 | the need for extensions (eg. ARMv6S-M Vs ARMv6-M with |
| 26868 | OS extension). */ |
| 26869 | if (have_ext_for_needed_feat_p (&known_arch_fset, |
| 26870 | &needed_ext_fset)) |
| 26871 | p_ver_ret = p_ver; |
| 26872 | } |
| 26873 | } |
| 26874 | } |
| 26875 | |
| 26876 | if (p_ver_ret == NULL) |
| 26877 | return -1; |
| 26878 | |
| 26879 | found: |
| 26880 | /* Tag_CPU_arch_profile. */ |
| 26881 | if (ARM_CPU_HAS_FEATURE (p_ver_ret->flags, arm_ext_v7a) |
| 26882 | || ARM_CPU_HAS_FEATURE (p_ver_ret->flags, arm_ext_v8) |
| 26883 | || (ARM_CPU_HAS_FEATURE (p_ver_ret->flags, arm_ext_atomics) |
| 26884 | && !ARM_CPU_HAS_FEATURE (p_ver_ret->flags, arm_ext_v8m_m_only))) |
| 26885 | *profile = 'A'; |
| 26886 | else if (ARM_CPU_HAS_FEATURE (p_ver_ret->flags, arm_ext_v7r)) |
| 26887 | *profile = 'R'; |
| 26888 | else if (ARM_CPU_HAS_FEATURE (p_ver_ret->flags, arm_ext_m)) |
| 26889 | *profile = 'M'; |
| 26890 | else |
| 26891 | *profile = '\0'; |
| 26892 | return p_ver_ret->val; |
| 26893 | } |
| 26894 | |
| 26895 | /* Set the public EABI object attributes. */ |
| 26896 | static void |
| 26897 | aeabi_set_public_attributes (void) |
| 26898 | { |
| 26899 | char profile; |
| 26900 | int arch = -1; |
| 26901 | int virt_sec = 0; |
| 26902 | int fp16_optional = 0; |
| 26903 | int skip_exact_match = 0; |
| 26904 | arm_feature_set flags, flags_arch, flags_ext; |
| 26905 | |
| 26906 | /* Autodetection mode, choose the architecture based the instructions |
| 26907 | actually used. */ |
| 26908 | if (no_cpu_selected ()) |
| 26909 | { |
| 26910 | ARM_MERGE_FEATURE_SETS (flags, arm_arch_used, thumb_arch_used); |
| 26911 | |
| 26912 | if (ARM_CPU_HAS_FEATURE (arm_arch_used, arm_arch_any)) |
| 26913 | ARM_MERGE_FEATURE_SETS (flags, flags, arm_ext_v1); |
| 26914 | |
| 26915 | if (ARM_CPU_HAS_FEATURE (thumb_arch_used, arm_arch_any)) |
| 26916 | ARM_MERGE_FEATURE_SETS (flags, flags, arm_ext_v4t); |
| 26917 | |
| 26918 | /* Code run during relaxation relies on selected_cpu being set. */ |
| 26919 | selected_cpu = flags; |
| 26920 | } |
| 26921 | /* Otherwise, choose the architecture based on the capabilities of the |
| 26922 | requested cpu. */ |
| 26923 | else |
| 26924 | flags = selected_cpu; |
| 26925 | ARM_MERGE_FEATURE_SETS (flags, flags, *mfpu_opt); |
| 26926 | |
| 26927 | /* Allow the user to override the reported architecture. */ |
| 26928 | if (object_arch) |
| 26929 | { |
| 26930 | ARM_CLEAR_FEATURE (flags_arch, *object_arch, fpu_any); |
| 26931 | flags_ext = arm_arch_none; |
| 26932 | } |
| 26933 | else |
| 26934 | { |
| 26935 | ARM_CLEAR_FEATURE (flags_arch, flags, fpu_any); |
| 26936 | flags_ext = dyn_mcpu_ext_opt ? *dyn_mcpu_ext_opt : arm_arch_none; |
| 26937 | skip_exact_match = ARM_FEATURE_EQUAL (selected_cpu, arm_arch_any); |
| 26938 | } |
| 26939 | |
| 26940 | /* When this function is run again after relaxation has happened there is no |
| 26941 | way to determine whether an architecture or CPU was specified by the user: |
| 26942 | - selected_cpu is set above for relaxation to work; |
| 26943 | - march_cpu_opt is not set if only -mcpu or .cpu is used; |
| 26944 | - mcpu_cpu_opt is set to arm_arch_any for autodetection. |
| 26945 | Therefore, if not in -march=all case we first try an exact match and fall |
| 26946 | back to autodetection. */ |
| 26947 | if (!skip_exact_match) |
| 26948 | arch = get_aeabi_cpu_arch_from_fset (&flags_arch, &flags_ext, &profile, 1); |
| 26949 | if (arch == -1) |
| 26950 | arch = get_aeabi_cpu_arch_from_fset (&flags_arch, &flags_ext, &profile, 0); |
| 26951 | if (arch == -1) |
| 26952 | as_bad (_("no architecture contains all the instructions used\n")); |
| 26953 | |
| 26954 | /* Tag_CPU_name. */ |
| 26955 | if (selected_cpu_name[0]) |
| 26956 | { |
| 26957 | char *q; |
| 26958 | |
| 26959 | q = selected_cpu_name; |
| 26960 | if (strncmp (q, "armv", 4) == 0) |
| 26961 | { |
| 26962 | int i; |
| 26963 | |
| 26964 | q += 4; |
| 26965 | for (i = 0; q[i]; i++) |
| 26966 | q[i] = TOUPPER (q[i]); |
| 26967 | } |
| 26968 | aeabi_set_attribute_string (Tag_CPU_name, q); |
| 26969 | } |
| 26970 | |
| 26971 | /* Tag_CPU_arch. */ |
| 26972 | aeabi_set_attribute_int (Tag_CPU_arch, arch); |
| 26973 | |
| 26974 | /* Tag_CPU_arch_profile. */ |
| 26975 | if (profile != '\0') |
| 26976 | aeabi_set_attribute_int (Tag_CPU_arch_profile, profile); |
| 26977 | |
| 26978 | /* Tag_DSP_extension. */ |
| 26979 | if (dyn_mcpu_ext_opt && ARM_CPU_HAS_FEATURE (*dyn_mcpu_ext_opt, arm_ext_dsp)) |
| 26980 | aeabi_set_attribute_int (Tag_DSP_extension, 1); |
| 26981 | |
| 26982 | ARM_CLEAR_FEATURE (flags_arch, flags, fpu_any); |
| 26983 | /* Tag_ARM_ISA_use. */ |
| 26984 | if (ARM_CPU_HAS_FEATURE (flags, arm_ext_v1) |
| 26985 | || ARM_FEATURE_ZERO (flags_arch)) |
| 26986 | aeabi_set_attribute_int (Tag_ARM_ISA_use, 1); |
| 26987 | |
| 26988 | /* Tag_THUMB_ISA_use. */ |
| 26989 | if (ARM_CPU_HAS_FEATURE (flags, arm_ext_v4t) |
| 26990 | || ARM_FEATURE_ZERO (flags_arch)) |
| 26991 | { |
| 26992 | int thumb_isa_use; |
| 26993 | |
| 26994 | if (!ARM_CPU_HAS_FEATURE (flags, arm_ext_v8) |
| 26995 | && ARM_CPU_HAS_FEATURE (flags, arm_ext_v8m_m_only)) |
| 26996 | thumb_isa_use = 3; |
| 26997 | else if (ARM_CPU_HAS_FEATURE (flags, arm_arch_t2)) |
| 26998 | thumb_isa_use = 2; |
| 26999 | else |
| 27000 | thumb_isa_use = 1; |
| 27001 | aeabi_set_attribute_int (Tag_THUMB_ISA_use, thumb_isa_use); |
| 27002 | } |
| 27003 | |
| 27004 | /* Tag_VFP_arch. */ |
| 27005 | if (ARM_CPU_HAS_FEATURE (flags, fpu_vfp_ext_armv8xd)) |
| 27006 | aeabi_set_attribute_int (Tag_VFP_arch, |
| 27007 | ARM_CPU_HAS_FEATURE (flags, fpu_vfp_ext_d32) |
| 27008 | ? 7 : 8); |
| 27009 | else if (ARM_CPU_HAS_FEATURE (flags, fpu_vfp_ext_fma)) |
| 27010 | aeabi_set_attribute_int (Tag_VFP_arch, |
| 27011 | ARM_CPU_HAS_FEATURE (flags, fpu_vfp_ext_d32) |
| 27012 | ? 5 : 6); |
| 27013 | else if (ARM_CPU_HAS_FEATURE (flags, fpu_vfp_ext_d32)) |
| 27014 | { |
| 27015 | fp16_optional = 1; |
| 27016 | aeabi_set_attribute_int (Tag_VFP_arch, 3); |
| 27017 | } |
| 27018 | else if (ARM_CPU_HAS_FEATURE (flags, fpu_vfp_ext_v3xd)) |
| 27019 | { |
| 27020 | aeabi_set_attribute_int (Tag_VFP_arch, 4); |
| 27021 | fp16_optional = 1; |
| 27022 | } |
| 27023 | else if (ARM_CPU_HAS_FEATURE (flags, fpu_vfp_ext_v2)) |
| 27024 | aeabi_set_attribute_int (Tag_VFP_arch, 2); |
| 27025 | else if (ARM_CPU_HAS_FEATURE (flags, fpu_vfp_ext_v1) |
| 27026 | || ARM_CPU_HAS_FEATURE (flags, fpu_vfp_ext_v1xd)) |
| 27027 | aeabi_set_attribute_int (Tag_VFP_arch, 1); |
| 27028 | |
| 27029 | /* Tag_ABI_HardFP_use. */ |
| 27030 | if (ARM_CPU_HAS_FEATURE (flags, fpu_vfp_ext_v1xd) |
| 27031 | && !ARM_CPU_HAS_FEATURE (flags, fpu_vfp_ext_v1)) |
| 27032 | aeabi_set_attribute_int (Tag_ABI_HardFP_use, 1); |
| 27033 | |
| 27034 | /* Tag_WMMX_arch. */ |
| 27035 | if (ARM_CPU_HAS_FEATURE (flags, arm_cext_iwmmxt2)) |
| 27036 | aeabi_set_attribute_int (Tag_WMMX_arch, 2); |
| 27037 | else if (ARM_CPU_HAS_FEATURE (flags, arm_cext_iwmmxt)) |
| 27038 | aeabi_set_attribute_int (Tag_WMMX_arch, 1); |
| 27039 | |
| 27040 | /* Tag_Advanced_SIMD_arch (formerly Tag_NEON_arch). */ |
| 27041 | if (ARM_CPU_HAS_FEATURE (flags, fpu_neon_ext_v8_1)) |
| 27042 | aeabi_set_attribute_int (Tag_Advanced_SIMD_arch, 4); |
| 27043 | else if (ARM_CPU_HAS_FEATURE (flags, fpu_neon_ext_armv8)) |
| 27044 | aeabi_set_attribute_int (Tag_Advanced_SIMD_arch, 3); |
| 27045 | else if (ARM_CPU_HAS_FEATURE (flags, fpu_neon_ext_v1)) |
| 27046 | { |
| 27047 | if (ARM_CPU_HAS_FEATURE (flags, fpu_neon_ext_fma)) |
| 27048 | { |
| 27049 | aeabi_set_attribute_int (Tag_Advanced_SIMD_arch, 2); |
| 27050 | } |
| 27051 | else |
| 27052 | { |
| 27053 | aeabi_set_attribute_int (Tag_Advanced_SIMD_arch, 1); |
| 27054 | fp16_optional = 1; |
| 27055 | } |
| 27056 | } |
| 27057 | |
| 27058 | /* Tag_VFP_HP_extension (formerly Tag_NEON_FP16_arch). */ |
| 27059 | if (ARM_CPU_HAS_FEATURE (flags, fpu_vfp_fp16) && fp16_optional) |
| 27060 | aeabi_set_attribute_int (Tag_VFP_HP_extension, 1); |
| 27061 | |
| 27062 | /* Tag_DIV_use. |
| 27063 | |
| 27064 | We set Tag_DIV_use to two when integer divide instructions have been used |
| 27065 | in ARM state, or when Thumb integer divide instructions have been used, |
| 27066 | but we have no architecture profile set, nor have we any ARM instructions. |
| 27067 | |
| 27068 | For ARMv8-A and ARMv8-M we set the tag to 0 as integer divide is implied |
| 27069 | by the base architecture. |
| 27070 | |
| 27071 | For new architectures we will have to check these tests. */ |
| 27072 | gas_assert (arch <= TAG_CPU_ARCH_V8M_MAIN); |
| 27073 | if (ARM_CPU_HAS_FEATURE (flags, arm_ext_v8) |
| 27074 | || ARM_CPU_HAS_FEATURE (flags, arm_ext_v8m)) |
| 27075 | aeabi_set_attribute_int (Tag_DIV_use, 0); |
| 27076 | else if (ARM_CPU_HAS_FEATURE (flags, arm_ext_adiv) |
| 27077 | || (profile == '\0' |
| 27078 | && ARM_CPU_HAS_FEATURE (flags, arm_ext_div) |
| 27079 | && !ARM_CPU_HAS_FEATURE (arm_arch_used, arm_arch_any))) |
| 27080 | aeabi_set_attribute_int (Tag_DIV_use, 2); |
| 27081 | |
| 27082 | /* Tag_MP_extension_use. */ |
| 27083 | if (ARM_CPU_HAS_FEATURE (flags, arm_ext_mp)) |
| 27084 | aeabi_set_attribute_int (Tag_MPextension_use, 1); |
| 27085 | |
| 27086 | /* Tag Virtualization_use. */ |
| 27087 | if (ARM_CPU_HAS_FEATURE (flags, arm_ext_sec)) |
| 27088 | virt_sec |= 1; |
| 27089 | if (ARM_CPU_HAS_FEATURE (flags, arm_ext_virt)) |
| 27090 | virt_sec |= 2; |
| 27091 | if (virt_sec != 0) |
| 27092 | aeabi_set_attribute_int (Tag_Virtualization_use, virt_sec); |
| 27093 | } |
| 27094 | |
| 27095 | /* Post relaxation hook. Recompute ARM attributes now that relaxation is |
| 27096 | finished and free extension feature bits which will not be used anymore. */ |
| 27097 | void |
| 27098 | arm_md_post_relax (void) |
| 27099 | { |
| 27100 | aeabi_set_public_attributes (); |
| 27101 | XDELETE (dyn_mcpu_ext_opt); |
| 27102 | dyn_mcpu_ext_opt = NULL; |
| 27103 | XDELETE (dyn_march_ext_opt); |
| 27104 | dyn_march_ext_opt = NULL; |
| 27105 | } |
| 27106 | |
| 27107 | /* Add the default contents for the .ARM.attributes section. */ |
| 27108 | void |
| 27109 | arm_md_end (void) |
| 27110 | { |
| 27111 | if (EF_ARM_EABI_VERSION (meabi_flags) < EF_ARM_EABI_VER4) |
| 27112 | return; |
| 27113 | |
| 27114 | aeabi_set_public_attributes (); |
| 27115 | } |
| 27116 | #endif /* OBJ_ELF */ |
| 27117 | |
| 27118 | |
| 27119 | /* Parse a .cpu directive. */ |
| 27120 | |
| 27121 | static void |
| 27122 | s_arm_cpu (int ignored ATTRIBUTE_UNUSED) |
| 27123 | { |
| 27124 | const struct arm_cpu_option_table *opt; |
| 27125 | char *name; |
| 27126 | char saved_char; |
| 27127 | |
| 27128 | name = input_line_pointer; |
| 27129 | while (*input_line_pointer && !ISSPACE (*input_line_pointer)) |
| 27130 | input_line_pointer++; |
| 27131 | saved_char = *input_line_pointer; |
| 27132 | *input_line_pointer = 0; |
| 27133 | |
| 27134 | /* Skip the first "all" entry. */ |
| 27135 | for (opt = arm_cpus + 1; opt->name != NULL; opt++) |
| 27136 | if (streq (opt->name, name)) |
| 27137 | { |
| 27138 | mcpu_cpu_opt = &opt->value; |
| 27139 | if (!dyn_mcpu_ext_opt) |
| 27140 | dyn_mcpu_ext_opt = XNEW (arm_feature_set); |
| 27141 | *dyn_mcpu_ext_opt = opt->ext; |
| 27142 | ARM_MERGE_FEATURE_SETS (selected_cpu, *mcpu_cpu_opt, *dyn_mcpu_ext_opt); |
| 27143 | if (opt->canonical_name) |
| 27144 | strcpy (selected_cpu_name, opt->canonical_name); |
| 27145 | else |
| 27146 | { |
| 27147 | int i; |
| 27148 | for (i = 0; opt->name[i]; i++) |
| 27149 | selected_cpu_name[i] = TOUPPER (opt->name[i]); |
| 27150 | |
| 27151 | selected_cpu_name[i] = 0; |
| 27152 | } |
| 27153 | ARM_MERGE_FEATURE_SETS (cpu_variant, *mcpu_cpu_opt, *mfpu_opt); |
| 27154 | if (dyn_mcpu_ext_opt) |
| 27155 | ARM_MERGE_FEATURE_SETS (cpu_variant, cpu_variant, *dyn_mcpu_ext_opt); |
| 27156 | *input_line_pointer = saved_char; |
| 27157 | demand_empty_rest_of_line (); |
| 27158 | return; |
| 27159 | } |
| 27160 | as_bad (_("unknown cpu `%s'"), name); |
| 27161 | *input_line_pointer = saved_char; |
| 27162 | ignore_rest_of_line (); |
| 27163 | } |
| 27164 | |
| 27165 | |
| 27166 | /* Parse a .arch directive. */ |
| 27167 | |
| 27168 | static void |
| 27169 | s_arm_arch (int ignored ATTRIBUTE_UNUSED) |
| 27170 | { |
| 27171 | const struct arm_arch_option_table *opt; |
| 27172 | char saved_char; |
| 27173 | char *name; |
| 27174 | |
| 27175 | name = input_line_pointer; |
| 27176 | while (*input_line_pointer && !ISSPACE (*input_line_pointer)) |
| 27177 | input_line_pointer++; |
| 27178 | saved_char = *input_line_pointer; |
| 27179 | *input_line_pointer = 0; |
| 27180 | |
| 27181 | /* Skip the first "all" entry. */ |
| 27182 | for (opt = arm_archs + 1; opt->name != NULL; opt++) |
| 27183 | if (streq (opt->name, name)) |
| 27184 | { |
| 27185 | mcpu_cpu_opt = &opt->value; |
| 27186 | XDELETE (dyn_mcpu_ext_opt); |
| 27187 | dyn_mcpu_ext_opt = NULL; |
| 27188 | selected_cpu = *mcpu_cpu_opt; |
| 27189 | strcpy (selected_cpu_name, opt->name); |
| 27190 | ARM_MERGE_FEATURE_SETS (cpu_variant, selected_cpu, *mfpu_opt); |
| 27191 | *input_line_pointer = saved_char; |
| 27192 | demand_empty_rest_of_line (); |
| 27193 | return; |
| 27194 | } |
| 27195 | |
| 27196 | as_bad (_("unknown architecture `%s'\n"), name); |
| 27197 | *input_line_pointer = saved_char; |
| 27198 | ignore_rest_of_line (); |
| 27199 | } |
| 27200 | |
| 27201 | |
| 27202 | /* Parse a .object_arch directive. */ |
| 27203 | |
| 27204 | static void |
| 27205 | s_arm_object_arch (int ignored ATTRIBUTE_UNUSED) |
| 27206 | { |
| 27207 | const struct arm_arch_option_table *opt; |
| 27208 | char saved_char; |
| 27209 | char *name; |
| 27210 | |
| 27211 | name = input_line_pointer; |
| 27212 | while (*input_line_pointer && !ISSPACE (*input_line_pointer)) |
| 27213 | input_line_pointer++; |
| 27214 | saved_char = *input_line_pointer; |
| 27215 | *input_line_pointer = 0; |
| 27216 | |
| 27217 | /* Skip the first "all" entry. */ |
| 27218 | for (opt = arm_archs + 1; opt->name != NULL; opt++) |
| 27219 | if (streq (opt->name, name)) |
| 27220 | { |
| 27221 | object_arch = &opt->value; |
| 27222 | *input_line_pointer = saved_char; |
| 27223 | demand_empty_rest_of_line (); |
| 27224 | return; |
| 27225 | } |
| 27226 | |
| 27227 | as_bad (_("unknown architecture `%s'\n"), name); |
| 27228 | *input_line_pointer = saved_char; |
| 27229 | ignore_rest_of_line (); |
| 27230 | } |
| 27231 | |
| 27232 | /* Parse a .arch_extension directive. */ |
| 27233 | |
| 27234 | static void |
| 27235 | s_arm_arch_extension (int ignored ATTRIBUTE_UNUSED) |
| 27236 | { |
| 27237 | const struct arm_option_extension_value_table *opt; |
| 27238 | const arm_feature_set arm_any = ARM_ANY; |
| 27239 | char saved_char; |
| 27240 | char *name; |
| 27241 | int adding_value = 1; |
| 27242 | |
| 27243 | name = input_line_pointer; |
| 27244 | while (*input_line_pointer && !ISSPACE (*input_line_pointer)) |
| 27245 | input_line_pointer++; |
| 27246 | saved_char = *input_line_pointer; |
| 27247 | *input_line_pointer = 0; |
| 27248 | |
| 27249 | if (strlen (name) >= 2 |
| 27250 | && strncmp (name, "no", 2) == 0) |
| 27251 | { |
| 27252 | adding_value = 0; |
| 27253 | name += 2; |
| 27254 | } |
| 27255 | |
| 27256 | for (opt = arm_extensions; opt->name != NULL; opt++) |
| 27257 | if (streq (opt->name, name)) |
| 27258 | { |
| 27259 | int i, nb_allowed_archs = |
| 27260 | sizeof (opt->allowed_archs) / sizeof (opt->allowed_archs[i]); |
| 27261 | for (i = 0; i < nb_allowed_archs; i++) |
| 27262 | { |
| 27263 | /* Empty entry. */ |
| 27264 | if (ARM_FEATURE_EQUAL (opt->allowed_archs[i], arm_any)) |
| 27265 | continue; |
| 27266 | if (ARM_FSET_CPU_SUBSET (opt->allowed_archs[i], *mcpu_cpu_opt)) |
| 27267 | break; |
| 27268 | } |
| 27269 | |
| 27270 | if (i == nb_allowed_archs) |
| 27271 | { |
| 27272 | as_bad (_("architectural extension `%s' is not allowed for the " |
| 27273 | "current base architecture"), name); |
| 27274 | break; |
| 27275 | } |
| 27276 | |
| 27277 | if (!dyn_mcpu_ext_opt) |
| 27278 | { |
| 27279 | dyn_mcpu_ext_opt = XNEW (arm_feature_set); |
| 27280 | *dyn_mcpu_ext_opt = arm_arch_none; |
| 27281 | } |
| 27282 | if (adding_value) |
| 27283 | ARM_MERGE_FEATURE_SETS (*dyn_mcpu_ext_opt, *dyn_mcpu_ext_opt, |
| 27284 | opt->merge_value); |
| 27285 | else |
| 27286 | ARM_CLEAR_FEATURE (*dyn_mcpu_ext_opt, *dyn_mcpu_ext_opt, |
| 27287 | opt->clear_value); |
| 27288 | |
| 27289 | ARM_MERGE_FEATURE_SETS (selected_cpu, *mcpu_cpu_opt, *dyn_mcpu_ext_opt); |
| 27290 | ARM_MERGE_FEATURE_SETS (cpu_variant, selected_cpu, *mfpu_opt); |
| 27291 | *input_line_pointer = saved_char; |
| 27292 | demand_empty_rest_of_line (); |
| 27293 | /* Allowing Thumb division instructions for ARMv7 in autodetection rely |
| 27294 | on this return so that duplicate extensions (extensions with the |
| 27295 | same name as a previous extension in the list) are not considered |
| 27296 | for command-line parsing. */ |
| 27297 | return; |
| 27298 | } |
| 27299 | |
| 27300 | if (opt->name == NULL) |
| 27301 | as_bad (_("unknown architecture extension `%s'\n"), name); |
| 27302 | |
| 27303 | *input_line_pointer = saved_char; |
| 27304 | ignore_rest_of_line (); |
| 27305 | } |
| 27306 | |
| 27307 | /* Parse a .fpu directive. */ |
| 27308 | |
| 27309 | static void |
| 27310 | s_arm_fpu (int ignored ATTRIBUTE_UNUSED) |
| 27311 | { |
| 27312 | const struct arm_option_fpu_value_table *opt; |
| 27313 | char saved_char; |
| 27314 | char *name; |
| 27315 | |
| 27316 | name = input_line_pointer; |
| 27317 | while (*input_line_pointer && !ISSPACE (*input_line_pointer)) |
| 27318 | input_line_pointer++; |
| 27319 | saved_char = *input_line_pointer; |
| 27320 | *input_line_pointer = 0; |
| 27321 | |
| 27322 | for (opt = arm_fpus; opt->name != NULL; opt++) |
| 27323 | if (streq (opt->name, name)) |
| 27324 | { |
| 27325 | mfpu_opt = &opt->value; |
| 27326 | ARM_MERGE_FEATURE_SETS (cpu_variant, *mcpu_cpu_opt, *mfpu_opt); |
| 27327 | if (dyn_mcpu_ext_opt) |
| 27328 | ARM_MERGE_FEATURE_SETS (cpu_variant, cpu_variant, *dyn_mcpu_ext_opt); |
| 27329 | *input_line_pointer = saved_char; |
| 27330 | demand_empty_rest_of_line (); |
| 27331 | return; |
| 27332 | } |
| 27333 | |
| 27334 | as_bad (_("unknown floating point format `%s'\n"), name); |
| 27335 | *input_line_pointer = saved_char; |
| 27336 | ignore_rest_of_line (); |
| 27337 | } |
| 27338 | |
| 27339 | /* Copy symbol information. */ |
| 27340 | |
| 27341 | void |
| 27342 | arm_copy_symbol_attributes (symbolS *dest, symbolS *src) |
| 27343 | { |
| 27344 | ARM_GET_FLAG (dest) = ARM_GET_FLAG (src); |
| 27345 | } |
| 27346 | |
| 27347 | #ifdef OBJ_ELF |
| 27348 | /* Given a symbolic attribute NAME, return the proper integer value. |
| 27349 | Returns -1 if the attribute is not known. */ |
| 27350 | |
| 27351 | int |
| 27352 | arm_convert_symbolic_attribute (const char *name) |
| 27353 | { |
| 27354 | static const struct |
| 27355 | { |
| 27356 | const char * name; |
| 27357 | const int tag; |
| 27358 | } |
| 27359 | attribute_table[] = |
| 27360 | { |
| 27361 | /* When you modify this table you should |
| 27362 | also modify the list in doc/c-arm.texi. */ |
| 27363 | #define T(tag) {#tag, tag} |
| 27364 | T (Tag_CPU_raw_name), |
| 27365 | T (Tag_CPU_name), |
| 27366 | T (Tag_CPU_arch), |
| 27367 | T (Tag_CPU_arch_profile), |
| 27368 | T (Tag_ARM_ISA_use), |
| 27369 | T (Tag_THUMB_ISA_use), |
| 27370 | T (Tag_FP_arch), |
| 27371 | T (Tag_VFP_arch), |
| 27372 | T (Tag_WMMX_arch), |
| 27373 | T (Tag_Advanced_SIMD_arch), |
| 27374 | T (Tag_PCS_config), |
| 27375 | T (Tag_ABI_PCS_R9_use), |
| 27376 | T (Tag_ABI_PCS_RW_data), |
| 27377 | T (Tag_ABI_PCS_RO_data), |
| 27378 | T (Tag_ABI_PCS_GOT_use), |
| 27379 | T (Tag_ABI_PCS_wchar_t), |
| 27380 | T (Tag_ABI_FP_rounding), |
| 27381 | T (Tag_ABI_FP_denormal), |
| 27382 | T (Tag_ABI_FP_exceptions), |
| 27383 | T (Tag_ABI_FP_user_exceptions), |
| 27384 | T (Tag_ABI_FP_number_model), |
| 27385 | T (Tag_ABI_align_needed), |
| 27386 | T (Tag_ABI_align8_needed), |
| 27387 | T (Tag_ABI_align_preserved), |
| 27388 | T (Tag_ABI_align8_preserved), |
| 27389 | T (Tag_ABI_enum_size), |
| 27390 | T (Tag_ABI_HardFP_use), |
| 27391 | T (Tag_ABI_VFP_args), |
| 27392 | T (Tag_ABI_WMMX_args), |
| 27393 | T (Tag_ABI_optimization_goals), |
| 27394 | T (Tag_ABI_FP_optimization_goals), |
| 27395 | T (Tag_compatibility), |
| 27396 | T (Tag_CPU_unaligned_access), |
| 27397 | T (Tag_FP_HP_extension), |
| 27398 | T (Tag_VFP_HP_extension), |
| 27399 | T (Tag_ABI_FP_16bit_format), |
| 27400 | T (Tag_MPextension_use), |
| 27401 | T (Tag_DIV_use), |
| 27402 | T (Tag_nodefaults), |
| 27403 | T (Tag_also_compatible_with), |
| 27404 | T (Tag_conformance), |
| 27405 | T (Tag_T2EE_use), |
| 27406 | T (Tag_Virtualization_use), |
| 27407 | T (Tag_DSP_extension), |
| 27408 | /* We deliberately do not include Tag_MPextension_use_legacy. */ |
| 27409 | #undef T |
| 27410 | }; |
| 27411 | unsigned int i; |
| 27412 | |
| 27413 | if (name == NULL) |
| 27414 | return -1; |
| 27415 | |
| 27416 | for (i = 0; i < ARRAY_SIZE (attribute_table); i++) |
| 27417 | if (streq (name, attribute_table[i].name)) |
| 27418 | return attribute_table[i].tag; |
| 27419 | |
| 27420 | return -1; |
| 27421 | } |
| 27422 | |
| 27423 | |
| 27424 | /* Apply sym value for relocations only in the case that they are for |
| 27425 | local symbols in the same segment as the fixup and you have the |
| 27426 | respective architectural feature for blx and simple switches. */ |
| 27427 | int |
| 27428 | arm_apply_sym_value (struct fix * fixP, segT this_seg) |
| 27429 | { |
| 27430 | if (fixP->fx_addsy |
| 27431 | && ARM_CPU_HAS_FEATURE (selected_cpu, arm_ext_v5t) |
| 27432 | /* PR 17444: If the local symbol is in a different section then a reloc |
| 27433 | will always be generated for it, so applying the symbol value now |
| 27434 | will result in a double offset being stored in the relocation. */ |
| 27435 | && (S_GET_SEGMENT (fixP->fx_addsy) == this_seg) |
| 27436 | && !S_FORCE_RELOC (fixP->fx_addsy, TRUE)) |
| 27437 | { |
| 27438 | switch (fixP->fx_r_type) |
| 27439 | { |
| 27440 | case BFD_RELOC_ARM_PCREL_BLX: |
| 27441 | case BFD_RELOC_THUMB_PCREL_BRANCH23: |
| 27442 | if (ARM_IS_FUNC (fixP->fx_addsy)) |
| 27443 | return 1; |
| 27444 | break; |
| 27445 | |
| 27446 | case BFD_RELOC_ARM_PCREL_CALL: |
| 27447 | case BFD_RELOC_THUMB_PCREL_BLX: |
| 27448 | if (THUMB_IS_FUNC (fixP->fx_addsy)) |
| 27449 | return 1; |
| 27450 | break; |
| 27451 | |
| 27452 | default: |
| 27453 | break; |
| 27454 | } |
| 27455 | |
| 27456 | } |
| 27457 | return 0; |
| 27458 | } |
| 27459 | #endif /* OBJ_ELF */ |