Add clwb instruction
[binutils-gdb.git] / gas / config / tc-i386.c
1 /* tc-i386.c -- Assemble code for the Intel 80386
2 Copyright (C) 1989-2014 Free Software Foundation, Inc.
3
4 This file is part of GAS, the GNU Assembler.
5
6 GAS is free software; you can redistribute it and/or modify
7 it under the terms of the GNU General Public License as published by
8 the Free Software Foundation; either version 3, or (at your option)
9 any later version.
10
11 GAS is distributed in the hope that it will be useful,
12 but WITHOUT ANY WARRANTY; without even the implied warranty of
13 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 GNU General Public License for more details.
15
16 You should have received a copy of the GNU General Public License
17 along with GAS; see the file COPYING. If not, write to the Free
18 Software Foundation, 51 Franklin Street - Fifth Floor, Boston, MA
19 02110-1301, USA. */
20
21 /* Intel 80386 machine specific gas.
22 Written by Eliot Dresselhaus (eliot@mgm.mit.edu).
23 x86_64 support by Jan Hubicka (jh@suse.cz)
24 VIA PadLock support by Michal Ludvig (mludvig@suse.cz)
25 Bugs & suggestions are completely welcome. This is free software.
26 Please help us make it better. */
27
28 #include "as.h"
29 #include "safe-ctype.h"
30 #include "subsegs.h"
31 #include "dwarf2dbg.h"
32 #include "dw2gencfi.h"
33 #include "elf/x86-64.h"
34 #include "opcodes/i386-init.h"
35
36 #ifndef REGISTER_WARNINGS
37 #define REGISTER_WARNINGS 1
38 #endif
39
40 #ifndef INFER_ADDR_PREFIX
41 #define INFER_ADDR_PREFIX 1
42 #endif
43
44 #ifndef DEFAULT_ARCH
45 #define DEFAULT_ARCH "i386"
46 #endif
47
48 #ifndef INLINE
49 #if __GNUC__ >= 2
50 #define INLINE __inline__
51 #else
52 #define INLINE
53 #endif
54 #endif
55
56 /* Prefixes will be emitted in the order defined below.
57 WAIT_PREFIX must be the first prefix since FWAIT is really is an
58 instruction, and so must come before any prefixes.
59 The preferred prefix order is SEG_PREFIX, ADDR_PREFIX, DATA_PREFIX,
60 REP_PREFIX/HLE_PREFIX, LOCK_PREFIX. */
61 #define WAIT_PREFIX 0
62 #define SEG_PREFIX 1
63 #define ADDR_PREFIX 2
64 #define DATA_PREFIX 3
65 #define REP_PREFIX 4
66 #define HLE_PREFIX REP_PREFIX
67 #define BND_PREFIX REP_PREFIX
68 #define LOCK_PREFIX 5
69 #define REX_PREFIX 6 /* must come last. */
70 #define MAX_PREFIXES 7 /* max prefixes per opcode */
71
72 /* we define the syntax here (modulo base,index,scale syntax) */
73 #define REGISTER_PREFIX '%'
74 #define IMMEDIATE_PREFIX '$'
75 #define ABSOLUTE_PREFIX '*'
76
77 /* these are the instruction mnemonic suffixes in AT&T syntax or
78 memory operand size in Intel syntax. */
79 #define WORD_MNEM_SUFFIX 'w'
80 #define BYTE_MNEM_SUFFIX 'b'
81 #define SHORT_MNEM_SUFFIX 's'
82 #define LONG_MNEM_SUFFIX 'l'
83 #define QWORD_MNEM_SUFFIX 'q'
84 #define XMMWORD_MNEM_SUFFIX 'x'
85 #define YMMWORD_MNEM_SUFFIX 'y'
86 #define ZMMWORD_MNEM_SUFFIX 'z'
87 /* Intel Syntax. Use a non-ascii letter since since it never appears
88 in instructions. */
89 #define LONG_DOUBLE_MNEM_SUFFIX '\1'
90
91 #define END_OF_INSN '\0'
92
93 /*
94 'templates' is for grouping together 'template' structures for opcodes
95 of the same name. This is only used for storing the insns in the grand
96 ole hash table of insns.
97 The templates themselves start at START and range up to (but not including)
98 END.
99 */
100 typedef struct
101 {
102 const insn_template *start;
103 const insn_template *end;
104 }
105 templates;
106
107 /* 386 operand encoding bytes: see 386 book for details of this. */
108 typedef struct
109 {
110 unsigned int regmem; /* codes register or memory operand */
111 unsigned int reg; /* codes register operand (or extended opcode) */
112 unsigned int mode; /* how to interpret regmem & reg */
113 }
114 modrm_byte;
115
116 /* x86-64 extension prefix. */
117 typedef int rex_byte;
118
119 /* 386 opcode byte to code indirect addressing. */
120 typedef struct
121 {
122 unsigned base;
123 unsigned index;
124 unsigned scale;
125 }
126 sib_byte;
127
128 /* x86 arch names, types and features */
129 typedef struct
130 {
131 const char *name; /* arch name */
132 unsigned int len; /* arch string length */
133 enum processor_type type; /* arch type */
134 i386_cpu_flags flags; /* cpu feature flags */
135 unsigned int skip; /* show_arch should skip this. */
136 unsigned int negated; /* turn off indicated flags. */
137 }
138 arch_entry;
139
140 static void update_code_flag (int, int);
141 static void set_code_flag (int);
142 static void set_16bit_gcc_code_flag (int);
143 static void set_intel_syntax (int);
144 static void set_intel_mnemonic (int);
145 static void set_allow_index_reg (int);
146 static void set_check (int);
147 static void set_cpu_arch (int);
148 #ifdef TE_PE
149 static void pe_directive_secrel (int);
150 #endif
151 static void signed_cons (int);
152 static char *output_invalid (int c);
153 static int i386_finalize_immediate (segT, expressionS *, i386_operand_type,
154 const char *);
155 static int i386_finalize_displacement (segT, expressionS *, i386_operand_type,
156 const char *);
157 static int i386_att_operand (char *);
158 static int i386_intel_operand (char *, int);
159 static int i386_intel_simplify (expressionS *);
160 static int i386_intel_parse_name (const char *, expressionS *);
161 static const reg_entry *parse_register (char *, char **);
162 static char *parse_insn (char *, char *);
163 static char *parse_operands (char *, const char *);
164 static void swap_operands (void);
165 static void swap_2_operands (int, int);
166 static void optimize_imm (void);
167 static void optimize_disp (void);
168 static const insn_template *match_template (void);
169 static int check_string (void);
170 static int process_suffix (void);
171 static int check_byte_reg (void);
172 static int check_long_reg (void);
173 static int check_qword_reg (void);
174 static int check_word_reg (void);
175 static int finalize_imm (void);
176 static int process_operands (void);
177 static const seg_entry *build_modrm_byte (void);
178 static void output_insn (void);
179 static void output_imm (fragS *, offsetT);
180 static void output_disp (fragS *, offsetT);
181 #ifndef I386COFF
182 static void s_bss (int);
183 #endif
184 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
185 static void handle_large_common (int small ATTRIBUTE_UNUSED);
186 #endif
187
188 static const char *default_arch = DEFAULT_ARCH;
189
190 /* This struct describes rounding control and SAE in the instruction. */
191 struct RC_Operation
192 {
193 enum rc_type
194 {
195 rne = 0,
196 rd,
197 ru,
198 rz,
199 saeonly
200 } type;
201 int operand;
202 };
203
204 static struct RC_Operation rc_op;
205
206 /* The struct describes masking, applied to OPERAND in the instruction.
207 MASK is a pointer to the corresponding mask register. ZEROING tells
208 whether merging or zeroing mask is used. */
209 struct Mask_Operation
210 {
211 const reg_entry *mask;
212 unsigned int zeroing;
213 /* The operand where this operation is associated. */
214 int operand;
215 };
216
217 static struct Mask_Operation mask_op;
218
219 /* The struct describes broadcasting, applied to OPERAND. FACTOR is
220 broadcast factor. */
221 struct Broadcast_Operation
222 {
223 /* Type of broadcast: no broadcast, {1to8}, or {1to16}. */
224 int type;
225
226 /* Index of broadcasted operand. */
227 int operand;
228 };
229
230 static struct Broadcast_Operation broadcast_op;
231
232 /* VEX prefix. */
233 typedef struct
234 {
235 /* VEX prefix is either 2 byte or 3 byte. EVEX is 4 byte. */
236 unsigned char bytes[4];
237 unsigned int length;
238 /* Destination or source register specifier. */
239 const reg_entry *register_specifier;
240 } vex_prefix;
241
242 /* 'md_assemble ()' gathers together information and puts it into a
243 i386_insn. */
244
245 union i386_op
246 {
247 expressionS *disps;
248 expressionS *imms;
249 const reg_entry *regs;
250 };
251
252 enum i386_error
253 {
254 operand_size_mismatch,
255 operand_type_mismatch,
256 register_type_mismatch,
257 number_of_operands_mismatch,
258 invalid_instruction_suffix,
259 bad_imm4,
260 old_gcc_only,
261 unsupported_with_intel_mnemonic,
262 unsupported_syntax,
263 unsupported,
264 invalid_vsib_address,
265 invalid_vector_register_set,
266 unsupported_vector_index_register,
267 unsupported_broadcast,
268 broadcast_not_on_src_operand,
269 broadcast_needed,
270 unsupported_masking,
271 mask_not_on_destination,
272 no_default_mask,
273 unsupported_rc_sae,
274 rc_sae_operand_not_last_imm,
275 invalid_register_operand,
276 try_vector_disp8
277 };
278
279 struct _i386_insn
280 {
281 /* TM holds the template for the insn were currently assembling. */
282 insn_template tm;
283
284 /* SUFFIX holds the instruction size suffix for byte, word, dword
285 or qword, if given. */
286 char suffix;
287
288 /* OPERANDS gives the number of given operands. */
289 unsigned int operands;
290
291 /* REG_OPERANDS, DISP_OPERANDS, MEM_OPERANDS, IMM_OPERANDS give the number
292 of given register, displacement, memory operands and immediate
293 operands. */
294 unsigned int reg_operands, disp_operands, mem_operands, imm_operands;
295
296 /* TYPES [i] is the type (see above #defines) which tells us how to
297 use OP[i] for the corresponding operand. */
298 i386_operand_type types[MAX_OPERANDS];
299
300 /* Displacement expression, immediate expression, or register for each
301 operand. */
302 union i386_op op[MAX_OPERANDS];
303
304 /* Flags for operands. */
305 unsigned int flags[MAX_OPERANDS];
306 #define Operand_PCrel 1
307
308 /* Relocation type for operand */
309 enum bfd_reloc_code_real reloc[MAX_OPERANDS];
310
311 /* BASE_REG, INDEX_REG, and LOG2_SCALE_FACTOR are used to encode
312 the base index byte below. */
313 const reg_entry *base_reg;
314 const reg_entry *index_reg;
315 unsigned int log2_scale_factor;
316
317 /* SEG gives the seg_entries of this insn. They are zero unless
318 explicit segment overrides are given. */
319 const seg_entry *seg[2];
320
321 /* PREFIX holds all the given prefix opcodes (usually null).
322 PREFIXES is the number of prefix opcodes. */
323 unsigned int prefixes;
324 unsigned char prefix[MAX_PREFIXES];
325
326 /* RM and SIB are the modrm byte and the sib byte where the
327 addressing modes of this insn are encoded. */
328 modrm_byte rm;
329 rex_byte rex;
330 rex_byte vrex;
331 sib_byte sib;
332 vex_prefix vex;
333
334 /* Masking attributes. */
335 struct Mask_Operation *mask;
336
337 /* Rounding control and SAE attributes. */
338 struct RC_Operation *rounding;
339
340 /* Broadcasting attributes. */
341 struct Broadcast_Operation *broadcast;
342
343 /* Compressed disp8*N attribute. */
344 unsigned int memshift;
345
346 /* Swap operand in encoding. */
347 unsigned int swap_operand;
348
349 /* Prefer 8bit or 32bit displacement in encoding. */
350 enum
351 {
352 disp_encoding_default = 0,
353 disp_encoding_8bit,
354 disp_encoding_32bit
355 } disp_encoding;
356
357 /* REP prefix. */
358 const char *rep_prefix;
359
360 /* HLE prefix. */
361 const char *hle_prefix;
362
363 /* Have BND prefix. */
364 const char *bnd_prefix;
365
366 /* Need VREX to support upper 16 registers. */
367 int need_vrex;
368
369 /* Error message. */
370 enum i386_error error;
371 };
372
373 typedef struct _i386_insn i386_insn;
374
375 /* Link RC type with corresponding string, that'll be looked for in
376 asm. */
377 struct RC_name
378 {
379 enum rc_type type;
380 const char *name;
381 unsigned int len;
382 };
383
384 static const struct RC_name RC_NamesTable[] =
385 {
386 { rne, STRING_COMMA_LEN ("rn-sae") },
387 { rd, STRING_COMMA_LEN ("rd-sae") },
388 { ru, STRING_COMMA_LEN ("ru-sae") },
389 { rz, STRING_COMMA_LEN ("rz-sae") },
390 { saeonly, STRING_COMMA_LEN ("sae") },
391 };
392
393 /* List of chars besides those in app.c:symbol_chars that can start an
394 operand. Used to prevent the scrubber eating vital white-space. */
395 const char extra_symbol_chars[] = "*%-([{"
396 #ifdef LEX_AT
397 "@"
398 #endif
399 #ifdef LEX_QM
400 "?"
401 #endif
402 ;
403
404 #if (defined (TE_I386AIX) \
405 || ((defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)) \
406 && !defined (TE_GNU) \
407 && !defined (TE_LINUX) \
408 && !defined (TE_NACL) \
409 && !defined (TE_NETWARE) \
410 && !defined (TE_FreeBSD) \
411 && !defined (TE_DragonFly) \
412 && !defined (TE_NetBSD)))
413 /* This array holds the chars that always start a comment. If the
414 pre-processor is disabled, these aren't very useful. The option
415 --divide will remove '/' from this list. */
416 const char *i386_comment_chars = "#/";
417 #define SVR4_COMMENT_CHARS 1
418 #define PREFIX_SEPARATOR '\\'
419
420 #else
421 const char *i386_comment_chars = "#";
422 #define PREFIX_SEPARATOR '/'
423 #endif
424
425 /* This array holds the chars that only start a comment at the beginning of
426 a line. If the line seems to have the form '# 123 filename'
427 .line and .file directives will appear in the pre-processed output.
428 Note that input_file.c hand checks for '#' at the beginning of the
429 first line of the input file. This is because the compiler outputs
430 #NO_APP at the beginning of its output.
431 Also note that comments started like this one will always work if
432 '/' isn't otherwise defined. */
433 const char line_comment_chars[] = "#/";
434
435 const char line_separator_chars[] = ";";
436
437 /* Chars that can be used to separate mant from exp in floating point
438 nums. */
439 const char EXP_CHARS[] = "eE";
440
441 /* Chars that mean this number is a floating point constant
442 As in 0f12.456
443 or 0d1.2345e12. */
444 const char FLT_CHARS[] = "fFdDxX";
445
446 /* Tables for lexical analysis. */
447 static char mnemonic_chars[256];
448 static char register_chars[256];
449 static char operand_chars[256];
450 static char identifier_chars[256];
451 static char digit_chars[256];
452
453 /* Lexical macros. */
454 #define is_mnemonic_char(x) (mnemonic_chars[(unsigned char) x])
455 #define is_operand_char(x) (operand_chars[(unsigned char) x])
456 #define is_register_char(x) (register_chars[(unsigned char) x])
457 #define is_space_char(x) ((x) == ' ')
458 #define is_identifier_char(x) (identifier_chars[(unsigned char) x])
459 #define is_digit_char(x) (digit_chars[(unsigned char) x])
460
461 /* All non-digit non-letter characters that may occur in an operand. */
462 static char operand_special_chars[] = "%$-+(,)*._~/<>|&^!:[@]";
463
464 /* md_assemble() always leaves the strings it's passed unaltered. To
465 effect this we maintain a stack of saved characters that we've smashed
466 with '\0's (indicating end of strings for various sub-fields of the
467 assembler instruction). */
468 static char save_stack[32];
469 static char *save_stack_p;
470 #define END_STRING_AND_SAVE(s) \
471 do { *save_stack_p++ = *(s); *(s) = '\0'; } while (0)
472 #define RESTORE_END_STRING(s) \
473 do { *(s) = *--save_stack_p; } while (0)
474
475 /* The instruction we're assembling. */
476 static i386_insn i;
477
478 /* Possible templates for current insn. */
479 static const templates *current_templates;
480
481 /* Per instruction expressionS buffers: max displacements & immediates. */
482 static expressionS disp_expressions[MAX_MEMORY_OPERANDS];
483 static expressionS im_expressions[MAX_IMMEDIATE_OPERANDS];
484
485 /* Current operand we are working on. */
486 static int this_operand = -1;
487
488 /* We support four different modes. FLAG_CODE variable is used to distinguish
489 these. */
490
491 enum flag_code {
492 CODE_32BIT,
493 CODE_16BIT,
494 CODE_64BIT };
495
496 static enum flag_code flag_code;
497 static unsigned int object_64bit;
498 static unsigned int disallow_64bit_reloc;
499 static int use_rela_relocations = 0;
500
501 #if ((defined (OBJ_MAYBE_COFF) && defined (OBJ_MAYBE_AOUT)) \
502 || defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) \
503 || defined (TE_PE) || defined (TE_PEP) || defined (OBJ_MACH_O))
504
505 /* The ELF ABI to use. */
506 enum x86_elf_abi
507 {
508 I386_ABI,
509 X86_64_ABI,
510 X86_64_X32_ABI
511 };
512
513 static enum x86_elf_abi x86_elf_abi = I386_ABI;
514 #endif
515
516 #if defined (TE_PE) || defined (TE_PEP)
517 /* Use big object file format. */
518 static int use_big_obj = 0;
519 #endif
520
521 /* 1 for intel syntax,
522 0 if att syntax. */
523 static int intel_syntax = 0;
524
525 /* 1 for intel mnemonic,
526 0 if att mnemonic. */
527 static int intel_mnemonic = !SYSV386_COMPAT;
528
529 /* 1 if support old (<= 2.8.1) versions of gcc. */
530 static int old_gcc = OLDGCC_COMPAT;
531
532 /* 1 if pseudo registers are permitted. */
533 static int allow_pseudo_reg = 0;
534
535 /* 1 if register prefix % not required. */
536 static int allow_naked_reg = 0;
537
538 /* 1 if the assembler should add BND prefix for all control-tranferring
539 instructions supporting it, even if this prefix wasn't specified
540 explicitly. */
541 static int add_bnd_prefix = 0;
542
543 /* 1 if pseudo index register, eiz/riz, is allowed . */
544 static int allow_index_reg = 0;
545
546 /* 1 if the assembler should ignore LOCK prefix, even if it was
547 specified explicitly. */
548 static int omit_lock_prefix = 0;
549
550 static enum check_kind
551 {
552 check_none = 0,
553 check_warning,
554 check_error
555 }
556 sse_check, operand_check = check_warning;
557
558 /* Register prefix used for error message. */
559 static const char *register_prefix = "%";
560
561 /* Used in 16 bit gcc mode to add an l suffix to call, ret, enter,
562 leave, push, and pop instructions so that gcc has the same stack
563 frame as in 32 bit mode. */
564 static char stackop_size = '\0';
565
566 /* Non-zero to optimize code alignment. */
567 int optimize_align_code = 1;
568
569 /* Non-zero to quieten some warnings. */
570 static int quiet_warnings = 0;
571
572 /* CPU name. */
573 static const char *cpu_arch_name = NULL;
574 static char *cpu_sub_arch_name = NULL;
575
576 /* CPU feature flags. */
577 static i386_cpu_flags cpu_arch_flags = CPU_UNKNOWN_FLAGS;
578
579 /* If we have selected a cpu we are generating instructions for. */
580 static int cpu_arch_tune_set = 0;
581
582 /* Cpu we are generating instructions for. */
583 enum processor_type cpu_arch_tune = PROCESSOR_UNKNOWN;
584
585 /* CPU feature flags of cpu we are generating instructions for. */
586 static i386_cpu_flags cpu_arch_tune_flags;
587
588 /* CPU instruction set architecture used. */
589 enum processor_type cpu_arch_isa = PROCESSOR_UNKNOWN;
590
591 /* CPU feature flags of instruction set architecture used. */
592 i386_cpu_flags cpu_arch_isa_flags;
593
594 /* If set, conditional jumps are not automatically promoted to handle
595 larger than a byte offset. */
596 static unsigned int no_cond_jump_promotion = 0;
597
598 /* Encode SSE instructions with VEX prefix. */
599 static unsigned int sse2avx;
600
601 /* Encode scalar AVX instructions with specific vector length. */
602 static enum
603 {
604 vex128 = 0,
605 vex256
606 } avxscalar;
607
608 /* Encode scalar EVEX LIG instructions with specific vector length. */
609 static enum
610 {
611 evexl128 = 0,
612 evexl256,
613 evexl512
614 } evexlig;
615
616 /* Encode EVEX WIG instructions with specific evex.w. */
617 static enum
618 {
619 evexw0 = 0,
620 evexw1
621 } evexwig;
622
623 /* Value to encode in EVEX RC bits, for SAE-only instructions. */
624 static enum rc_type evexrcig = rne;
625
626 /* Pre-defined "_GLOBAL_OFFSET_TABLE_". */
627 static symbolS *GOT_symbol;
628
629 /* The dwarf2 return column, adjusted for 32 or 64 bit. */
630 unsigned int x86_dwarf2_return_column;
631
632 /* The dwarf2 data alignment, adjusted for 32 or 64 bit. */
633 int x86_cie_data_alignment;
634
635 /* Interface to relax_segment.
636 There are 3 major relax states for 386 jump insns because the
637 different types of jumps add different sizes to frags when we're
638 figuring out what sort of jump to choose to reach a given label. */
639
640 /* Types. */
641 #define UNCOND_JUMP 0
642 #define COND_JUMP 1
643 #define COND_JUMP86 2
644
645 /* Sizes. */
646 #define CODE16 1
647 #define SMALL 0
648 #define SMALL16 (SMALL | CODE16)
649 #define BIG 2
650 #define BIG16 (BIG | CODE16)
651
652 #ifndef INLINE
653 #ifdef __GNUC__
654 #define INLINE __inline__
655 #else
656 #define INLINE
657 #endif
658 #endif
659
660 #define ENCODE_RELAX_STATE(type, size) \
661 ((relax_substateT) (((type) << 2) | (size)))
662 #define TYPE_FROM_RELAX_STATE(s) \
663 ((s) >> 2)
664 #define DISP_SIZE_FROM_RELAX_STATE(s) \
665 ((((s) & 3) == BIG ? 4 : (((s) & 3) == BIG16 ? 2 : 1)))
666
667 /* This table is used by relax_frag to promote short jumps to long
668 ones where necessary. SMALL (short) jumps may be promoted to BIG
669 (32 bit long) ones, and SMALL16 jumps to BIG16 (16 bit long). We
670 don't allow a short jump in a 32 bit code segment to be promoted to
671 a 16 bit offset jump because it's slower (requires data size
672 prefix), and doesn't work, unless the destination is in the bottom
673 64k of the code segment (The top 16 bits of eip are zeroed). */
674
675 const relax_typeS md_relax_table[] =
676 {
677 /* The fields are:
678 1) most positive reach of this state,
679 2) most negative reach of this state,
680 3) how many bytes this mode will have in the variable part of the frag
681 4) which index into the table to try if we can't fit into this one. */
682
683 /* UNCOND_JUMP states. */
684 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (UNCOND_JUMP, BIG)},
685 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (UNCOND_JUMP, BIG16)},
686 /* dword jmp adds 4 bytes to frag:
687 0 extra opcode bytes, 4 displacement bytes. */
688 {0, 0, 4, 0},
689 /* word jmp adds 2 byte2 to frag:
690 0 extra opcode bytes, 2 displacement bytes. */
691 {0, 0, 2, 0},
692
693 /* COND_JUMP states. */
694 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (COND_JUMP, BIG)},
695 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (COND_JUMP, BIG16)},
696 /* dword conditionals adds 5 bytes to frag:
697 1 extra opcode byte, 4 displacement bytes. */
698 {0, 0, 5, 0},
699 /* word conditionals add 3 bytes to frag:
700 1 extra opcode byte, 2 displacement bytes. */
701 {0, 0, 3, 0},
702
703 /* COND_JUMP86 states. */
704 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (COND_JUMP86, BIG)},
705 {127 + 1, -128 + 1, 1, ENCODE_RELAX_STATE (COND_JUMP86, BIG16)},
706 /* dword conditionals adds 5 bytes to frag:
707 1 extra opcode byte, 4 displacement bytes. */
708 {0, 0, 5, 0},
709 /* word conditionals add 4 bytes to frag:
710 1 displacement byte and a 3 byte long branch insn. */
711 {0, 0, 4, 0}
712 };
713
714 static const arch_entry cpu_arch[] =
715 {
716 /* Do not replace the first two entries - i386_target_format()
717 relies on them being there in this order. */
718 { STRING_COMMA_LEN ("generic32"), PROCESSOR_GENERIC32,
719 CPU_GENERIC32_FLAGS, 0, 0 },
720 { STRING_COMMA_LEN ("generic64"), PROCESSOR_GENERIC64,
721 CPU_GENERIC64_FLAGS, 0, 0 },
722 { STRING_COMMA_LEN ("i8086"), PROCESSOR_UNKNOWN,
723 CPU_NONE_FLAGS, 0, 0 },
724 { STRING_COMMA_LEN ("i186"), PROCESSOR_UNKNOWN,
725 CPU_I186_FLAGS, 0, 0 },
726 { STRING_COMMA_LEN ("i286"), PROCESSOR_UNKNOWN,
727 CPU_I286_FLAGS, 0, 0 },
728 { STRING_COMMA_LEN ("i386"), PROCESSOR_I386,
729 CPU_I386_FLAGS, 0, 0 },
730 { STRING_COMMA_LEN ("i486"), PROCESSOR_I486,
731 CPU_I486_FLAGS, 0, 0 },
732 { STRING_COMMA_LEN ("i586"), PROCESSOR_PENTIUM,
733 CPU_I586_FLAGS, 0, 0 },
734 { STRING_COMMA_LEN ("i686"), PROCESSOR_PENTIUMPRO,
735 CPU_I686_FLAGS, 0, 0 },
736 { STRING_COMMA_LEN ("pentium"), PROCESSOR_PENTIUM,
737 CPU_I586_FLAGS, 0, 0 },
738 { STRING_COMMA_LEN ("pentiumpro"), PROCESSOR_PENTIUMPRO,
739 CPU_PENTIUMPRO_FLAGS, 0, 0 },
740 { STRING_COMMA_LEN ("pentiumii"), PROCESSOR_PENTIUMPRO,
741 CPU_P2_FLAGS, 0, 0 },
742 { STRING_COMMA_LEN ("pentiumiii"),PROCESSOR_PENTIUMPRO,
743 CPU_P3_FLAGS, 0, 0 },
744 { STRING_COMMA_LEN ("pentium4"), PROCESSOR_PENTIUM4,
745 CPU_P4_FLAGS, 0, 0 },
746 { STRING_COMMA_LEN ("prescott"), PROCESSOR_NOCONA,
747 CPU_CORE_FLAGS, 0, 0 },
748 { STRING_COMMA_LEN ("nocona"), PROCESSOR_NOCONA,
749 CPU_NOCONA_FLAGS, 0, 0 },
750 { STRING_COMMA_LEN ("yonah"), PROCESSOR_CORE,
751 CPU_CORE_FLAGS, 1, 0 },
752 { STRING_COMMA_LEN ("core"), PROCESSOR_CORE,
753 CPU_CORE_FLAGS, 0, 0 },
754 { STRING_COMMA_LEN ("merom"), PROCESSOR_CORE2,
755 CPU_CORE2_FLAGS, 1, 0 },
756 { STRING_COMMA_LEN ("core2"), PROCESSOR_CORE2,
757 CPU_CORE2_FLAGS, 0, 0 },
758 { STRING_COMMA_LEN ("corei7"), PROCESSOR_COREI7,
759 CPU_COREI7_FLAGS, 0, 0 },
760 { STRING_COMMA_LEN ("l1om"), PROCESSOR_L1OM,
761 CPU_L1OM_FLAGS, 0, 0 },
762 { STRING_COMMA_LEN ("k1om"), PROCESSOR_K1OM,
763 CPU_K1OM_FLAGS, 0, 0 },
764 { STRING_COMMA_LEN ("k6"), PROCESSOR_K6,
765 CPU_K6_FLAGS, 0, 0 },
766 { STRING_COMMA_LEN ("k6_2"), PROCESSOR_K6,
767 CPU_K6_2_FLAGS, 0, 0 },
768 { STRING_COMMA_LEN ("athlon"), PROCESSOR_ATHLON,
769 CPU_ATHLON_FLAGS, 0, 0 },
770 { STRING_COMMA_LEN ("sledgehammer"), PROCESSOR_K8,
771 CPU_K8_FLAGS, 1, 0 },
772 { STRING_COMMA_LEN ("opteron"), PROCESSOR_K8,
773 CPU_K8_FLAGS, 0, 0 },
774 { STRING_COMMA_LEN ("k8"), PROCESSOR_K8,
775 CPU_K8_FLAGS, 0, 0 },
776 { STRING_COMMA_LEN ("amdfam10"), PROCESSOR_AMDFAM10,
777 CPU_AMDFAM10_FLAGS, 0, 0 },
778 { STRING_COMMA_LEN ("bdver1"), PROCESSOR_BD,
779 CPU_BDVER1_FLAGS, 0, 0 },
780 { STRING_COMMA_LEN ("bdver2"), PROCESSOR_BD,
781 CPU_BDVER2_FLAGS, 0, 0 },
782 { STRING_COMMA_LEN ("bdver3"), PROCESSOR_BD,
783 CPU_BDVER3_FLAGS, 0, 0 },
784 { STRING_COMMA_LEN ("bdver4"), PROCESSOR_BD,
785 CPU_BDVER4_FLAGS, 0, 0 },
786 { STRING_COMMA_LEN ("btver1"), PROCESSOR_BT,
787 CPU_BTVER1_FLAGS, 0, 0 },
788 { STRING_COMMA_LEN ("btver2"), PROCESSOR_BT,
789 CPU_BTVER2_FLAGS, 0, 0 },
790 { STRING_COMMA_LEN (".8087"), PROCESSOR_UNKNOWN,
791 CPU_8087_FLAGS, 0, 0 },
792 { STRING_COMMA_LEN (".287"), PROCESSOR_UNKNOWN,
793 CPU_287_FLAGS, 0, 0 },
794 { STRING_COMMA_LEN (".387"), PROCESSOR_UNKNOWN,
795 CPU_387_FLAGS, 0, 0 },
796 { STRING_COMMA_LEN (".no87"), PROCESSOR_UNKNOWN,
797 CPU_ANY87_FLAGS, 0, 1 },
798 { STRING_COMMA_LEN (".mmx"), PROCESSOR_UNKNOWN,
799 CPU_MMX_FLAGS, 0, 0 },
800 { STRING_COMMA_LEN (".nommx"), PROCESSOR_UNKNOWN,
801 CPU_3DNOWA_FLAGS, 0, 1 },
802 { STRING_COMMA_LEN (".sse"), PROCESSOR_UNKNOWN,
803 CPU_SSE_FLAGS, 0, 0 },
804 { STRING_COMMA_LEN (".sse2"), PROCESSOR_UNKNOWN,
805 CPU_SSE2_FLAGS, 0, 0 },
806 { STRING_COMMA_LEN (".sse3"), PROCESSOR_UNKNOWN,
807 CPU_SSE3_FLAGS, 0, 0 },
808 { STRING_COMMA_LEN (".ssse3"), PROCESSOR_UNKNOWN,
809 CPU_SSSE3_FLAGS, 0, 0 },
810 { STRING_COMMA_LEN (".sse4.1"), PROCESSOR_UNKNOWN,
811 CPU_SSE4_1_FLAGS, 0, 0 },
812 { STRING_COMMA_LEN (".sse4.2"), PROCESSOR_UNKNOWN,
813 CPU_SSE4_2_FLAGS, 0, 0 },
814 { STRING_COMMA_LEN (".sse4"), PROCESSOR_UNKNOWN,
815 CPU_SSE4_2_FLAGS, 0, 0 },
816 { STRING_COMMA_LEN (".nosse"), PROCESSOR_UNKNOWN,
817 CPU_ANY_SSE_FLAGS, 0, 1 },
818 { STRING_COMMA_LEN (".avx"), PROCESSOR_UNKNOWN,
819 CPU_AVX_FLAGS, 0, 0 },
820 { STRING_COMMA_LEN (".avx2"), PROCESSOR_UNKNOWN,
821 CPU_AVX2_FLAGS, 0, 0 },
822 { STRING_COMMA_LEN (".avx512f"), PROCESSOR_UNKNOWN,
823 CPU_AVX512F_FLAGS, 0, 0 },
824 { STRING_COMMA_LEN (".avx512cd"), PROCESSOR_UNKNOWN,
825 CPU_AVX512CD_FLAGS, 0, 0 },
826 { STRING_COMMA_LEN (".avx512er"), PROCESSOR_UNKNOWN,
827 CPU_AVX512ER_FLAGS, 0, 0 },
828 { STRING_COMMA_LEN (".avx512pf"), PROCESSOR_UNKNOWN,
829 CPU_AVX512PF_FLAGS, 0, 0 },
830 { STRING_COMMA_LEN (".avx512dq"), PROCESSOR_UNKNOWN,
831 CPU_AVX512DQ_FLAGS, 0, 0 },
832 { STRING_COMMA_LEN (".avx512bw"), PROCESSOR_UNKNOWN,
833 CPU_AVX512BW_FLAGS, 0, 0 },
834 { STRING_COMMA_LEN (".avx512vl"), PROCESSOR_UNKNOWN,
835 CPU_AVX512VL_FLAGS, 0, 0 },
836 { STRING_COMMA_LEN (".noavx"), PROCESSOR_UNKNOWN,
837 CPU_ANY_AVX_FLAGS, 0, 1 },
838 { STRING_COMMA_LEN (".vmx"), PROCESSOR_UNKNOWN,
839 CPU_VMX_FLAGS, 0, 0 },
840 { STRING_COMMA_LEN (".vmfunc"), PROCESSOR_UNKNOWN,
841 CPU_VMFUNC_FLAGS, 0, 0 },
842 { STRING_COMMA_LEN (".smx"), PROCESSOR_UNKNOWN,
843 CPU_SMX_FLAGS, 0, 0 },
844 { STRING_COMMA_LEN (".xsave"), PROCESSOR_UNKNOWN,
845 CPU_XSAVE_FLAGS, 0, 0 },
846 { STRING_COMMA_LEN (".xsaveopt"), PROCESSOR_UNKNOWN,
847 CPU_XSAVEOPT_FLAGS, 0, 0 },
848 { STRING_COMMA_LEN (".xsavec"), PROCESSOR_UNKNOWN,
849 CPU_XSAVEC_FLAGS, 0, 0 },
850 { STRING_COMMA_LEN (".xsaves"), PROCESSOR_UNKNOWN,
851 CPU_XSAVES_FLAGS, 0, 0 },
852 { STRING_COMMA_LEN (".aes"), PROCESSOR_UNKNOWN,
853 CPU_AES_FLAGS, 0, 0 },
854 { STRING_COMMA_LEN (".pclmul"), PROCESSOR_UNKNOWN,
855 CPU_PCLMUL_FLAGS, 0, 0 },
856 { STRING_COMMA_LEN (".clmul"), PROCESSOR_UNKNOWN,
857 CPU_PCLMUL_FLAGS, 1, 0 },
858 { STRING_COMMA_LEN (".fsgsbase"), PROCESSOR_UNKNOWN,
859 CPU_FSGSBASE_FLAGS, 0, 0 },
860 { STRING_COMMA_LEN (".rdrnd"), PROCESSOR_UNKNOWN,
861 CPU_RDRND_FLAGS, 0, 0 },
862 { STRING_COMMA_LEN (".f16c"), PROCESSOR_UNKNOWN,
863 CPU_F16C_FLAGS, 0, 0 },
864 { STRING_COMMA_LEN (".bmi2"), PROCESSOR_UNKNOWN,
865 CPU_BMI2_FLAGS, 0, 0 },
866 { STRING_COMMA_LEN (".fma"), PROCESSOR_UNKNOWN,
867 CPU_FMA_FLAGS, 0, 0 },
868 { STRING_COMMA_LEN (".fma4"), PROCESSOR_UNKNOWN,
869 CPU_FMA4_FLAGS, 0, 0 },
870 { STRING_COMMA_LEN (".xop"), PROCESSOR_UNKNOWN,
871 CPU_XOP_FLAGS, 0, 0 },
872 { STRING_COMMA_LEN (".lwp"), PROCESSOR_UNKNOWN,
873 CPU_LWP_FLAGS, 0, 0 },
874 { STRING_COMMA_LEN (".movbe"), PROCESSOR_UNKNOWN,
875 CPU_MOVBE_FLAGS, 0, 0 },
876 { STRING_COMMA_LEN (".cx16"), PROCESSOR_UNKNOWN,
877 CPU_CX16_FLAGS, 0, 0 },
878 { STRING_COMMA_LEN (".ept"), PROCESSOR_UNKNOWN,
879 CPU_EPT_FLAGS, 0, 0 },
880 { STRING_COMMA_LEN (".lzcnt"), PROCESSOR_UNKNOWN,
881 CPU_LZCNT_FLAGS, 0, 0 },
882 { STRING_COMMA_LEN (".hle"), PROCESSOR_UNKNOWN,
883 CPU_HLE_FLAGS, 0, 0 },
884 { STRING_COMMA_LEN (".rtm"), PROCESSOR_UNKNOWN,
885 CPU_RTM_FLAGS, 0, 0 },
886 { STRING_COMMA_LEN (".invpcid"), PROCESSOR_UNKNOWN,
887 CPU_INVPCID_FLAGS, 0, 0 },
888 { STRING_COMMA_LEN (".clflush"), PROCESSOR_UNKNOWN,
889 CPU_CLFLUSH_FLAGS, 0, 0 },
890 { STRING_COMMA_LEN (".nop"), PROCESSOR_UNKNOWN,
891 CPU_NOP_FLAGS, 0, 0 },
892 { STRING_COMMA_LEN (".syscall"), PROCESSOR_UNKNOWN,
893 CPU_SYSCALL_FLAGS, 0, 0 },
894 { STRING_COMMA_LEN (".rdtscp"), PROCESSOR_UNKNOWN,
895 CPU_RDTSCP_FLAGS, 0, 0 },
896 { STRING_COMMA_LEN (".3dnow"), PROCESSOR_UNKNOWN,
897 CPU_3DNOW_FLAGS, 0, 0 },
898 { STRING_COMMA_LEN (".3dnowa"), PROCESSOR_UNKNOWN,
899 CPU_3DNOWA_FLAGS, 0, 0 },
900 { STRING_COMMA_LEN (".padlock"), PROCESSOR_UNKNOWN,
901 CPU_PADLOCK_FLAGS, 0, 0 },
902 { STRING_COMMA_LEN (".pacifica"), PROCESSOR_UNKNOWN,
903 CPU_SVME_FLAGS, 1, 0 },
904 { STRING_COMMA_LEN (".svme"), PROCESSOR_UNKNOWN,
905 CPU_SVME_FLAGS, 0, 0 },
906 { STRING_COMMA_LEN (".sse4a"), PROCESSOR_UNKNOWN,
907 CPU_SSE4A_FLAGS, 0, 0 },
908 { STRING_COMMA_LEN (".abm"), PROCESSOR_UNKNOWN,
909 CPU_ABM_FLAGS, 0, 0 },
910 { STRING_COMMA_LEN (".bmi"), PROCESSOR_UNKNOWN,
911 CPU_BMI_FLAGS, 0, 0 },
912 { STRING_COMMA_LEN (".tbm"), PROCESSOR_UNKNOWN,
913 CPU_TBM_FLAGS, 0, 0 },
914 { STRING_COMMA_LEN (".adx"), PROCESSOR_UNKNOWN,
915 CPU_ADX_FLAGS, 0, 0 },
916 { STRING_COMMA_LEN (".rdseed"), PROCESSOR_UNKNOWN,
917 CPU_RDSEED_FLAGS, 0, 0 },
918 { STRING_COMMA_LEN (".prfchw"), PROCESSOR_UNKNOWN,
919 CPU_PRFCHW_FLAGS, 0, 0 },
920 { STRING_COMMA_LEN (".smap"), PROCESSOR_UNKNOWN,
921 CPU_SMAP_FLAGS, 0, 0 },
922 { STRING_COMMA_LEN (".mpx"), PROCESSOR_UNKNOWN,
923 CPU_MPX_FLAGS, 0, 0 },
924 { STRING_COMMA_LEN (".sha"), PROCESSOR_UNKNOWN,
925 CPU_SHA_FLAGS, 0, 0 },
926 { STRING_COMMA_LEN (".clflushopt"), PROCESSOR_UNKNOWN,
927 CPU_CLFLUSHOPT_FLAGS, 0, 0 },
928 { STRING_COMMA_LEN (".prefetchwt1"), PROCESSOR_UNKNOWN,
929 CPU_PREFETCHWT1_FLAGS, 0, 0 },
930 { STRING_COMMA_LEN (".se1"), PROCESSOR_UNKNOWN,
931 CPU_SE1_FLAGS, 0, 0 },
932 { STRING_COMMA_LEN (".clwb"), PROCESSOR_UNKNOWN,
933 CPU_CLWB_FLAGS, 0, 0 },
934 };
935
936 #ifdef I386COFF
937 /* Like s_lcomm_internal in gas/read.c but the alignment string
938 is allowed to be optional. */
939
940 static symbolS *
941 pe_lcomm_internal (int needs_align, symbolS *symbolP, addressT size)
942 {
943 addressT align = 0;
944
945 SKIP_WHITESPACE ();
946
947 if (needs_align
948 && *input_line_pointer == ',')
949 {
950 align = parse_align (needs_align - 1);
951
952 if (align == (addressT) -1)
953 return NULL;
954 }
955 else
956 {
957 if (size >= 8)
958 align = 3;
959 else if (size >= 4)
960 align = 2;
961 else if (size >= 2)
962 align = 1;
963 else
964 align = 0;
965 }
966
967 bss_alloc (symbolP, size, align);
968 return symbolP;
969 }
970
971 static void
972 pe_lcomm (int needs_align)
973 {
974 s_comm_internal (needs_align * 2, pe_lcomm_internal);
975 }
976 #endif
977
978 const pseudo_typeS md_pseudo_table[] =
979 {
980 #if !defined(OBJ_AOUT) && !defined(USE_ALIGN_PTWO)
981 {"align", s_align_bytes, 0},
982 #else
983 {"align", s_align_ptwo, 0},
984 #endif
985 {"arch", set_cpu_arch, 0},
986 #ifndef I386COFF
987 {"bss", s_bss, 0},
988 #else
989 {"lcomm", pe_lcomm, 1},
990 #endif
991 {"ffloat", float_cons, 'f'},
992 {"dfloat", float_cons, 'd'},
993 {"tfloat", float_cons, 'x'},
994 {"value", cons, 2},
995 {"slong", signed_cons, 4},
996 {"noopt", s_ignore, 0},
997 {"optim", s_ignore, 0},
998 {"code16gcc", set_16bit_gcc_code_flag, CODE_16BIT},
999 {"code16", set_code_flag, CODE_16BIT},
1000 {"code32", set_code_flag, CODE_32BIT},
1001 {"code64", set_code_flag, CODE_64BIT},
1002 {"intel_syntax", set_intel_syntax, 1},
1003 {"att_syntax", set_intel_syntax, 0},
1004 {"intel_mnemonic", set_intel_mnemonic, 1},
1005 {"att_mnemonic", set_intel_mnemonic, 0},
1006 {"allow_index_reg", set_allow_index_reg, 1},
1007 {"disallow_index_reg", set_allow_index_reg, 0},
1008 {"sse_check", set_check, 0},
1009 {"operand_check", set_check, 1},
1010 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
1011 {"largecomm", handle_large_common, 0},
1012 #else
1013 {"file", (void (*) (int)) dwarf2_directive_file, 0},
1014 {"loc", dwarf2_directive_loc, 0},
1015 {"loc_mark_labels", dwarf2_directive_loc_mark_labels, 0},
1016 #endif
1017 #ifdef TE_PE
1018 {"secrel32", pe_directive_secrel, 0},
1019 #endif
1020 {0, 0, 0}
1021 };
1022
1023 /* For interface with expression (). */
1024 extern char *input_line_pointer;
1025
1026 /* Hash table for instruction mnemonic lookup. */
1027 static struct hash_control *op_hash;
1028
1029 /* Hash table for register lookup. */
1030 static struct hash_control *reg_hash;
1031 \f
1032 void
1033 i386_align_code (fragS *fragP, int count)
1034 {
1035 /* Various efficient no-op patterns for aligning code labels.
1036 Note: Don't try to assemble the instructions in the comments.
1037 0L and 0w are not legal. */
1038 static const char f32_1[] =
1039 {0x90}; /* nop */
1040 static const char f32_2[] =
1041 {0x66,0x90}; /* xchg %ax,%ax */
1042 static const char f32_3[] =
1043 {0x8d,0x76,0x00}; /* leal 0(%esi),%esi */
1044 static const char f32_4[] =
1045 {0x8d,0x74,0x26,0x00}; /* leal 0(%esi,1),%esi */
1046 static const char f32_5[] =
1047 {0x90, /* nop */
1048 0x8d,0x74,0x26,0x00}; /* leal 0(%esi,1),%esi */
1049 static const char f32_6[] =
1050 {0x8d,0xb6,0x00,0x00,0x00,0x00}; /* leal 0L(%esi),%esi */
1051 static const char f32_7[] =
1052 {0x8d,0xb4,0x26,0x00,0x00,0x00,0x00}; /* leal 0L(%esi,1),%esi */
1053 static const char f32_8[] =
1054 {0x90, /* nop */
1055 0x8d,0xb4,0x26,0x00,0x00,0x00,0x00}; /* leal 0L(%esi,1),%esi */
1056 static const char f32_9[] =
1057 {0x89,0xf6, /* movl %esi,%esi */
1058 0x8d,0xbc,0x27,0x00,0x00,0x00,0x00}; /* leal 0L(%edi,1),%edi */
1059 static const char f32_10[] =
1060 {0x8d,0x76,0x00, /* leal 0(%esi),%esi */
1061 0x8d,0xbc,0x27,0x00,0x00,0x00,0x00}; /* leal 0L(%edi,1),%edi */
1062 static const char f32_11[] =
1063 {0x8d,0x74,0x26,0x00, /* leal 0(%esi,1),%esi */
1064 0x8d,0xbc,0x27,0x00,0x00,0x00,0x00}; /* leal 0L(%edi,1),%edi */
1065 static const char f32_12[] =
1066 {0x8d,0xb6,0x00,0x00,0x00,0x00, /* leal 0L(%esi),%esi */
1067 0x8d,0xbf,0x00,0x00,0x00,0x00}; /* leal 0L(%edi),%edi */
1068 static const char f32_13[] =
1069 {0x8d,0xb6,0x00,0x00,0x00,0x00, /* leal 0L(%esi),%esi */
1070 0x8d,0xbc,0x27,0x00,0x00,0x00,0x00}; /* leal 0L(%edi,1),%edi */
1071 static const char f32_14[] =
1072 {0x8d,0xb4,0x26,0x00,0x00,0x00,0x00, /* leal 0L(%esi,1),%esi */
1073 0x8d,0xbc,0x27,0x00,0x00,0x00,0x00}; /* leal 0L(%edi,1),%edi */
1074 static const char f16_3[] =
1075 {0x8d,0x74,0x00}; /* lea 0(%esi),%esi */
1076 static const char f16_4[] =
1077 {0x8d,0xb4,0x00,0x00}; /* lea 0w(%si),%si */
1078 static const char f16_5[] =
1079 {0x90, /* nop */
1080 0x8d,0xb4,0x00,0x00}; /* lea 0w(%si),%si */
1081 static const char f16_6[] =
1082 {0x89,0xf6, /* mov %si,%si */
1083 0x8d,0xbd,0x00,0x00}; /* lea 0w(%di),%di */
1084 static const char f16_7[] =
1085 {0x8d,0x74,0x00, /* lea 0(%si),%si */
1086 0x8d,0xbd,0x00,0x00}; /* lea 0w(%di),%di */
1087 static const char f16_8[] =
1088 {0x8d,0xb4,0x00,0x00, /* lea 0w(%si),%si */
1089 0x8d,0xbd,0x00,0x00}; /* lea 0w(%di),%di */
1090 static const char jump_31[] =
1091 {0xeb,0x1d,0x90,0x90,0x90,0x90,0x90, /* jmp .+31; lotsa nops */
1092 0x90,0x90,0x90,0x90,0x90,0x90,0x90,0x90,
1093 0x90,0x90,0x90,0x90,0x90,0x90,0x90,0x90,
1094 0x90,0x90,0x90,0x90,0x90,0x90,0x90,0x90};
1095 static const char *const f32_patt[] = {
1096 f32_1, f32_2, f32_3, f32_4, f32_5, f32_6, f32_7, f32_8,
1097 f32_9, f32_10, f32_11, f32_12, f32_13, f32_14
1098 };
1099 static const char *const f16_patt[] = {
1100 f32_1, f32_2, f16_3, f16_4, f16_5, f16_6, f16_7, f16_8
1101 };
1102 /* nopl (%[re]ax) */
1103 static const char alt_3[] =
1104 {0x0f,0x1f,0x00};
1105 /* nopl 0(%[re]ax) */
1106 static const char alt_4[] =
1107 {0x0f,0x1f,0x40,0x00};
1108 /* nopl 0(%[re]ax,%[re]ax,1) */
1109 static const char alt_5[] =
1110 {0x0f,0x1f,0x44,0x00,0x00};
1111 /* nopw 0(%[re]ax,%[re]ax,1) */
1112 static const char alt_6[] =
1113 {0x66,0x0f,0x1f,0x44,0x00,0x00};
1114 /* nopl 0L(%[re]ax) */
1115 static const char alt_7[] =
1116 {0x0f,0x1f,0x80,0x00,0x00,0x00,0x00};
1117 /* nopl 0L(%[re]ax,%[re]ax,1) */
1118 static const char alt_8[] =
1119 {0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
1120 /* nopw 0L(%[re]ax,%[re]ax,1) */
1121 static const char alt_9[] =
1122 {0x66,0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
1123 /* nopw %cs:0L(%[re]ax,%[re]ax,1) */
1124 static const char alt_10[] =
1125 {0x66,0x2e,0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
1126 /* data16
1127 nopw %cs:0L(%[re]ax,%[re]ax,1) */
1128 static const char alt_long_11[] =
1129 {0x66,
1130 0x66,0x2e,0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
1131 /* data16
1132 data16
1133 nopw %cs:0L(%[re]ax,%[re]ax,1) */
1134 static const char alt_long_12[] =
1135 {0x66,
1136 0x66,
1137 0x66,0x2e,0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
1138 /* data16
1139 data16
1140 data16
1141 nopw %cs:0L(%[re]ax,%[re]ax,1) */
1142 static const char alt_long_13[] =
1143 {0x66,
1144 0x66,
1145 0x66,
1146 0x66,0x2e,0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
1147 /* data16
1148 data16
1149 data16
1150 data16
1151 nopw %cs:0L(%[re]ax,%[re]ax,1) */
1152 static const char alt_long_14[] =
1153 {0x66,
1154 0x66,
1155 0x66,
1156 0x66,
1157 0x66,0x2e,0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
1158 /* data16
1159 data16
1160 data16
1161 data16
1162 data16
1163 nopw %cs:0L(%[re]ax,%[re]ax,1) */
1164 static const char alt_long_15[] =
1165 {0x66,
1166 0x66,
1167 0x66,
1168 0x66,
1169 0x66,
1170 0x66,0x2e,0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
1171 /* nopl 0(%[re]ax,%[re]ax,1)
1172 nopw 0(%[re]ax,%[re]ax,1) */
1173 static const char alt_short_11[] =
1174 {0x0f,0x1f,0x44,0x00,0x00,
1175 0x66,0x0f,0x1f,0x44,0x00,0x00};
1176 /* nopw 0(%[re]ax,%[re]ax,1)
1177 nopw 0(%[re]ax,%[re]ax,1) */
1178 static const char alt_short_12[] =
1179 {0x66,0x0f,0x1f,0x44,0x00,0x00,
1180 0x66,0x0f,0x1f,0x44,0x00,0x00};
1181 /* nopw 0(%[re]ax,%[re]ax,1)
1182 nopl 0L(%[re]ax) */
1183 static const char alt_short_13[] =
1184 {0x66,0x0f,0x1f,0x44,0x00,0x00,
1185 0x0f,0x1f,0x80,0x00,0x00,0x00,0x00};
1186 /* nopl 0L(%[re]ax)
1187 nopl 0L(%[re]ax) */
1188 static const char alt_short_14[] =
1189 {0x0f,0x1f,0x80,0x00,0x00,0x00,0x00,
1190 0x0f,0x1f,0x80,0x00,0x00,0x00,0x00};
1191 /* nopl 0L(%[re]ax)
1192 nopl 0L(%[re]ax,%[re]ax,1) */
1193 static const char alt_short_15[] =
1194 {0x0f,0x1f,0x80,0x00,0x00,0x00,0x00,
1195 0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00};
1196 static const char *const alt_short_patt[] = {
1197 f32_1, f32_2, alt_3, alt_4, alt_5, alt_6, alt_7, alt_8,
1198 alt_9, alt_10, alt_short_11, alt_short_12, alt_short_13,
1199 alt_short_14, alt_short_15
1200 };
1201 static const char *const alt_long_patt[] = {
1202 f32_1, f32_2, alt_3, alt_4, alt_5, alt_6, alt_7, alt_8,
1203 alt_9, alt_10, alt_long_11, alt_long_12, alt_long_13,
1204 alt_long_14, alt_long_15
1205 };
1206
1207 /* Only align for at least a positive non-zero boundary. */
1208 if (count <= 0 || count > MAX_MEM_FOR_RS_ALIGN_CODE)
1209 return;
1210
1211 /* We need to decide which NOP sequence to use for 32bit and
1212 64bit. When -mtune= is used:
1213
1214 1. For PROCESSOR_I386, PROCESSOR_I486, PROCESSOR_PENTIUM and
1215 PROCESSOR_GENERIC32, f32_patt will be used.
1216 2. For PROCESSOR_PENTIUMPRO, PROCESSOR_PENTIUM4, PROCESSOR_NOCONA,
1217 PROCESSOR_CORE, PROCESSOR_CORE2, PROCESSOR_COREI7, and
1218 PROCESSOR_GENERIC64, alt_long_patt will be used.
1219 3. For PROCESSOR_ATHLON, PROCESSOR_K6, PROCESSOR_K8 and
1220 PROCESSOR_AMDFAM10, PROCESSOR_BD and PROCESSOR_BT, alt_short_patt
1221 will be used.
1222
1223 When -mtune= isn't used, alt_long_patt will be used if
1224 cpu_arch_isa_flags has CpuNop. Otherwise, f32_patt will
1225 be used.
1226
1227 When -march= or .arch is used, we can't use anything beyond
1228 cpu_arch_isa_flags. */
1229
1230 if (flag_code == CODE_16BIT)
1231 {
1232 if (count > 8)
1233 {
1234 memcpy (fragP->fr_literal + fragP->fr_fix,
1235 jump_31, count);
1236 /* Adjust jump offset. */
1237 fragP->fr_literal[fragP->fr_fix + 1] = count - 2;
1238 }
1239 else
1240 memcpy (fragP->fr_literal + fragP->fr_fix,
1241 f16_patt[count - 1], count);
1242 }
1243 else
1244 {
1245 const char *const *patt = NULL;
1246
1247 if (fragP->tc_frag_data.isa == PROCESSOR_UNKNOWN)
1248 {
1249 /* PROCESSOR_UNKNOWN means that all ISAs may be used. */
1250 switch (cpu_arch_tune)
1251 {
1252 case PROCESSOR_UNKNOWN:
1253 /* We use cpu_arch_isa_flags to check if we SHOULD
1254 optimize with nops. */
1255 if (fragP->tc_frag_data.isa_flags.bitfield.cpunop)
1256 patt = alt_long_patt;
1257 else
1258 patt = f32_patt;
1259 break;
1260 case PROCESSOR_PENTIUM4:
1261 case PROCESSOR_NOCONA:
1262 case PROCESSOR_CORE:
1263 case PROCESSOR_CORE2:
1264 case PROCESSOR_COREI7:
1265 case PROCESSOR_L1OM:
1266 case PROCESSOR_K1OM:
1267 case PROCESSOR_GENERIC64:
1268 patt = alt_long_patt;
1269 break;
1270 case PROCESSOR_K6:
1271 case PROCESSOR_ATHLON:
1272 case PROCESSOR_K8:
1273 case PROCESSOR_AMDFAM10:
1274 case PROCESSOR_BD:
1275 case PROCESSOR_BT:
1276 patt = alt_short_patt;
1277 break;
1278 case PROCESSOR_I386:
1279 case PROCESSOR_I486:
1280 case PROCESSOR_PENTIUM:
1281 case PROCESSOR_PENTIUMPRO:
1282 case PROCESSOR_GENERIC32:
1283 patt = f32_patt;
1284 break;
1285 }
1286 }
1287 else
1288 {
1289 switch (fragP->tc_frag_data.tune)
1290 {
1291 case PROCESSOR_UNKNOWN:
1292 /* When cpu_arch_isa is set, cpu_arch_tune shouldn't be
1293 PROCESSOR_UNKNOWN. */
1294 abort ();
1295 break;
1296
1297 case PROCESSOR_I386:
1298 case PROCESSOR_I486:
1299 case PROCESSOR_PENTIUM:
1300 case PROCESSOR_K6:
1301 case PROCESSOR_ATHLON:
1302 case PROCESSOR_K8:
1303 case PROCESSOR_AMDFAM10:
1304 case PROCESSOR_BD:
1305 case PROCESSOR_BT:
1306 case PROCESSOR_GENERIC32:
1307 /* We use cpu_arch_isa_flags to check if we CAN optimize
1308 with nops. */
1309 if (fragP->tc_frag_data.isa_flags.bitfield.cpunop)
1310 patt = alt_short_patt;
1311 else
1312 patt = f32_patt;
1313 break;
1314 case PROCESSOR_PENTIUMPRO:
1315 case PROCESSOR_PENTIUM4:
1316 case PROCESSOR_NOCONA:
1317 case PROCESSOR_CORE:
1318 case PROCESSOR_CORE2:
1319 case PROCESSOR_COREI7:
1320 case PROCESSOR_L1OM:
1321 case PROCESSOR_K1OM:
1322 if (fragP->tc_frag_data.isa_flags.bitfield.cpunop)
1323 patt = alt_long_patt;
1324 else
1325 patt = f32_patt;
1326 break;
1327 case PROCESSOR_GENERIC64:
1328 patt = alt_long_patt;
1329 break;
1330 }
1331 }
1332
1333 if (patt == f32_patt)
1334 {
1335 /* If the padding is less than 15 bytes, we use the normal
1336 ones. Otherwise, we use a jump instruction and adjust
1337 its offset. */
1338 int limit;
1339
1340 /* For 64bit, the limit is 3 bytes. */
1341 if (flag_code == CODE_64BIT
1342 && fragP->tc_frag_data.isa_flags.bitfield.cpulm)
1343 limit = 3;
1344 else
1345 limit = 15;
1346 if (count < limit)
1347 memcpy (fragP->fr_literal + fragP->fr_fix,
1348 patt[count - 1], count);
1349 else
1350 {
1351 memcpy (fragP->fr_literal + fragP->fr_fix,
1352 jump_31, count);
1353 /* Adjust jump offset. */
1354 fragP->fr_literal[fragP->fr_fix + 1] = count - 2;
1355 }
1356 }
1357 else
1358 {
1359 /* Maximum length of an instruction is 15 byte. If the
1360 padding is greater than 15 bytes and we don't use jump,
1361 we have to break it into smaller pieces. */
1362 int padding = count;
1363 while (padding > 15)
1364 {
1365 padding -= 15;
1366 memcpy (fragP->fr_literal + fragP->fr_fix + padding,
1367 patt [14], 15);
1368 }
1369
1370 if (padding)
1371 memcpy (fragP->fr_literal + fragP->fr_fix,
1372 patt [padding - 1], padding);
1373 }
1374 }
1375 fragP->fr_var = count;
1376 }
1377
1378 static INLINE int
1379 operand_type_all_zero (const union i386_operand_type *x)
1380 {
1381 switch (ARRAY_SIZE(x->array))
1382 {
1383 case 3:
1384 if (x->array[2])
1385 return 0;
1386 case 2:
1387 if (x->array[1])
1388 return 0;
1389 case 1:
1390 return !x->array[0];
1391 default:
1392 abort ();
1393 }
1394 }
1395
1396 static INLINE void
1397 operand_type_set (union i386_operand_type *x, unsigned int v)
1398 {
1399 switch (ARRAY_SIZE(x->array))
1400 {
1401 case 3:
1402 x->array[2] = v;
1403 case 2:
1404 x->array[1] = v;
1405 case 1:
1406 x->array[0] = v;
1407 break;
1408 default:
1409 abort ();
1410 }
1411 }
1412
1413 static INLINE int
1414 operand_type_equal (const union i386_operand_type *x,
1415 const union i386_operand_type *y)
1416 {
1417 switch (ARRAY_SIZE(x->array))
1418 {
1419 case 3:
1420 if (x->array[2] != y->array[2])
1421 return 0;
1422 case 2:
1423 if (x->array[1] != y->array[1])
1424 return 0;
1425 case 1:
1426 return x->array[0] == y->array[0];
1427 break;
1428 default:
1429 abort ();
1430 }
1431 }
1432
1433 static INLINE int
1434 cpu_flags_all_zero (const union i386_cpu_flags *x)
1435 {
1436 switch (ARRAY_SIZE(x->array))
1437 {
1438 case 3:
1439 if (x->array[2])
1440 return 0;
1441 case 2:
1442 if (x->array[1])
1443 return 0;
1444 case 1:
1445 return !x->array[0];
1446 default:
1447 abort ();
1448 }
1449 }
1450
1451 static INLINE void
1452 cpu_flags_set (union i386_cpu_flags *x, unsigned int v)
1453 {
1454 switch (ARRAY_SIZE(x->array))
1455 {
1456 case 3:
1457 x->array[2] = v;
1458 case 2:
1459 x->array[1] = v;
1460 case 1:
1461 x->array[0] = v;
1462 break;
1463 default:
1464 abort ();
1465 }
1466 }
1467
1468 static INLINE int
1469 cpu_flags_equal (const union i386_cpu_flags *x,
1470 const union i386_cpu_flags *y)
1471 {
1472 switch (ARRAY_SIZE(x->array))
1473 {
1474 case 3:
1475 if (x->array[2] != y->array[2])
1476 return 0;
1477 case 2:
1478 if (x->array[1] != y->array[1])
1479 return 0;
1480 case 1:
1481 return x->array[0] == y->array[0];
1482 break;
1483 default:
1484 abort ();
1485 }
1486 }
1487
1488 static INLINE int
1489 cpu_flags_check_cpu64 (i386_cpu_flags f)
1490 {
1491 return !((flag_code == CODE_64BIT && f.bitfield.cpuno64)
1492 || (flag_code != CODE_64BIT && f.bitfield.cpu64));
1493 }
1494
1495 static INLINE i386_cpu_flags
1496 cpu_flags_and (i386_cpu_flags x, i386_cpu_flags y)
1497 {
1498 switch (ARRAY_SIZE (x.array))
1499 {
1500 case 3:
1501 x.array [2] &= y.array [2];
1502 case 2:
1503 x.array [1] &= y.array [1];
1504 case 1:
1505 x.array [0] &= y.array [0];
1506 break;
1507 default:
1508 abort ();
1509 }
1510 return x;
1511 }
1512
1513 static INLINE i386_cpu_flags
1514 cpu_flags_or (i386_cpu_flags x, i386_cpu_flags y)
1515 {
1516 switch (ARRAY_SIZE (x.array))
1517 {
1518 case 3:
1519 x.array [2] |= y.array [2];
1520 case 2:
1521 x.array [1] |= y.array [1];
1522 case 1:
1523 x.array [0] |= y.array [0];
1524 break;
1525 default:
1526 abort ();
1527 }
1528 return x;
1529 }
1530
1531 static INLINE i386_cpu_flags
1532 cpu_flags_and_not (i386_cpu_flags x, i386_cpu_flags y)
1533 {
1534 switch (ARRAY_SIZE (x.array))
1535 {
1536 case 3:
1537 x.array [2] &= ~y.array [2];
1538 case 2:
1539 x.array [1] &= ~y.array [1];
1540 case 1:
1541 x.array [0] &= ~y.array [0];
1542 break;
1543 default:
1544 abort ();
1545 }
1546 return x;
1547 }
1548
1549 #define CPU_FLAGS_ARCH_MATCH 0x1
1550 #define CPU_FLAGS_64BIT_MATCH 0x2
1551 #define CPU_FLAGS_AES_MATCH 0x4
1552 #define CPU_FLAGS_PCLMUL_MATCH 0x8
1553 #define CPU_FLAGS_AVX_MATCH 0x10
1554
1555 #define CPU_FLAGS_32BIT_MATCH \
1556 (CPU_FLAGS_ARCH_MATCH | CPU_FLAGS_AES_MATCH \
1557 | CPU_FLAGS_PCLMUL_MATCH | CPU_FLAGS_AVX_MATCH)
1558 #define CPU_FLAGS_PERFECT_MATCH \
1559 (CPU_FLAGS_32BIT_MATCH | CPU_FLAGS_64BIT_MATCH)
1560
1561 /* Return CPU flags match bits. */
1562
1563 static int
1564 cpu_flags_match (const insn_template *t)
1565 {
1566 i386_cpu_flags x = t->cpu_flags;
1567 int match = cpu_flags_check_cpu64 (x) ? CPU_FLAGS_64BIT_MATCH : 0;
1568
1569 x.bitfield.cpu64 = 0;
1570 x.bitfield.cpuno64 = 0;
1571
1572 if (cpu_flags_all_zero (&x))
1573 {
1574 /* This instruction is available on all archs. */
1575 match |= CPU_FLAGS_32BIT_MATCH;
1576 }
1577 else
1578 {
1579 /* This instruction is available only on some archs. */
1580 i386_cpu_flags cpu = cpu_arch_flags;
1581
1582 cpu.bitfield.cpu64 = 0;
1583 cpu.bitfield.cpuno64 = 0;
1584 cpu = cpu_flags_and (x, cpu);
1585 if (!cpu_flags_all_zero (&cpu))
1586 {
1587 if (x.bitfield.cpuavx)
1588 {
1589 /* We only need to check AES/PCLMUL/SSE2AVX with AVX. */
1590 if (cpu.bitfield.cpuavx)
1591 {
1592 /* Check SSE2AVX. */
1593 if (!t->opcode_modifier.sse2avx|| sse2avx)
1594 {
1595 match |= (CPU_FLAGS_ARCH_MATCH
1596 | CPU_FLAGS_AVX_MATCH);
1597 /* Check AES. */
1598 if (!x.bitfield.cpuaes || cpu.bitfield.cpuaes)
1599 match |= CPU_FLAGS_AES_MATCH;
1600 /* Check PCLMUL. */
1601 if (!x.bitfield.cpupclmul
1602 || cpu.bitfield.cpupclmul)
1603 match |= CPU_FLAGS_PCLMUL_MATCH;
1604 }
1605 }
1606 else
1607 match |= CPU_FLAGS_ARCH_MATCH;
1608 }
1609 else
1610 match |= CPU_FLAGS_32BIT_MATCH;
1611 }
1612 }
1613 return match;
1614 }
1615
1616 static INLINE i386_operand_type
1617 operand_type_and (i386_operand_type x, i386_operand_type y)
1618 {
1619 switch (ARRAY_SIZE (x.array))
1620 {
1621 case 3:
1622 x.array [2] &= y.array [2];
1623 case 2:
1624 x.array [1] &= y.array [1];
1625 case 1:
1626 x.array [0] &= y.array [0];
1627 break;
1628 default:
1629 abort ();
1630 }
1631 return x;
1632 }
1633
1634 static INLINE i386_operand_type
1635 operand_type_or (i386_operand_type x, i386_operand_type y)
1636 {
1637 switch (ARRAY_SIZE (x.array))
1638 {
1639 case 3:
1640 x.array [2] |= y.array [2];
1641 case 2:
1642 x.array [1] |= y.array [1];
1643 case 1:
1644 x.array [0] |= y.array [0];
1645 break;
1646 default:
1647 abort ();
1648 }
1649 return x;
1650 }
1651
1652 static INLINE i386_operand_type
1653 operand_type_xor (i386_operand_type x, i386_operand_type y)
1654 {
1655 switch (ARRAY_SIZE (x.array))
1656 {
1657 case 3:
1658 x.array [2] ^= y.array [2];
1659 case 2:
1660 x.array [1] ^= y.array [1];
1661 case 1:
1662 x.array [0] ^= y.array [0];
1663 break;
1664 default:
1665 abort ();
1666 }
1667 return x;
1668 }
1669
1670 static const i386_operand_type acc32 = OPERAND_TYPE_ACC32;
1671 static const i386_operand_type acc64 = OPERAND_TYPE_ACC64;
1672 static const i386_operand_type control = OPERAND_TYPE_CONTROL;
1673 static const i386_operand_type inoutportreg
1674 = OPERAND_TYPE_INOUTPORTREG;
1675 static const i386_operand_type reg16_inoutportreg
1676 = OPERAND_TYPE_REG16_INOUTPORTREG;
1677 static const i386_operand_type disp16 = OPERAND_TYPE_DISP16;
1678 static const i386_operand_type disp32 = OPERAND_TYPE_DISP32;
1679 static const i386_operand_type disp32s = OPERAND_TYPE_DISP32S;
1680 static const i386_operand_type disp16_32 = OPERAND_TYPE_DISP16_32;
1681 static const i386_operand_type anydisp
1682 = OPERAND_TYPE_ANYDISP;
1683 static const i386_operand_type regxmm = OPERAND_TYPE_REGXMM;
1684 static const i386_operand_type regymm = OPERAND_TYPE_REGYMM;
1685 static const i386_operand_type regzmm = OPERAND_TYPE_REGZMM;
1686 static const i386_operand_type regmask = OPERAND_TYPE_REGMASK;
1687 static const i386_operand_type imm8 = OPERAND_TYPE_IMM8;
1688 static const i386_operand_type imm8s = OPERAND_TYPE_IMM8S;
1689 static const i386_operand_type imm16 = OPERAND_TYPE_IMM16;
1690 static const i386_operand_type imm32 = OPERAND_TYPE_IMM32;
1691 static const i386_operand_type imm32s = OPERAND_TYPE_IMM32S;
1692 static const i386_operand_type imm64 = OPERAND_TYPE_IMM64;
1693 static const i386_operand_type imm16_32 = OPERAND_TYPE_IMM16_32;
1694 static const i386_operand_type imm16_32s = OPERAND_TYPE_IMM16_32S;
1695 static const i386_operand_type imm16_32_32s = OPERAND_TYPE_IMM16_32_32S;
1696 static const i386_operand_type vec_imm4 = OPERAND_TYPE_VEC_IMM4;
1697
1698 enum operand_type
1699 {
1700 reg,
1701 imm,
1702 disp,
1703 anymem
1704 };
1705
1706 static INLINE int
1707 operand_type_check (i386_operand_type t, enum operand_type c)
1708 {
1709 switch (c)
1710 {
1711 case reg:
1712 return (t.bitfield.reg8
1713 || t.bitfield.reg16
1714 || t.bitfield.reg32
1715 || t.bitfield.reg64);
1716
1717 case imm:
1718 return (t.bitfield.imm8
1719 || t.bitfield.imm8s
1720 || t.bitfield.imm16
1721 || t.bitfield.imm32
1722 || t.bitfield.imm32s
1723 || t.bitfield.imm64);
1724
1725 case disp:
1726 return (t.bitfield.disp8
1727 || t.bitfield.disp16
1728 || t.bitfield.disp32
1729 || t.bitfield.disp32s
1730 || t.bitfield.disp64);
1731
1732 case anymem:
1733 return (t.bitfield.disp8
1734 || t.bitfield.disp16
1735 || t.bitfield.disp32
1736 || t.bitfield.disp32s
1737 || t.bitfield.disp64
1738 || t.bitfield.baseindex);
1739
1740 default:
1741 abort ();
1742 }
1743
1744 return 0;
1745 }
1746
1747 /* Return 1 if there is no conflict in 8bit/16bit/32bit/64bit on
1748 operand J for instruction template T. */
1749
1750 static INLINE int
1751 match_reg_size (const insn_template *t, unsigned int j)
1752 {
1753 return !((i.types[j].bitfield.byte
1754 && !t->operand_types[j].bitfield.byte)
1755 || (i.types[j].bitfield.word
1756 && !t->operand_types[j].bitfield.word)
1757 || (i.types[j].bitfield.dword
1758 && !t->operand_types[j].bitfield.dword)
1759 || (i.types[j].bitfield.qword
1760 && !t->operand_types[j].bitfield.qword));
1761 }
1762
1763 /* Return 1 if there is no conflict in any size on operand J for
1764 instruction template T. */
1765
1766 static INLINE int
1767 match_mem_size (const insn_template *t, unsigned int j)
1768 {
1769 return (match_reg_size (t, j)
1770 && !((i.types[j].bitfield.unspecified
1771 && !t->operand_types[j].bitfield.unspecified)
1772 || (i.types[j].bitfield.fword
1773 && !t->operand_types[j].bitfield.fword)
1774 || (i.types[j].bitfield.tbyte
1775 && !t->operand_types[j].bitfield.tbyte)
1776 || (i.types[j].bitfield.xmmword
1777 && !t->operand_types[j].bitfield.xmmword)
1778 || (i.types[j].bitfield.ymmword
1779 && !t->operand_types[j].bitfield.ymmword)
1780 || (i.types[j].bitfield.zmmword
1781 && !t->operand_types[j].bitfield.zmmword)));
1782 }
1783
1784 /* Return 1 if there is no size conflict on any operands for
1785 instruction template T. */
1786
1787 static INLINE int
1788 operand_size_match (const insn_template *t)
1789 {
1790 unsigned int j;
1791 int match = 1;
1792
1793 /* Don't check jump instructions. */
1794 if (t->opcode_modifier.jump
1795 || t->opcode_modifier.jumpbyte
1796 || t->opcode_modifier.jumpdword
1797 || t->opcode_modifier.jumpintersegment)
1798 return match;
1799
1800 /* Check memory and accumulator operand size. */
1801 for (j = 0; j < i.operands; j++)
1802 {
1803 if (t->operand_types[j].bitfield.anysize)
1804 continue;
1805
1806 if (t->operand_types[j].bitfield.acc && !match_reg_size (t, j))
1807 {
1808 match = 0;
1809 break;
1810 }
1811
1812 if (i.types[j].bitfield.mem && !match_mem_size (t, j))
1813 {
1814 match = 0;
1815 break;
1816 }
1817 }
1818
1819 if (match)
1820 return match;
1821 else if (!t->opcode_modifier.d && !t->opcode_modifier.floatd)
1822 {
1823 mismatch:
1824 i.error = operand_size_mismatch;
1825 return 0;
1826 }
1827
1828 /* Check reverse. */
1829 gas_assert (i.operands == 2);
1830
1831 match = 1;
1832 for (j = 0; j < 2; j++)
1833 {
1834 if (t->operand_types[j].bitfield.acc
1835 && !match_reg_size (t, j ? 0 : 1))
1836 goto mismatch;
1837
1838 if (i.types[j].bitfield.mem
1839 && !match_mem_size (t, j ? 0 : 1))
1840 goto mismatch;
1841 }
1842
1843 return match;
1844 }
1845
1846 static INLINE int
1847 operand_type_match (i386_operand_type overlap,
1848 i386_operand_type given)
1849 {
1850 i386_operand_type temp = overlap;
1851
1852 temp.bitfield.jumpabsolute = 0;
1853 temp.bitfield.unspecified = 0;
1854 temp.bitfield.byte = 0;
1855 temp.bitfield.word = 0;
1856 temp.bitfield.dword = 0;
1857 temp.bitfield.fword = 0;
1858 temp.bitfield.qword = 0;
1859 temp.bitfield.tbyte = 0;
1860 temp.bitfield.xmmword = 0;
1861 temp.bitfield.ymmword = 0;
1862 temp.bitfield.zmmword = 0;
1863 if (operand_type_all_zero (&temp))
1864 goto mismatch;
1865
1866 if (given.bitfield.baseindex == overlap.bitfield.baseindex
1867 && given.bitfield.jumpabsolute == overlap.bitfield.jumpabsolute)
1868 return 1;
1869
1870 mismatch:
1871 i.error = operand_type_mismatch;
1872 return 0;
1873 }
1874
1875 /* If given types g0 and g1 are registers they must be of the same type
1876 unless the expected operand type register overlap is null.
1877 Note that Acc in a template matches every size of reg. */
1878
1879 static INLINE int
1880 operand_type_register_match (i386_operand_type m0,
1881 i386_operand_type g0,
1882 i386_operand_type t0,
1883 i386_operand_type m1,
1884 i386_operand_type g1,
1885 i386_operand_type t1)
1886 {
1887 if (!operand_type_check (g0, reg))
1888 return 1;
1889
1890 if (!operand_type_check (g1, reg))
1891 return 1;
1892
1893 if (g0.bitfield.reg8 == g1.bitfield.reg8
1894 && g0.bitfield.reg16 == g1.bitfield.reg16
1895 && g0.bitfield.reg32 == g1.bitfield.reg32
1896 && g0.bitfield.reg64 == g1.bitfield.reg64)
1897 return 1;
1898
1899 if (m0.bitfield.acc)
1900 {
1901 t0.bitfield.reg8 = 1;
1902 t0.bitfield.reg16 = 1;
1903 t0.bitfield.reg32 = 1;
1904 t0.bitfield.reg64 = 1;
1905 }
1906
1907 if (m1.bitfield.acc)
1908 {
1909 t1.bitfield.reg8 = 1;
1910 t1.bitfield.reg16 = 1;
1911 t1.bitfield.reg32 = 1;
1912 t1.bitfield.reg64 = 1;
1913 }
1914
1915 if (!(t0.bitfield.reg8 & t1.bitfield.reg8)
1916 && !(t0.bitfield.reg16 & t1.bitfield.reg16)
1917 && !(t0.bitfield.reg32 & t1.bitfield.reg32)
1918 && !(t0.bitfield.reg64 & t1.bitfield.reg64))
1919 return 1;
1920
1921 i.error = register_type_mismatch;
1922
1923 return 0;
1924 }
1925
1926 static INLINE unsigned int
1927 register_number (const reg_entry *r)
1928 {
1929 unsigned int nr = r->reg_num;
1930
1931 if (r->reg_flags & RegRex)
1932 nr += 8;
1933
1934 return nr;
1935 }
1936
1937 static INLINE unsigned int
1938 mode_from_disp_size (i386_operand_type t)
1939 {
1940 if (t.bitfield.disp8 || t.bitfield.vec_disp8)
1941 return 1;
1942 else if (t.bitfield.disp16
1943 || t.bitfield.disp32
1944 || t.bitfield.disp32s)
1945 return 2;
1946 else
1947 return 0;
1948 }
1949
1950 static INLINE int
1951 fits_in_signed_byte (addressT num)
1952 {
1953 return num + 0x80 <= 0xff;
1954 }
1955
1956 static INLINE int
1957 fits_in_unsigned_byte (addressT num)
1958 {
1959 return num <= 0xff;
1960 }
1961
1962 static INLINE int
1963 fits_in_unsigned_word (addressT num)
1964 {
1965 return num <= 0xffff;
1966 }
1967
1968 static INLINE int
1969 fits_in_signed_word (addressT num)
1970 {
1971 return num + 0x8000 <= 0xffff;
1972 }
1973
1974 static INLINE int
1975 fits_in_signed_long (addressT num ATTRIBUTE_UNUSED)
1976 {
1977 #ifndef BFD64
1978 return 1;
1979 #else
1980 return num + 0x80000000 <= 0xffffffff;
1981 #endif
1982 } /* fits_in_signed_long() */
1983
1984 static INLINE int
1985 fits_in_unsigned_long (addressT num ATTRIBUTE_UNUSED)
1986 {
1987 #ifndef BFD64
1988 return 1;
1989 #else
1990 return num <= 0xffffffff;
1991 #endif
1992 } /* fits_in_unsigned_long() */
1993
1994 static INLINE int
1995 fits_in_vec_disp8 (offsetT num)
1996 {
1997 int shift = i.memshift;
1998 unsigned int mask;
1999
2000 if (shift == -1)
2001 abort ();
2002
2003 mask = (1 << shift) - 1;
2004
2005 /* Return 0 if NUM isn't properly aligned. */
2006 if ((num & mask))
2007 return 0;
2008
2009 /* Check if NUM will fit in 8bit after shift. */
2010 return fits_in_signed_byte (num >> shift);
2011 }
2012
2013 static INLINE int
2014 fits_in_imm4 (offsetT num)
2015 {
2016 return (num & 0xf) == num;
2017 }
2018
2019 static i386_operand_type
2020 smallest_imm_type (offsetT num)
2021 {
2022 i386_operand_type t;
2023
2024 operand_type_set (&t, 0);
2025 t.bitfield.imm64 = 1;
2026
2027 if (cpu_arch_tune != PROCESSOR_I486 && num == 1)
2028 {
2029 /* This code is disabled on the 486 because all the Imm1 forms
2030 in the opcode table are slower on the i486. They're the
2031 versions with the implicitly specified single-position
2032 displacement, which has another syntax if you really want to
2033 use that form. */
2034 t.bitfield.imm1 = 1;
2035 t.bitfield.imm8 = 1;
2036 t.bitfield.imm8s = 1;
2037 t.bitfield.imm16 = 1;
2038 t.bitfield.imm32 = 1;
2039 t.bitfield.imm32s = 1;
2040 }
2041 else if (fits_in_signed_byte (num))
2042 {
2043 t.bitfield.imm8 = 1;
2044 t.bitfield.imm8s = 1;
2045 t.bitfield.imm16 = 1;
2046 t.bitfield.imm32 = 1;
2047 t.bitfield.imm32s = 1;
2048 }
2049 else if (fits_in_unsigned_byte (num))
2050 {
2051 t.bitfield.imm8 = 1;
2052 t.bitfield.imm16 = 1;
2053 t.bitfield.imm32 = 1;
2054 t.bitfield.imm32s = 1;
2055 }
2056 else if (fits_in_signed_word (num) || fits_in_unsigned_word (num))
2057 {
2058 t.bitfield.imm16 = 1;
2059 t.bitfield.imm32 = 1;
2060 t.bitfield.imm32s = 1;
2061 }
2062 else if (fits_in_signed_long (num))
2063 {
2064 t.bitfield.imm32 = 1;
2065 t.bitfield.imm32s = 1;
2066 }
2067 else if (fits_in_unsigned_long (num))
2068 t.bitfield.imm32 = 1;
2069
2070 return t;
2071 }
2072
2073 static offsetT
2074 offset_in_range (offsetT val, int size)
2075 {
2076 addressT mask;
2077
2078 switch (size)
2079 {
2080 case 1: mask = ((addressT) 1 << 8) - 1; break;
2081 case 2: mask = ((addressT) 1 << 16) - 1; break;
2082 case 4: mask = ((addressT) 2 << 31) - 1; break;
2083 #ifdef BFD64
2084 case 8: mask = ((addressT) 2 << 63) - 1; break;
2085 #endif
2086 default: abort ();
2087 }
2088
2089 #ifdef BFD64
2090 /* If BFD64, sign extend val for 32bit address mode. */
2091 if (flag_code != CODE_64BIT
2092 || i.prefix[ADDR_PREFIX])
2093 if ((val & ~(((addressT) 2 << 31) - 1)) == 0)
2094 val = (val ^ ((addressT) 1 << 31)) - ((addressT) 1 << 31);
2095 #endif
2096
2097 if ((val & ~mask) != 0 && (val & ~mask) != ~mask)
2098 {
2099 char buf1[40], buf2[40];
2100
2101 sprint_value (buf1, val);
2102 sprint_value (buf2, val & mask);
2103 as_warn (_("%s shortened to %s"), buf1, buf2);
2104 }
2105 return val & mask;
2106 }
2107
2108 enum PREFIX_GROUP
2109 {
2110 PREFIX_EXIST = 0,
2111 PREFIX_LOCK,
2112 PREFIX_REP,
2113 PREFIX_OTHER
2114 };
2115
2116 /* Returns
2117 a. PREFIX_EXIST if attempting to add a prefix where one from the
2118 same class already exists.
2119 b. PREFIX_LOCK if lock prefix is added.
2120 c. PREFIX_REP if rep/repne prefix is added.
2121 d. PREFIX_OTHER if other prefix is added.
2122 */
2123
2124 static enum PREFIX_GROUP
2125 add_prefix (unsigned int prefix)
2126 {
2127 enum PREFIX_GROUP ret = PREFIX_OTHER;
2128 unsigned int q;
2129
2130 if (prefix >= REX_OPCODE && prefix < REX_OPCODE + 16
2131 && flag_code == CODE_64BIT)
2132 {
2133 if ((i.prefix[REX_PREFIX] & prefix & REX_W)
2134 || ((i.prefix[REX_PREFIX] & (REX_R | REX_X | REX_B))
2135 && (prefix & (REX_R | REX_X | REX_B))))
2136 ret = PREFIX_EXIST;
2137 q = REX_PREFIX;
2138 }
2139 else
2140 {
2141 switch (prefix)
2142 {
2143 default:
2144 abort ();
2145
2146 case CS_PREFIX_OPCODE:
2147 case DS_PREFIX_OPCODE:
2148 case ES_PREFIX_OPCODE:
2149 case FS_PREFIX_OPCODE:
2150 case GS_PREFIX_OPCODE:
2151 case SS_PREFIX_OPCODE:
2152 q = SEG_PREFIX;
2153 break;
2154
2155 case REPNE_PREFIX_OPCODE:
2156 case REPE_PREFIX_OPCODE:
2157 q = REP_PREFIX;
2158 ret = PREFIX_REP;
2159 break;
2160
2161 case LOCK_PREFIX_OPCODE:
2162 q = LOCK_PREFIX;
2163 ret = PREFIX_LOCK;
2164 break;
2165
2166 case FWAIT_OPCODE:
2167 q = WAIT_PREFIX;
2168 break;
2169
2170 case ADDR_PREFIX_OPCODE:
2171 q = ADDR_PREFIX;
2172 break;
2173
2174 case DATA_PREFIX_OPCODE:
2175 q = DATA_PREFIX;
2176 break;
2177 }
2178 if (i.prefix[q] != 0)
2179 ret = PREFIX_EXIST;
2180 }
2181
2182 if (ret)
2183 {
2184 if (!i.prefix[q])
2185 ++i.prefixes;
2186 i.prefix[q] |= prefix;
2187 }
2188 else
2189 as_bad (_("same type of prefix used twice"));
2190
2191 return ret;
2192 }
2193
2194 static void
2195 update_code_flag (int value, int check)
2196 {
2197 PRINTF_LIKE ((*as_error));
2198
2199 flag_code = (enum flag_code) value;
2200 if (flag_code == CODE_64BIT)
2201 {
2202 cpu_arch_flags.bitfield.cpu64 = 1;
2203 cpu_arch_flags.bitfield.cpuno64 = 0;
2204 }
2205 else
2206 {
2207 cpu_arch_flags.bitfield.cpu64 = 0;
2208 cpu_arch_flags.bitfield.cpuno64 = 1;
2209 }
2210 if (value == CODE_64BIT && !cpu_arch_flags.bitfield.cpulm )
2211 {
2212 if (check)
2213 as_error = as_fatal;
2214 else
2215 as_error = as_bad;
2216 (*as_error) (_("64bit mode not supported on `%s'."),
2217 cpu_arch_name ? cpu_arch_name : default_arch);
2218 }
2219 if (value == CODE_32BIT && !cpu_arch_flags.bitfield.cpui386)
2220 {
2221 if (check)
2222 as_error = as_fatal;
2223 else
2224 as_error = as_bad;
2225 (*as_error) (_("32bit mode not supported on `%s'."),
2226 cpu_arch_name ? cpu_arch_name : default_arch);
2227 }
2228 stackop_size = '\0';
2229 }
2230
2231 static void
2232 set_code_flag (int value)
2233 {
2234 update_code_flag (value, 0);
2235 }
2236
2237 static void
2238 set_16bit_gcc_code_flag (int new_code_flag)
2239 {
2240 flag_code = (enum flag_code) new_code_flag;
2241 if (flag_code != CODE_16BIT)
2242 abort ();
2243 cpu_arch_flags.bitfield.cpu64 = 0;
2244 cpu_arch_flags.bitfield.cpuno64 = 1;
2245 stackop_size = LONG_MNEM_SUFFIX;
2246 }
2247
2248 static void
2249 set_intel_syntax (int syntax_flag)
2250 {
2251 /* Find out if register prefixing is specified. */
2252 int ask_naked_reg = 0;
2253
2254 SKIP_WHITESPACE ();
2255 if (!is_end_of_line[(unsigned char) *input_line_pointer])
2256 {
2257 char *string = input_line_pointer;
2258 int e = get_symbol_end ();
2259
2260 if (strcmp (string, "prefix") == 0)
2261 ask_naked_reg = 1;
2262 else if (strcmp (string, "noprefix") == 0)
2263 ask_naked_reg = -1;
2264 else
2265 as_bad (_("bad argument to syntax directive."));
2266 *input_line_pointer = e;
2267 }
2268 demand_empty_rest_of_line ();
2269
2270 intel_syntax = syntax_flag;
2271
2272 if (ask_naked_reg == 0)
2273 allow_naked_reg = (intel_syntax
2274 && (bfd_get_symbol_leading_char (stdoutput) != '\0'));
2275 else
2276 allow_naked_reg = (ask_naked_reg < 0);
2277
2278 expr_set_rank (O_full_ptr, syntax_flag ? 10 : 0);
2279
2280 identifier_chars['%'] = intel_syntax && allow_naked_reg ? '%' : 0;
2281 identifier_chars['$'] = intel_syntax ? '$' : 0;
2282 register_prefix = allow_naked_reg ? "" : "%";
2283 }
2284
2285 static void
2286 set_intel_mnemonic (int mnemonic_flag)
2287 {
2288 intel_mnemonic = mnemonic_flag;
2289 }
2290
2291 static void
2292 set_allow_index_reg (int flag)
2293 {
2294 allow_index_reg = flag;
2295 }
2296
2297 static void
2298 set_check (int what)
2299 {
2300 enum check_kind *kind;
2301 const char *str;
2302
2303 if (what)
2304 {
2305 kind = &operand_check;
2306 str = "operand";
2307 }
2308 else
2309 {
2310 kind = &sse_check;
2311 str = "sse";
2312 }
2313
2314 SKIP_WHITESPACE ();
2315
2316 if (!is_end_of_line[(unsigned char) *input_line_pointer])
2317 {
2318 char *string = input_line_pointer;
2319 int e = get_symbol_end ();
2320
2321 if (strcmp (string, "none") == 0)
2322 *kind = check_none;
2323 else if (strcmp (string, "warning") == 0)
2324 *kind = check_warning;
2325 else if (strcmp (string, "error") == 0)
2326 *kind = check_error;
2327 else
2328 as_bad (_("bad argument to %s_check directive."), str);
2329 *input_line_pointer = e;
2330 }
2331 else
2332 as_bad (_("missing argument for %s_check directive"), str);
2333
2334 demand_empty_rest_of_line ();
2335 }
2336
2337 static void
2338 check_cpu_arch_compatible (const char *name ATTRIBUTE_UNUSED,
2339 i386_cpu_flags new_flag ATTRIBUTE_UNUSED)
2340 {
2341 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
2342 static const char *arch;
2343
2344 /* Intel LIOM is only supported on ELF. */
2345 if (!IS_ELF)
2346 return;
2347
2348 if (!arch)
2349 {
2350 /* Use cpu_arch_name if it is set in md_parse_option. Otherwise
2351 use default_arch. */
2352 arch = cpu_arch_name;
2353 if (!arch)
2354 arch = default_arch;
2355 }
2356
2357 /* If we are targeting Intel L1OM, we must enable it. */
2358 if (get_elf_backend_data (stdoutput)->elf_machine_code != EM_L1OM
2359 || new_flag.bitfield.cpul1om)
2360 return;
2361
2362 /* If we are targeting Intel K1OM, we must enable it. */
2363 if (get_elf_backend_data (stdoutput)->elf_machine_code != EM_K1OM
2364 || new_flag.bitfield.cpuk1om)
2365 return;
2366
2367 as_bad (_("`%s' is not supported on `%s'"), name, arch);
2368 #endif
2369 }
2370
2371 static void
2372 set_cpu_arch (int dummy ATTRIBUTE_UNUSED)
2373 {
2374 SKIP_WHITESPACE ();
2375
2376 if (!is_end_of_line[(unsigned char) *input_line_pointer])
2377 {
2378 char *string = input_line_pointer;
2379 int e = get_symbol_end ();
2380 unsigned int j;
2381 i386_cpu_flags flags;
2382
2383 for (j = 0; j < ARRAY_SIZE (cpu_arch); j++)
2384 {
2385 if (strcmp (string, cpu_arch[j].name) == 0)
2386 {
2387 check_cpu_arch_compatible (string, cpu_arch[j].flags);
2388
2389 if (*string != '.')
2390 {
2391 cpu_arch_name = cpu_arch[j].name;
2392 cpu_sub_arch_name = NULL;
2393 cpu_arch_flags = cpu_arch[j].flags;
2394 if (flag_code == CODE_64BIT)
2395 {
2396 cpu_arch_flags.bitfield.cpu64 = 1;
2397 cpu_arch_flags.bitfield.cpuno64 = 0;
2398 }
2399 else
2400 {
2401 cpu_arch_flags.bitfield.cpu64 = 0;
2402 cpu_arch_flags.bitfield.cpuno64 = 1;
2403 }
2404 cpu_arch_isa = cpu_arch[j].type;
2405 cpu_arch_isa_flags = cpu_arch[j].flags;
2406 if (!cpu_arch_tune_set)
2407 {
2408 cpu_arch_tune = cpu_arch_isa;
2409 cpu_arch_tune_flags = cpu_arch_isa_flags;
2410 }
2411 break;
2412 }
2413
2414 if (!cpu_arch[j].negated)
2415 flags = cpu_flags_or (cpu_arch_flags,
2416 cpu_arch[j].flags);
2417 else
2418 flags = cpu_flags_and_not (cpu_arch_flags,
2419 cpu_arch[j].flags);
2420 if (!cpu_flags_equal (&flags, &cpu_arch_flags))
2421 {
2422 if (cpu_sub_arch_name)
2423 {
2424 char *name = cpu_sub_arch_name;
2425 cpu_sub_arch_name = concat (name,
2426 cpu_arch[j].name,
2427 (const char *) NULL);
2428 free (name);
2429 }
2430 else
2431 cpu_sub_arch_name = xstrdup (cpu_arch[j].name);
2432 cpu_arch_flags = flags;
2433 cpu_arch_isa_flags = flags;
2434 }
2435 *input_line_pointer = e;
2436 demand_empty_rest_of_line ();
2437 return;
2438 }
2439 }
2440 if (j >= ARRAY_SIZE (cpu_arch))
2441 as_bad (_("no such architecture: `%s'"), string);
2442
2443 *input_line_pointer = e;
2444 }
2445 else
2446 as_bad (_("missing cpu architecture"));
2447
2448 no_cond_jump_promotion = 0;
2449 if (*input_line_pointer == ','
2450 && !is_end_of_line[(unsigned char) input_line_pointer[1]])
2451 {
2452 char *string = ++input_line_pointer;
2453 int e = get_symbol_end ();
2454
2455 if (strcmp (string, "nojumps") == 0)
2456 no_cond_jump_promotion = 1;
2457 else if (strcmp (string, "jumps") == 0)
2458 ;
2459 else
2460 as_bad (_("no such architecture modifier: `%s'"), string);
2461
2462 *input_line_pointer = e;
2463 }
2464
2465 demand_empty_rest_of_line ();
2466 }
2467
2468 enum bfd_architecture
2469 i386_arch (void)
2470 {
2471 if (cpu_arch_isa == PROCESSOR_L1OM)
2472 {
2473 if (OUTPUT_FLAVOR != bfd_target_elf_flavour
2474 || flag_code != CODE_64BIT)
2475 as_fatal (_("Intel L1OM is 64bit ELF only"));
2476 return bfd_arch_l1om;
2477 }
2478 else if (cpu_arch_isa == PROCESSOR_K1OM)
2479 {
2480 if (OUTPUT_FLAVOR != bfd_target_elf_flavour
2481 || flag_code != CODE_64BIT)
2482 as_fatal (_("Intel K1OM is 64bit ELF only"));
2483 return bfd_arch_k1om;
2484 }
2485 else
2486 return bfd_arch_i386;
2487 }
2488
2489 unsigned long
2490 i386_mach (void)
2491 {
2492 if (!strncmp (default_arch, "x86_64", 6))
2493 {
2494 if (cpu_arch_isa == PROCESSOR_L1OM)
2495 {
2496 if (OUTPUT_FLAVOR != bfd_target_elf_flavour
2497 || default_arch[6] != '\0')
2498 as_fatal (_("Intel L1OM is 64bit ELF only"));
2499 return bfd_mach_l1om;
2500 }
2501 else if (cpu_arch_isa == PROCESSOR_K1OM)
2502 {
2503 if (OUTPUT_FLAVOR != bfd_target_elf_flavour
2504 || default_arch[6] != '\0')
2505 as_fatal (_("Intel K1OM is 64bit ELF only"));
2506 return bfd_mach_k1om;
2507 }
2508 else if (default_arch[6] == '\0')
2509 return bfd_mach_x86_64;
2510 else
2511 return bfd_mach_x64_32;
2512 }
2513 else if (!strcmp (default_arch, "i386"))
2514 return bfd_mach_i386_i386;
2515 else
2516 as_fatal (_("unknown architecture"));
2517 }
2518 \f
2519 void
2520 md_begin (void)
2521 {
2522 const char *hash_err;
2523
2524 /* Initialize op_hash hash table. */
2525 op_hash = hash_new ();
2526
2527 {
2528 const insn_template *optab;
2529 templates *core_optab;
2530
2531 /* Setup for loop. */
2532 optab = i386_optab;
2533 core_optab = (templates *) xmalloc (sizeof (templates));
2534 core_optab->start = optab;
2535
2536 while (1)
2537 {
2538 ++optab;
2539 if (optab->name == NULL
2540 || strcmp (optab->name, (optab - 1)->name) != 0)
2541 {
2542 /* different name --> ship out current template list;
2543 add to hash table; & begin anew. */
2544 core_optab->end = optab;
2545 hash_err = hash_insert (op_hash,
2546 (optab - 1)->name,
2547 (void *) core_optab);
2548 if (hash_err)
2549 {
2550 as_fatal (_("can't hash %s: %s"),
2551 (optab - 1)->name,
2552 hash_err);
2553 }
2554 if (optab->name == NULL)
2555 break;
2556 core_optab = (templates *) xmalloc (sizeof (templates));
2557 core_optab->start = optab;
2558 }
2559 }
2560 }
2561
2562 /* Initialize reg_hash hash table. */
2563 reg_hash = hash_new ();
2564 {
2565 const reg_entry *regtab;
2566 unsigned int regtab_size = i386_regtab_size;
2567
2568 for (regtab = i386_regtab; regtab_size--; regtab++)
2569 {
2570 hash_err = hash_insert (reg_hash, regtab->reg_name, (void *) regtab);
2571 if (hash_err)
2572 as_fatal (_("can't hash %s: %s"),
2573 regtab->reg_name,
2574 hash_err);
2575 }
2576 }
2577
2578 /* Fill in lexical tables: mnemonic_chars, operand_chars. */
2579 {
2580 int c;
2581 char *p;
2582
2583 for (c = 0; c < 256; c++)
2584 {
2585 if (ISDIGIT (c))
2586 {
2587 digit_chars[c] = c;
2588 mnemonic_chars[c] = c;
2589 register_chars[c] = c;
2590 operand_chars[c] = c;
2591 }
2592 else if (ISLOWER (c))
2593 {
2594 mnemonic_chars[c] = c;
2595 register_chars[c] = c;
2596 operand_chars[c] = c;
2597 }
2598 else if (ISUPPER (c))
2599 {
2600 mnemonic_chars[c] = TOLOWER (c);
2601 register_chars[c] = mnemonic_chars[c];
2602 operand_chars[c] = c;
2603 }
2604 else if (c == '{' || c == '}')
2605 operand_chars[c] = c;
2606
2607 if (ISALPHA (c) || ISDIGIT (c))
2608 identifier_chars[c] = c;
2609 else if (c >= 128)
2610 {
2611 identifier_chars[c] = c;
2612 operand_chars[c] = c;
2613 }
2614 }
2615
2616 #ifdef LEX_AT
2617 identifier_chars['@'] = '@';
2618 #endif
2619 #ifdef LEX_QM
2620 identifier_chars['?'] = '?';
2621 operand_chars['?'] = '?';
2622 #endif
2623 digit_chars['-'] = '-';
2624 mnemonic_chars['_'] = '_';
2625 mnemonic_chars['-'] = '-';
2626 mnemonic_chars['.'] = '.';
2627 identifier_chars['_'] = '_';
2628 identifier_chars['.'] = '.';
2629
2630 for (p = operand_special_chars; *p != '\0'; p++)
2631 operand_chars[(unsigned char) *p] = *p;
2632 }
2633
2634 if (flag_code == CODE_64BIT)
2635 {
2636 #if defined (OBJ_COFF) && defined (TE_PE)
2637 x86_dwarf2_return_column = (OUTPUT_FLAVOR == bfd_target_coff_flavour
2638 ? 32 : 16);
2639 #else
2640 x86_dwarf2_return_column = 16;
2641 #endif
2642 x86_cie_data_alignment = -8;
2643 }
2644 else
2645 {
2646 x86_dwarf2_return_column = 8;
2647 x86_cie_data_alignment = -4;
2648 }
2649 }
2650
2651 void
2652 i386_print_statistics (FILE *file)
2653 {
2654 hash_print_statistics (file, "i386 opcode", op_hash);
2655 hash_print_statistics (file, "i386 register", reg_hash);
2656 }
2657 \f
2658 #ifdef DEBUG386
2659
2660 /* Debugging routines for md_assemble. */
2661 static void pte (insn_template *);
2662 static void pt (i386_operand_type);
2663 static void pe (expressionS *);
2664 static void ps (symbolS *);
2665
2666 static void
2667 pi (char *line, i386_insn *x)
2668 {
2669 unsigned int j;
2670
2671 fprintf (stdout, "%s: template ", line);
2672 pte (&x->tm);
2673 fprintf (stdout, " address: base %s index %s scale %x\n",
2674 x->base_reg ? x->base_reg->reg_name : "none",
2675 x->index_reg ? x->index_reg->reg_name : "none",
2676 x->log2_scale_factor);
2677 fprintf (stdout, " modrm: mode %x reg %x reg/mem %x\n",
2678 x->rm.mode, x->rm.reg, x->rm.regmem);
2679 fprintf (stdout, " sib: base %x index %x scale %x\n",
2680 x->sib.base, x->sib.index, x->sib.scale);
2681 fprintf (stdout, " rex: 64bit %x extX %x extY %x extZ %x\n",
2682 (x->rex & REX_W) != 0,
2683 (x->rex & REX_R) != 0,
2684 (x->rex & REX_X) != 0,
2685 (x->rex & REX_B) != 0);
2686 for (j = 0; j < x->operands; j++)
2687 {
2688 fprintf (stdout, " #%d: ", j + 1);
2689 pt (x->types[j]);
2690 fprintf (stdout, "\n");
2691 if (x->types[j].bitfield.reg8
2692 || x->types[j].bitfield.reg16
2693 || x->types[j].bitfield.reg32
2694 || x->types[j].bitfield.reg64
2695 || x->types[j].bitfield.regmmx
2696 || x->types[j].bitfield.regxmm
2697 || x->types[j].bitfield.regymm
2698 || x->types[j].bitfield.regzmm
2699 || x->types[j].bitfield.sreg2
2700 || x->types[j].bitfield.sreg3
2701 || x->types[j].bitfield.control
2702 || x->types[j].bitfield.debug
2703 || x->types[j].bitfield.test)
2704 fprintf (stdout, "%s\n", x->op[j].regs->reg_name);
2705 if (operand_type_check (x->types[j], imm))
2706 pe (x->op[j].imms);
2707 if (operand_type_check (x->types[j], disp))
2708 pe (x->op[j].disps);
2709 }
2710 }
2711
2712 static void
2713 pte (insn_template *t)
2714 {
2715 unsigned int j;
2716 fprintf (stdout, " %d operands ", t->operands);
2717 fprintf (stdout, "opcode %x ", t->base_opcode);
2718 if (t->extension_opcode != None)
2719 fprintf (stdout, "ext %x ", t->extension_opcode);
2720 if (t->opcode_modifier.d)
2721 fprintf (stdout, "D");
2722 if (t->opcode_modifier.w)
2723 fprintf (stdout, "W");
2724 fprintf (stdout, "\n");
2725 for (j = 0; j < t->operands; j++)
2726 {
2727 fprintf (stdout, " #%d type ", j + 1);
2728 pt (t->operand_types[j]);
2729 fprintf (stdout, "\n");
2730 }
2731 }
2732
2733 static void
2734 pe (expressionS *e)
2735 {
2736 fprintf (stdout, " operation %d\n", e->X_op);
2737 fprintf (stdout, " add_number %ld (%lx)\n",
2738 (long) e->X_add_number, (long) e->X_add_number);
2739 if (e->X_add_symbol)
2740 {
2741 fprintf (stdout, " add_symbol ");
2742 ps (e->X_add_symbol);
2743 fprintf (stdout, "\n");
2744 }
2745 if (e->X_op_symbol)
2746 {
2747 fprintf (stdout, " op_symbol ");
2748 ps (e->X_op_symbol);
2749 fprintf (stdout, "\n");
2750 }
2751 }
2752
2753 static void
2754 ps (symbolS *s)
2755 {
2756 fprintf (stdout, "%s type %s%s",
2757 S_GET_NAME (s),
2758 S_IS_EXTERNAL (s) ? "EXTERNAL " : "",
2759 segment_name (S_GET_SEGMENT (s)));
2760 }
2761
2762 static struct type_name
2763 {
2764 i386_operand_type mask;
2765 const char *name;
2766 }
2767 const type_names[] =
2768 {
2769 { OPERAND_TYPE_REG8, "r8" },
2770 { OPERAND_TYPE_REG16, "r16" },
2771 { OPERAND_TYPE_REG32, "r32" },
2772 { OPERAND_TYPE_REG64, "r64" },
2773 { OPERAND_TYPE_IMM8, "i8" },
2774 { OPERAND_TYPE_IMM8, "i8s" },
2775 { OPERAND_TYPE_IMM16, "i16" },
2776 { OPERAND_TYPE_IMM32, "i32" },
2777 { OPERAND_TYPE_IMM32S, "i32s" },
2778 { OPERAND_TYPE_IMM64, "i64" },
2779 { OPERAND_TYPE_IMM1, "i1" },
2780 { OPERAND_TYPE_BASEINDEX, "BaseIndex" },
2781 { OPERAND_TYPE_DISP8, "d8" },
2782 { OPERAND_TYPE_DISP16, "d16" },
2783 { OPERAND_TYPE_DISP32, "d32" },
2784 { OPERAND_TYPE_DISP32S, "d32s" },
2785 { OPERAND_TYPE_DISP64, "d64" },
2786 { OPERAND_TYPE_VEC_DISP8, "Vector d8" },
2787 { OPERAND_TYPE_INOUTPORTREG, "InOutPortReg" },
2788 { OPERAND_TYPE_SHIFTCOUNT, "ShiftCount" },
2789 { OPERAND_TYPE_CONTROL, "control reg" },
2790 { OPERAND_TYPE_TEST, "test reg" },
2791 { OPERAND_TYPE_DEBUG, "debug reg" },
2792 { OPERAND_TYPE_FLOATREG, "FReg" },
2793 { OPERAND_TYPE_FLOATACC, "FAcc" },
2794 { OPERAND_TYPE_SREG2, "SReg2" },
2795 { OPERAND_TYPE_SREG3, "SReg3" },
2796 { OPERAND_TYPE_ACC, "Acc" },
2797 { OPERAND_TYPE_JUMPABSOLUTE, "Jump Absolute" },
2798 { OPERAND_TYPE_REGMMX, "rMMX" },
2799 { OPERAND_TYPE_REGXMM, "rXMM" },
2800 { OPERAND_TYPE_REGYMM, "rYMM" },
2801 { OPERAND_TYPE_REGZMM, "rZMM" },
2802 { OPERAND_TYPE_REGMASK, "Mask reg" },
2803 { OPERAND_TYPE_ESSEG, "es" },
2804 };
2805
2806 static void
2807 pt (i386_operand_type t)
2808 {
2809 unsigned int j;
2810 i386_operand_type a;
2811
2812 for (j = 0; j < ARRAY_SIZE (type_names); j++)
2813 {
2814 a = operand_type_and (t, type_names[j].mask);
2815 if (!operand_type_all_zero (&a))
2816 fprintf (stdout, "%s, ", type_names[j].name);
2817 }
2818 fflush (stdout);
2819 }
2820
2821 #endif /* DEBUG386 */
2822 \f
2823 static bfd_reloc_code_real_type
2824 reloc (unsigned int size,
2825 int pcrel,
2826 int sign,
2827 int bnd_prefix,
2828 bfd_reloc_code_real_type other)
2829 {
2830 if (other != NO_RELOC)
2831 {
2832 reloc_howto_type *rel;
2833
2834 if (size == 8)
2835 switch (other)
2836 {
2837 case BFD_RELOC_X86_64_GOT32:
2838 return BFD_RELOC_X86_64_GOT64;
2839 break;
2840 case BFD_RELOC_X86_64_GOTPLT64:
2841 return BFD_RELOC_X86_64_GOTPLT64;
2842 break;
2843 case BFD_RELOC_X86_64_PLTOFF64:
2844 return BFD_RELOC_X86_64_PLTOFF64;
2845 break;
2846 case BFD_RELOC_X86_64_GOTPC32:
2847 other = BFD_RELOC_X86_64_GOTPC64;
2848 break;
2849 case BFD_RELOC_X86_64_GOTPCREL:
2850 other = BFD_RELOC_X86_64_GOTPCREL64;
2851 break;
2852 case BFD_RELOC_X86_64_TPOFF32:
2853 other = BFD_RELOC_X86_64_TPOFF64;
2854 break;
2855 case BFD_RELOC_X86_64_DTPOFF32:
2856 other = BFD_RELOC_X86_64_DTPOFF64;
2857 break;
2858 default:
2859 break;
2860 }
2861
2862 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
2863 if (other == BFD_RELOC_SIZE32)
2864 {
2865 if (size == 8)
2866 other = BFD_RELOC_SIZE64;
2867 if (pcrel)
2868 {
2869 as_bad (_("there are no pc-relative size relocations"));
2870 return NO_RELOC;
2871 }
2872 }
2873 #endif
2874
2875 /* Sign-checking 4-byte relocations in 16-/32-bit code is pointless. */
2876 if (size == 4 && (flag_code != CODE_64BIT || disallow_64bit_reloc))
2877 sign = -1;
2878
2879 rel = bfd_reloc_type_lookup (stdoutput, other);
2880 if (!rel)
2881 as_bad (_("unknown relocation (%u)"), other);
2882 else if (size != bfd_get_reloc_size (rel))
2883 as_bad (_("%u-byte relocation cannot be applied to %u-byte field"),
2884 bfd_get_reloc_size (rel),
2885 size);
2886 else if (pcrel && !rel->pc_relative)
2887 as_bad (_("non-pc-relative relocation for pc-relative field"));
2888 else if ((rel->complain_on_overflow == complain_overflow_signed
2889 && !sign)
2890 || (rel->complain_on_overflow == complain_overflow_unsigned
2891 && sign > 0))
2892 as_bad (_("relocated field and relocation type differ in signedness"));
2893 else
2894 return other;
2895 return NO_RELOC;
2896 }
2897
2898 if (pcrel)
2899 {
2900 if (!sign)
2901 as_bad (_("there are no unsigned pc-relative relocations"));
2902 switch (size)
2903 {
2904 case 1: return BFD_RELOC_8_PCREL;
2905 case 2: return BFD_RELOC_16_PCREL;
2906 case 4: return (bnd_prefix && object_64bit
2907 ? BFD_RELOC_X86_64_PC32_BND
2908 : BFD_RELOC_32_PCREL);
2909 case 8: return BFD_RELOC_64_PCREL;
2910 }
2911 as_bad (_("cannot do %u byte pc-relative relocation"), size);
2912 }
2913 else
2914 {
2915 if (sign > 0)
2916 switch (size)
2917 {
2918 case 4: return BFD_RELOC_X86_64_32S;
2919 }
2920 else
2921 switch (size)
2922 {
2923 case 1: return BFD_RELOC_8;
2924 case 2: return BFD_RELOC_16;
2925 case 4: return BFD_RELOC_32;
2926 case 8: return BFD_RELOC_64;
2927 }
2928 as_bad (_("cannot do %s %u byte relocation"),
2929 sign > 0 ? "signed" : "unsigned", size);
2930 }
2931
2932 return NO_RELOC;
2933 }
2934
2935 /* Here we decide which fixups can be adjusted to make them relative to
2936 the beginning of the section instead of the symbol. Basically we need
2937 to make sure that the dynamic relocations are done correctly, so in
2938 some cases we force the original symbol to be used. */
2939
2940 int
2941 tc_i386_fix_adjustable (fixS *fixP ATTRIBUTE_UNUSED)
2942 {
2943 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
2944 if (!IS_ELF)
2945 return 1;
2946
2947 /* Don't adjust pc-relative references to merge sections in 64-bit
2948 mode. */
2949 if (use_rela_relocations
2950 && (S_GET_SEGMENT (fixP->fx_addsy)->flags & SEC_MERGE) != 0
2951 && fixP->fx_pcrel)
2952 return 0;
2953
2954 /* The x86_64 GOTPCREL are represented as 32bit PCrel relocations
2955 and changed later by validate_fix. */
2956 if (GOT_symbol && fixP->fx_subsy == GOT_symbol
2957 && fixP->fx_r_type == BFD_RELOC_32_PCREL)
2958 return 0;
2959
2960 /* Adjust_reloc_syms doesn't know about the GOT. Need to keep symbol
2961 for size relocations. */
2962 if (fixP->fx_r_type == BFD_RELOC_SIZE32
2963 || fixP->fx_r_type == BFD_RELOC_SIZE64
2964 || fixP->fx_r_type == BFD_RELOC_386_GOTOFF
2965 || fixP->fx_r_type == BFD_RELOC_386_PLT32
2966 || fixP->fx_r_type == BFD_RELOC_386_GOT32
2967 || fixP->fx_r_type == BFD_RELOC_386_TLS_GD
2968 || fixP->fx_r_type == BFD_RELOC_386_TLS_LDM
2969 || fixP->fx_r_type == BFD_RELOC_386_TLS_LDO_32
2970 || fixP->fx_r_type == BFD_RELOC_386_TLS_IE_32
2971 || fixP->fx_r_type == BFD_RELOC_386_TLS_IE
2972 || fixP->fx_r_type == BFD_RELOC_386_TLS_GOTIE
2973 || fixP->fx_r_type == BFD_RELOC_386_TLS_LE_32
2974 || fixP->fx_r_type == BFD_RELOC_386_TLS_LE
2975 || fixP->fx_r_type == BFD_RELOC_386_TLS_GOTDESC
2976 || fixP->fx_r_type == BFD_RELOC_386_TLS_DESC_CALL
2977 || fixP->fx_r_type == BFD_RELOC_X86_64_PLT32
2978 || fixP->fx_r_type == BFD_RELOC_X86_64_GOT32
2979 || fixP->fx_r_type == BFD_RELOC_X86_64_GOTPCREL
2980 || fixP->fx_r_type == BFD_RELOC_X86_64_TLSGD
2981 || fixP->fx_r_type == BFD_RELOC_X86_64_TLSLD
2982 || fixP->fx_r_type == BFD_RELOC_X86_64_DTPOFF32
2983 || fixP->fx_r_type == BFD_RELOC_X86_64_DTPOFF64
2984 || fixP->fx_r_type == BFD_RELOC_X86_64_GOTTPOFF
2985 || fixP->fx_r_type == BFD_RELOC_X86_64_TPOFF32
2986 || fixP->fx_r_type == BFD_RELOC_X86_64_TPOFF64
2987 || fixP->fx_r_type == BFD_RELOC_X86_64_GOTOFF64
2988 || fixP->fx_r_type == BFD_RELOC_X86_64_GOTPC32_TLSDESC
2989 || fixP->fx_r_type == BFD_RELOC_X86_64_TLSDESC_CALL
2990 || fixP->fx_r_type == BFD_RELOC_VTABLE_INHERIT
2991 || fixP->fx_r_type == BFD_RELOC_VTABLE_ENTRY)
2992 return 0;
2993 #endif
2994 return 1;
2995 }
2996
2997 static int
2998 intel_float_operand (const char *mnemonic)
2999 {
3000 /* Note that the value returned is meaningful only for opcodes with (memory)
3001 operands, hence the code here is free to improperly handle opcodes that
3002 have no operands (for better performance and smaller code). */
3003
3004 if (mnemonic[0] != 'f')
3005 return 0; /* non-math */
3006
3007 switch (mnemonic[1])
3008 {
3009 /* fclex, fdecstp, fdisi, femms, feni, fincstp, finit, fsetpm, and
3010 the fs segment override prefix not currently handled because no
3011 call path can make opcodes without operands get here */
3012 case 'i':
3013 return 2 /* integer op */;
3014 case 'l':
3015 if (mnemonic[2] == 'd' && (mnemonic[3] == 'c' || mnemonic[3] == 'e'))
3016 return 3; /* fldcw/fldenv */
3017 break;
3018 case 'n':
3019 if (mnemonic[2] != 'o' /* fnop */)
3020 return 3; /* non-waiting control op */
3021 break;
3022 case 'r':
3023 if (mnemonic[2] == 's')
3024 return 3; /* frstor/frstpm */
3025 break;
3026 case 's':
3027 if (mnemonic[2] == 'a')
3028 return 3; /* fsave */
3029 if (mnemonic[2] == 't')
3030 {
3031 switch (mnemonic[3])
3032 {
3033 case 'c': /* fstcw */
3034 case 'd': /* fstdw */
3035 case 'e': /* fstenv */
3036 case 's': /* fsts[gw] */
3037 return 3;
3038 }
3039 }
3040 break;
3041 case 'x':
3042 if (mnemonic[2] == 'r' || mnemonic[2] == 's')
3043 return 0; /* fxsave/fxrstor are not really math ops */
3044 break;
3045 }
3046
3047 return 1;
3048 }
3049
3050 /* Build the VEX prefix. */
3051
3052 static void
3053 build_vex_prefix (const insn_template *t)
3054 {
3055 unsigned int register_specifier;
3056 unsigned int implied_prefix;
3057 unsigned int vector_length;
3058
3059 /* Check register specifier. */
3060 if (i.vex.register_specifier)
3061 {
3062 register_specifier =
3063 ~register_number (i.vex.register_specifier) & 0xf;
3064 gas_assert ((i.vex.register_specifier->reg_flags & RegVRex) == 0);
3065 }
3066 else
3067 register_specifier = 0xf;
3068
3069 /* Use 2-byte VEX prefix by swappping destination and source
3070 operand. */
3071 if (!i.swap_operand
3072 && i.operands == i.reg_operands
3073 && i.tm.opcode_modifier.vexopcode == VEX0F
3074 && i.tm.opcode_modifier.s
3075 && i.rex == REX_B)
3076 {
3077 unsigned int xchg = i.operands - 1;
3078 union i386_op temp_op;
3079 i386_operand_type temp_type;
3080
3081 temp_type = i.types[xchg];
3082 i.types[xchg] = i.types[0];
3083 i.types[0] = temp_type;
3084 temp_op = i.op[xchg];
3085 i.op[xchg] = i.op[0];
3086 i.op[0] = temp_op;
3087
3088 gas_assert (i.rm.mode == 3);
3089
3090 i.rex = REX_R;
3091 xchg = i.rm.regmem;
3092 i.rm.regmem = i.rm.reg;
3093 i.rm.reg = xchg;
3094
3095 /* Use the next insn. */
3096 i.tm = t[1];
3097 }
3098
3099 if (i.tm.opcode_modifier.vex == VEXScalar)
3100 vector_length = avxscalar;
3101 else
3102 vector_length = i.tm.opcode_modifier.vex == VEX256 ? 1 : 0;
3103
3104 switch ((i.tm.base_opcode >> 8) & 0xff)
3105 {
3106 case 0:
3107 implied_prefix = 0;
3108 break;
3109 case DATA_PREFIX_OPCODE:
3110 implied_prefix = 1;
3111 break;
3112 case REPE_PREFIX_OPCODE:
3113 implied_prefix = 2;
3114 break;
3115 case REPNE_PREFIX_OPCODE:
3116 implied_prefix = 3;
3117 break;
3118 default:
3119 abort ();
3120 }
3121
3122 /* Use 2-byte VEX prefix if possible. */
3123 if (i.tm.opcode_modifier.vexopcode == VEX0F
3124 && i.tm.opcode_modifier.vexw != VEXW1
3125 && (i.rex & (REX_W | REX_X | REX_B)) == 0)
3126 {
3127 /* 2-byte VEX prefix. */
3128 unsigned int r;
3129
3130 i.vex.length = 2;
3131 i.vex.bytes[0] = 0xc5;
3132
3133 /* Check the REX.R bit. */
3134 r = (i.rex & REX_R) ? 0 : 1;
3135 i.vex.bytes[1] = (r << 7
3136 | register_specifier << 3
3137 | vector_length << 2
3138 | implied_prefix);
3139 }
3140 else
3141 {
3142 /* 3-byte VEX prefix. */
3143 unsigned int m, w;
3144
3145 i.vex.length = 3;
3146
3147 switch (i.tm.opcode_modifier.vexopcode)
3148 {
3149 case VEX0F:
3150 m = 0x1;
3151 i.vex.bytes[0] = 0xc4;
3152 break;
3153 case VEX0F38:
3154 m = 0x2;
3155 i.vex.bytes[0] = 0xc4;
3156 break;
3157 case VEX0F3A:
3158 m = 0x3;
3159 i.vex.bytes[0] = 0xc4;
3160 break;
3161 case XOP08:
3162 m = 0x8;
3163 i.vex.bytes[0] = 0x8f;
3164 break;
3165 case XOP09:
3166 m = 0x9;
3167 i.vex.bytes[0] = 0x8f;
3168 break;
3169 case XOP0A:
3170 m = 0xa;
3171 i.vex.bytes[0] = 0x8f;
3172 break;
3173 default:
3174 abort ();
3175 }
3176
3177 /* The high 3 bits of the second VEX byte are 1's compliment
3178 of RXB bits from REX. */
3179 i.vex.bytes[1] = (~i.rex & 0x7) << 5 | m;
3180
3181 /* Check the REX.W bit. */
3182 w = (i.rex & REX_W) ? 1 : 0;
3183 if (i.tm.opcode_modifier.vexw == VEXW1)
3184 w = 1;
3185
3186 i.vex.bytes[2] = (w << 7
3187 | register_specifier << 3
3188 | vector_length << 2
3189 | implied_prefix);
3190 }
3191 }
3192
3193 /* Build the EVEX prefix. */
3194
3195 static void
3196 build_evex_prefix (void)
3197 {
3198 unsigned int register_specifier;
3199 unsigned int implied_prefix;
3200 unsigned int m, w;
3201 rex_byte vrex_used = 0;
3202
3203 /* Check register specifier. */
3204 if (i.vex.register_specifier)
3205 {
3206 gas_assert ((i.vrex & REX_X) == 0);
3207
3208 register_specifier = i.vex.register_specifier->reg_num;
3209 if ((i.vex.register_specifier->reg_flags & RegRex))
3210 register_specifier += 8;
3211 /* The upper 16 registers are encoded in the fourth byte of the
3212 EVEX prefix. */
3213 if (!(i.vex.register_specifier->reg_flags & RegVRex))
3214 i.vex.bytes[3] = 0x8;
3215 register_specifier = ~register_specifier & 0xf;
3216 }
3217 else
3218 {
3219 register_specifier = 0xf;
3220
3221 /* Encode upper 16 vector index register in the fourth byte of
3222 the EVEX prefix. */
3223 if (!(i.vrex & REX_X))
3224 i.vex.bytes[3] = 0x8;
3225 else
3226 vrex_used |= REX_X;
3227 }
3228
3229 switch ((i.tm.base_opcode >> 8) & 0xff)
3230 {
3231 case 0:
3232 implied_prefix = 0;
3233 break;
3234 case DATA_PREFIX_OPCODE:
3235 implied_prefix = 1;
3236 break;
3237 case REPE_PREFIX_OPCODE:
3238 implied_prefix = 2;
3239 break;
3240 case REPNE_PREFIX_OPCODE:
3241 implied_prefix = 3;
3242 break;
3243 default:
3244 abort ();
3245 }
3246
3247 /* 4 byte EVEX prefix. */
3248 i.vex.length = 4;
3249 i.vex.bytes[0] = 0x62;
3250
3251 /* mmmm bits. */
3252 switch (i.tm.opcode_modifier.vexopcode)
3253 {
3254 case VEX0F:
3255 m = 1;
3256 break;
3257 case VEX0F38:
3258 m = 2;
3259 break;
3260 case VEX0F3A:
3261 m = 3;
3262 break;
3263 default:
3264 abort ();
3265 break;
3266 }
3267
3268 /* The high 3 bits of the second EVEX byte are 1's compliment of RXB
3269 bits from REX. */
3270 i.vex.bytes[1] = (~i.rex & 0x7) << 5 | m;
3271
3272 /* The fifth bit of the second EVEX byte is 1's compliment of the
3273 REX_R bit in VREX. */
3274 if (!(i.vrex & REX_R))
3275 i.vex.bytes[1] |= 0x10;
3276 else
3277 vrex_used |= REX_R;
3278
3279 if ((i.reg_operands + i.imm_operands) == i.operands)
3280 {
3281 /* When all operands are registers, the REX_X bit in REX is not
3282 used. We reuse it to encode the upper 16 registers, which is
3283 indicated by the REX_B bit in VREX. The REX_X bit is encoded
3284 as 1's compliment. */
3285 if ((i.vrex & REX_B))
3286 {
3287 vrex_used |= REX_B;
3288 i.vex.bytes[1] &= ~0x40;
3289 }
3290 }
3291
3292 /* EVEX instructions shouldn't need the REX prefix. */
3293 i.vrex &= ~vrex_used;
3294 gas_assert (i.vrex == 0);
3295
3296 /* Check the REX.W bit. */
3297 w = (i.rex & REX_W) ? 1 : 0;
3298 if (i.tm.opcode_modifier.vexw)
3299 {
3300 if (i.tm.opcode_modifier.vexw == VEXW1)
3301 w = 1;
3302 }
3303 /* If w is not set it means we are dealing with WIG instruction. */
3304 else if (!w)
3305 {
3306 if (evexwig == evexw1)
3307 w = 1;
3308 }
3309
3310 /* Encode the U bit. */
3311 implied_prefix |= 0x4;
3312
3313 /* The third byte of the EVEX prefix. */
3314 i.vex.bytes[2] = (w << 7 | register_specifier << 3 | implied_prefix);
3315
3316 /* The fourth byte of the EVEX prefix. */
3317 /* The zeroing-masking bit. */
3318 if (i.mask && i.mask->zeroing)
3319 i.vex.bytes[3] |= 0x80;
3320
3321 /* Don't always set the broadcast bit if there is no RC. */
3322 if (!i.rounding)
3323 {
3324 /* Encode the vector length. */
3325 unsigned int vec_length;
3326
3327 switch (i.tm.opcode_modifier.evex)
3328 {
3329 case EVEXLIG: /* LL' is ignored */
3330 vec_length = evexlig << 5;
3331 break;
3332 case EVEX128:
3333 vec_length = 0 << 5;
3334 break;
3335 case EVEX256:
3336 vec_length = 1 << 5;
3337 break;
3338 case EVEX512:
3339 vec_length = 2 << 5;
3340 break;
3341 default:
3342 abort ();
3343 break;
3344 }
3345 i.vex.bytes[3] |= vec_length;
3346 /* Encode the broadcast bit. */
3347 if (i.broadcast)
3348 i.vex.bytes[3] |= 0x10;
3349 }
3350 else
3351 {
3352 if (i.rounding->type != saeonly)
3353 i.vex.bytes[3] |= 0x10 | (i.rounding->type << 5);
3354 else
3355 i.vex.bytes[3] |= 0x10 | (evexrcig << 5);
3356 }
3357
3358 if (i.mask && i.mask->mask)
3359 i.vex.bytes[3] |= i.mask->mask->reg_num;
3360 }
3361
3362 static void
3363 process_immext (void)
3364 {
3365 expressionS *exp;
3366
3367 if ((i.tm.cpu_flags.bitfield.cpusse3 || i.tm.cpu_flags.bitfield.cpusvme)
3368 && i.operands > 0)
3369 {
3370 /* MONITOR/MWAIT as well as SVME instructions have fixed operands
3371 with an opcode suffix which is coded in the same place as an
3372 8-bit immediate field would be.
3373 Here we check those operands and remove them afterwards. */
3374 unsigned int x;
3375
3376 for (x = 0; x < i.operands; x++)
3377 if (register_number (i.op[x].regs) != x)
3378 as_bad (_("can't use register '%s%s' as operand %d in '%s'."),
3379 register_prefix, i.op[x].regs->reg_name, x + 1,
3380 i.tm.name);
3381
3382 i.operands = 0;
3383 }
3384
3385 /* These AMD 3DNow! and SSE2 instructions have an opcode suffix
3386 which is coded in the same place as an 8-bit immediate field
3387 would be. Here we fake an 8-bit immediate operand from the
3388 opcode suffix stored in tm.extension_opcode.
3389
3390 AVX instructions also use this encoding, for some of
3391 3 argument instructions. */
3392
3393 gas_assert (i.imm_operands <= 1
3394 && (i.operands <= 2
3395 || ((i.tm.opcode_modifier.vex
3396 || i.tm.opcode_modifier.evex)
3397 && i.operands <= 4)));
3398
3399 exp = &im_expressions[i.imm_operands++];
3400 i.op[i.operands].imms = exp;
3401 i.types[i.operands] = imm8;
3402 i.operands++;
3403 exp->X_op = O_constant;
3404 exp->X_add_number = i.tm.extension_opcode;
3405 i.tm.extension_opcode = None;
3406 }
3407
3408
3409 static int
3410 check_hle (void)
3411 {
3412 switch (i.tm.opcode_modifier.hleprefixok)
3413 {
3414 default:
3415 abort ();
3416 case HLEPrefixNone:
3417 as_bad (_("invalid instruction `%s' after `%s'"),
3418 i.tm.name, i.hle_prefix);
3419 return 0;
3420 case HLEPrefixLock:
3421 if (i.prefix[LOCK_PREFIX])
3422 return 1;
3423 as_bad (_("missing `lock' with `%s'"), i.hle_prefix);
3424 return 0;
3425 case HLEPrefixAny:
3426 return 1;
3427 case HLEPrefixRelease:
3428 if (i.prefix[HLE_PREFIX] != XRELEASE_PREFIX_OPCODE)
3429 {
3430 as_bad (_("instruction `%s' after `xacquire' not allowed"),
3431 i.tm.name);
3432 return 0;
3433 }
3434 if (i.mem_operands == 0
3435 || !operand_type_check (i.types[i.operands - 1], anymem))
3436 {
3437 as_bad (_("memory destination needed for instruction `%s'"
3438 " after `xrelease'"), i.tm.name);
3439 return 0;
3440 }
3441 return 1;
3442 }
3443 }
3444
3445 /* This is the guts of the machine-dependent assembler. LINE points to a
3446 machine dependent instruction. This function is supposed to emit
3447 the frags/bytes it assembles to. */
3448
3449 void
3450 md_assemble (char *line)
3451 {
3452 unsigned int j;
3453 char mnemonic[MAX_MNEM_SIZE];
3454 const insn_template *t;
3455
3456 /* Initialize globals. */
3457 memset (&i, '\0', sizeof (i));
3458 for (j = 0; j < MAX_OPERANDS; j++)
3459 i.reloc[j] = NO_RELOC;
3460 memset (disp_expressions, '\0', sizeof (disp_expressions));
3461 memset (im_expressions, '\0', sizeof (im_expressions));
3462 save_stack_p = save_stack;
3463
3464 /* First parse an instruction mnemonic & call i386_operand for the operands.
3465 We assume that the scrubber has arranged it so that line[0] is the valid
3466 start of a (possibly prefixed) mnemonic. */
3467
3468 line = parse_insn (line, mnemonic);
3469 if (line == NULL)
3470 return;
3471
3472 line = parse_operands (line, mnemonic);
3473 this_operand = -1;
3474 if (line == NULL)
3475 return;
3476
3477 /* Now we've parsed the mnemonic into a set of templates, and have the
3478 operands at hand. */
3479
3480 /* All intel opcodes have reversed operands except for "bound" and
3481 "enter". We also don't reverse intersegment "jmp" and "call"
3482 instructions with 2 immediate operands so that the immediate segment
3483 precedes the offset, as it does when in AT&T mode. */
3484 if (intel_syntax
3485 && i.operands > 1
3486 && (strcmp (mnemonic, "bound") != 0)
3487 && (strcmp (mnemonic, "invlpga") != 0)
3488 && !(operand_type_check (i.types[0], imm)
3489 && operand_type_check (i.types[1], imm)))
3490 swap_operands ();
3491
3492 /* The order of the immediates should be reversed
3493 for 2 immediates extrq and insertq instructions */
3494 if (i.imm_operands == 2
3495 && (strcmp (mnemonic, "extrq") == 0
3496 || strcmp (mnemonic, "insertq") == 0))
3497 swap_2_operands (0, 1);
3498
3499 if (i.imm_operands)
3500 optimize_imm ();
3501
3502 /* Don't optimize displacement for movabs since it only takes 64bit
3503 displacement. */
3504 if (i.disp_operands
3505 && i.disp_encoding != disp_encoding_32bit
3506 && (flag_code != CODE_64BIT
3507 || strcmp (mnemonic, "movabs") != 0))
3508 optimize_disp ();
3509
3510 /* Next, we find a template that matches the given insn,
3511 making sure the overlap of the given operands types is consistent
3512 with the template operand types. */
3513
3514 if (!(t = match_template ()))
3515 return;
3516
3517 if (sse_check != check_none
3518 && !i.tm.opcode_modifier.noavx
3519 && (i.tm.cpu_flags.bitfield.cpusse
3520 || i.tm.cpu_flags.bitfield.cpusse2
3521 || i.tm.cpu_flags.bitfield.cpusse3
3522 || i.tm.cpu_flags.bitfield.cpussse3
3523 || i.tm.cpu_flags.bitfield.cpusse4_1
3524 || i.tm.cpu_flags.bitfield.cpusse4_2))
3525 {
3526 (sse_check == check_warning
3527 ? as_warn
3528 : as_bad) (_("SSE instruction `%s' is used"), i.tm.name);
3529 }
3530
3531 /* Zap movzx and movsx suffix. The suffix has been set from
3532 "word ptr" or "byte ptr" on the source operand in Intel syntax
3533 or extracted from mnemonic in AT&T syntax. But we'll use
3534 the destination register to choose the suffix for encoding. */
3535 if ((i.tm.base_opcode & ~9) == 0x0fb6)
3536 {
3537 /* In Intel syntax, there must be a suffix. In AT&T syntax, if
3538 there is no suffix, the default will be byte extension. */
3539 if (i.reg_operands != 2
3540 && !i.suffix
3541 && intel_syntax)
3542 as_bad (_("ambiguous operand size for `%s'"), i.tm.name);
3543
3544 i.suffix = 0;
3545 }
3546
3547 if (i.tm.opcode_modifier.fwait)
3548 if (!add_prefix (FWAIT_OPCODE))
3549 return;
3550
3551 /* Check if REP prefix is OK. */
3552 if (i.rep_prefix && !i.tm.opcode_modifier.repprefixok)
3553 {
3554 as_bad (_("invalid instruction `%s' after `%s'"),
3555 i.tm.name, i.rep_prefix);
3556 return;
3557 }
3558
3559 /* Check for lock without a lockable instruction. Destination operand
3560 must be memory unless it is xchg (0x86). */
3561 if (i.prefix[LOCK_PREFIX]
3562 && (!i.tm.opcode_modifier.islockable
3563 || i.mem_operands == 0
3564 || (i.tm.base_opcode != 0x86
3565 && !operand_type_check (i.types[i.operands - 1], anymem))))
3566 {
3567 as_bad (_("expecting lockable instruction after `lock'"));
3568 return;
3569 }
3570
3571 /* Check if HLE prefix is OK. */
3572 if (i.hle_prefix && !check_hle ())
3573 return;
3574
3575 /* Check BND prefix. */
3576 if (i.bnd_prefix && !i.tm.opcode_modifier.bndprefixok)
3577 as_bad (_("expecting valid branch instruction after `bnd'"));
3578
3579 if (i.tm.cpu_flags.bitfield.cpumpx
3580 && flag_code == CODE_64BIT
3581 && i.prefix[ADDR_PREFIX])
3582 as_bad (_("32-bit address isn't allowed in 64-bit MPX instructions."));
3583
3584 /* Insert BND prefix. */
3585 if (add_bnd_prefix
3586 && i.tm.opcode_modifier.bndprefixok
3587 && !i.prefix[BND_PREFIX])
3588 add_prefix (BND_PREFIX_OPCODE);
3589
3590 /* Check string instruction segment overrides. */
3591 if (i.tm.opcode_modifier.isstring && i.mem_operands != 0)
3592 {
3593 if (!check_string ())
3594 return;
3595 i.disp_operands = 0;
3596 }
3597
3598 if (!process_suffix ())
3599 return;
3600
3601 /* Update operand types. */
3602 for (j = 0; j < i.operands; j++)
3603 i.types[j] = operand_type_and (i.types[j], i.tm.operand_types[j]);
3604
3605 /* Make still unresolved immediate matches conform to size of immediate
3606 given in i.suffix. */
3607 if (!finalize_imm ())
3608 return;
3609
3610 if (i.types[0].bitfield.imm1)
3611 i.imm_operands = 0; /* kludge for shift insns. */
3612
3613 /* We only need to check those implicit registers for instructions
3614 with 3 operands or less. */
3615 if (i.operands <= 3)
3616 for (j = 0; j < i.operands; j++)
3617 if (i.types[j].bitfield.inoutportreg
3618 || i.types[j].bitfield.shiftcount
3619 || i.types[j].bitfield.acc
3620 || i.types[j].bitfield.floatacc)
3621 i.reg_operands--;
3622
3623 /* ImmExt should be processed after SSE2AVX. */
3624 if (!i.tm.opcode_modifier.sse2avx
3625 && i.tm.opcode_modifier.immext)
3626 process_immext ();
3627
3628 /* For insns with operands there are more diddles to do to the opcode. */
3629 if (i.operands)
3630 {
3631 if (!process_operands ())
3632 return;
3633 }
3634 else if (!quiet_warnings && i.tm.opcode_modifier.ugh)
3635 {
3636 /* UnixWare fsub no args is alias for fsubp, fadd -> faddp, etc. */
3637 as_warn (_("translating to `%sp'"), i.tm.name);
3638 }
3639
3640 if (i.tm.opcode_modifier.vex || i.tm.opcode_modifier.evex)
3641 {
3642 if (flag_code == CODE_16BIT)
3643 {
3644 as_bad (_("instruction `%s' isn't supported in 16-bit mode."),
3645 i.tm.name);
3646 return;
3647 }
3648
3649 if (i.tm.opcode_modifier.vex)
3650 build_vex_prefix (t);
3651 else
3652 build_evex_prefix ();
3653 }
3654
3655 /* Handle conversion of 'int $3' --> special int3 insn. XOP or FMA4
3656 instructions may define INT_OPCODE as well, so avoid this corner
3657 case for those instructions that use MODRM. */
3658 if (i.tm.base_opcode == INT_OPCODE
3659 && !i.tm.opcode_modifier.modrm
3660 && i.op[0].imms->X_add_number == 3)
3661 {
3662 i.tm.base_opcode = INT3_OPCODE;
3663 i.imm_operands = 0;
3664 }
3665
3666 if ((i.tm.opcode_modifier.jump
3667 || i.tm.opcode_modifier.jumpbyte
3668 || i.tm.opcode_modifier.jumpdword)
3669 && i.op[0].disps->X_op == O_constant)
3670 {
3671 /* Convert "jmp constant" (and "call constant") to a jump (call) to
3672 the absolute address given by the constant. Since ix86 jumps and
3673 calls are pc relative, we need to generate a reloc. */
3674 i.op[0].disps->X_add_symbol = &abs_symbol;
3675 i.op[0].disps->X_op = O_symbol;
3676 }
3677
3678 if (i.tm.opcode_modifier.rex64)
3679 i.rex |= REX_W;
3680
3681 /* For 8 bit registers we need an empty rex prefix. Also if the
3682 instruction already has a prefix, we need to convert old
3683 registers to new ones. */
3684
3685 if ((i.types[0].bitfield.reg8
3686 && (i.op[0].regs->reg_flags & RegRex64) != 0)
3687 || (i.types[1].bitfield.reg8
3688 && (i.op[1].regs->reg_flags & RegRex64) != 0)
3689 || ((i.types[0].bitfield.reg8
3690 || i.types[1].bitfield.reg8)
3691 && i.rex != 0))
3692 {
3693 int x;
3694
3695 i.rex |= REX_OPCODE;
3696 for (x = 0; x < 2; x++)
3697 {
3698 /* Look for 8 bit operand that uses old registers. */
3699 if (i.types[x].bitfield.reg8
3700 && (i.op[x].regs->reg_flags & RegRex64) == 0)
3701 {
3702 /* In case it is "hi" register, give up. */
3703 if (i.op[x].regs->reg_num > 3)
3704 as_bad (_("can't encode register '%s%s' in an "
3705 "instruction requiring REX prefix."),
3706 register_prefix, i.op[x].regs->reg_name);
3707
3708 /* Otherwise it is equivalent to the extended register.
3709 Since the encoding doesn't change this is merely
3710 cosmetic cleanup for debug output. */
3711
3712 i.op[x].regs = i.op[x].regs + 8;
3713 }
3714 }
3715 }
3716
3717 if (i.rex != 0)
3718 add_prefix (REX_OPCODE | i.rex);
3719
3720 /* We are ready to output the insn. */
3721 output_insn ();
3722 }
3723
3724 static char *
3725 parse_insn (char *line, char *mnemonic)
3726 {
3727 char *l = line;
3728 char *token_start = l;
3729 char *mnem_p;
3730 int supported;
3731 const insn_template *t;
3732 char *dot_p = NULL;
3733
3734 while (1)
3735 {
3736 mnem_p = mnemonic;
3737 while ((*mnem_p = mnemonic_chars[(unsigned char) *l]) != 0)
3738 {
3739 if (*mnem_p == '.')
3740 dot_p = mnem_p;
3741 mnem_p++;
3742 if (mnem_p >= mnemonic + MAX_MNEM_SIZE)
3743 {
3744 as_bad (_("no such instruction: `%s'"), token_start);
3745 return NULL;
3746 }
3747 l++;
3748 }
3749 if (!is_space_char (*l)
3750 && *l != END_OF_INSN
3751 && (intel_syntax
3752 || (*l != PREFIX_SEPARATOR
3753 && *l != ',')))
3754 {
3755 as_bad (_("invalid character %s in mnemonic"),
3756 output_invalid (*l));
3757 return NULL;
3758 }
3759 if (token_start == l)
3760 {
3761 if (!intel_syntax && *l == PREFIX_SEPARATOR)
3762 as_bad (_("expecting prefix; got nothing"));
3763 else
3764 as_bad (_("expecting mnemonic; got nothing"));
3765 return NULL;
3766 }
3767
3768 /* Look up instruction (or prefix) via hash table. */
3769 current_templates = (const templates *) hash_find (op_hash, mnemonic);
3770
3771 if (*l != END_OF_INSN
3772 && (!is_space_char (*l) || l[1] != END_OF_INSN)
3773 && current_templates
3774 && current_templates->start->opcode_modifier.isprefix)
3775 {
3776 if (!cpu_flags_check_cpu64 (current_templates->start->cpu_flags))
3777 {
3778 as_bad ((flag_code != CODE_64BIT
3779 ? _("`%s' is only supported in 64-bit mode")
3780 : _("`%s' is not supported in 64-bit mode")),
3781 current_templates->start->name);
3782 return NULL;
3783 }
3784 /* If we are in 16-bit mode, do not allow addr16 or data16.
3785 Similarly, in 32-bit mode, do not allow addr32 or data32. */
3786 if ((current_templates->start->opcode_modifier.size16
3787 || current_templates->start->opcode_modifier.size32)
3788 && flag_code != CODE_64BIT
3789 && (current_templates->start->opcode_modifier.size32
3790 ^ (flag_code == CODE_16BIT)))
3791 {
3792 as_bad (_("redundant %s prefix"),
3793 current_templates->start->name);
3794 return NULL;
3795 }
3796 /* Add prefix, checking for repeated prefixes. */
3797 switch (add_prefix (current_templates->start->base_opcode))
3798 {
3799 case PREFIX_EXIST:
3800 return NULL;
3801 case PREFIX_REP:
3802 if (current_templates->start->cpu_flags.bitfield.cpuhle)
3803 i.hle_prefix = current_templates->start->name;
3804 else if (current_templates->start->cpu_flags.bitfield.cpumpx)
3805 i.bnd_prefix = current_templates->start->name;
3806 else
3807 i.rep_prefix = current_templates->start->name;
3808 break;
3809 default:
3810 break;
3811 }
3812 /* Skip past PREFIX_SEPARATOR and reset token_start. */
3813 token_start = ++l;
3814 }
3815 else
3816 break;
3817 }
3818
3819 if (!current_templates)
3820 {
3821 /* Check if we should swap operand or force 32bit displacement in
3822 encoding. */
3823 if (mnem_p - 2 == dot_p && dot_p[1] == 's')
3824 i.swap_operand = 1;
3825 else if (mnem_p - 3 == dot_p
3826 && dot_p[1] == 'd'
3827 && dot_p[2] == '8')
3828 i.disp_encoding = disp_encoding_8bit;
3829 else if (mnem_p - 4 == dot_p
3830 && dot_p[1] == 'd'
3831 && dot_p[2] == '3'
3832 && dot_p[3] == '2')
3833 i.disp_encoding = disp_encoding_32bit;
3834 else
3835 goto check_suffix;
3836 mnem_p = dot_p;
3837 *dot_p = '\0';
3838 current_templates = (const templates *) hash_find (op_hash, mnemonic);
3839 }
3840
3841 if (!current_templates)
3842 {
3843 check_suffix:
3844 /* See if we can get a match by trimming off a suffix. */
3845 switch (mnem_p[-1])
3846 {
3847 case WORD_MNEM_SUFFIX:
3848 if (intel_syntax && (intel_float_operand (mnemonic) & 2))
3849 i.suffix = SHORT_MNEM_SUFFIX;
3850 else
3851 case BYTE_MNEM_SUFFIX:
3852 case QWORD_MNEM_SUFFIX:
3853 i.suffix = mnem_p[-1];
3854 mnem_p[-1] = '\0';
3855 current_templates = (const templates *) hash_find (op_hash,
3856 mnemonic);
3857 break;
3858 case SHORT_MNEM_SUFFIX:
3859 case LONG_MNEM_SUFFIX:
3860 if (!intel_syntax)
3861 {
3862 i.suffix = mnem_p[-1];
3863 mnem_p[-1] = '\0';
3864 current_templates = (const templates *) hash_find (op_hash,
3865 mnemonic);
3866 }
3867 break;
3868
3869 /* Intel Syntax. */
3870 case 'd':
3871 if (intel_syntax)
3872 {
3873 if (intel_float_operand (mnemonic) == 1)
3874 i.suffix = SHORT_MNEM_SUFFIX;
3875 else
3876 i.suffix = LONG_MNEM_SUFFIX;
3877 mnem_p[-1] = '\0';
3878 current_templates = (const templates *) hash_find (op_hash,
3879 mnemonic);
3880 }
3881 break;
3882 }
3883 if (!current_templates)
3884 {
3885 as_bad (_("no such instruction: `%s'"), token_start);
3886 return NULL;
3887 }
3888 }
3889
3890 if (current_templates->start->opcode_modifier.jump
3891 || current_templates->start->opcode_modifier.jumpbyte)
3892 {
3893 /* Check for a branch hint. We allow ",pt" and ",pn" for
3894 predict taken and predict not taken respectively.
3895 I'm not sure that branch hints actually do anything on loop
3896 and jcxz insns (JumpByte) for current Pentium4 chips. They
3897 may work in the future and it doesn't hurt to accept them
3898 now. */
3899 if (l[0] == ',' && l[1] == 'p')
3900 {
3901 if (l[2] == 't')
3902 {
3903 if (!add_prefix (DS_PREFIX_OPCODE))
3904 return NULL;
3905 l += 3;
3906 }
3907 else if (l[2] == 'n')
3908 {
3909 if (!add_prefix (CS_PREFIX_OPCODE))
3910 return NULL;
3911 l += 3;
3912 }
3913 }
3914 }
3915 /* Any other comma loses. */
3916 if (*l == ',')
3917 {
3918 as_bad (_("invalid character %s in mnemonic"),
3919 output_invalid (*l));
3920 return NULL;
3921 }
3922
3923 /* Check if instruction is supported on specified architecture. */
3924 supported = 0;
3925 for (t = current_templates->start; t < current_templates->end; ++t)
3926 {
3927 supported |= cpu_flags_match (t);
3928 if (supported == CPU_FLAGS_PERFECT_MATCH)
3929 goto skip;
3930 }
3931
3932 if (!(supported & CPU_FLAGS_64BIT_MATCH))
3933 {
3934 as_bad (flag_code == CODE_64BIT
3935 ? _("`%s' is not supported in 64-bit mode")
3936 : _("`%s' is only supported in 64-bit mode"),
3937 current_templates->start->name);
3938 return NULL;
3939 }
3940 if (supported != CPU_FLAGS_PERFECT_MATCH)
3941 {
3942 as_bad (_("`%s' is not supported on `%s%s'"),
3943 current_templates->start->name,
3944 cpu_arch_name ? cpu_arch_name : default_arch,
3945 cpu_sub_arch_name ? cpu_sub_arch_name : "");
3946 return NULL;
3947 }
3948
3949 skip:
3950 if (!cpu_arch_flags.bitfield.cpui386
3951 && (flag_code != CODE_16BIT))
3952 {
3953 as_warn (_("use .code16 to ensure correct addressing mode"));
3954 }
3955
3956 return l;
3957 }
3958
3959 static char *
3960 parse_operands (char *l, const char *mnemonic)
3961 {
3962 char *token_start;
3963
3964 /* 1 if operand is pending after ','. */
3965 unsigned int expecting_operand = 0;
3966
3967 /* Non-zero if operand parens not balanced. */
3968 unsigned int paren_not_balanced;
3969
3970 while (*l != END_OF_INSN)
3971 {
3972 /* Skip optional white space before operand. */
3973 if (is_space_char (*l))
3974 ++l;
3975 if (!is_operand_char (*l) && *l != END_OF_INSN)
3976 {
3977 as_bad (_("invalid character %s before operand %d"),
3978 output_invalid (*l),
3979 i.operands + 1);
3980 return NULL;
3981 }
3982 token_start = l; /* after white space */
3983 paren_not_balanced = 0;
3984 while (paren_not_balanced || *l != ',')
3985 {
3986 if (*l == END_OF_INSN)
3987 {
3988 if (paren_not_balanced)
3989 {
3990 if (!intel_syntax)
3991 as_bad (_("unbalanced parenthesis in operand %d."),
3992 i.operands + 1);
3993 else
3994 as_bad (_("unbalanced brackets in operand %d."),
3995 i.operands + 1);
3996 return NULL;
3997 }
3998 else
3999 break; /* we are done */
4000 }
4001 else if (!is_operand_char (*l) && !is_space_char (*l))
4002 {
4003 as_bad (_("invalid character %s in operand %d"),
4004 output_invalid (*l),
4005 i.operands + 1);
4006 return NULL;
4007 }
4008 if (!intel_syntax)
4009 {
4010 if (*l == '(')
4011 ++paren_not_balanced;
4012 if (*l == ')')
4013 --paren_not_balanced;
4014 }
4015 else
4016 {
4017 if (*l == '[')
4018 ++paren_not_balanced;
4019 if (*l == ']')
4020 --paren_not_balanced;
4021 }
4022 l++;
4023 }
4024 if (l != token_start)
4025 { /* Yes, we've read in another operand. */
4026 unsigned int operand_ok;
4027 this_operand = i.operands++;
4028 i.types[this_operand].bitfield.unspecified = 1;
4029 if (i.operands > MAX_OPERANDS)
4030 {
4031 as_bad (_("spurious operands; (%d operands/instruction max)"),
4032 MAX_OPERANDS);
4033 return NULL;
4034 }
4035 /* Now parse operand adding info to 'i' as we go along. */
4036 END_STRING_AND_SAVE (l);
4037
4038 if (intel_syntax)
4039 operand_ok =
4040 i386_intel_operand (token_start,
4041 intel_float_operand (mnemonic));
4042 else
4043 operand_ok = i386_att_operand (token_start);
4044
4045 RESTORE_END_STRING (l);
4046 if (!operand_ok)
4047 return NULL;
4048 }
4049 else
4050 {
4051 if (expecting_operand)
4052 {
4053 expecting_operand_after_comma:
4054 as_bad (_("expecting operand after ','; got nothing"));
4055 return NULL;
4056 }
4057 if (*l == ',')
4058 {
4059 as_bad (_("expecting operand before ','; got nothing"));
4060 return NULL;
4061 }
4062 }
4063
4064 /* Now *l must be either ',' or END_OF_INSN. */
4065 if (*l == ',')
4066 {
4067 if (*++l == END_OF_INSN)
4068 {
4069 /* Just skip it, if it's \n complain. */
4070 goto expecting_operand_after_comma;
4071 }
4072 expecting_operand = 1;
4073 }
4074 }
4075 return l;
4076 }
4077
4078 static void
4079 swap_2_operands (int xchg1, int xchg2)
4080 {
4081 union i386_op temp_op;
4082 i386_operand_type temp_type;
4083 enum bfd_reloc_code_real temp_reloc;
4084
4085 temp_type = i.types[xchg2];
4086 i.types[xchg2] = i.types[xchg1];
4087 i.types[xchg1] = temp_type;
4088 temp_op = i.op[xchg2];
4089 i.op[xchg2] = i.op[xchg1];
4090 i.op[xchg1] = temp_op;
4091 temp_reloc = i.reloc[xchg2];
4092 i.reloc[xchg2] = i.reloc[xchg1];
4093 i.reloc[xchg1] = temp_reloc;
4094
4095 if (i.mask)
4096 {
4097 if (i.mask->operand == xchg1)
4098 i.mask->operand = xchg2;
4099 else if (i.mask->operand == xchg2)
4100 i.mask->operand = xchg1;
4101 }
4102 if (i.broadcast)
4103 {
4104 if (i.broadcast->operand == xchg1)
4105 i.broadcast->operand = xchg2;
4106 else if (i.broadcast->operand == xchg2)
4107 i.broadcast->operand = xchg1;
4108 }
4109 if (i.rounding)
4110 {
4111 if (i.rounding->operand == xchg1)
4112 i.rounding->operand = xchg2;
4113 else if (i.rounding->operand == xchg2)
4114 i.rounding->operand = xchg1;
4115 }
4116 }
4117
4118 static void
4119 swap_operands (void)
4120 {
4121 switch (i.operands)
4122 {
4123 case 5:
4124 case 4:
4125 swap_2_operands (1, i.operands - 2);
4126 case 3:
4127 case 2:
4128 swap_2_operands (0, i.operands - 1);
4129 break;
4130 default:
4131 abort ();
4132 }
4133
4134 if (i.mem_operands == 2)
4135 {
4136 const seg_entry *temp_seg;
4137 temp_seg = i.seg[0];
4138 i.seg[0] = i.seg[1];
4139 i.seg[1] = temp_seg;
4140 }
4141 }
4142
4143 /* Try to ensure constant immediates are represented in the smallest
4144 opcode possible. */
4145 static void
4146 optimize_imm (void)
4147 {
4148 char guess_suffix = 0;
4149 int op;
4150
4151 if (i.suffix)
4152 guess_suffix = i.suffix;
4153 else if (i.reg_operands)
4154 {
4155 /* Figure out a suffix from the last register operand specified.
4156 We can't do this properly yet, ie. excluding InOutPortReg,
4157 but the following works for instructions with immediates.
4158 In any case, we can't set i.suffix yet. */
4159 for (op = i.operands; --op >= 0;)
4160 if (i.types[op].bitfield.reg8)
4161 {
4162 guess_suffix = BYTE_MNEM_SUFFIX;
4163 break;
4164 }
4165 else if (i.types[op].bitfield.reg16)
4166 {
4167 guess_suffix = WORD_MNEM_SUFFIX;
4168 break;
4169 }
4170 else if (i.types[op].bitfield.reg32)
4171 {
4172 guess_suffix = LONG_MNEM_SUFFIX;
4173 break;
4174 }
4175 else if (i.types[op].bitfield.reg64)
4176 {
4177 guess_suffix = QWORD_MNEM_SUFFIX;
4178 break;
4179 }
4180 }
4181 else if ((flag_code == CODE_16BIT) ^ (i.prefix[DATA_PREFIX] != 0))
4182 guess_suffix = WORD_MNEM_SUFFIX;
4183
4184 for (op = i.operands; --op >= 0;)
4185 if (operand_type_check (i.types[op], imm))
4186 {
4187 switch (i.op[op].imms->X_op)
4188 {
4189 case O_constant:
4190 /* If a suffix is given, this operand may be shortened. */
4191 switch (guess_suffix)
4192 {
4193 case LONG_MNEM_SUFFIX:
4194 i.types[op].bitfield.imm32 = 1;
4195 i.types[op].bitfield.imm64 = 1;
4196 break;
4197 case WORD_MNEM_SUFFIX:
4198 i.types[op].bitfield.imm16 = 1;
4199 i.types[op].bitfield.imm32 = 1;
4200 i.types[op].bitfield.imm32s = 1;
4201 i.types[op].bitfield.imm64 = 1;
4202 break;
4203 case BYTE_MNEM_SUFFIX:
4204 i.types[op].bitfield.imm8 = 1;
4205 i.types[op].bitfield.imm8s = 1;
4206 i.types[op].bitfield.imm16 = 1;
4207 i.types[op].bitfield.imm32 = 1;
4208 i.types[op].bitfield.imm32s = 1;
4209 i.types[op].bitfield.imm64 = 1;
4210 break;
4211 }
4212
4213 /* If this operand is at most 16 bits, convert it
4214 to a signed 16 bit number before trying to see
4215 whether it will fit in an even smaller size.
4216 This allows a 16-bit operand such as $0xffe0 to
4217 be recognised as within Imm8S range. */
4218 if ((i.types[op].bitfield.imm16)
4219 && (i.op[op].imms->X_add_number & ~(offsetT) 0xffff) == 0)
4220 {
4221 i.op[op].imms->X_add_number =
4222 (((i.op[op].imms->X_add_number & 0xffff) ^ 0x8000) - 0x8000);
4223 }
4224 if ((i.types[op].bitfield.imm32)
4225 && ((i.op[op].imms->X_add_number & ~(((offsetT) 2 << 31) - 1))
4226 == 0))
4227 {
4228 i.op[op].imms->X_add_number = ((i.op[op].imms->X_add_number
4229 ^ ((offsetT) 1 << 31))
4230 - ((offsetT) 1 << 31));
4231 }
4232 i.types[op]
4233 = operand_type_or (i.types[op],
4234 smallest_imm_type (i.op[op].imms->X_add_number));
4235
4236 /* We must avoid matching of Imm32 templates when 64bit
4237 only immediate is available. */
4238 if (guess_suffix == QWORD_MNEM_SUFFIX)
4239 i.types[op].bitfield.imm32 = 0;
4240 break;
4241
4242 case O_absent:
4243 case O_register:
4244 abort ();
4245
4246 /* Symbols and expressions. */
4247 default:
4248 /* Convert symbolic operand to proper sizes for matching, but don't
4249 prevent matching a set of insns that only supports sizes other
4250 than those matching the insn suffix. */
4251 {
4252 i386_operand_type mask, allowed;
4253 const insn_template *t;
4254
4255 operand_type_set (&mask, 0);
4256 operand_type_set (&allowed, 0);
4257
4258 for (t = current_templates->start;
4259 t < current_templates->end;
4260 ++t)
4261 allowed = operand_type_or (allowed,
4262 t->operand_types[op]);
4263 switch (guess_suffix)
4264 {
4265 case QWORD_MNEM_SUFFIX:
4266 mask.bitfield.imm64 = 1;
4267 mask.bitfield.imm32s = 1;
4268 break;
4269 case LONG_MNEM_SUFFIX:
4270 mask.bitfield.imm32 = 1;
4271 break;
4272 case WORD_MNEM_SUFFIX:
4273 mask.bitfield.imm16 = 1;
4274 break;
4275 case BYTE_MNEM_SUFFIX:
4276 mask.bitfield.imm8 = 1;
4277 break;
4278 default:
4279 break;
4280 }
4281 allowed = operand_type_and (mask, allowed);
4282 if (!operand_type_all_zero (&allowed))
4283 i.types[op] = operand_type_and (i.types[op], mask);
4284 }
4285 break;
4286 }
4287 }
4288 }
4289
4290 /* Try to use the smallest displacement type too. */
4291 static void
4292 optimize_disp (void)
4293 {
4294 int op;
4295
4296 for (op = i.operands; --op >= 0;)
4297 if (operand_type_check (i.types[op], disp))
4298 {
4299 if (i.op[op].disps->X_op == O_constant)
4300 {
4301 offsetT op_disp = i.op[op].disps->X_add_number;
4302
4303 if (i.types[op].bitfield.disp16
4304 && (op_disp & ~(offsetT) 0xffff) == 0)
4305 {
4306 /* If this operand is at most 16 bits, convert
4307 to a signed 16 bit number and don't use 64bit
4308 displacement. */
4309 op_disp = (((op_disp & 0xffff) ^ 0x8000) - 0x8000);
4310 i.types[op].bitfield.disp64 = 0;
4311 }
4312 if (i.types[op].bitfield.disp32
4313 && (op_disp & ~(((offsetT) 2 << 31) - 1)) == 0)
4314 {
4315 /* If this operand is at most 32 bits, convert
4316 to a signed 32 bit number and don't use 64bit
4317 displacement. */
4318 op_disp &= (((offsetT) 2 << 31) - 1);
4319 op_disp = (op_disp ^ ((offsetT) 1 << 31)) - ((addressT) 1 << 31);
4320 i.types[op].bitfield.disp64 = 0;
4321 }
4322 if (!op_disp && i.types[op].bitfield.baseindex)
4323 {
4324 i.types[op].bitfield.disp8 = 0;
4325 i.types[op].bitfield.disp16 = 0;
4326 i.types[op].bitfield.disp32 = 0;
4327 i.types[op].bitfield.disp32s = 0;
4328 i.types[op].bitfield.disp64 = 0;
4329 i.op[op].disps = 0;
4330 i.disp_operands--;
4331 }
4332 else if (flag_code == CODE_64BIT)
4333 {
4334 if (fits_in_signed_long (op_disp))
4335 {
4336 i.types[op].bitfield.disp64 = 0;
4337 i.types[op].bitfield.disp32s = 1;
4338 }
4339 if (i.prefix[ADDR_PREFIX]
4340 && fits_in_unsigned_long (op_disp))
4341 i.types[op].bitfield.disp32 = 1;
4342 }
4343 if ((i.types[op].bitfield.disp32
4344 || i.types[op].bitfield.disp32s
4345 || i.types[op].bitfield.disp16)
4346 && fits_in_signed_byte (op_disp))
4347 i.types[op].bitfield.disp8 = 1;
4348 }
4349 else if (i.reloc[op] == BFD_RELOC_386_TLS_DESC_CALL
4350 || i.reloc[op] == BFD_RELOC_X86_64_TLSDESC_CALL)
4351 {
4352 fix_new_exp (frag_now, frag_more (0) - frag_now->fr_literal, 0,
4353 i.op[op].disps, 0, i.reloc[op]);
4354 i.types[op].bitfield.disp8 = 0;
4355 i.types[op].bitfield.disp16 = 0;
4356 i.types[op].bitfield.disp32 = 0;
4357 i.types[op].bitfield.disp32s = 0;
4358 i.types[op].bitfield.disp64 = 0;
4359 }
4360 else
4361 /* We only support 64bit displacement on constants. */
4362 i.types[op].bitfield.disp64 = 0;
4363 }
4364 }
4365
4366 /* Check if operands are valid for the instruction. */
4367
4368 static int
4369 check_VecOperands (const insn_template *t)
4370 {
4371 unsigned int op;
4372
4373 /* Without VSIB byte, we can't have a vector register for index. */
4374 if (!t->opcode_modifier.vecsib
4375 && i.index_reg
4376 && (i.index_reg->reg_type.bitfield.regxmm
4377 || i.index_reg->reg_type.bitfield.regymm
4378 || i.index_reg->reg_type.bitfield.regzmm))
4379 {
4380 i.error = unsupported_vector_index_register;
4381 return 1;
4382 }
4383
4384 /* Check if default mask is allowed. */
4385 if (t->opcode_modifier.nodefmask
4386 && (!i.mask || i.mask->mask->reg_num == 0))
4387 {
4388 i.error = no_default_mask;
4389 return 1;
4390 }
4391
4392 /* For VSIB byte, we need a vector register for index, and all vector
4393 registers must be distinct. */
4394 if (t->opcode_modifier.vecsib)
4395 {
4396 if (!i.index_reg
4397 || !((t->opcode_modifier.vecsib == VecSIB128
4398 && i.index_reg->reg_type.bitfield.regxmm)
4399 || (t->opcode_modifier.vecsib == VecSIB256
4400 && i.index_reg->reg_type.bitfield.regymm)
4401 || (t->opcode_modifier.vecsib == VecSIB512
4402 && i.index_reg->reg_type.bitfield.regzmm)))
4403 {
4404 i.error = invalid_vsib_address;
4405 return 1;
4406 }
4407
4408 gas_assert (i.reg_operands == 2 || i.mask);
4409 if (i.reg_operands == 2 && !i.mask)
4410 {
4411 gas_assert (i.types[0].bitfield.regxmm
4412 || i.types[0].bitfield.regymm);
4413 gas_assert (i.types[2].bitfield.regxmm
4414 || i.types[2].bitfield.regymm);
4415 if (operand_check == check_none)
4416 return 0;
4417 if (register_number (i.op[0].regs)
4418 != register_number (i.index_reg)
4419 && register_number (i.op[2].regs)
4420 != register_number (i.index_reg)
4421 && register_number (i.op[0].regs)
4422 != register_number (i.op[2].regs))
4423 return 0;
4424 if (operand_check == check_error)
4425 {
4426 i.error = invalid_vector_register_set;
4427 return 1;
4428 }
4429 as_warn (_("mask, index, and destination registers should be distinct"));
4430 }
4431 else if (i.reg_operands == 1 && i.mask)
4432 {
4433 if ((i.types[1].bitfield.regymm
4434 || i.types[1].bitfield.regzmm)
4435 && (register_number (i.op[1].regs)
4436 == register_number (i.index_reg)))
4437 {
4438 if (operand_check == check_error)
4439 {
4440 i.error = invalid_vector_register_set;
4441 return 1;
4442 }
4443 if (operand_check != check_none)
4444 as_warn (_("index and destination registers should be distinct"));
4445 }
4446 }
4447 }
4448
4449 /* Check if broadcast is supported by the instruction and is applied
4450 to the memory operand. */
4451 if (i.broadcast)
4452 {
4453 int broadcasted_opnd_size;
4454
4455 /* Check if specified broadcast is supported in this instruction,
4456 and it's applied to memory operand of DWORD or QWORD type,
4457 depending on VecESize. */
4458 if (i.broadcast->type != t->opcode_modifier.broadcast
4459 || !i.types[i.broadcast->operand].bitfield.mem
4460 || (t->opcode_modifier.vecesize == 0
4461 && !i.types[i.broadcast->operand].bitfield.dword
4462 && !i.types[i.broadcast->operand].bitfield.unspecified)
4463 || (t->opcode_modifier.vecesize == 1
4464 && !i.types[i.broadcast->operand].bitfield.qword
4465 && !i.types[i.broadcast->operand].bitfield.unspecified))
4466 goto bad_broadcast;
4467
4468 broadcasted_opnd_size = t->opcode_modifier.vecesize ? 64 : 32;
4469 if (i.broadcast->type == BROADCAST_1TO16)
4470 broadcasted_opnd_size <<= 4; /* Broadcast 1to16. */
4471 else if (i.broadcast->type == BROADCAST_1TO8)
4472 broadcasted_opnd_size <<= 3; /* Broadcast 1to8. */
4473 else if (i.broadcast->type == BROADCAST_1TO4)
4474 broadcasted_opnd_size <<= 2; /* Broadcast 1to4. */
4475 else if (i.broadcast->type == BROADCAST_1TO2)
4476 broadcasted_opnd_size <<= 1; /* Broadcast 1to2. */
4477 else
4478 goto bad_broadcast;
4479
4480 if ((broadcasted_opnd_size == 256
4481 && !t->operand_types[i.broadcast->operand].bitfield.ymmword)
4482 || (broadcasted_opnd_size == 512
4483 && !t->operand_types[i.broadcast->operand].bitfield.zmmword))
4484 {
4485 bad_broadcast:
4486 i.error = unsupported_broadcast;
4487 return 1;
4488 }
4489 }
4490 /* If broadcast is supported in this instruction, we need to check if
4491 operand of one-element size isn't specified without broadcast. */
4492 else if (t->opcode_modifier.broadcast && i.mem_operands)
4493 {
4494 /* Find memory operand. */
4495 for (op = 0; op < i.operands; op++)
4496 if (operand_type_check (i.types[op], anymem))
4497 break;
4498 gas_assert (op < i.operands);
4499 /* Check size of the memory operand. */
4500 if ((t->opcode_modifier.vecesize == 0
4501 && i.types[op].bitfield.dword)
4502 || (t->opcode_modifier.vecesize == 1
4503 && i.types[op].bitfield.qword))
4504 {
4505 i.error = broadcast_needed;
4506 return 1;
4507 }
4508 }
4509
4510 /* Check if requested masking is supported. */
4511 if (i.mask
4512 && (!t->opcode_modifier.masking
4513 || (i.mask->zeroing
4514 && t->opcode_modifier.masking == MERGING_MASKING)))
4515 {
4516 i.error = unsupported_masking;
4517 return 1;
4518 }
4519
4520 /* Check if masking is applied to dest operand. */
4521 if (i.mask && (i.mask->operand != (int) (i.operands - 1)))
4522 {
4523 i.error = mask_not_on_destination;
4524 return 1;
4525 }
4526
4527 /* Check RC/SAE. */
4528 if (i.rounding)
4529 {
4530 if ((i.rounding->type != saeonly
4531 && !t->opcode_modifier.staticrounding)
4532 || (i.rounding->type == saeonly
4533 && (t->opcode_modifier.staticrounding
4534 || !t->opcode_modifier.sae)))
4535 {
4536 i.error = unsupported_rc_sae;
4537 return 1;
4538 }
4539 /* If the instruction has several immediate operands and one of
4540 them is rounding, the rounding operand should be the last
4541 immediate operand. */
4542 if (i.imm_operands > 1
4543 && i.rounding->operand != (int) (i.imm_operands - 1))
4544 {
4545 i.error = rc_sae_operand_not_last_imm;
4546 return 1;
4547 }
4548 }
4549
4550 /* Check vector Disp8 operand. */
4551 if (t->opcode_modifier.disp8memshift)
4552 {
4553 if (i.broadcast)
4554 i.memshift = t->opcode_modifier.vecesize ? 3 : 2;
4555 else
4556 i.memshift = t->opcode_modifier.disp8memshift;
4557
4558 for (op = 0; op < i.operands; op++)
4559 if (operand_type_check (i.types[op], disp)
4560 && i.op[op].disps->X_op == O_constant)
4561 {
4562 offsetT value = i.op[op].disps->X_add_number;
4563 int vec_disp8_ok = fits_in_vec_disp8 (value);
4564 if (t->operand_types [op].bitfield.vec_disp8)
4565 {
4566 if (vec_disp8_ok)
4567 i.types[op].bitfield.vec_disp8 = 1;
4568 else
4569 {
4570 /* Vector insn can only have Vec_Disp8/Disp32 in
4571 32/64bit modes, and Vec_Disp8/Disp16 in 16bit
4572 mode. */
4573 i.types[op].bitfield.disp8 = 0;
4574 if (flag_code != CODE_16BIT)
4575 i.types[op].bitfield.disp16 = 0;
4576 }
4577 }
4578 else if (flag_code != CODE_16BIT)
4579 {
4580 /* One form of this instruction supports vector Disp8.
4581 Try vector Disp8 if we need to use Disp32. */
4582 if (vec_disp8_ok && !fits_in_signed_byte (value))
4583 {
4584 i.error = try_vector_disp8;
4585 return 1;
4586 }
4587 }
4588 }
4589 }
4590 else
4591 i.memshift = -1;
4592
4593 return 0;
4594 }
4595
4596 /* Check if operands are valid for the instruction. Update VEX
4597 operand types. */
4598
4599 static int
4600 VEX_check_operands (const insn_template *t)
4601 {
4602 /* VREX is only valid with EVEX prefix. */
4603 if (i.need_vrex && !t->opcode_modifier.evex)
4604 {
4605 i.error = invalid_register_operand;
4606 return 1;
4607 }
4608
4609 if (!t->opcode_modifier.vex)
4610 return 0;
4611
4612 /* Only check VEX_Imm4, which must be the first operand. */
4613 if (t->operand_types[0].bitfield.vec_imm4)
4614 {
4615 if (i.op[0].imms->X_op != O_constant
4616 || !fits_in_imm4 (i.op[0].imms->X_add_number))
4617 {
4618 i.error = bad_imm4;
4619 return 1;
4620 }
4621
4622 /* Turn off Imm8 so that update_imm won't complain. */
4623 i.types[0] = vec_imm4;
4624 }
4625
4626 return 0;
4627 }
4628
4629 static const insn_template *
4630 match_template (void)
4631 {
4632 /* Points to template once we've found it. */
4633 const insn_template *t;
4634 i386_operand_type overlap0, overlap1, overlap2, overlap3;
4635 i386_operand_type overlap4;
4636 unsigned int found_reverse_match;
4637 i386_opcode_modifier suffix_check;
4638 i386_operand_type operand_types [MAX_OPERANDS];
4639 int addr_prefix_disp;
4640 unsigned int j;
4641 unsigned int found_cpu_match;
4642 unsigned int check_register;
4643 enum i386_error specific_error = 0;
4644
4645 #if MAX_OPERANDS != 5
4646 # error "MAX_OPERANDS must be 5."
4647 #endif
4648
4649 found_reverse_match = 0;
4650 addr_prefix_disp = -1;
4651
4652 memset (&suffix_check, 0, sizeof (suffix_check));
4653 if (i.suffix == BYTE_MNEM_SUFFIX)
4654 suffix_check.no_bsuf = 1;
4655 else if (i.suffix == WORD_MNEM_SUFFIX)
4656 suffix_check.no_wsuf = 1;
4657 else if (i.suffix == SHORT_MNEM_SUFFIX)
4658 suffix_check.no_ssuf = 1;
4659 else if (i.suffix == LONG_MNEM_SUFFIX)
4660 suffix_check.no_lsuf = 1;
4661 else if (i.suffix == QWORD_MNEM_SUFFIX)
4662 suffix_check.no_qsuf = 1;
4663 else if (i.suffix == LONG_DOUBLE_MNEM_SUFFIX)
4664 suffix_check.no_ldsuf = 1;
4665
4666 /* Must have right number of operands. */
4667 i.error = number_of_operands_mismatch;
4668
4669 for (t = current_templates->start; t < current_templates->end; t++)
4670 {
4671 addr_prefix_disp = -1;
4672
4673 if (i.operands != t->operands)
4674 continue;
4675
4676 /* Check processor support. */
4677 i.error = unsupported;
4678 found_cpu_match = (cpu_flags_match (t)
4679 == CPU_FLAGS_PERFECT_MATCH);
4680 if (!found_cpu_match)
4681 continue;
4682
4683 /* Check old gcc support. */
4684 i.error = old_gcc_only;
4685 if (!old_gcc && t->opcode_modifier.oldgcc)
4686 continue;
4687
4688 /* Check AT&T mnemonic. */
4689 i.error = unsupported_with_intel_mnemonic;
4690 if (intel_mnemonic && t->opcode_modifier.attmnemonic)
4691 continue;
4692
4693 /* Check AT&T/Intel syntax. */
4694 i.error = unsupported_syntax;
4695 if ((intel_syntax && t->opcode_modifier.attsyntax)
4696 || (!intel_syntax && t->opcode_modifier.intelsyntax))
4697 continue;
4698
4699 /* Check the suffix, except for some instructions in intel mode. */
4700 i.error = invalid_instruction_suffix;
4701 if ((!intel_syntax || !t->opcode_modifier.ignoresize)
4702 && ((t->opcode_modifier.no_bsuf && suffix_check.no_bsuf)
4703 || (t->opcode_modifier.no_wsuf && suffix_check.no_wsuf)
4704 || (t->opcode_modifier.no_lsuf && suffix_check.no_lsuf)
4705 || (t->opcode_modifier.no_ssuf && suffix_check.no_ssuf)
4706 || (t->opcode_modifier.no_qsuf && suffix_check.no_qsuf)
4707 || (t->opcode_modifier.no_ldsuf && suffix_check.no_ldsuf)))
4708 continue;
4709
4710 if (!operand_size_match (t))
4711 continue;
4712
4713 for (j = 0; j < MAX_OPERANDS; j++)
4714 operand_types[j] = t->operand_types[j];
4715
4716 /* In general, don't allow 64-bit operands in 32-bit mode. */
4717 if (i.suffix == QWORD_MNEM_SUFFIX
4718 && flag_code != CODE_64BIT
4719 && (intel_syntax
4720 ? (!t->opcode_modifier.ignoresize
4721 && !intel_float_operand (t->name))
4722 : intel_float_operand (t->name) != 2)
4723 && ((!operand_types[0].bitfield.regmmx
4724 && !operand_types[0].bitfield.regxmm
4725 && !operand_types[0].bitfield.regymm
4726 && !operand_types[0].bitfield.regzmm)
4727 || (!operand_types[t->operands > 1].bitfield.regmmx
4728 && operand_types[t->operands > 1].bitfield.regxmm
4729 && operand_types[t->operands > 1].bitfield.regymm
4730 && operand_types[t->operands > 1].bitfield.regzmm))
4731 && (t->base_opcode != 0x0fc7
4732 || t->extension_opcode != 1 /* cmpxchg8b */))
4733 continue;
4734
4735 /* In general, don't allow 32-bit operands on pre-386. */
4736 else if (i.suffix == LONG_MNEM_SUFFIX
4737 && !cpu_arch_flags.bitfield.cpui386
4738 && (intel_syntax
4739 ? (!t->opcode_modifier.ignoresize
4740 && !intel_float_operand (t->name))
4741 : intel_float_operand (t->name) != 2)
4742 && ((!operand_types[0].bitfield.regmmx
4743 && !operand_types[0].bitfield.regxmm)
4744 || (!operand_types[t->operands > 1].bitfield.regmmx
4745 && operand_types[t->operands > 1].bitfield.regxmm)))
4746 continue;
4747
4748 /* Do not verify operands when there are none. */
4749 else
4750 {
4751 if (!t->operands)
4752 /* We've found a match; break out of loop. */
4753 break;
4754 }
4755
4756 /* Address size prefix will turn Disp64/Disp32/Disp16 operand
4757 into Disp32/Disp16/Disp32 operand. */
4758 if (i.prefix[ADDR_PREFIX] != 0)
4759 {
4760 /* There should be only one Disp operand. */
4761 switch (flag_code)
4762 {
4763 case CODE_16BIT:
4764 for (j = 0; j < MAX_OPERANDS; j++)
4765 {
4766 if (operand_types[j].bitfield.disp16)
4767 {
4768 addr_prefix_disp = j;
4769 operand_types[j].bitfield.disp32 = 1;
4770 operand_types[j].bitfield.disp16 = 0;
4771 break;
4772 }
4773 }
4774 break;
4775 case CODE_32BIT:
4776 for (j = 0; j < MAX_OPERANDS; j++)
4777 {
4778 if (operand_types[j].bitfield.disp32)
4779 {
4780 addr_prefix_disp = j;
4781 operand_types[j].bitfield.disp32 = 0;
4782 operand_types[j].bitfield.disp16 = 1;
4783 break;
4784 }
4785 }
4786 break;
4787 case CODE_64BIT:
4788 for (j = 0; j < MAX_OPERANDS; j++)
4789 {
4790 if (operand_types[j].bitfield.disp64)
4791 {
4792 addr_prefix_disp = j;
4793 operand_types[j].bitfield.disp64 = 0;
4794 operand_types[j].bitfield.disp32 = 1;
4795 break;
4796 }
4797 }
4798 break;
4799 }
4800 }
4801
4802 /* We check register size if needed. */
4803 check_register = t->opcode_modifier.checkregsize;
4804 overlap0 = operand_type_and (i.types[0], operand_types[0]);
4805 switch (t->operands)
4806 {
4807 case 1:
4808 if (!operand_type_match (overlap0, i.types[0]))
4809 continue;
4810 break;
4811 case 2:
4812 /* xchg %eax, %eax is a special case. It is an aliase for nop
4813 only in 32bit mode and we can use opcode 0x90. In 64bit
4814 mode, we can't use 0x90 for xchg %eax, %eax since it should
4815 zero-extend %eax to %rax. */
4816 if (flag_code == CODE_64BIT
4817 && t->base_opcode == 0x90
4818 && operand_type_equal (&i.types [0], &acc32)
4819 && operand_type_equal (&i.types [1], &acc32))
4820 continue;
4821 if (i.swap_operand)
4822 {
4823 /* If we swap operand in encoding, we either match
4824 the next one or reverse direction of operands. */
4825 if (t->opcode_modifier.s)
4826 continue;
4827 else if (t->opcode_modifier.d)
4828 goto check_reverse;
4829 }
4830
4831 case 3:
4832 /* If we swap operand in encoding, we match the next one. */
4833 if (i.swap_operand && t->opcode_modifier.s)
4834 continue;
4835 case 4:
4836 case 5:
4837 overlap1 = operand_type_and (i.types[1], operand_types[1]);
4838 if (!operand_type_match (overlap0, i.types[0])
4839 || !operand_type_match (overlap1, i.types[1])
4840 || (check_register
4841 && !operand_type_register_match (overlap0, i.types[0],
4842 operand_types[0],
4843 overlap1, i.types[1],
4844 operand_types[1])))
4845 {
4846 /* Check if other direction is valid ... */
4847 if (!t->opcode_modifier.d && !t->opcode_modifier.floatd)
4848 continue;
4849
4850 check_reverse:
4851 /* Try reversing direction of operands. */
4852 overlap0 = operand_type_and (i.types[0], operand_types[1]);
4853 overlap1 = operand_type_and (i.types[1], operand_types[0]);
4854 if (!operand_type_match (overlap0, i.types[0])
4855 || !operand_type_match (overlap1, i.types[1])
4856 || (check_register
4857 && !operand_type_register_match (overlap0,
4858 i.types[0],
4859 operand_types[1],
4860 overlap1,
4861 i.types[1],
4862 operand_types[0])))
4863 {
4864 /* Does not match either direction. */
4865 continue;
4866 }
4867 /* found_reverse_match holds which of D or FloatDR
4868 we've found. */
4869 if (t->opcode_modifier.d)
4870 found_reverse_match = Opcode_D;
4871 else if (t->opcode_modifier.floatd)
4872 found_reverse_match = Opcode_FloatD;
4873 else
4874 found_reverse_match = 0;
4875 if (t->opcode_modifier.floatr)
4876 found_reverse_match |= Opcode_FloatR;
4877 }
4878 else
4879 {
4880 /* Found a forward 2 operand match here. */
4881 switch (t->operands)
4882 {
4883 case 5:
4884 overlap4 = operand_type_and (i.types[4],
4885 operand_types[4]);
4886 case 4:
4887 overlap3 = operand_type_and (i.types[3],
4888 operand_types[3]);
4889 case 3:
4890 overlap2 = operand_type_and (i.types[2],
4891 operand_types[2]);
4892 break;
4893 }
4894
4895 switch (t->operands)
4896 {
4897 case 5:
4898 if (!operand_type_match (overlap4, i.types[4])
4899 || !operand_type_register_match (overlap3,
4900 i.types[3],
4901 operand_types[3],
4902 overlap4,
4903 i.types[4],
4904 operand_types[4]))
4905 continue;
4906 case 4:
4907 if (!operand_type_match (overlap3, i.types[3])
4908 || (check_register
4909 && !operand_type_register_match (overlap2,
4910 i.types[2],
4911 operand_types[2],
4912 overlap3,
4913 i.types[3],
4914 operand_types[3])))
4915 continue;
4916 case 3:
4917 /* Here we make use of the fact that there are no
4918 reverse match 3 operand instructions, and all 3
4919 operand instructions only need to be checked for
4920 register consistency between operands 2 and 3. */
4921 if (!operand_type_match (overlap2, i.types[2])
4922 || (check_register
4923 && !operand_type_register_match (overlap1,
4924 i.types[1],
4925 operand_types[1],
4926 overlap2,
4927 i.types[2],
4928 operand_types[2])))
4929 continue;
4930 break;
4931 }
4932 }
4933 /* Found either forward/reverse 2, 3 or 4 operand match here:
4934 slip through to break. */
4935 }
4936 if (!found_cpu_match)
4937 {
4938 found_reverse_match = 0;
4939 continue;
4940 }
4941
4942 /* Check if vector and VEX operands are valid. */
4943 if (check_VecOperands (t) || VEX_check_operands (t))
4944 {
4945 specific_error = i.error;
4946 continue;
4947 }
4948
4949 /* We've found a match; break out of loop. */
4950 break;
4951 }
4952
4953 if (t == current_templates->end)
4954 {
4955 /* We found no match. */
4956 const char *err_msg;
4957 switch (specific_error ? specific_error : i.error)
4958 {
4959 default:
4960 abort ();
4961 case operand_size_mismatch:
4962 err_msg = _("operand size mismatch");
4963 break;
4964 case operand_type_mismatch:
4965 err_msg = _("operand type mismatch");
4966 break;
4967 case register_type_mismatch:
4968 err_msg = _("register type mismatch");
4969 break;
4970 case number_of_operands_mismatch:
4971 err_msg = _("number of operands mismatch");
4972 break;
4973 case invalid_instruction_suffix:
4974 err_msg = _("invalid instruction suffix");
4975 break;
4976 case bad_imm4:
4977 err_msg = _("constant doesn't fit in 4 bits");
4978 break;
4979 case old_gcc_only:
4980 err_msg = _("only supported with old gcc");
4981 break;
4982 case unsupported_with_intel_mnemonic:
4983 err_msg = _("unsupported with Intel mnemonic");
4984 break;
4985 case unsupported_syntax:
4986 err_msg = _("unsupported syntax");
4987 break;
4988 case unsupported:
4989 as_bad (_("unsupported instruction `%s'"),
4990 current_templates->start->name);
4991 return NULL;
4992 case invalid_vsib_address:
4993 err_msg = _("invalid VSIB address");
4994 break;
4995 case invalid_vector_register_set:
4996 err_msg = _("mask, index, and destination registers must be distinct");
4997 break;
4998 case unsupported_vector_index_register:
4999 err_msg = _("unsupported vector index register");
5000 break;
5001 case unsupported_broadcast:
5002 err_msg = _("unsupported broadcast");
5003 break;
5004 case broadcast_not_on_src_operand:
5005 err_msg = _("broadcast not on source memory operand");
5006 break;
5007 case broadcast_needed:
5008 err_msg = _("broadcast is needed for operand of such type");
5009 break;
5010 case unsupported_masking:
5011 err_msg = _("unsupported masking");
5012 break;
5013 case mask_not_on_destination:
5014 err_msg = _("mask not on destination operand");
5015 break;
5016 case no_default_mask:
5017 err_msg = _("default mask isn't allowed");
5018 break;
5019 case unsupported_rc_sae:
5020 err_msg = _("unsupported static rounding/sae");
5021 break;
5022 case rc_sae_operand_not_last_imm:
5023 if (intel_syntax)
5024 err_msg = _("RC/SAE operand must precede immediate operands");
5025 else
5026 err_msg = _("RC/SAE operand must follow immediate operands");
5027 break;
5028 case invalid_register_operand:
5029 err_msg = _("invalid register operand");
5030 break;
5031 }
5032 as_bad (_("%s for `%s'"), err_msg,
5033 current_templates->start->name);
5034 return NULL;
5035 }
5036
5037 if (!quiet_warnings)
5038 {
5039 if (!intel_syntax
5040 && (i.types[0].bitfield.jumpabsolute
5041 != operand_types[0].bitfield.jumpabsolute))
5042 {
5043 as_warn (_("indirect %s without `*'"), t->name);
5044 }
5045
5046 if (t->opcode_modifier.isprefix
5047 && t->opcode_modifier.ignoresize)
5048 {
5049 /* Warn them that a data or address size prefix doesn't
5050 affect assembly of the next line of code. */
5051 as_warn (_("stand-alone `%s' prefix"), t->name);
5052 }
5053 }
5054
5055 /* Copy the template we found. */
5056 i.tm = *t;
5057
5058 if (addr_prefix_disp != -1)
5059 i.tm.operand_types[addr_prefix_disp]
5060 = operand_types[addr_prefix_disp];
5061
5062 if (found_reverse_match)
5063 {
5064 /* If we found a reverse match we must alter the opcode
5065 direction bit. found_reverse_match holds bits to change
5066 (different for int & float insns). */
5067
5068 i.tm.base_opcode ^= found_reverse_match;
5069
5070 i.tm.operand_types[0] = operand_types[1];
5071 i.tm.operand_types[1] = operand_types[0];
5072 }
5073
5074 return t;
5075 }
5076
5077 static int
5078 check_string (void)
5079 {
5080 int mem_op = operand_type_check (i.types[0], anymem) ? 0 : 1;
5081 if (i.tm.operand_types[mem_op].bitfield.esseg)
5082 {
5083 if (i.seg[0] != NULL && i.seg[0] != &es)
5084 {
5085 as_bad (_("`%s' operand %d must use `%ses' segment"),
5086 i.tm.name,
5087 mem_op + 1,
5088 register_prefix);
5089 return 0;
5090 }
5091 /* There's only ever one segment override allowed per instruction.
5092 This instruction possibly has a legal segment override on the
5093 second operand, so copy the segment to where non-string
5094 instructions store it, allowing common code. */
5095 i.seg[0] = i.seg[1];
5096 }
5097 else if (i.tm.operand_types[mem_op + 1].bitfield.esseg)
5098 {
5099 if (i.seg[1] != NULL && i.seg[1] != &es)
5100 {
5101 as_bad (_("`%s' operand %d must use `%ses' segment"),
5102 i.tm.name,
5103 mem_op + 2,
5104 register_prefix);
5105 return 0;
5106 }
5107 }
5108 return 1;
5109 }
5110
5111 static int
5112 process_suffix (void)
5113 {
5114 /* If matched instruction specifies an explicit instruction mnemonic
5115 suffix, use it. */
5116 if (i.tm.opcode_modifier.size16)
5117 i.suffix = WORD_MNEM_SUFFIX;
5118 else if (i.tm.opcode_modifier.size32)
5119 i.suffix = LONG_MNEM_SUFFIX;
5120 else if (i.tm.opcode_modifier.size64)
5121 i.suffix = QWORD_MNEM_SUFFIX;
5122 else if (i.reg_operands)
5123 {
5124 /* If there's no instruction mnemonic suffix we try to invent one
5125 based on register operands. */
5126 if (!i.suffix)
5127 {
5128 /* We take i.suffix from the last register operand specified,
5129 Destination register type is more significant than source
5130 register type. crc32 in SSE4.2 prefers source register
5131 type. */
5132 if (i.tm.base_opcode == 0xf20f38f1)
5133 {
5134 if (i.types[0].bitfield.reg16)
5135 i.suffix = WORD_MNEM_SUFFIX;
5136 else if (i.types[0].bitfield.reg32)
5137 i.suffix = LONG_MNEM_SUFFIX;
5138 else if (i.types[0].bitfield.reg64)
5139 i.suffix = QWORD_MNEM_SUFFIX;
5140 }
5141 else if (i.tm.base_opcode == 0xf20f38f0)
5142 {
5143 if (i.types[0].bitfield.reg8)
5144 i.suffix = BYTE_MNEM_SUFFIX;
5145 }
5146
5147 if (!i.suffix)
5148 {
5149 int op;
5150
5151 if (i.tm.base_opcode == 0xf20f38f1
5152 || i.tm.base_opcode == 0xf20f38f0)
5153 {
5154 /* We have to know the operand size for crc32. */
5155 as_bad (_("ambiguous memory operand size for `%s`"),
5156 i.tm.name);
5157 return 0;
5158 }
5159
5160 for (op = i.operands; --op >= 0;)
5161 if (!i.tm.operand_types[op].bitfield.inoutportreg)
5162 {
5163 if (i.types[op].bitfield.reg8)
5164 {
5165 i.suffix = BYTE_MNEM_SUFFIX;
5166 break;
5167 }
5168 else if (i.types[op].bitfield.reg16)
5169 {
5170 i.suffix = WORD_MNEM_SUFFIX;
5171 break;
5172 }
5173 else if (i.types[op].bitfield.reg32)
5174 {
5175 i.suffix = LONG_MNEM_SUFFIX;
5176 break;
5177 }
5178 else if (i.types[op].bitfield.reg64)
5179 {
5180 i.suffix = QWORD_MNEM_SUFFIX;
5181 break;
5182 }
5183 }
5184 }
5185 }
5186 else if (i.suffix == BYTE_MNEM_SUFFIX)
5187 {
5188 if (intel_syntax
5189 && i.tm.opcode_modifier.ignoresize
5190 && i.tm.opcode_modifier.no_bsuf)
5191 i.suffix = 0;
5192 else if (!check_byte_reg ())
5193 return 0;
5194 }
5195 else if (i.suffix == LONG_MNEM_SUFFIX)
5196 {
5197 if (intel_syntax
5198 && i.tm.opcode_modifier.ignoresize
5199 && i.tm.opcode_modifier.no_lsuf)
5200 i.suffix = 0;
5201 else if (!check_long_reg ())
5202 return 0;
5203 }
5204 else if (i.suffix == QWORD_MNEM_SUFFIX)
5205 {
5206 if (intel_syntax
5207 && i.tm.opcode_modifier.ignoresize
5208 && i.tm.opcode_modifier.no_qsuf)
5209 i.suffix = 0;
5210 else if (!check_qword_reg ())
5211 return 0;
5212 }
5213 else if (i.suffix == WORD_MNEM_SUFFIX)
5214 {
5215 if (intel_syntax
5216 && i.tm.opcode_modifier.ignoresize
5217 && i.tm.opcode_modifier.no_wsuf)
5218 i.suffix = 0;
5219 else if (!check_word_reg ())
5220 return 0;
5221 }
5222 else if (i.suffix == XMMWORD_MNEM_SUFFIX
5223 || i.suffix == YMMWORD_MNEM_SUFFIX
5224 || i.suffix == ZMMWORD_MNEM_SUFFIX)
5225 {
5226 /* Skip if the instruction has x/y/z suffix. match_template
5227 should check if it is a valid suffix. */
5228 }
5229 else if (intel_syntax && i.tm.opcode_modifier.ignoresize)
5230 /* Do nothing if the instruction is going to ignore the prefix. */
5231 ;
5232 else
5233 abort ();
5234 }
5235 else if (i.tm.opcode_modifier.defaultsize
5236 && !i.suffix
5237 /* exclude fldenv/frstor/fsave/fstenv */
5238 && i.tm.opcode_modifier.no_ssuf)
5239 {
5240 i.suffix = stackop_size;
5241 }
5242 else if (intel_syntax
5243 && !i.suffix
5244 && (i.tm.operand_types[0].bitfield.jumpabsolute
5245 || i.tm.opcode_modifier.jumpbyte
5246 || i.tm.opcode_modifier.jumpintersegment
5247 || (i.tm.base_opcode == 0x0f01 /* [ls][gi]dt */
5248 && i.tm.extension_opcode <= 3)))
5249 {
5250 switch (flag_code)
5251 {
5252 case CODE_64BIT:
5253 if (!i.tm.opcode_modifier.no_qsuf)
5254 {
5255 i.suffix = QWORD_MNEM_SUFFIX;
5256 break;
5257 }
5258 case CODE_32BIT:
5259 if (!i.tm.opcode_modifier.no_lsuf)
5260 i.suffix = LONG_MNEM_SUFFIX;
5261 break;
5262 case CODE_16BIT:
5263 if (!i.tm.opcode_modifier.no_wsuf)
5264 i.suffix = WORD_MNEM_SUFFIX;
5265 break;
5266 }
5267 }
5268
5269 if (!i.suffix)
5270 {
5271 if (!intel_syntax)
5272 {
5273 if (i.tm.opcode_modifier.w)
5274 {
5275 as_bad (_("no instruction mnemonic suffix given and "
5276 "no register operands; can't size instruction"));
5277 return 0;
5278 }
5279 }
5280 else
5281 {
5282 unsigned int suffixes;
5283
5284 suffixes = !i.tm.opcode_modifier.no_bsuf;
5285 if (!i.tm.opcode_modifier.no_wsuf)
5286 suffixes |= 1 << 1;
5287 if (!i.tm.opcode_modifier.no_lsuf)
5288 suffixes |= 1 << 2;
5289 if (!i.tm.opcode_modifier.no_ldsuf)
5290 suffixes |= 1 << 3;
5291 if (!i.tm.opcode_modifier.no_ssuf)
5292 suffixes |= 1 << 4;
5293 if (!i.tm.opcode_modifier.no_qsuf)
5294 suffixes |= 1 << 5;
5295
5296 /* There are more than suffix matches. */
5297 if (i.tm.opcode_modifier.w
5298 || ((suffixes & (suffixes - 1))
5299 && !i.tm.opcode_modifier.defaultsize
5300 && !i.tm.opcode_modifier.ignoresize))
5301 {
5302 as_bad (_("ambiguous operand size for `%s'"), i.tm.name);
5303 return 0;
5304 }
5305 }
5306 }
5307
5308 /* Change the opcode based on the operand size given by i.suffix;
5309 We don't need to change things for byte insns. */
5310
5311 if (i.suffix
5312 && i.suffix != BYTE_MNEM_SUFFIX
5313 && i.suffix != XMMWORD_MNEM_SUFFIX
5314 && i.suffix != YMMWORD_MNEM_SUFFIX
5315 && i.suffix != ZMMWORD_MNEM_SUFFIX)
5316 {
5317 /* It's not a byte, select word/dword operation. */
5318 if (i.tm.opcode_modifier.w)
5319 {
5320 if (i.tm.opcode_modifier.shortform)
5321 i.tm.base_opcode |= 8;
5322 else
5323 i.tm.base_opcode |= 1;
5324 }
5325
5326 /* Now select between word & dword operations via the operand
5327 size prefix, except for instructions that will ignore this
5328 prefix anyway. */
5329 if (i.tm.opcode_modifier.addrprefixop0)
5330 {
5331 /* The address size override prefix changes the size of the
5332 first operand. */
5333 if ((flag_code == CODE_32BIT
5334 && i.op->regs[0].reg_type.bitfield.reg16)
5335 || (flag_code != CODE_32BIT
5336 && i.op->regs[0].reg_type.bitfield.reg32))
5337 if (!add_prefix (ADDR_PREFIX_OPCODE))
5338 return 0;
5339 }
5340 else if (i.suffix != QWORD_MNEM_SUFFIX
5341 && i.suffix != LONG_DOUBLE_MNEM_SUFFIX
5342 && !i.tm.opcode_modifier.ignoresize
5343 && !i.tm.opcode_modifier.floatmf
5344 && ((i.suffix == LONG_MNEM_SUFFIX) == (flag_code == CODE_16BIT)
5345 || (flag_code == CODE_64BIT
5346 && i.tm.opcode_modifier.jumpbyte)))
5347 {
5348 unsigned int prefix = DATA_PREFIX_OPCODE;
5349
5350 if (i.tm.opcode_modifier.jumpbyte) /* jcxz, loop */
5351 prefix = ADDR_PREFIX_OPCODE;
5352
5353 if (!add_prefix (prefix))
5354 return 0;
5355 }
5356
5357 /* Set mode64 for an operand. */
5358 if (i.suffix == QWORD_MNEM_SUFFIX
5359 && flag_code == CODE_64BIT
5360 && !i.tm.opcode_modifier.norex64)
5361 {
5362 /* Special case for xchg %rax,%rax. It is NOP and doesn't
5363 need rex64. cmpxchg8b is also a special case. */
5364 if (! (i.operands == 2
5365 && i.tm.base_opcode == 0x90
5366 && i.tm.extension_opcode == None
5367 && operand_type_equal (&i.types [0], &acc64)
5368 && operand_type_equal (&i.types [1], &acc64))
5369 && ! (i.operands == 1
5370 && i.tm.base_opcode == 0xfc7
5371 && i.tm.extension_opcode == 1
5372 && !operand_type_check (i.types [0], reg)
5373 && operand_type_check (i.types [0], anymem)))
5374 i.rex |= REX_W;
5375 }
5376
5377 /* Size floating point instruction. */
5378 if (i.suffix == LONG_MNEM_SUFFIX)
5379 if (i.tm.opcode_modifier.floatmf)
5380 i.tm.base_opcode ^= 4;
5381 }
5382
5383 return 1;
5384 }
5385
5386 static int
5387 check_byte_reg (void)
5388 {
5389 int op;
5390
5391 for (op = i.operands; --op >= 0;)
5392 {
5393 /* If this is an eight bit register, it's OK. If it's the 16 or
5394 32 bit version of an eight bit register, we will just use the
5395 low portion, and that's OK too. */
5396 if (i.types[op].bitfield.reg8)
5397 continue;
5398
5399 /* I/O port address operands are OK too. */
5400 if (i.tm.operand_types[op].bitfield.inoutportreg)
5401 continue;
5402
5403 /* crc32 doesn't generate this warning. */
5404 if (i.tm.base_opcode == 0xf20f38f0)
5405 continue;
5406
5407 if ((i.types[op].bitfield.reg16
5408 || i.types[op].bitfield.reg32
5409 || i.types[op].bitfield.reg64)
5410 && i.op[op].regs->reg_num < 4
5411 /* Prohibit these changes in 64bit mode, since the lowering
5412 would be more complicated. */
5413 && flag_code != CODE_64BIT)
5414 {
5415 #if REGISTER_WARNINGS
5416 if (!quiet_warnings)
5417 as_warn (_("using `%s%s' instead of `%s%s' due to `%c' suffix"),
5418 register_prefix,
5419 (i.op[op].regs + (i.types[op].bitfield.reg16
5420 ? REGNAM_AL - REGNAM_AX
5421 : REGNAM_AL - REGNAM_EAX))->reg_name,
5422 register_prefix,
5423 i.op[op].regs->reg_name,
5424 i.suffix);
5425 #endif
5426 continue;
5427 }
5428 /* Any other register is bad. */
5429 if (i.types[op].bitfield.reg16
5430 || i.types[op].bitfield.reg32
5431 || i.types[op].bitfield.reg64
5432 || i.types[op].bitfield.regmmx
5433 || i.types[op].bitfield.regxmm
5434 || i.types[op].bitfield.regymm
5435 || i.types[op].bitfield.regzmm
5436 || i.types[op].bitfield.sreg2
5437 || i.types[op].bitfield.sreg3
5438 || i.types[op].bitfield.control
5439 || i.types[op].bitfield.debug
5440 || i.types[op].bitfield.test
5441 || i.types[op].bitfield.floatreg
5442 || i.types[op].bitfield.floatacc)
5443 {
5444 as_bad (_("`%s%s' not allowed with `%s%c'"),
5445 register_prefix,
5446 i.op[op].regs->reg_name,
5447 i.tm.name,
5448 i.suffix);
5449 return 0;
5450 }
5451 }
5452 return 1;
5453 }
5454
5455 static int
5456 check_long_reg (void)
5457 {
5458 int op;
5459
5460 for (op = i.operands; --op >= 0;)
5461 /* Reject eight bit registers, except where the template requires
5462 them. (eg. movzb) */
5463 if (i.types[op].bitfield.reg8
5464 && (i.tm.operand_types[op].bitfield.reg16
5465 || i.tm.operand_types[op].bitfield.reg32
5466 || i.tm.operand_types[op].bitfield.acc))
5467 {
5468 as_bad (_("`%s%s' not allowed with `%s%c'"),
5469 register_prefix,
5470 i.op[op].regs->reg_name,
5471 i.tm.name,
5472 i.suffix);
5473 return 0;
5474 }
5475 /* Warn if the e prefix on a general reg is missing. */
5476 else if ((!quiet_warnings || flag_code == CODE_64BIT)
5477 && i.types[op].bitfield.reg16
5478 && (i.tm.operand_types[op].bitfield.reg32
5479 || i.tm.operand_types[op].bitfield.acc))
5480 {
5481 /* Prohibit these changes in the 64bit mode, since the
5482 lowering is more complicated. */
5483 if (flag_code == CODE_64BIT)
5484 {
5485 as_bad (_("incorrect register `%s%s' used with `%c' suffix"),
5486 register_prefix, i.op[op].regs->reg_name,
5487 i.suffix);
5488 return 0;
5489 }
5490 #if REGISTER_WARNINGS
5491 as_warn (_("using `%s%s' instead of `%s%s' due to `%c' suffix"),
5492 register_prefix,
5493 (i.op[op].regs + REGNAM_EAX - REGNAM_AX)->reg_name,
5494 register_prefix, i.op[op].regs->reg_name, i.suffix);
5495 #endif
5496 }
5497 /* Warn if the r prefix on a general reg is present. */
5498 else if (i.types[op].bitfield.reg64
5499 && (i.tm.operand_types[op].bitfield.reg32
5500 || i.tm.operand_types[op].bitfield.acc))
5501 {
5502 if (intel_syntax
5503 && i.tm.opcode_modifier.toqword
5504 && !i.types[0].bitfield.regxmm)
5505 {
5506 /* Convert to QWORD. We want REX byte. */
5507 i.suffix = QWORD_MNEM_SUFFIX;
5508 }
5509 else
5510 {
5511 as_bad (_("incorrect register `%s%s' used with `%c' suffix"),
5512 register_prefix, i.op[op].regs->reg_name,
5513 i.suffix);
5514 return 0;
5515 }
5516 }
5517 return 1;
5518 }
5519
5520 static int
5521 check_qword_reg (void)
5522 {
5523 int op;
5524
5525 for (op = i.operands; --op >= 0; )
5526 /* Reject eight bit registers, except where the template requires
5527 them. (eg. movzb) */
5528 if (i.types[op].bitfield.reg8
5529 && (i.tm.operand_types[op].bitfield.reg16
5530 || i.tm.operand_types[op].bitfield.reg32
5531 || i.tm.operand_types[op].bitfield.acc))
5532 {
5533 as_bad (_("`%s%s' not allowed with `%s%c'"),
5534 register_prefix,
5535 i.op[op].regs->reg_name,
5536 i.tm.name,
5537 i.suffix);
5538 return 0;
5539 }
5540 /* Warn if the r prefix on a general reg is missing. */
5541 else if ((i.types[op].bitfield.reg16
5542 || i.types[op].bitfield.reg32)
5543 && (i.tm.operand_types[op].bitfield.reg32
5544 || i.tm.operand_types[op].bitfield.acc))
5545 {
5546 /* Prohibit these changes in the 64bit mode, since the
5547 lowering is more complicated. */
5548 if (intel_syntax
5549 && i.tm.opcode_modifier.todword
5550 && !i.types[0].bitfield.regxmm)
5551 {
5552 /* Convert to DWORD. We don't want REX byte. */
5553 i.suffix = LONG_MNEM_SUFFIX;
5554 }
5555 else
5556 {
5557 as_bad (_("incorrect register `%s%s' used with `%c' suffix"),
5558 register_prefix, i.op[op].regs->reg_name,
5559 i.suffix);
5560 return 0;
5561 }
5562 }
5563 return 1;
5564 }
5565
5566 static int
5567 check_word_reg (void)
5568 {
5569 int op;
5570 for (op = i.operands; --op >= 0;)
5571 /* Reject eight bit registers, except where the template requires
5572 them. (eg. movzb) */
5573 if (i.types[op].bitfield.reg8
5574 && (i.tm.operand_types[op].bitfield.reg16
5575 || i.tm.operand_types[op].bitfield.reg32
5576 || i.tm.operand_types[op].bitfield.acc))
5577 {
5578 as_bad (_("`%s%s' not allowed with `%s%c'"),
5579 register_prefix,
5580 i.op[op].regs->reg_name,
5581 i.tm.name,
5582 i.suffix);
5583 return 0;
5584 }
5585 /* Warn if the e or r prefix on a general reg is present. */
5586 else if ((!quiet_warnings || flag_code == CODE_64BIT)
5587 && (i.types[op].bitfield.reg32
5588 || i.types[op].bitfield.reg64)
5589 && (i.tm.operand_types[op].bitfield.reg16
5590 || i.tm.operand_types[op].bitfield.acc))
5591 {
5592 /* Prohibit these changes in the 64bit mode, since the
5593 lowering is more complicated. */
5594 if (flag_code == CODE_64BIT)
5595 {
5596 as_bad (_("incorrect register `%s%s' used with `%c' suffix"),
5597 register_prefix, i.op[op].regs->reg_name,
5598 i.suffix);
5599 return 0;
5600 }
5601 #if REGISTER_WARNINGS
5602 as_warn (_("using `%s%s' instead of `%s%s' due to `%c' suffix"),
5603 register_prefix,
5604 (i.op[op].regs + REGNAM_AX - REGNAM_EAX)->reg_name,
5605 register_prefix, i.op[op].regs->reg_name, i.suffix);
5606 #endif
5607 }
5608 return 1;
5609 }
5610
5611 static int
5612 update_imm (unsigned int j)
5613 {
5614 i386_operand_type overlap = i.types[j];
5615 if ((overlap.bitfield.imm8
5616 || overlap.bitfield.imm8s
5617 || overlap.bitfield.imm16
5618 || overlap.bitfield.imm32
5619 || overlap.bitfield.imm32s
5620 || overlap.bitfield.imm64)
5621 && !operand_type_equal (&overlap, &imm8)
5622 && !operand_type_equal (&overlap, &imm8s)
5623 && !operand_type_equal (&overlap, &imm16)
5624 && !operand_type_equal (&overlap, &imm32)
5625 && !operand_type_equal (&overlap, &imm32s)
5626 && !operand_type_equal (&overlap, &imm64))
5627 {
5628 if (i.suffix)
5629 {
5630 i386_operand_type temp;
5631
5632 operand_type_set (&temp, 0);
5633 if (i.suffix == BYTE_MNEM_SUFFIX)
5634 {
5635 temp.bitfield.imm8 = overlap.bitfield.imm8;
5636 temp.bitfield.imm8s = overlap.bitfield.imm8s;
5637 }
5638 else if (i.suffix == WORD_MNEM_SUFFIX)
5639 temp.bitfield.imm16 = overlap.bitfield.imm16;
5640 else if (i.suffix == QWORD_MNEM_SUFFIX)
5641 {
5642 temp.bitfield.imm64 = overlap.bitfield.imm64;
5643 temp.bitfield.imm32s = overlap.bitfield.imm32s;
5644 }
5645 else
5646 temp.bitfield.imm32 = overlap.bitfield.imm32;
5647 overlap = temp;
5648 }
5649 else if (operand_type_equal (&overlap, &imm16_32_32s)
5650 || operand_type_equal (&overlap, &imm16_32)
5651 || operand_type_equal (&overlap, &imm16_32s))
5652 {
5653 if ((flag_code == CODE_16BIT) ^ (i.prefix[DATA_PREFIX] != 0))
5654 overlap = imm16;
5655 else
5656 overlap = imm32s;
5657 }
5658 if (!operand_type_equal (&overlap, &imm8)
5659 && !operand_type_equal (&overlap, &imm8s)
5660 && !operand_type_equal (&overlap, &imm16)
5661 && !operand_type_equal (&overlap, &imm32)
5662 && !operand_type_equal (&overlap, &imm32s)
5663 && !operand_type_equal (&overlap, &imm64))
5664 {
5665 as_bad (_("no instruction mnemonic suffix given; "
5666 "can't determine immediate size"));
5667 return 0;
5668 }
5669 }
5670 i.types[j] = overlap;
5671
5672 return 1;
5673 }
5674
5675 static int
5676 finalize_imm (void)
5677 {
5678 unsigned int j, n;
5679
5680 /* Update the first 2 immediate operands. */
5681 n = i.operands > 2 ? 2 : i.operands;
5682 if (n)
5683 {
5684 for (j = 0; j < n; j++)
5685 if (update_imm (j) == 0)
5686 return 0;
5687
5688 /* The 3rd operand can't be immediate operand. */
5689 gas_assert (operand_type_check (i.types[2], imm) == 0);
5690 }
5691
5692 return 1;
5693 }
5694
5695 static int
5696 bad_implicit_operand (int xmm)
5697 {
5698 const char *ireg = xmm ? "xmm0" : "ymm0";
5699
5700 if (intel_syntax)
5701 as_bad (_("the last operand of `%s' must be `%s%s'"),
5702 i.tm.name, register_prefix, ireg);
5703 else
5704 as_bad (_("the first operand of `%s' must be `%s%s'"),
5705 i.tm.name, register_prefix, ireg);
5706 return 0;
5707 }
5708
5709 static int
5710 process_operands (void)
5711 {
5712 /* Default segment register this instruction will use for memory
5713 accesses. 0 means unknown. This is only for optimizing out
5714 unnecessary segment overrides. */
5715 const seg_entry *default_seg = 0;
5716
5717 if (i.tm.opcode_modifier.sse2avx && i.tm.opcode_modifier.vexvvvv)
5718 {
5719 unsigned int dupl = i.operands;
5720 unsigned int dest = dupl - 1;
5721 unsigned int j;
5722
5723 /* The destination must be an xmm register. */
5724 gas_assert (i.reg_operands
5725 && MAX_OPERANDS > dupl
5726 && operand_type_equal (&i.types[dest], &regxmm));
5727
5728 if (i.tm.opcode_modifier.firstxmm0)
5729 {
5730 /* The first operand is implicit and must be xmm0. */
5731 gas_assert (operand_type_equal (&i.types[0], &regxmm));
5732 if (register_number (i.op[0].regs) != 0)
5733 return bad_implicit_operand (1);
5734
5735 if (i.tm.opcode_modifier.vexsources == VEX3SOURCES)
5736 {
5737 /* Keep xmm0 for instructions with VEX prefix and 3
5738 sources. */
5739 goto duplicate;
5740 }
5741 else
5742 {
5743 /* We remove the first xmm0 and keep the number of
5744 operands unchanged, which in fact duplicates the
5745 destination. */
5746 for (j = 1; j < i.operands; j++)
5747 {
5748 i.op[j - 1] = i.op[j];
5749 i.types[j - 1] = i.types[j];
5750 i.tm.operand_types[j - 1] = i.tm.operand_types[j];
5751 }
5752 }
5753 }
5754 else if (i.tm.opcode_modifier.implicit1stxmm0)
5755 {
5756 gas_assert ((MAX_OPERANDS - 1) > dupl
5757 && (i.tm.opcode_modifier.vexsources
5758 == VEX3SOURCES));
5759
5760 /* Add the implicit xmm0 for instructions with VEX prefix
5761 and 3 sources. */
5762 for (j = i.operands; j > 0; j--)
5763 {
5764 i.op[j] = i.op[j - 1];
5765 i.types[j] = i.types[j - 1];
5766 i.tm.operand_types[j] = i.tm.operand_types[j - 1];
5767 }
5768 i.op[0].regs
5769 = (const reg_entry *) hash_find (reg_hash, "xmm0");
5770 i.types[0] = regxmm;
5771 i.tm.operand_types[0] = regxmm;
5772
5773 i.operands += 2;
5774 i.reg_operands += 2;
5775 i.tm.operands += 2;
5776
5777 dupl++;
5778 dest++;
5779 i.op[dupl] = i.op[dest];
5780 i.types[dupl] = i.types[dest];
5781 i.tm.operand_types[dupl] = i.tm.operand_types[dest];
5782 }
5783 else
5784 {
5785 duplicate:
5786 i.operands++;
5787 i.reg_operands++;
5788 i.tm.operands++;
5789
5790 i.op[dupl] = i.op[dest];
5791 i.types[dupl] = i.types[dest];
5792 i.tm.operand_types[dupl] = i.tm.operand_types[dest];
5793 }
5794
5795 if (i.tm.opcode_modifier.immext)
5796 process_immext ();
5797 }
5798 else if (i.tm.opcode_modifier.firstxmm0)
5799 {
5800 unsigned int j;
5801
5802 /* The first operand is implicit and must be xmm0/ymm0/zmm0. */
5803 gas_assert (i.reg_operands
5804 && (operand_type_equal (&i.types[0], &regxmm)
5805 || operand_type_equal (&i.types[0], &regymm)
5806 || operand_type_equal (&i.types[0], &regzmm)));
5807 if (register_number (i.op[0].regs) != 0)
5808 return bad_implicit_operand (i.types[0].bitfield.regxmm);
5809
5810 for (j = 1; j < i.operands; j++)
5811 {
5812 i.op[j - 1] = i.op[j];
5813 i.types[j - 1] = i.types[j];
5814
5815 /* We need to adjust fields in i.tm since they are used by
5816 build_modrm_byte. */
5817 i.tm.operand_types [j - 1] = i.tm.operand_types [j];
5818 }
5819
5820 i.operands--;
5821 i.reg_operands--;
5822 i.tm.operands--;
5823 }
5824 else if (i.tm.opcode_modifier.regkludge)
5825 {
5826 /* The imul $imm, %reg instruction is converted into
5827 imul $imm, %reg, %reg, and the clr %reg instruction
5828 is converted into xor %reg, %reg. */
5829
5830 unsigned int first_reg_op;
5831
5832 if (operand_type_check (i.types[0], reg))
5833 first_reg_op = 0;
5834 else
5835 first_reg_op = 1;
5836 /* Pretend we saw the extra register operand. */
5837 gas_assert (i.reg_operands == 1
5838 && i.op[first_reg_op + 1].regs == 0);
5839 i.op[first_reg_op + 1].regs = i.op[first_reg_op].regs;
5840 i.types[first_reg_op + 1] = i.types[first_reg_op];
5841 i.operands++;
5842 i.reg_operands++;
5843 }
5844
5845 if (i.tm.opcode_modifier.shortform)
5846 {
5847 if (i.types[0].bitfield.sreg2
5848 || i.types[0].bitfield.sreg3)
5849 {
5850 if (i.tm.base_opcode == POP_SEG_SHORT
5851 && i.op[0].regs->reg_num == 1)
5852 {
5853 as_bad (_("you can't `pop %scs'"), register_prefix);
5854 return 0;
5855 }
5856 i.tm.base_opcode |= (i.op[0].regs->reg_num << 3);
5857 if ((i.op[0].regs->reg_flags & RegRex) != 0)
5858 i.rex |= REX_B;
5859 }
5860 else
5861 {
5862 /* The register or float register operand is in operand
5863 0 or 1. */
5864 unsigned int op;
5865
5866 if (i.types[0].bitfield.floatreg
5867 || operand_type_check (i.types[0], reg))
5868 op = 0;
5869 else
5870 op = 1;
5871 /* Register goes in low 3 bits of opcode. */
5872 i.tm.base_opcode |= i.op[op].regs->reg_num;
5873 if ((i.op[op].regs->reg_flags & RegRex) != 0)
5874 i.rex |= REX_B;
5875 if (!quiet_warnings && i.tm.opcode_modifier.ugh)
5876 {
5877 /* Warn about some common errors, but press on regardless.
5878 The first case can be generated by gcc (<= 2.8.1). */
5879 if (i.operands == 2)
5880 {
5881 /* Reversed arguments on faddp, fsubp, etc. */
5882 as_warn (_("translating to `%s %s%s,%s%s'"), i.tm.name,
5883 register_prefix, i.op[!intel_syntax].regs->reg_name,
5884 register_prefix, i.op[intel_syntax].regs->reg_name);
5885 }
5886 else
5887 {
5888 /* Extraneous `l' suffix on fp insn. */
5889 as_warn (_("translating to `%s %s%s'"), i.tm.name,
5890 register_prefix, i.op[0].regs->reg_name);
5891 }
5892 }
5893 }
5894 }
5895 else if (i.tm.opcode_modifier.modrm)
5896 {
5897 /* The opcode is completed (modulo i.tm.extension_opcode which
5898 must be put into the modrm byte). Now, we make the modrm and
5899 index base bytes based on all the info we've collected. */
5900
5901 default_seg = build_modrm_byte ();
5902 }
5903 else if ((i.tm.base_opcode & ~0x3) == MOV_AX_DISP32)
5904 {
5905 default_seg = &ds;
5906 }
5907 else if (i.tm.opcode_modifier.isstring)
5908 {
5909 /* For the string instructions that allow a segment override
5910 on one of their operands, the default segment is ds. */
5911 default_seg = &ds;
5912 }
5913
5914 if (i.tm.base_opcode == 0x8d /* lea */
5915 && i.seg[0]
5916 && !quiet_warnings)
5917 as_warn (_("segment override on `%s' is ineffectual"), i.tm.name);
5918
5919 /* If a segment was explicitly specified, and the specified segment
5920 is not the default, use an opcode prefix to select it. If we
5921 never figured out what the default segment is, then default_seg
5922 will be zero at this point, and the specified segment prefix will
5923 always be used. */
5924 if ((i.seg[0]) && (i.seg[0] != default_seg))
5925 {
5926 if (!add_prefix (i.seg[0]->seg_prefix))
5927 return 0;
5928 }
5929 return 1;
5930 }
5931
5932 static const seg_entry *
5933 build_modrm_byte (void)
5934 {
5935 const seg_entry *default_seg = 0;
5936 unsigned int source, dest;
5937 int vex_3_sources;
5938
5939 /* The first operand of instructions with VEX prefix and 3 sources
5940 must be VEX_Imm4. */
5941 vex_3_sources = i.tm.opcode_modifier.vexsources == VEX3SOURCES;
5942 if (vex_3_sources)
5943 {
5944 unsigned int nds, reg_slot;
5945 expressionS *exp;
5946
5947 if (i.tm.opcode_modifier.veximmext
5948 && i.tm.opcode_modifier.immext)
5949 {
5950 dest = i.operands - 2;
5951 gas_assert (dest == 3);
5952 }
5953 else
5954 dest = i.operands - 1;
5955 nds = dest - 1;
5956
5957 /* There are 2 kinds of instructions:
5958 1. 5 operands: 4 register operands or 3 register operands
5959 plus 1 memory operand plus one Vec_Imm4 operand, VexXDS, and
5960 VexW0 or VexW1. The destination must be either XMM, YMM or
5961 ZMM register.
5962 2. 4 operands: 4 register operands or 3 register operands
5963 plus 1 memory operand, VexXDS, and VexImmExt */
5964 gas_assert ((i.reg_operands == 4
5965 || (i.reg_operands == 3 && i.mem_operands == 1))
5966 && i.tm.opcode_modifier.vexvvvv == VEXXDS
5967 && (i.tm.opcode_modifier.veximmext
5968 || (i.imm_operands == 1
5969 && i.types[0].bitfield.vec_imm4
5970 && (i.tm.opcode_modifier.vexw == VEXW0
5971 || i.tm.opcode_modifier.vexw == VEXW1)
5972 && (operand_type_equal (&i.tm.operand_types[dest], &regxmm)
5973 || operand_type_equal (&i.tm.operand_types[dest], &regymm)
5974 || operand_type_equal (&i.tm.operand_types[dest], &regzmm)))));
5975
5976 if (i.imm_operands == 0)
5977 {
5978 /* When there is no immediate operand, generate an 8bit
5979 immediate operand to encode the first operand. */
5980 exp = &im_expressions[i.imm_operands++];
5981 i.op[i.operands].imms = exp;
5982 i.types[i.operands] = imm8;
5983 i.operands++;
5984 /* If VexW1 is set, the first operand is the source and
5985 the second operand is encoded in the immediate operand. */
5986 if (i.tm.opcode_modifier.vexw == VEXW1)
5987 {
5988 source = 0;
5989 reg_slot = 1;
5990 }
5991 else
5992 {
5993 source = 1;
5994 reg_slot = 0;
5995 }
5996
5997 /* FMA swaps REG and NDS. */
5998 if (i.tm.cpu_flags.bitfield.cpufma)
5999 {
6000 unsigned int tmp;
6001 tmp = reg_slot;
6002 reg_slot = nds;
6003 nds = tmp;
6004 }
6005
6006 gas_assert (operand_type_equal (&i.tm.operand_types[reg_slot],
6007 &regxmm)
6008 || operand_type_equal (&i.tm.operand_types[reg_slot],
6009 &regymm)
6010 || operand_type_equal (&i.tm.operand_types[reg_slot],
6011 &regzmm));
6012 exp->X_op = O_constant;
6013 exp->X_add_number = register_number (i.op[reg_slot].regs) << 4;
6014 gas_assert ((i.op[reg_slot].regs->reg_flags & RegVRex) == 0);
6015 }
6016 else
6017 {
6018 unsigned int imm_slot;
6019
6020 if (i.tm.opcode_modifier.vexw == VEXW0)
6021 {
6022 /* If VexW0 is set, the third operand is the source and
6023 the second operand is encoded in the immediate
6024 operand. */
6025 source = 2;
6026 reg_slot = 1;
6027 }
6028 else
6029 {
6030 /* VexW1 is set, the second operand is the source and
6031 the third operand is encoded in the immediate
6032 operand. */
6033 source = 1;
6034 reg_slot = 2;
6035 }
6036
6037 if (i.tm.opcode_modifier.immext)
6038 {
6039 /* When ImmExt is set, the immdiate byte is the last
6040 operand. */
6041 imm_slot = i.operands - 1;
6042 source--;
6043 reg_slot--;
6044 }
6045 else
6046 {
6047 imm_slot = 0;
6048
6049 /* Turn on Imm8 so that output_imm will generate it. */
6050 i.types[imm_slot].bitfield.imm8 = 1;
6051 }
6052
6053 gas_assert (operand_type_equal (&i.tm.operand_types[reg_slot],
6054 &regxmm)
6055 || operand_type_equal (&i.tm.operand_types[reg_slot],
6056 &regymm)
6057 || operand_type_equal (&i.tm.operand_types[reg_slot],
6058 &regzmm));
6059 i.op[imm_slot].imms->X_add_number
6060 |= register_number (i.op[reg_slot].regs) << 4;
6061 gas_assert ((i.op[reg_slot].regs->reg_flags & RegVRex) == 0);
6062 }
6063
6064 gas_assert (operand_type_equal (&i.tm.operand_types[nds], &regxmm)
6065 || operand_type_equal (&i.tm.operand_types[nds],
6066 &regymm)
6067 || operand_type_equal (&i.tm.operand_types[nds],
6068 &regzmm));
6069 i.vex.register_specifier = i.op[nds].regs;
6070 }
6071 else
6072 source = dest = 0;
6073
6074 /* i.reg_operands MUST be the number of real register operands;
6075 implicit registers do not count. If there are 3 register
6076 operands, it must be a instruction with VexNDS. For a
6077 instruction with VexNDD, the destination register is encoded
6078 in VEX prefix. If there are 4 register operands, it must be
6079 a instruction with VEX prefix and 3 sources. */
6080 if (i.mem_operands == 0
6081 && ((i.reg_operands == 2
6082 && i.tm.opcode_modifier.vexvvvv <= VEXXDS)
6083 || (i.reg_operands == 3
6084 && i.tm.opcode_modifier.vexvvvv == VEXXDS)
6085 || (i.reg_operands == 4 && vex_3_sources)))
6086 {
6087 switch (i.operands)
6088 {
6089 case 2:
6090 source = 0;
6091 break;
6092 case 3:
6093 /* When there are 3 operands, one of them may be immediate,
6094 which may be the first or the last operand. Otherwise,
6095 the first operand must be shift count register (cl) or it
6096 is an instruction with VexNDS. */
6097 gas_assert (i.imm_operands == 1
6098 || (i.imm_operands == 0
6099 && (i.tm.opcode_modifier.vexvvvv == VEXXDS
6100 || i.types[0].bitfield.shiftcount)));
6101 if (operand_type_check (i.types[0], imm)
6102 || i.types[0].bitfield.shiftcount)
6103 source = 1;
6104 else
6105 source = 0;
6106 break;
6107 case 4:
6108 /* When there are 4 operands, the first two must be 8bit
6109 immediate operands. The source operand will be the 3rd
6110 one.
6111
6112 For instructions with VexNDS, if the first operand
6113 an imm8, the source operand is the 2nd one. If the last
6114 operand is imm8, the source operand is the first one. */
6115 gas_assert ((i.imm_operands == 2
6116 && i.types[0].bitfield.imm8
6117 && i.types[1].bitfield.imm8)
6118 || (i.tm.opcode_modifier.vexvvvv == VEXXDS
6119 && i.imm_operands == 1
6120 && (i.types[0].bitfield.imm8
6121 || i.types[i.operands - 1].bitfield.imm8
6122 || i.rounding)));
6123 if (i.imm_operands == 2)
6124 source = 2;
6125 else
6126 {
6127 if (i.types[0].bitfield.imm8)
6128 source = 1;
6129 else
6130 source = 0;
6131 }
6132 break;
6133 case 5:
6134 if (i.tm.opcode_modifier.evex)
6135 {
6136 /* For EVEX instructions, when there are 5 operands, the
6137 first one must be immediate operand. If the second one
6138 is immediate operand, the source operand is the 3th
6139 one. If the last one is immediate operand, the source
6140 operand is the 2nd one. */
6141 gas_assert (i.imm_operands == 2
6142 && i.tm.opcode_modifier.sae
6143 && operand_type_check (i.types[0], imm));
6144 if (operand_type_check (i.types[1], imm))
6145 source = 2;
6146 else if (operand_type_check (i.types[4], imm))
6147 source = 1;
6148 else
6149 abort ();
6150 }
6151 break;
6152 default:
6153 abort ();
6154 }
6155
6156 if (!vex_3_sources)
6157 {
6158 dest = source + 1;
6159
6160 /* RC/SAE operand could be between DEST and SRC. That happens
6161 when one operand is GPR and the other one is XMM/YMM/ZMM
6162 register. */
6163 if (i.rounding && i.rounding->operand == (int) dest)
6164 dest++;
6165
6166 if (i.tm.opcode_modifier.vexvvvv == VEXXDS)
6167 {
6168 /* For instructions with VexNDS, the register-only source
6169 operand must be 32/64bit integer, XMM, YMM or ZMM
6170 register. It is encoded in VEX prefix. We need to
6171 clear RegMem bit before calling operand_type_equal. */
6172
6173 i386_operand_type op;
6174 unsigned int vvvv;
6175
6176 /* Check register-only source operand when two source
6177 operands are swapped. */
6178 if (!i.tm.operand_types[source].bitfield.baseindex
6179 && i.tm.operand_types[dest].bitfield.baseindex)
6180 {
6181 vvvv = source;
6182 source = dest;
6183 }
6184 else
6185 vvvv = dest;
6186
6187 op = i.tm.operand_types[vvvv];
6188 op.bitfield.regmem = 0;
6189 if ((dest + 1) >= i.operands
6190 || (!op.bitfield.reg32
6191 && op.bitfield.reg64
6192 && !operand_type_equal (&op, &regxmm)
6193 && !operand_type_equal (&op, &regymm)
6194 && !operand_type_equal (&op, &regzmm)
6195 && !operand_type_equal (&op, &regmask)))
6196 abort ();
6197 i.vex.register_specifier = i.op[vvvv].regs;
6198 dest++;
6199 }
6200 }
6201
6202 i.rm.mode = 3;
6203 /* One of the register operands will be encoded in the i.tm.reg
6204 field, the other in the combined i.tm.mode and i.tm.regmem
6205 fields. If no form of this instruction supports a memory
6206 destination operand, then we assume the source operand may
6207 sometimes be a memory operand and so we need to store the
6208 destination in the i.rm.reg field. */
6209 if (!i.tm.operand_types[dest].bitfield.regmem
6210 && operand_type_check (i.tm.operand_types[dest], anymem) == 0)
6211 {
6212 i.rm.reg = i.op[dest].regs->reg_num;
6213 i.rm.regmem = i.op[source].regs->reg_num;
6214 if ((i.op[dest].regs->reg_flags & RegRex) != 0)
6215 i.rex |= REX_R;
6216 if ((i.op[dest].regs->reg_flags & RegVRex) != 0)
6217 i.vrex |= REX_R;
6218 if ((i.op[source].regs->reg_flags & RegRex) != 0)
6219 i.rex |= REX_B;
6220 if ((i.op[source].regs->reg_flags & RegVRex) != 0)
6221 i.vrex |= REX_B;
6222 }
6223 else
6224 {
6225 i.rm.reg = i.op[source].regs->reg_num;
6226 i.rm.regmem = i.op[dest].regs->reg_num;
6227 if ((i.op[dest].regs->reg_flags & RegRex) != 0)
6228 i.rex |= REX_B;
6229 if ((i.op[dest].regs->reg_flags & RegVRex) != 0)
6230 i.vrex |= REX_B;
6231 if ((i.op[source].regs->reg_flags & RegRex) != 0)
6232 i.rex |= REX_R;
6233 if ((i.op[source].regs->reg_flags & RegVRex) != 0)
6234 i.vrex |= REX_R;
6235 }
6236 if (flag_code != CODE_64BIT && (i.rex & (REX_R | REX_B)))
6237 {
6238 if (!i.types[0].bitfield.control
6239 && !i.types[1].bitfield.control)
6240 abort ();
6241 i.rex &= ~(REX_R | REX_B);
6242 add_prefix (LOCK_PREFIX_OPCODE);
6243 }
6244 }
6245 else
6246 { /* If it's not 2 reg operands... */
6247 unsigned int mem;
6248
6249 if (i.mem_operands)
6250 {
6251 unsigned int fake_zero_displacement = 0;
6252 unsigned int op;
6253
6254 for (op = 0; op < i.operands; op++)
6255 if (operand_type_check (i.types[op], anymem))
6256 break;
6257 gas_assert (op < i.operands);
6258
6259 if (i.tm.opcode_modifier.vecsib)
6260 {
6261 if (i.index_reg->reg_num == RegEiz
6262 || i.index_reg->reg_num == RegRiz)
6263 abort ();
6264
6265 i.rm.regmem = ESCAPE_TO_TWO_BYTE_ADDRESSING;
6266 if (!i.base_reg)
6267 {
6268 i.sib.base = NO_BASE_REGISTER;
6269 i.sib.scale = i.log2_scale_factor;
6270 /* No Vec_Disp8 if there is no base. */
6271 i.types[op].bitfield.vec_disp8 = 0;
6272 i.types[op].bitfield.disp8 = 0;
6273 i.types[op].bitfield.disp16 = 0;
6274 i.types[op].bitfield.disp64 = 0;
6275 if (flag_code != CODE_64BIT)
6276 {
6277 /* Must be 32 bit */
6278 i.types[op].bitfield.disp32 = 1;
6279 i.types[op].bitfield.disp32s = 0;
6280 }
6281 else
6282 {
6283 i.types[op].bitfield.disp32 = 0;
6284 i.types[op].bitfield.disp32s = 1;
6285 }
6286 }
6287 i.sib.index = i.index_reg->reg_num;
6288 if ((i.index_reg->reg_flags & RegRex) != 0)
6289 i.rex |= REX_X;
6290 if ((i.index_reg->reg_flags & RegVRex) != 0)
6291 i.vrex |= REX_X;
6292 }
6293
6294 default_seg = &ds;
6295
6296 if (i.base_reg == 0)
6297 {
6298 i.rm.mode = 0;
6299 if (!i.disp_operands)
6300 {
6301 fake_zero_displacement = 1;
6302 /* Instructions with VSIB byte need 32bit displacement
6303 if there is no base register. */
6304 if (i.tm.opcode_modifier.vecsib)
6305 i.types[op].bitfield.disp32 = 1;
6306 }
6307 if (i.index_reg == 0)
6308 {
6309 gas_assert (!i.tm.opcode_modifier.vecsib);
6310 /* Operand is just <disp> */
6311 if (flag_code == CODE_64BIT)
6312 {
6313 /* 64bit mode overwrites the 32bit absolute
6314 addressing by RIP relative addressing and
6315 absolute addressing is encoded by one of the
6316 redundant SIB forms. */
6317 i.rm.regmem = ESCAPE_TO_TWO_BYTE_ADDRESSING;
6318 i.sib.base = NO_BASE_REGISTER;
6319 i.sib.index = NO_INDEX_REGISTER;
6320 i.types[op] = ((i.prefix[ADDR_PREFIX] == 0)
6321 ? disp32s : disp32);
6322 }
6323 else if ((flag_code == CODE_16BIT)
6324 ^ (i.prefix[ADDR_PREFIX] != 0))
6325 {
6326 i.rm.regmem = NO_BASE_REGISTER_16;
6327 i.types[op] = disp16;
6328 }
6329 else
6330 {
6331 i.rm.regmem = NO_BASE_REGISTER;
6332 i.types[op] = disp32;
6333 }
6334 }
6335 else if (!i.tm.opcode_modifier.vecsib)
6336 {
6337 /* !i.base_reg && i.index_reg */
6338 if (i.index_reg->reg_num == RegEiz
6339 || i.index_reg->reg_num == RegRiz)
6340 i.sib.index = NO_INDEX_REGISTER;
6341 else
6342 i.sib.index = i.index_reg->reg_num;
6343 i.sib.base = NO_BASE_REGISTER;
6344 i.sib.scale = i.log2_scale_factor;
6345 i.rm.regmem = ESCAPE_TO_TWO_BYTE_ADDRESSING;
6346 /* No Vec_Disp8 if there is no base. */
6347 i.types[op].bitfield.vec_disp8 = 0;
6348 i.types[op].bitfield.disp8 = 0;
6349 i.types[op].bitfield.disp16 = 0;
6350 i.types[op].bitfield.disp64 = 0;
6351 if (flag_code != CODE_64BIT)
6352 {
6353 /* Must be 32 bit */
6354 i.types[op].bitfield.disp32 = 1;
6355 i.types[op].bitfield.disp32s = 0;
6356 }
6357 else
6358 {
6359 i.types[op].bitfield.disp32 = 0;
6360 i.types[op].bitfield.disp32s = 1;
6361 }
6362 if ((i.index_reg->reg_flags & RegRex) != 0)
6363 i.rex |= REX_X;
6364 }
6365 }
6366 /* RIP addressing for 64bit mode. */
6367 else if (i.base_reg->reg_num == RegRip ||
6368 i.base_reg->reg_num == RegEip)
6369 {
6370 gas_assert (!i.tm.opcode_modifier.vecsib);
6371 i.rm.regmem = NO_BASE_REGISTER;
6372 i.types[op].bitfield.disp8 = 0;
6373 i.types[op].bitfield.disp16 = 0;
6374 i.types[op].bitfield.disp32 = 0;
6375 i.types[op].bitfield.disp32s = 1;
6376 i.types[op].bitfield.disp64 = 0;
6377 i.types[op].bitfield.vec_disp8 = 0;
6378 i.flags[op] |= Operand_PCrel;
6379 if (! i.disp_operands)
6380 fake_zero_displacement = 1;
6381 }
6382 else if (i.base_reg->reg_type.bitfield.reg16)
6383 {
6384 gas_assert (!i.tm.opcode_modifier.vecsib);
6385 switch (i.base_reg->reg_num)
6386 {
6387 case 3: /* (%bx) */
6388 if (i.index_reg == 0)
6389 i.rm.regmem = 7;
6390 else /* (%bx,%si) -> 0, or (%bx,%di) -> 1 */
6391 i.rm.regmem = i.index_reg->reg_num - 6;
6392 break;
6393 case 5: /* (%bp) */
6394 default_seg = &ss;
6395 if (i.index_reg == 0)
6396 {
6397 i.rm.regmem = 6;
6398 if (operand_type_check (i.types[op], disp) == 0)
6399 {
6400 /* fake (%bp) into 0(%bp) */
6401 if (i.tm.operand_types[op].bitfield.vec_disp8)
6402 i.types[op].bitfield.vec_disp8 = 1;
6403 else
6404 i.types[op].bitfield.disp8 = 1;
6405 fake_zero_displacement = 1;
6406 }
6407 }
6408 else /* (%bp,%si) -> 2, or (%bp,%di) -> 3 */
6409 i.rm.regmem = i.index_reg->reg_num - 6 + 2;
6410 break;
6411 default: /* (%si) -> 4 or (%di) -> 5 */
6412 i.rm.regmem = i.base_reg->reg_num - 6 + 4;
6413 }
6414 i.rm.mode = mode_from_disp_size (i.types[op]);
6415 }
6416 else /* i.base_reg and 32/64 bit mode */
6417 {
6418 if (flag_code == CODE_64BIT
6419 && operand_type_check (i.types[op], disp))
6420 {
6421 i386_operand_type temp;
6422 operand_type_set (&temp, 0);
6423 temp.bitfield.disp8 = i.types[op].bitfield.disp8;
6424 temp.bitfield.vec_disp8
6425 = i.types[op].bitfield.vec_disp8;
6426 i.types[op] = temp;
6427 if (i.prefix[ADDR_PREFIX] == 0)
6428 i.types[op].bitfield.disp32s = 1;
6429 else
6430 i.types[op].bitfield.disp32 = 1;
6431 }
6432
6433 if (!i.tm.opcode_modifier.vecsib)
6434 i.rm.regmem = i.base_reg->reg_num;
6435 if ((i.base_reg->reg_flags & RegRex) != 0)
6436 i.rex |= REX_B;
6437 i.sib.base = i.base_reg->reg_num;
6438 /* x86-64 ignores REX prefix bit here to avoid decoder
6439 complications. */
6440 if (!(i.base_reg->reg_flags & RegRex)
6441 && (i.base_reg->reg_num == EBP_REG_NUM
6442 || i.base_reg->reg_num == ESP_REG_NUM))
6443 default_seg = &ss;
6444 if (i.base_reg->reg_num == 5 && i.disp_operands == 0)
6445 {
6446 fake_zero_displacement = 1;
6447 if (i.tm.operand_types [op].bitfield.vec_disp8)
6448 i.types[op].bitfield.vec_disp8 = 1;
6449 else
6450 i.types[op].bitfield.disp8 = 1;
6451 }
6452 i.sib.scale = i.log2_scale_factor;
6453 if (i.index_reg == 0)
6454 {
6455 gas_assert (!i.tm.opcode_modifier.vecsib);
6456 /* <disp>(%esp) becomes two byte modrm with no index
6457 register. We've already stored the code for esp
6458 in i.rm.regmem ie. ESCAPE_TO_TWO_BYTE_ADDRESSING.
6459 Any base register besides %esp will not use the
6460 extra modrm byte. */
6461 i.sib.index = NO_INDEX_REGISTER;
6462 }
6463 else if (!i.tm.opcode_modifier.vecsib)
6464 {
6465 if (i.index_reg->reg_num == RegEiz
6466 || i.index_reg->reg_num == RegRiz)
6467 i.sib.index = NO_INDEX_REGISTER;
6468 else
6469 i.sib.index = i.index_reg->reg_num;
6470 i.rm.regmem = ESCAPE_TO_TWO_BYTE_ADDRESSING;
6471 if ((i.index_reg->reg_flags & RegRex) != 0)
6472 i.rex |= REX_X;
6473 }
6474
6475 if (i.disp_operands
6476 && (i.reloc[op] == BFD_RELOC_386_TLS_DESC_CALL
6477 || i.reloc[op] == BFD_RELOC_X86_64_TLSDESC_CALL))
6478 i.rm.mode = 0;
6479 else
6480 {
6481 if (!fake_zero_displacement
6482 && !i.disp_operands
6483 && i.disp_encoding)
6484 {
6485 fake_zero_displacement = 1;
6486 if (i.disp_encoding == disp_encoding_8bit)
6487 i.types[op].bitfield.disp8 = 1;
6488 else
6489 i.types[op].bitfield.disp32 = 1;
6490 }
6491 i.rm.mode = mode_from_disp_size (i.types[op]);
6492 }
6493 }
6494
6495 if (fake_zero_displacement)
6496 {
6497 /* Fakes a zero displacement assuming that i.types[op]
6498 holds the correct displacement size. */
6499 expressionS *exp;
6500
6501 gas_assert (i.op[op].disps == 0);
6502 exp = &disp_expressions[i.disp_operands++];
6503 i.op[op].disps = exp;
6504 exp->X_op = O_constant;
6505 exp->X_add_number = 0;
6506 exp->X_add_symbol = (symbolS *) 0;
6507 exp->X_op_symbol = (symbolS *) 0;
6508 }
6509
6510 mem = op;
6511 }
6512 else
6513 mem = ~0;
6514
6515 if (i.tm.opcode_modifier.vexsources == XOP2SOURCES)
6516 {
6517 if (operand_type_check (i.types[0], imm))
6518 i.vex.register_specifier = NULL;
6519 else
6520 {
6521 /* VEX.vvvv encodes one of the sources when the first
6522 operand is not an immediate. */
6523 if (i.tm.opcode_modifier.vexw == VEXW0)
6524 i.vex.register_specifier = i.op[0].regs;
6525 else
6526 i.vex.register_specifier = i.op[1].regs;
6527 }
6528
6529 /* Destination is a XMM register encoded in the ModRM.reg
6530 and VEX.R bit. */
6531 i.rm.reg = i.op[2].regs->reg_num;
6532 if ((i.op[2].regs->reg_flags & RegRex) != 0)
6533 i.rex |= REX_R;
6534
6535 /* ModRM.rm and VEX.B encodes the other source. */
6536 if (!i.mem_operands)
6537 {
6538 i.rm.mode = 3;
6539
6540 if (i.tm.opcode_modifier.vexw == VEXW0)
6541 i.rm.regmem = i.op[1].regs->reg_num;
6542 else
6543 i.rm.regmem = i.op[0].regs->reg_num;
6544
6545 if ((i.op[1].regs->reg_flags & RegRex) != 0)
6546 i.rex |= REX_B;
6547 }
6548 }
6549 else if (i.tm.opcode_modifier.vexvvvv == VEXLWP)
6550 {
6551 i.vex.register_specifier = i.op[2].regs;
6552 if (!i.mem_operands)
6553 {
6554 i.rm.mode = 3;
6555 i.rm.regmem = i.op[1].regs->reg_num;
6556 if ((i.op[1].regs->reg_flags & RegRex) != 0)
6557 i.rex |= REX_B;
6558 }
6559 }
6560 /* Fill in i.rm.reg or i.rm.regmem field with register operand
6561 (if any) based on i.tm.extension_opcode. Again, we must be
6562 careful to make sure that segment/control/debug/test/MMX
6563 registers are coded into the i.rm.reg field. */
6564 else if (i.reg_operands)
6565 {
6566 unsigned int op;
6567 unsigned int vex_reg = ~0;
6568
6569 for (op = 0; op < i.operands; op++)
6570 if (i.types[op].bitfield.reg8
6571 || i.types[op].bitfield.reg16
6572 || i.types[op].bitfield.reg32
6573 || i.types[op].bitfield.reg64
6574 || i.types[op].bitfield.regmmx
6575 || i.types[op].bitfield.regxmm
6576 || i.types[op].bitfield.regymm
6577 || i.types[op].bitfield.regbnd
6578 || i.types[op].bitfield.regzmm
6579 || i.types[op].bitfield.regmask
6580 || i.types[op].bitfield.sreg2
6581 || i.types[op].bitfield.sreg3
6582 || i.types[op].bitfield.control
6583 || i.types[op].bitfield.debug
6584 || i.types[op].bitfield.test)
6585 break;
6586
6587 if (vex_3_sources)
6588 op = dest;
6589 else if (i.tm.opcode_modifier.vexvvvv == VEXXDS)
6590 {
6591 /* For instructions with VexNDS, the register-only
6592 source operand is encoded in VEX prefix. */
6593 gas_assert (mem != (unsigned int) ~0);
6594
6595 if (op > mem)
6596 {
6597 vex_reg = op++;
6598 gas_assert (op < i.operands);
6599 }
6600 else
6601 {
6602 /* Check register-only source operand when two source
6603 operands are swapped. */
6604 if (!i.tm.operand_types[op].bitfield.baseindex
6605 && i.tm.operand_types[op + 1].bitfield.baseindex)
6606 {
6607 vex_reg = op;
6608 op += 2;
6609 gas_assert (mem == (vex_reg + 1)
6610 && op < i.operands);
6611 }
6612 else
6613 {
6614 vex_reg = op + 1;
6615 gas_assert (vex_reg < i.operands);
6616 }
6617 }
6618 }
6619 else if (i.tm.opcode_modifier.vexvvvv == VEXNDD)
6620 {
6621 /* For instructions with VexNDD, the register destination
6622 is encoded in VEX prefix. */
6623 if (i.mem_operands == 0)
6624 {
6625 /* There is no memory operand. */
6626 gas_assert ((op + 2) == i.operands);
6627 vex_reg = op + 1;
6628 }
6629 else
6630 {
6631 /* There are only 2 operands. */
6632 gas_assert (op < 2 && i.operands == 2);
6633 vex_reg = 1;
6634 }
6635 }
6636 else
6637 gas_assert (op < i.operands);
6638
6639 if (vex_reg != (unsigned int) ~0)
6640 {
6641 i386_operand_type *type = &i.tm.operand_types[vex_reg];
6642
6643 if (type->bitfield.reg32 != 1
6644 && type->bitfield.reg64 != 1
6645 && !operand_type_equal (type, &regxmm)
6646 && !operand_type_equal (type, &regymm)
6647 && !operand_type_equal (type, &regzmm)
6648 && !operand_type_equal (type, &regmask))
6649 abort ();
6650
6651 i.vex.register_specifier = i.op[vex_reg].regs;
6652 }
6653
6654 /* Don't set OP operand twice. */
6655 if (vex_reg != op)
6656 {
6657 /* If there is an extension opcode to put here, the
6658 register number must be put into the regmem field. */
6659 if (i.tm.extension_opcode != None)
6660 {
6661 i.rm.regmem = i.op[op].regs->reg_num;
6662 if ((i.op[op].regs->reg_flags & RegRex) != 0)
6663 i.rex |= REX_B;
6664 if ((i.op[op].regs->reg_flags & RegVRex) != 0)
6665 i.vrex |= REX_B;
6666 }
6667 else
6668 {
6669 i.rm.reg = i.op[op].regs->reg_num;
6670 if ((i.op[op].regs->reg_flags & RegRex) != 0)
6671 i.rex |= REX_R;
6672 if ((i.op[op].regs->reg_flags & RegVRex) != 0)
6673 i.vrex |= REX_R;
6674 }
6675 }
6676
6677 /* Now, if no memory operand has set i.rm.mode = 0, 1, 2 we
6678 must set it to 3 to indicate this is a register operand
6679 in the regmem field. */
6680 if (!i.mem_operands)
6681 i.rm.mode = 3;
6682 }
6683
6684 /* Fill in i.rm.reg field with extension opcode (if any). */
6685 if (i.tm.extension_opcode != None)
6686 i.rm.reg = i.tm.extension_opcode;
6687 }
6688 return default_seg;
6689 }
6690
6691 static void
6692 output_branch (void)
6693 {
6694 char *p;
6695 int size;
6696 int code16;
6697 int prefix;
6698 relax_substateT subtype;
6699 symbolS *sym;
6700 offsetT off;
6701
6702 code16 = flag_code == CODE_16BIT ? CODE16 : 0;
6703 size = i.disp_encoding == disp_encoding_32bit ? BIG : SMALL;
6704
6705 prefix = 0;
6706 if (i.prefix[DATA_PREFIX] != 0)
6707 {
6708 prefix = 1;
6709 i.prefixes -= 1;
6710 code16 ^= CODE16;
6711 }
6712 /* Pentium4 branch hints. */
6713 if (i.prefix[SEG_PREFIX] == CS_PREFIX_OPCODE /* not taken */
6714 || i.prefix[SEG_PREFIX] == DS_PREFIX_OPCODE /* taken */)
6715 {
6716 prefix++;
6717 i.prefixes--;
6718 }
6719 if (i.prefix[REX_PREFIX] != 0)
6720 {
6721 prefix++;
6722 i.prefixes--;
6723 }
6724
6725 /* BND prefixed jump. */
6726 if (i.prefix[BND_PREFIX] != 0)
6727 {
6728 FRAG_APPEND_1_CHAR (i.prefix[BND_PREFIX]);
6729 i.prefixes -= 1;
6730 }
6731
6732 if (i.prefixes != 0 && !intel_syntax)
6733 as_warn (_("skipping prefixes on this instruction"));
6734
6735 /* It's always a symbol; End frag & setup for relax.
6736 Make sure there is enough room in this frag for the largest
6737 instruction we may generate in md_convert_frag. This is 2
6738 bytes for the opcode and room for the prefix and largest
6739 displacement. */
6740 frag_grow (prefix + 2 + 4);
6741 /* Prefix and 1 opcode byte go in fr_fix. */
6742 p = frag_more (prefix + 1);
6743 if (i.prefix[DATA_PREFIX] != 0)
6744 *p++ = DATA_PREFIX_OPCODE;
6745 if (i.prefix[SEG_PREFIX] == CS_PREFIX_OPCODE
6746 || i.prefix[SEG_PREFIX] == DS_PREFIX_OPCODE)
6747 *p++ = i.prefix[SEG_PREFIX];
6748 if (i.prefix[REX_PREFIX] != 0)
6749 *p++ = i.prefix[REX_PREFIX];
6750 *p = i.tm.base_opcode;
6751
6752 if ((unsigned char) *p == JUMP_PC_RELATIVE)
6753 subtype = ENCODE_RELAX_STATE (UNCOND_JUMP, size);
6754 else if (cpu_arch_flags.bitfield.cpui386)
6755 subtype = ENCODE_RELAX_STATE (COND_JUMP, size);
6756 else
6757 subtype = ENCODE_RELAX_STATE (COND_JUMP86, size);
6758 subtype |= code16;
6759
6760 sym = i.op[0].disps->X_add_symbol;
6761 off = i.op[0].disps->X_add_number;
6762
6763 if (i.op[0].disps->X_op != O_constant
6764 && i.op[0].disps->X_op != O_symbol)
6765 {
6766 /* Handle complex expressions. */
6767 sym = make_expr_symbol (i.op[0].disps);
6768 off = 0;
6769 }
6770
6771 /* 1 possible extra opcode + 4 byte displacement go in var part.
6772 Pass reloc in fr_var. */
6773 frag_var (rs_machine_dependent, 5,
6774 ((!object_64bit
6775 || i.reloc[0] != NO_RELOC
6776 || (i.bnd_prefix == NULL && !add_bnd_prefix))
6777 ? i.reloc[0]
6778 : BFD_RELOC_X86_64_PC32_BND),
6779 subtype, sym, off, p);
6780 }
6781
6782 static void
6783 output_jump (void)
6784 {
6785 char *p;
6786 int size;
6787 fixS *fixP;
6788
6789 if (i.tm.opcode_modifier.jumpbyte)
6790 {
6791 /* This is a loop or jecxz type instruction. */
6792 size = 1;
6793 if (i.prefix[ADDR_PREFIX] != 0)
6794 {
6795 FRAG_APPEND_1_CHAR (ADDR_PREFIX_OPCODE);
6796 i.prefixes -= 1;
6797 }
6798 /* Pentium4 branch hints. */
6799 if (i.prefix[SEG_PREFIX] == CS_PREFIX_OPCODE /* not taken */
6800 || i.prefix[SEG_PREFIX] == DS_PREFIX_OPCODE /* taken */)
6801 {
6802 FRAG_APPEND_1_CHAR (i.prefix[SEG_PREFIX]);
6803 i.prefixes--;
6804 }
6805 }
6806 else
6807 {
6808 int code16;
6809
6810 code16 = 0;
6811 if (flag_code == CODE_16BIT)
6812 code16 = CODE16;
6813
6814 if (i.prefix[DATA_PREFIX] != 0)
6815 {
6816 FRAG_APPEND_1_CHAR (DATA_PREFIX_OPCODE);
6817 i.prefixes -= 1;
6818 code16 ^= CODE16;
6819 }
6820
6821 size = 4;
6822 if (code16)
6823 size = 2;
6824 }
6825
6826 if (i.prefix[REX_PREFIX] != 0)
6827 {
6828 FRAG_APPEND_1_CHAR (i.prefix[REX_PREFIX]);
6829 i.prefixes -= 1;
6830 }
6831
6832 /* BND prefixed jump. */
6833 if (i.prefix[BND_PREFIX] != 0)
6834 {
6835 FRAG_APPEND_1_CHAR (i.prefix[BND_PREFIX]);
6836 i.prefixes -= 1;
6837 }
6838
6839 if (i.prefixes != 0 && !intel_syntax)
6840 as_warn (_("skipping prefixes on this instruction"));
6841
6842 p = frag_more (i.tm.opcode_length + size);
6843 switch (i.tm.opcode_length)
6844 {
6845 case 2:
6846 *p++ = i.tm.base_opcode >> 8;
6847 case 1:
6848 *p++ = i.tm.base_opcode;
6849 break;
6850 default:
6851 abort ();
6852 }
6853
6854 fixP = fix_new_exp (frag_now, p - frag_now->fr_literal, size,
6855 i.op[0].disps, 1, reloc (size, 1, 1,
6856 (i.bnd_prefix != NULL
6857 || add_bnd_prefix),
6858 i.reloc[0]));
6859
6860 /* All jumps handled here are signed, but don't use a signed limit
6861 check for 32 and 16 bit jumps as we want to allow wrap around at
6862 4G and 64k respectively. */
6863 if (size == 1)
6864 fixP->fx_signed = 1;
6865 }
6866
6867 static void
6868 output_interseg_jump (void)
6869 {
6870 char *p;
6871 int size;
6872 int prefix;
6873 int code16;
6874
6875 code16 = 0;
6876 if (flag_code == CODE_16BIT)
6877 code16 = CODE16;
6878
6879 prefix = 0;
6880 if (i.prefix[DATA_PREFIX] != 0)
6881 {
6882 prefix = 1;
6883 i.prefixes -= 1;
6884 code16 ^= CODE16;
6885 }
6886 if (i.prefix[REX_PREFIX] != 0)
6887 {
6888 prefix++;
6889 i.prefixes -= 1;
6890 }
6891
6892 size = 4;
6893 if (code16)
6894 size = 2;
6895
6896 if (i.prefixes != 0 && !intel_syntax)
6897 as_warn (_("skipping prefixes on this instruction"));
6898
6899 /* 1 opcode; 2 segment; offset */
6900 p = frag_more (prefix + 1 + 2 + size);
6901
6902 if (i.prefix[DATA_PREFIX] != 0)
6903 *p++ = DATA_PREFIX_OPCODE;
6904
6905 if (i.prefix[REX_PREFIX] != 0)
6906 *p++ = i.prefix[REX_PREFIX];
6907
6908 *p++ = i.tm.base_opcode;
6909 if (i.op[1].imms->X_op == O_constant)
6910 {
6911 offsetT n = i.op[1].imms->X_add_number;
6912
6913 if (size == 2
6914 && !fits_in_unsigned_word (n)
6915 && !fits_in_signed_word (n))
6916 {
6917 as_bad (_("16-bit jump out of range"));
6918 return;
6919 }
6920 md_number_to_chars (p, n, size);
6921 }
6922 else
6923 fix_new_exp (frag_now, p - frag_now->fr_literal, size,
6924 i.op[1].imms, 0, reloc (size, 0, 0, 0, i.reloc[1]));
6925 if (i.op[0].imms->X_op != O_constant)
6926 as_bad (_("can't handle non absolute segment in `%s'"),
6927 i.tm.name);
6928 md_number_to_chars (p + size, (valueT) i.op[0].imms->X_add_number, 2);
6929 }
6930
6931 static void
6932 output_insn (void)
6933 {
6934 fragS *insn_start_frag;
6935 offsetT insn_start_off;
6936
6937 /* Tie dwarf2 debug info to the address at the start of the insn.
6938 We can't do this after the insn has been output as the current
6939 frag may have been closed off. eg. by frag_var. */
6940 dwarf2_emit_insn (0);
6941
6942 insn_start_frag = frag_now;
6943 insn_start_off = frag_now_fix ();
6944
6945 /* Output jumps. */
6946 if (i.tm.opcode_modifier.jump)
6947 output_branch ();
6948 else if (i.tm.opcode_modifier.jumpbyte
6949 || i.tm.opcode_modifier.jumpdword)
6950 output_jump ();
6951 else if (i.tm.opcode_modifier.jumpintersegment)
6952 output_interseg_jump ();
6953 else
6954 {
6955 /* Output normal instructions here. */
6956 char *p;
6957 unsigned char *q;
6958 unsigned int j;
6959 unsigned int prefix;
6960
6961 /* Some processors fail on LOCK prefix. This options makes
6962 assembler ignore LOCK prefix and serves as a workaround. */
6963 if (omit_lock_prefix)
6964 {
6965 if (i.tm.base_opcode == LOCK_PREFIX_OPCODE)
6966 return;
6967 i.prefix[LOCK_PREFIX] = 0;
6968 }
6969
6970 /* Since the VEX/EVEX prefix contains the implicit prefix, we
6971 don't need the explicit prefix. */
6972 if (!i.tm.opcode_modifier.vex && !i.tm.opcode_modifier.evex)
6973 {
6974 switch (i.tm.opcode_length)
6975 {
6976 case 3:
6977 if (i.tm.base_opcode & 0xff000000)
6978 {
6979 prefix = (i.tm.base_opcode >> 24) & 0xff;
6980 goto check_prefix;
6981 }
6982 break;
6983 case 2:
6984 if ((i.tm.base_opcode & 0xff0000) != 0)
6985 {
6986 prefix = (i.tm.base_opcode >> 16) & 0xff;
6987 if (i.tm.cpu_flags.bitfield.cpupadlock)
6988 {
6989 check_prefix:
6990 if (prefix != REPE_PREFIX_OPCODE
6991 || (i.prefix[REP_PREFIX]
6992 != REPE_PREFIX_OPCODE))
6993 add_prefix (prefix);
6994 }
6995 else
6996 add_prefix (prefix);
6997 }
6998 break;
6999 case 1:
7000 break;
7001 default:
7002 abort ();
7003 }
7004
7005 #if defined (OBJ_MAYBE_ELF) || defined (OBJ_ELF)
7006 /* For x32, add a dummy REX_OPCODE prefix for mov/add with
7007 R_X86_64_GOTTPOFF relocation so that linker can safely
7008 perform IE->LE optimization. */
7009 if (x86_elf_abi == X86_64_X32_ABI
7010 && i.operands == 2
7011 && i.reloc[0] == BFD_RELOC_X86_64_GOTTPOFF
7012 && i.prefix[REX_PREFIX] == 0)
7013 add_prefix (REX_OPCODE);
7014 #endif
7015
7016 /* The prefix bytes. */
7017 for (j = ARRAY_SIZE (i.prefix), q = i.prefix; j > 0; j--, q++)
7018 if (*q)
7019 FRAG_APPEND_1_CHAR (*q);
7020 }
7021 else
7022 {
7023 for (j = 0, q = i.prefix; j < ARRAY_SIZE (i.prefix); j++, q++)
7024 if (*q)
7025 switch (j)
7026 {
7027 case REX_PREFIX:
7028 /* REX byte is encoded in VEX prefix. */
7029 break;
7030 case SEG_PREFIX:
7031 case ADDR_PREFIX:
7032 FRAG_APPEND_1_CHAR (*q);
7033 break;
7034 default:
7035 /* There should be no other prefixes for instructions
7036 with VEX prefix. */
7037 abort ();
7038 }
7039
7040 /* For EVEX instructions i.vrex should become 0 after
7041 build_evex_prefix. For VEX instructions upper 16 registers
7042 aren't available, so VREX should be 0. */
7043 if (i.vrex)
7044 abort ();
7045 /* Now the VEX prefix. */
7046 p = frag_more (i.vex.length);
7047 for (j = 0; j < i.vex.length; j++)
7048 p[j] = i.vex.bytes[j];
7049 }
7050
7051 /* Now the opcode; be careful about word order here! */
7052 if (i.tm.opcode_length == 1)
7053 {
7054 FRAG_APPEND_1_CHAR (i.tm.base_opcode);
7055 }
7056 else
7057 {
7058 switch (i.tm.opcode_length)
7059 {
7060 case 4:
7061 p = frag_more (4);
7062 *p++ = (i.tm.base_opcode >> 24) & 0xff;
7063 *p++ = (i.tm.base_opcode >> 16) & 0xff;
7064 break;
7065 case 3:
7066 p = frag_more (3);
7067 *p++ = (i.tm.base_opcode >> 16) & 0xff;
7068 break;
7069 case 2:
7070 p = frag_more (2);
7071 break;
7072 default:
7073 abort ();
7074 break;
7075 }
7076
7077 /* Put out high byte first: can't use md_number_to_chars! */
7078 *p++ = (i.tm.base_opcode >> 8) & 0xff;
7079 *p = i.tm.base_opcode & 0xff;
7080 }
7081
7082 /* Now the modrm byte and sib byte (if present). */
7083 if (i.tm.opcode_modifier.modrm)
7084 {
7085 FRAG_APPEND_1_CHAR ((i.rm.regmem << 0
7086 | i.rm.reg << 3
7087 | i.rm.mode << 6));
7088 /* If i.rm.regmem == ESP (4)
7089 && i.rm.mode != (Register mode)
7090 && not 16 bit
7091 ==> need second modrm byte. */
7092 if (i.rm.regmem == ESCAPE_TO_TWO_BYTE_ADDRESSING
7093 && i.rm.mode != 3
7094 && !(i.base_reg && i.base_reg->reg_type.bitfield.reg16))
7095 FRAG_APPEND_1_CHAR ((i.sib.base << 0
7096 | i.sib.index << 3
7097 | i.sib.scale << 6));
7098 }
7099
7100 if (i.disp_operands)
7101 output_disp (insn_start_frag, insn_start_off);
7102
7103 if (i.imm_operands)
7104 output_imm (insn_start_frag, insn_start_off);
7105 }
7106
7107 #ifdef DEBUG386
7108 if (flag_debug)
7109 {
7110 pi ("" /*line*/, &i);
7111 }
7112 #endif /* DEBUG386 */
7113 }
7114
7115 /* Return the size of the displacement operand N. */
7116
7117 static int
7118 disp_size (unsigned int n)
7119 {
7120 int size = 4;
7121
7122 /* Vec_Disp8 has to be 8bit. */
7123 if (i.types[n].bitfield.vec_disp8)
7124 size = 1;
7125 else if (i.types[n].bitfield.disp64)
7126 size = 8;
7127 else if (i.types[n].bitfield.disp8)
7128 size = 1;
7129 else if (i.types[n].bitfield.disp16)
7130 size = 2;
7131 return size;
7132 }
7133
7134 /* Return the size of the immediate operand N. */
7135
7136 static int
7137 imm_size (unsigned int n)
7138 {
7139 int size = 4;
7140 if (i.types[n].bitfield.imm64)
7141 size = 8;
7142 else if (i.types[n].bitfield.imm8 || i.types[n].bitfield.imm8s)
7143 size = 1;
7144 else if (i.types[n].bitfield.imm16)
7145 size = 2;
7146 return size;
7147 }
7148
7149 static void
7150 output_disp (fragS *insn_start_frag, offsetT insn_start_off)
7151 {
7152 char *p;
7153 unsigned int n;
7154
7155 for (n = 0; n < i.operands; n++)
7156 {
7157 if (i.types[n].bitfield.vec_disp8
7158 || operand_type_check (i.types[n], disp))
7159 {
7160 if (i.op[n].disps->X_op == O_constant)
7161 {
7162 int size = disp_size (n);
7163 offsetT val = i.op[n].disps->X_add_number;
7164
7165 if (i.types[n].bitfield.vec_disp8)
7166 val >>= i.memshift;
7167 val = offset_in_range (val, size);
7168 p = frag_more (size);
7169 md_number_to_chars (p, val, size);
7170 }
7171 else
7172 {
7173 enum bfd_reloc_code_real reloc_type;
7174 int size = disp_size (n);
7175 int sign = i.types[n].bitfield.disp32s;
7176 int pcrel = (i.flags[n] & Operand_PCrel) != 0;
7177
7178 /* We can't have 8 bit displacement here. */
7179 gas_assert (!i.types[n].bitfield.disp8);
7180
7181 /* The PC relative address is computed relative
7182 to the instruction boundary, so in case immediate
7183 fields follows, we need to adjust the value. */
7184 if (pcrel && i.imm_operands)
7185 {
7186 unsigned int n1;
7187 int sz = 0;
7188
7189 for (n1 = 0; n1 < i.operands; n1++)
7190 if (operand_type_check (i.types[n1], imm))
7191 {
7192 /* Only one immediate is allowed for PC
7193 relative address. */
7194 gas_assert (sz == 0);
7195 sz = imm_size (n1);
7196 i.op[n].disps->X_add_number -= sz;
7197 }
7198 /* We should find the immediate. */
7199 gas_assert (sz != 0);
7200 }
7201
7202 p = frag_more (size);
7203 reloc_type = reloc (size, pcrel, sign,
7204 (i.bnd_prefix != NULL
7205 || add_bnd_prefix),
7206 i.reloc[n]);
7207 if (GOT_symbol
7208 && GOT_symbol == i.op[n].disps->X_add_symbol
7209 && (((reloc_type == BFD_RELOC_32
7210 || reloc_type == BFD_RELOC_X86_64_32S
7211 || (reloc_type == BFD_RELOC_64
7212 && object_64bit))
7213 && (i.op[n].disps->X_op == O_symbol
7214 || (i.op[n].disps->X_op == O_add
7215 && ((symbol_get_value_expression
7216 (i.op[n].disps->X_op_symbol)->X_op)
7217 == O_subtract))))
7218 || reloc_type == BFD_RELOC_32_PCREL))
7219 {
7220 offsetT add;
7221
7222 if (insn_start_frag == frag_now)
7223 add = (p - frag_now->fr_literal) - insn_start_off;
7224 else
7225 {
7226 fragS *fr;
7227
7228 add = insn_start_frag->fr_fix - insn_start_off;
7229 for (fr = insn_start_frag->fr_next;
7230 fr && fr != frag_now; fr = fr->fr_next)
7231 add += fr->fr_fix;
7232 add += p - frag_now->fr_literal;
7233 }
7234
7235 if (!object_64bit)
7236 {
7237 reloc_type = BFD_RELOC_386_GOTPC;
7238 i.op[n].imms->X_add_number += add;
7239 }
7240 else if (reloc_type == BFD_RELOC_64)
7241 reloc_type = BFD_RELOC_X86_64_GOTPC64;
7242 else
7243 /* Don't do the adjustment for x86-64, as there
7244 the pcrel addressing is relative to the _next_
7245 insn, and that is taken care of in other code. */
7246 reloc_type = BFD_RELOC_X86_64_GOTPC32;
7247 }
7248 fix_new_exp (frag_now, p - frag_now->fr_literal, size,
7249 i.op[n].disps, pcrel, reloc_type);
7250 }
7251 }
7252 }
7253 }
7254
7255 static void
7256 output_imm (fragS *insn_start_frag, offsetT insn_start_off)
7257 {
7258 char *p;
7259 unsigned int n;
7260
7261 for (n = 0; n < i.operands; n++)
7262 {
7263 /* Skip SAE/RC Imm operand in EVEX. They are already handled. */
7264 if (i.rounding && (int) n == i.rounding->operand)
7265 continue;
7266
7267 if (operand_type_check (i.types[n], imm))
7268 {
7269 if (i.op[n].imms->X_op == O_constant)
7270 {
7271 int size = imm_size (n);
7272 offsetT val;
7273
7274 val = offset_in_range (i.op[n].imms->X_add_number,
7275 size);
7276 p = frag_more (size);
7277 md_number_to_chars (p, val, size);
7278 }
7279 else
7280 {
7281 /* Not absolute_section.
7282 Need a 32-bit fixup (don't support 8bit
7283 non-absolute imms). Try to support other
7284 sizes ... */
7285 enum bfd_reloc_code_real reloc_type;
7286 int size = imm_size (n);
7287 int sign;
7288
7289 if (i.types[n].bitfield.imm32s
7290 && (i.suffix == QWORD_MNEM_SUFFIX
7291 || (!i.suffix && i.tm.opcode_modifier.no_lsuf)))
7292 sign = 1;
7293 else
7294 sign = 0;
7295
7296 p = frag_more (size);
7297 reloc_type = reloc (size, 0, sign, 0, i.reloc[n]);
7298
7299 /* This is tough to explain. We end up with this one if we
7300 * have operands that look like
7301 * "_GLOBAL_OFFSET_TABLE_+[.-.L284]". The goal here is to
7302 * obtain the absolute address of the GOT, and it is strongly
7303 * preferable from a performance point of view to avoid using
7304 * a runtime relocation for this. The actual sequence of
7305 * instructions often look something like:
7306 *
7307 * call .L66
7308 * .L66:
7309 * popl %ebx
7310 * addl $_GLOBAL_OFFSET_TABLE_+[.-.L66],%ebx
7311 *
7312 * The call and pop essentially return the absolute address
7313 * of the label .L66 and store it in %ebx. The linker itself
7314 * will ultimately change the first operand of the addl so
7315 * that %ebx points to the GOT, but to keep things simple, the
7316 * .o file must have this operand set so that it generates not
7317 * the absolute address of .L66, but the absolute address of
7318 * itself. This allows the linker itself simply treat a GOTPC
7319 * relocation as asking for a pcrel offset to the GOT to be
7320 * added in, and the addend of the relocation is stored in the
7321 * operand field for the instruction itself.
7322 *
7323 * Our job here is to fix the operand so that it would add
7324 * the correct offset so that %ebx would point to itself. The
7325 * thing that is tricky is that .-.L66 will point to the
7326 * beginning of the instruction, so we need to further modify
7327 * the operand so that it will point to itself. There are
7328 * other cases where you have something like:
7329 *
7330 * .long $_GLOBAL_OFFSET_TABLE_+[.-.L66]
7331 *
7332 * and here no correction would be required. Internally in
7333 * the assembler we treat operands of this form as not being
7334 * pcrel since the '.' is explicitly mentioned, and I wonder
7335 * whether it would simplify matters to do it this way. Who
7336 * knows. In earlier versions of the PIC patches, the
7337 * pcrel_adjust field was used to store the correction, but
7338 * since the expression is not pcrel, I felt it would be
7339 * confusing to do it this way. */
7340
7341 if ((reloc_type == BFD_RELOC_32
7342 || reloc_type == BFD_RELOC_X86_64_32S
7343 || reloc_type == BFD_RELOC_64)
7344 && GOT_symbol
7345 && GOT_symbol == i.op[n].imms->X_add_symbol
7346 && (i.op[n].imms->X_op == O_symbol
7347 || (i.op[n].imms->X_op == O_add
7348 && ((symbol_get_value_expression
7349 (i.op[n].imms->X_op_symbol)->X_op)
7350 == O_subtract))))
7351 {
7352 offsetT add;
7353
7354 if (insn_start_frag == frag_now)
7355 add = (p - frag_now->fr_literal) - insn_start_off;
7356 else
7357 {
7358 fragS *fr;
7359
7360 add = insn_start_frag->fr_fix - insn_start_off;
7361 for (fr = insn_start_frag->fr_next;
7362 fr && fr != frag_now; fr = fr->fr_next)
7363 add += fr->fr_fix;
7364 add += p - frag_now->fr_literal;
7365 }
7366
7367 if (!object_64bit)
7368 reloc_type = BFD_RELOC_386_GOTPC;
7369 else if (size == 4)
7370 reloc_type = BFD_RELOC_X86_64_GOTPC32;
7371 else if (size == 8)
7372 reloc_type = BFD_RELOC_X86_64_GOTPC64;
7373 i.op[n].imms->X_add_number += add;
7374 }
7375 fix_new_exp (frag_now, p - frag_now->fr_literal, size,
7376 i.op[n].imms, 0, reloc_type);
7377 }
7378 }
7379 }
7380 }
7381 \f
7382 /* x86_cons_fix_new is called via the expression parsing code when a
7383 reloc is needed. We use this hook to get the correct .got reloc. */
7384 static int cons_sign = -1;
7385
7386 void
7387 x86_cons_fix_new (fragS *frag, unsigned int off, unsigned int len,
7388 expressionS *exp, bfd_reloc_code_real_type r)
7389 {
7390 r = reloc (len, 0, cons_sign, 0, r);
7391
7392 #ifdef TE_PE
7393 if (exp->X_op == O_secrel)
7394 {
7395 exp->X_op = O_symbol;
7396 r = BFD_RELOC_32_SECREL;
7397 }
7398 #endif
7399
7400 fix_new_exp (frag, off, len, exp, 0, r);
7401 }
7402
7403 /* Export the ABI address size for use by TC_ADDRESS_BYTES for the
7404 purpose of the `.dc.a' internal pseudo-op. */
7405
7406 int
7407 x86_address_bytes (void)
7408 {
7409 if ((stdoutput->arch_info->mach & bfd_mach_x64_32))
7410 return 4;
7411 return stdoutput->arch_info->bits_per_address / 8;
7412 }
7413
7414 #if !(defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) || defined (OBJ_MACH_O)) \
7415 || defined (LEX_AT)
7416 # define lex_got(reloc, adjust, types, bnd_prefix) NULL
7417 #else
7418 /* Parse operands of the form
7419 <symbol>@GOTOFF+<nnn>
7420 and similar .plt or .got references.
7421
7422 If we find one, set up the correct relocation in RELOC and copy the
7423 input string, minus the `@GOTOFF' into a malloc'd buffer for
7424 parsing by the calling routine. Return this buffer, and if ADJUST
7425 is non-null set it to the length of the string we removed from the
7426 input line. Otherwise return NULL. */
7427 static char *
7428 lex_got (enum bfd_reloc_code_real *rel,
7429 int *adjust,
7430 i386_operand_type *types,
7431 int bnd_prefix)
7432 {
7433 /* Some of the relocations depend on the size of what field is to
7434 be relocated. But in our callers i386_immediate and i386_displacement
7435 we don't yet know the operand size (this will be set by insn
7436 matching). Hence we record the word32 relocation here,
7437 and adjust the reloc according to the real size in reloc(). */
7438 static const struct {
7439 const char *str;
7440 int len;
7441 const enum bfd_reloc_code_real rel[2];
7442 const i386_operand_type types64;
7443 } gotrel[] = {
7444 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
7445 { STRING_COMMA_LEN ("SIZE"), { BFD_RELOC_SIZE32,
7446 BFD_RELOC_SIZE32 },
7447 OPERAND_TYPE_IMM32_64 },
7448 #endif
7449 { STRING_COMMA_LEN ("PLTOFF"), { _dummy_first_bfd_reloc_code_real,
7450 BFD_RELOC_X86_64_PLTOFF64 },
7451 OPERAND_TYPE_IMM64 },
7452 { STRING_COMMA_LEN ("PLT"), { BFD_RELOC_386_PLT32,
7453 BFD_RELOC_X86_64_PLT32 },
7454 OPERAND_TYPE_IMM32_32S_DISP32 },
7455 { STRING_COMMA_LEN ("GOTPLT"), { _dummy_first_bfd_reloc_code_real,
7456 BFD_RELOC_X86_64_GOTPLT64 },
7457 OPERAND_TYPE_IMM64_DISP64 },
7458 { STRING_COMMA_LEN ("GOTOFF"), { BFD_RELOC_386_GOTOFF,
7459 BFD_RELOC_X86_64_GOTOFF64 },
7460 OPERAND_TYPE_IMM64_DISP64 },
7461 { STRING_COMMA_LEN ("GOTPCREL"), { _dummy_first_bfd_reloc_code_real,
7462 BFD_RELOC_X86_64_GOTPCREL },
7463 OPERAND_TYPE_IMM32_32S_DISP32 },
7464 { STRING_COMMA_LEN ("TLSGD"), { BFD_RELOC_386_TLS_GD,
7465 BFD_RELOC_X86_64_TLSGD },
7466 OPERAND_TYPE_IMM32_32S_DISP32 },
7467 { STRING_COMMA_LEN ("TLSLDM"), { BFD_RELOC_386_TLS_LDM,
7468 _dummy_first_bfd_reloc_code_real },
7469 OPERAND_TYPE_NONE },
7470 { STRING_COMMA_LEN ("TLSLD"), { _dummy_first_bfd_reloc_code_real,
7471 BFD_RELOC_X86_64_TLSLD },
7472 OPERAND_TYPE_IMM32_32S_DISP32 },
7473 { STRING_COMMA_LEN ("GOTTPOFF"), { BFD_RELOC_386_TLS_IE_32,
7474 BFD_RELOC_X86_64_GOTTPOFF },
7475 OPERAND_TYPE_IMM32_32S_DISP32 },
7476 { STRING_COMMA_LEN ("TPOFF"), { BFD_RELOC_386_TLS_LE_32,
7477 BFD_RELOC_X86_64_TPOFF32 },
7478 OPERAND_TYPE_IMM32_32S_64_DISP32_64 },
7479 { STRING_COMMA_LEN ("NTPOFF"), { BFD_RELOC_386_TLS_LE,
7480 _dummy_first_bfd_reloc_code_real },
7481 OPERAND_TYPE_NONE },
7482 { STRING_COMMA_LEN ("DTPOFF"), { BFD_RELOC_386_TLS_LDO_32,
7483 BFD_RELOC_X86_64_DTPOFF32 },
7484 OPERAND_TYPE_IMM32_32S_64_DISP32_64 },
7485 { STRING_COMMA_LEN ("GOTNTPOFF"),{ BFD_RELOC_386_TLS_GOTIE,
7486 _dummy_first_bfd_reloc_code_real },
7487 OPERAND_TYPE_NONE },
7488 { STRING_COMMA_LEN ("INDNTPOFF"),{ BFD_RELOC_386_TLS_IE,
7489 _dummy_first_bfd_reloc_code_real },
7490 OPERAND_TYPE_NONE },
7491 { STRING_COMMA_LEN ("GOT"), { BFD_RELOC_386_GOT32,
7492 BFD_RELOC_X86_64_GOT32 },
7493 OPERAND_TYPE_IMM32_32S_64_DISP32 },
7494 { STRING_COMMA_LEN ("TLSDESC"), { BFD_RELOC_386_TLS_GOTDESC,
7495 BFD_RELOC_X86_64_GOTPC32_TLSDESC },
7496 OPERAND_TYPE_IMM32_32S_DISP32 },
7497 { STRING_COMMA_LEN ("TLSCALL"), { BFD_RELOC_386_TLS_DESC_CALL,
7498 BFD_RELOC_X86_64_TLSDESC_CALL },
7499 OPERAND_TYPE_IMM32_32S_DISP32 },
7500 };
7501 char *cp;
7502 unsigned int j;
7503
7504 #if defined (OBJ_MAYBE_ELF)
7505 if (!IS_ELF)
7506 return NULL;
7507 #endif
7508
7509 for (cp = input_line_pointer; *cp != '@'; cp++)
7510 if (is_end_of_line[(unsigned char) *cp] || *cp == ',')
7511 return NULL;
7512
7513 for (j = 0; j < ARRAY_SIZE (gotrel); j++)
7514 {
7515 int len = gotrel[j].len;
7516 if (strncasecmp (cp + 1, gotrel[j].str, len) == 0)
7517 {
7518 if (gotrel[j].rel[object_64bit] != 0)
7519 {
7520 int first, second;
7521 char *tmpbuf, *past_reloc;
7522
7523 *rel = gotrel[j].rel[object_64bit];
7524
7525 if (types)
7526 {
7527 if (flag_code != CODE_64BIT)
7528 {
7529 types->bitfield.imm32 = 1;
7530 types->bitfield.disp32 = 1;
7531 }
7532 else
7533 *types = gotrel[j].types64;
7534 }
7535
7536 if (j != 0 && GOT_symbol == NULL)
7537 GOT_symbol = symbol_find_or_make (GLOBAL_OFFSET_TABLE_NAME);
7538
7539 /* The length of the first part of our input line. */
7540 first = cp - input_line_pointer;
7541
7542 /* The second part goes from after the reloc token until
7543 (and including) an end_of_line char or comma. */
7544 past_reloc = cp + 1 + len;
7545 cp = past_reloc;
7546 while (!is_end_of_line[(unsigned char) *cp] && *cp != ',')
7547 ++cp;
7548 second = cp + 1 - past_reloc;
7549
7550 /* Allocate and copy string. The trailing NUL shouldn't
7551 be necessary, but be safe. */
7552 tmpbuf = (char *) xmalloc (first + second + 2);
7553 memcpy (tmpbuf, input_line_pointer, first);
7554 if (second != 0 && *past_reloc != ' ')
7555 /* Replace the relocation token with ' ', so that
7556 errors like foo@GOTOFF1 will be detected. */
7557 tmpbuf[first++] = ' ';
7558 else
7559 /* Increment length by 1 if the relocation token is
7560 removed. */
7561 len++;
7562 if (adjust)
7563 *adjust = len;
7564 memcpy (tmpbuf + first, past_reloc, second);
7565 tmpbuf[first + second] = '\0';
7566 if (bnd_prefix && *rel == BFD_RELOC_X86_64_PLT32)
7567 *rel = BFD_RELOC_X86_64_PLT32_BND;
7568 return tmpbuf;
7569 }
7570
7571 as_bad (_("@%s reloc is not supported with %d-bit output format"),
7572 gotrel[j].str, 1 << (5 + object_64bit));
7573 return NULL;
7574 }
7575 }
7576
7577 /* Might be a symbol version string. Don't as_bad here. */
7578 return NULL;
7579 }
7580 #endif
7581
7582 #ifdef TE_PE
7583 #ifdef lex_got
7584 #undef lex_got
7585 #endif
7586 /* Parse operands of the form
7587 <symbol>@SECREL32+<nnn>
7588
7589 If we find one, set up the correct relocation in RELOC and copy the
7590 input string, minus the `@SECREL32' into a malloc'd buffer for
7591 parsing by the calling routine. Return this buffer, and if ADJUST
7592 is non-null set it to the length of the string we removed from the
7593 input line. Otherwise return NULL.
7594
7595 This function is copied from the ELF version above adjusted for PE targets. */
7596
7597 static char *
7598 lex_got (enum bfd_reloc_code_real *rel ATTRIBUTE_UNUSED,
7599 int *adjust ATTRIBUTE_UNUSED,
7600 i386_operand_type *types,
7601 int bnd_prefix ATTRIBUTE_UNUSED)
7602 {
7603 static const struct
7604 {
7605 const char *str;
7606 int len;
7607 const enum bfd_reloc_code_real rel[2];
7608 const i386_operand_type types64;
7609 }
7610 gotrel[] =
7611 {
7612 { STRING_COMMA_LEN ("SECREL32"), { BFD_RELOC_32_SECREL,
7613 BFD_RELOC_32_SECREL },
7614 OPERAND_TYPE_IMM32_32S_64_DISP32_64 },
7615 };
7616
7617 char *cp;
7618 unsigned j;
7619
7620 for (cp = input_line_pointer; *cp != '@'; cp++)
7621 if (is_end_of_line[(unsigned char) *cp] || *cp == ',')
7622 return NULL;
7623
7624 for (j = 0; j < ARRAY_SIZE (gotrel); j++)
7625 {
7626 int len = gotrel[j].len;
7627
7628 if (strncasecmp (cp + 1, gotrel[j].str, len) == 0)
7629 {
7630 if (gotrel[j].rel[object_64bit] != 0)
7631 {
7632 int first, second;
7633 char *tmpbuf, *past_reloc;
7634
7635 *rel = gotrel[j].rel[object_64bit];
7636 if (adjust)
7637 *adjust = len;
7638
7639 if (types)
7640 {
7641 if (flag_code != CODE_64BIT)
7642 {
7643 types->bitfield.imm32 = 1;
7644 types->bitfield.disp32 = 1;
7645 }
7646 else
7647 *types = gotrel[j].types64;
7648 }
7649
7650 /* The length of the first part of our input line. */
7651 first = cp - input_line_pointer;
7652
7653 /* The second part goes from after the reloc token until
7654 (and including) an end_of_line char or comma. */
7655 past_reloc = cp + 1 + len;
7656 cp = past_reloc;
7657 while (!is_end_of_line[(unsigned char) *cp] && *cp != ',')
7658 ++cp;
7659 second = cp + 1 - past_reloc;
7660
7661 /* Allocate and copy string. The trailing NUL shouldn't
7662 be necessary, but be safe. */
7663 tmpbuf = (char *) xmalloc (first + second + 2);
7664 memcpy (tmpbuf, input_line_pointer, first);
7665 if (second != 0 && *past_reloc != ' ')
7666 /* Replace the relocation token with ' ', so that
7667 errors like foo@SECLREL321 will be detected. */
7668 tmpbuf[first++] = ' ';
7669 memcpy (tmpbuf + first, past_reloc, second);
7670 tmpbuf[first + second] = '\0';
7671 return tmpbuf;
7672 }
7673
7674 as_bad (_("@%s reloc is not supported with %d-bit output format"),
7675 gotrel[j].str, 1 << (5 + object_64bit));
7676 return NULL;
7677 }
7678 }
7679
7680 /* Might be a symbol version string. Don't as_bad here. */
7681 return NULL;
7682 }
7683
7684 #endif /* TE_PE */
7685
7686 bfd_reloc_code_real_type
7687 x86_cons (expressionS *exp, int size)
7688 {
7689 bfd_reloc_code_real_type got_reloc = NO_RELOC;
7690
7691 intel_syntax = -intel_syntax;
7692
7693 exp->X_md = 0;
7694 if (size == 4 || (object_64bit && size == 8))
7695 {
7696 /* Handle @GOTOFF and the like in an expression. */
7697 char *save;
7698 char *gotfree_input_line;
7699 int adjust = 0;
7700
7701 save = input_line_pointer;
7702 gotfree_input_line = lex_got (&got_reloc, &adjust, NULL, 0);
7703 if (gotfree_input_line)
7704 input_line_pointer = gotfree_input_line;
7705
7706 expression (exp);
7707
7708 if (gotfree_input_line)
7709 {
7710 /* expression () has merrily parsed up to the end of line,
7711 or a comma - in the wrong buffer. Transfer how far
7712 input_line_pointer has moved to the right buffer. */
7713 input_line_pointer = (save
7714 + (input_line_pointer - gotfree_input_line)
7715 + adjust);
7716 free (gotfree_input_line);
7717 if (exp->X_op == O_constant
7718 || exp->X_op == O_absent
7719 || exp->X_op == O_illegal
7720 || exp->X_op == O_register
7721 || exp->X_op == O_big)
7722 {
7723 char c = *input_line_pointer;
7724 *input_line_pointer = 0;
7725 as_bad (_("missing or invalid expression `%s'"), save);
7726 *input_line_pointer = c;
7727 }
7728 }
7729 }
7730 else
7731 expression (exp);
7732
7733 intel_syntax = -intel_syntax;
7734
7735 if (intel_syntax)
7736 i386_intel_simplify (exp);
7737
7738 return got_reloc;
7739 }
7740
7741 static void
7742 signed_cons (int size)
7743 {
7744 if (flag_code == CODE_64BIT)
7745 cons_sign = 1;
7746 cons (size);
7747 cons_sign = -1;
7748 }
7749
7750 #ifdef TE_PE
7751 static void
7752 pe_directive_secrel (int dummy ATTRIBUTE_UNUSED)
7753 {
7754 expressionS exp;
7755
7756 do
7757 {
7758 expression (&exp);
7759 if (exp.X_op == O_symbol)
7760 exp.X_op = O_secrel;
7761
7762 emit_expr (&exp, 4);
7763 }
7764 while (*input_line_pointer++ == ',');
7765
7766 input_line_pointer--;
7767 demand_empty_rest_of_line ();
7768 }
7769 #endif
7770
7771 /* Handle Vector operations. */
7772
7773 static char *
7774 check_VecOperations (char *op_string, char *op_end)
7775 {
7776 const reg_entry *mask;
7777 const char *saved;
7778 char *end_op;
7779
7780 while (*op_string
7781 && (op_end == NULL || op_string < op_end))
7782 {
7783 saved = op_string;
7784 if (*op_string == '{')
7785 {
7786 op_string++;
7787
7788 /* Check broadcasts. */
7789 if (strncmp (op_string, "1to", 3) == 0)
7790 {
7791 int bcst_type;
7792
7793 if (i.broadcast)
7794 goto duplicated_vec_op;
7795
7796 op_string += 3;
7797 if (*op_string == '8')
7798 bcst_type = BROADCAST_1TO8;
7799 else if (*op_string == '4')
7800 bcst_type = BROADCAST_1TO4;
7801 else if (*op_string == '2')
7802 bcst_type = BROADCAST_1TO2;
7803 else if (*op_string == '1'
7804 && *(op_string+1) == '6')
7805 {
7806 bcst_type = BROADCAST_1TO16;
7807 op_string++;
7808 }
7809 else
7810 {
7811 as_bad (_("Unsupported broadcast: `%s'"), saved);
7812 return NULL;
7813 }
7814 op_string++;
7815
7816 broadcast_op.type = bcst_type;
7817 broadcast_op.operand = this_operand;
7818 i.broadcast = &broadcast_op;
7819 }
7820 /* Check masking operation. */
7821 else if ((mask = parse_register (op_string, &end_op)) != NULL)
7822 {
7823 /* k0 can't be used for write mask. */
7824 if (mask->reg_num == 0)
7825 {
7826 as_bad (_("`%s' can't be used for write mask"),
7827 op_string);
7828 return NULL;
7829 }
7830
7831 if (!i.mask)
7832 {
7833 mask_op.mask = mask;
7834 mask_op.zeroing = 0;
7835 mask_op.operand = this_operand;
7836 i.mask = &mask_op;
7837 }
7838 else
7839 {
7840 if (i.mask->mask)
7841 goto duplicated_vec_op;
7842
7843 i.mask->mask = mask;
7844
7845 /* Only "{z}" is allowed here. No need to check
7846 zeroing mask explicitly. */
7847 if (i.mask->operand != this_operand)
7848 {
7849 as_bad (_("invalid write mask `%s'"), saved);
7850 return NULL;
7851 }
7852 }
7853
7854 op_string = end_op;
7855 }
7856 /* Check zeroing-flag for masking operation. */
7857 else if (*op_string == 'z')
7858 {
7859 if (!i.mask)
7860 {
7861 mask_op.mask = NULL;
7862 mask_op.zeroing = 1;
7863 mask_op.operand = this_operand;
7864 i.mask = &mask_op;
7865 }
7866 else
7867 {
7868 if (i.mask->zeroing)
7869 {
7870 duplicated_vec_op:
7871 as_bad (_("duplicated `%s'"), saved);
7872 return NULL;
7873 }
7874
7875 i.mask->zeroing = 1;
7876
7877 /* Only "{%k}" is allowed here. No need to check mask
7878 register explicitly. */
7879 if (i.mask->operand != this_operand)
7880 {
7881 as_bad (_("invalid zeroing-masking `%s'"),
7882 saved);
7883 return NULL;
7884 }
7885 }
7886
7887 op_string++;
7888 }
7889 else
7890 goto unknown_vec_op;
7891
7892 if (*op_string != '}')
7893 {
7894 as_bad (_("missing `}' in `%s'"), saved);
7895 return NULL;
7896 }
7897 op_string++;
7898 continue;
7899 }
7900 unknown_vec_op:
7901 /* We don't know this one. */
7902 as_bad (_("unknown vector operation: `%s'"), saved);
7903 return NULL;
7904 }
7905
7906 return op_string;
7907 }
7908
7909 static int
7910 i386_immediate (char *imm_start)
7911 {
7912 char *save_input_line_pointer;
7913 char *gotfree_input_line;
7914 segT exp_seg = 0;
7915 expressionS *exp;
7916 i386_operand_type types;
7917
7918 operand_type_set (&types, ~0);
7919
7920 if (i.imm_operands == MAX_IMMEDIATE_OPERANDS)
7921 {
7922 as_bad (_("at most %d immediate operands are allowed"),
7923 MAX_IMMEDIATE_OPERANDS);
7924 return 0;
7925 }
7926
7927 exp = &im_expressions[i.imm_operands++];
7928 i.op[this_operand].imms = exp;
7929
7930 if (is_space_char (*imm_start))
7931 ++imm_start;
7932
7933 save_input_line_pointer = input_line_pointer;
7934 input_line_pointer = imm_start;
7935
7936 gotfree_input_line = lex_got (&i.reloc[this_operand], NULL, &types,
7937 (i.bnd_prefix != NULL
7938 || add_bnd_prefix));
7939 if (gotfree_input_line)
7940 input_line_pointer = gotfree_input_line;
7941
7942 exp_seg = expression (exp);
7943
7944 SKIP_WHITESPACE ();
7945
7946 /* Handle vector operations. */
7947 if (*input_line_pointer == '{')
7948 {
7949 input_line_pointer = check_VecOperations (input_line_pointer,
7950 NULL);
7951 if (input_line_pointer == NULL)
7952 return 0;
7953 }
7954
7955 if (*input_line_pointer)
7956 as_bad (_("junk `%s' after expression"), input_line_pointer);
7957
7958 input_line_pointer = save_input_line_pointer;
7959 if (gotfree_input_line)
7960 {
7961 free (gotfree_input_line);
7962
7963 if (exp->X_op == O_constant || exp->X_op == O_register)
7964 exp->X_op = O_illegal;
7965 }
7966
7967 return i386_finalize_immediate (exp_seg, exp, types, imm_start);
7968 }
7969
7970 static int
7971 i386_finalize_immediate (segT exp_seg ATTRIBUTE_UNUSED, expressionS *exp,
7972 i386_operand_type types, const char *imm_start)
7973 {
7974 if (exp->X_op == O_absent || exp->X_op == O_illegal || exp->X_op == O_big)
7975 {
7976 if (imm_start)
7977 as_bad (_("missing or invalid immediate expression `%s'"),
7978 imm_start);
7979 return 0;
7980 }
7981 else if (exp->X_op == O_constant)
7982 {
7983 /* Size it properly later. */
7984 i.types[this_operand].bitfield.imm64 = 1;
7985 /* If not 64bit, sign extend val. */
7986 if (flag_code != CODE_64BIT
7987 && (exp->X_add_number & ~(((addressT) 2 << 31) - 1)) == 0)
7988 exp->X_add_number
7989 = (exp->X_add_number ^ ((addressT) 1 << 31)) - ((addressT) 1 << 31);
7990 }
7991 #if (defined (OBJ_AOUT) || defined (OBJ_MAYBE_AOUT))
7992 else if (OUTPUT_FLAVOR == bfd_target_aout_flavour
7993 && exp_seg != absolute_section
7994 && exp_seg != text_section
7995 && exp_seg != data_section
7996 && exp_seg != bss_section
7997 && exp_seg != undefined_section
7998 && !bfd_is_com_section (exp_seg))
7999 {
8000 as_bad (_("unimplemented segment %s in operand"), exp_seg->name);
8001 return 0;
8002 }
8003 #endif
8004 else if (!intel_syntax && exp_seg == reg_section)
8005 {
8006 if (imm_start)
8007 as_bad (_("illegal immediate register operand %s"), imm_start);
8008 return 0;
8009 }
8010 else
8011 {
8012 /* This is an address. The size of the address will be
8013 determined later, depending on destination register,
8014 suffix, or the default for the section. */
8015 i.types[this_operand].bitfield.imm8 = 1;
8016 i.types[this_operand].bitfield.imm16 = 1;
8017 i.types[this_operand].bitfield.imm32 = 1;
8018 i.types[this_operand].bitfield.imm32s = 1;
8019 i.types[this_operand].bitfield.imm64 = 1;
8020 i.types[this_operand] = operand_type_and (i.types[this_operand],
8021 types);
8022 }
8023
8024 return 1;
8025 }
8026
8027 static char *
8028 i386_scale (char *scale)
8029 {
8030 offsetT val;
8031 char *save = input_line_pointer;
8032
8033 input_line_pointer = scale;
8034 val = get_absolute_expression ();
8035
8036 switch (val)
8037 {
8038 case 1:
8039 i.log2_scale_factor = 0;
8040 break;
8041 case 2:
8042 i.log2_scale_factor = 1;
8043 break;
8044 case 4:
8045 i.log2_scale_factor = 2;
8046 break;
8047 case 8:
8048 i.log2_scale_factor = 3;
8049 break;
8050 default:
8051 {
8052 char sep = *input_line_pointer;
8053
8054 *input_line_pointer = '\0';
8055 as_bad (_("expecting scale factor of 1, 2, 4, or 8: got `%s'"),
8056 scale);
8057 *input_line_pointer = sep;
8058 input_line_pointer = save;
8059 return NULL;
8060 }
8061 }
8062 if (i.log2_scale_factor != 0 && i.index_reg == 0)
8063 {
8064 as_warn (_("scale factor of %d without an index register"),
8065 1 << i.log2_scale_factor);
8066 i.log2_scale_factor = 0;
8067 }
8068 scale = input_line_pointer;
8069 input_line_pointer = save;
8070 return scale;
8071 }
8072
8073 static int
8074 i386_displacement (char *disp_start, char *disp_end)
8075 {
8076 expressionS *exp;
8077 segT exp_seg = 0;
8078 char *save_input_line_pointer;
8079 char *gotfree_input_line;
8080 int override;
8081 i386_operand_type bigdisp, types = anydisp;
8082 int ret;
8083
8084 if (i.disp_operands == MAX_MEMORY_OPERANDS)
8085 {
8086 as_bad (_("at most %d displacement operands are allowed"),
8087 MAX_MEMORY_OPERANDS);
8088 return 0;
8089 }
8090
8091 operand_type_set (&bigdisp, 0);
8092 if ((i.types[this_operand].bitfield.jumpabsolute)
8093 || (!current_templates->start->opcode_modifier.jump
8094 && !current_templates->start->opcode_modifier.jumpdword))
8095 {
8096 bigdisp.bitfield.disp32 = 1;
8097 override = (i.prefix[ADDR_PREFIX] != 0);
8098 if (flag_code == CODE_64BIT)
8099 {
8100 if (!override)
8101 {
8102 bigdisp.bitfield.disp32s = 1;
8103 bigdisp.bitfield.disp64 = 1;
8104 }
8105 }
8106 else if ((flag_code == CODE_16BIT) ^ override)
8107 {
8108 bigdisp.bitfield.disp32 = 0;
8109 bigdisp.bitfield.disp16 = 1;
8110 }
8111 }
8112 else
8113 {
8114 /* For PC-relative branches, the width of the displacement
8115 is dependent upon data size, not address size. */
8116 override = (i.prefix[DATA_PREFIX] != 0);
8117 if (flag_code == CODE_64BIT)
8118 {
8119 if (override || i.suffix == WORD_MNEM_SUFFIX)
8120 bigdisp.bitfield.disp16 = 1;
8121 else
8122 {
8123 bigdisp.bitfield.disp32 = 1;
8124 bigdisp.bitfield.disp32s = 1;
8125 }
8126 }
8127 else
8128 {
8129 if (!override)
8130 override = (i.suffix == (flag_code != CODE_16BIT
8131 ? WORD_MNEM_SUFFIX
8132 : LONG_MNEM_SUFFIX));
8133 bigdisp.bitfield.disp32 = 1;
8134 if ((flag_code == CODE_16BIT) ^ override)
8135 {
8136 bigdisp.bitfield.disp32 = 0;
8137 bigdisp.bitfield.disp16 = 1;
8138 }
8139 }
8140 }
8141 i.types[this_operand] = operand_type_or (i.types[this_operand],
8142 bigdisp);
8143
8144 exp = &disp_expressions[i.disp_operands];
8145 i.op[this_operand].disps = exp;
8146 i.disp_operands++;
8147 save_input_line_pointer = input_line_pointer;
8148 input_line_pointer = disp_start;
8149 END_STRING_AND_SAVE (disp_end);
8150
8151 #ifndef GCC_ASM_O_HACK
8152 #define GCC_ASM_O_HACK 0
8153 #endif
8154 #if GCC_ASM_O_HACK
8155 END_STRING_AND_SAVE (disp_end + 1);
8156 if (i.types[this_operand].bitfield.baseIndex
8157 && displacement_string_end[-1] == '+')
8158 {
8159 /* This hack is to avoid a warning when using the "o"
8160 constraint within gcc asm statements.
8161 For instance:
8162
8163 #define _set_tssldt_desc(n,addr,limit,type) \
8164 __asm__ __volatile__ ( \
8165 "movw %w2,%0\n\t" \
8166 "movw %w1,2+%0\n\t" \
8167 "rorl $16,%1\n\t" \
8168 "movb %b1,4+%0\n\t" \
8169 "movb %4,5+%0\n\t" \
8170 "movb $0,6+%0\n\t" \
8171 "movb %h1,7+%0\n\t" \
8172 "rorl $16,%1" \
8173 : "=o"(*(n)) : "q" (addr), "ri"(limit), "i"(type))
8174
8175 This works great except that the output assembler ends
8176 up looking a bit weird if it turns out that there is
8177 no offset. You end up producing code that looks like:
8178
8179 #APP
8180 movw $235,(%eax)
8181 movw %dx,2+(%eax)
8182 rorl $16,%edx
8183 movb %dl,4+(%eax)
8184 movb $137,5+(%eax)
8185 movb $0,6+(%eax)
8186 movb %dh,7+(%eax)
8187 rorl $16,%edx
8188 #NO_APP
8189
8190 So here we provide the missing zero. */
8191
8192 *displacement_string_end = '0';
8193 }
8194 #endif
8195 gotfree_input_line = lex_got (&i.reloc[this_operand], NULL, &types,
8196 (i.bnd_prefix != NULL
8197 || add_bnd_prefix));
8198 if (gotfree_input_line)
8199 input_line_pointer = gotfree_input_line;
8200
8201 exp_seg = expression (exp);
8202
8203 SKIP_WHITESPACE ();
8204 if (*input_line_pointer)
8205 as_bad (_("junk `%s' after expression"), input_line_pointer);
8206 #if GCC_ASM_O_HACK
8207 RESTORE_END_STRING (disp_end + 1);
8208 #endif
8209 input_line_pointer = save_input_line_pointer;
8210 if (gotfree_input_line)
8211 {
8212 free (gotfree_input_line);
8213
8214 if (exp->X_op == O_constant || exp->X_op == O_register)
8215 exp->X_op = O_illegal;
8216 }
8217
8218 ret = i386_finalize_displacement (exp_seg, exp, types, disp_start);
8219
8220 RESTORE_END_STRING (disp_end);
8221
8222 return ret;
8223 }
8224
8225 static int
8226 i386_finalize_displacement (segT exp_seg ATTRIBUTE_UNUSED, expressionS *exp,
8227 i386_operand_type types, const char *disp_start)
8228 {
8229 i386_operand_type bigdisp;
8230 int ret = 1;
8231
8232 /* We do this to make sure that the section symbol is in
8233 the symbol table. We will ultimately change the relocation
8234 to be relative to the beginning of the section. */
8235 if (i.reloc[this_operand] == BFD_RELOC_386_GOTOFF
8236 || i.reloc[this_operand] == BFD_RELOC_X86_64_GOTPCREL
8237 || i.reloc[this_operand] == BFD_RELOC_X86_64_GOTOFF64)
8238 {
8239 if (exp->X_op != O_symbol)
8240 goto inv_disp;
8241
8242 if (S_IS_LOCAL (exp->X_add_symbol)
8243 && S_GET_SEGMENT (exp->X_add_symbol) != undefined_section
8244 && S_GET_SEGMENT (exp->X_add_symbol) != expr_section)
8245 section_symbol (S_GET_SEGMENT (exp->X_add_symbol));
8246 exp->X_op = O_subtract;
8247 exp->X_op_symbol = GOT_symbol;
8248 if (i.reloc[this_operand] == BFD_RELOC_X86_64_GOTPCREL)
8249 i.reloc[this_operand] = BFD_RELOC_32_PCREL;
8250 else if (i.reloc[this_operand] == BFD_RELOC_X86_64_GOTOFF64)
8251 i.reloc[this_operand] = BFD_RELOC_64;
8252 else
8253 i.reloc[this_operand] = BFD_RELOC_32;
8254 }
8255
8256 else if (exp->X_op == O_absent
8257 || exp->X_op == O_illegal
8258 || exp->X_op == O_big)
8259 {
8260 inv_disp:
8261 as_bad (_("missing or invalid displacement expression `%s'"),
8262 disp_start);
8263 ret = 0;
8264 }
8265
8266 else if (flag_code == CODE_64BIT
8267 && !i.prefix[ADDR_PREFIX]
8268 && exp->X_op == O_constant)
8269 {
8270 /* Since displacement is signed extended to 64bit, don't allow
8271 disp32 and turn off disp32s if they are out of range. */
8272 i.types[this_operand].bitfield.disp32 = 0;
8273 if (!fits_in_signed_long (exp->X_add_number))
8274 {
8275 i.types[this_operand].bitfield.disp32s = 0;
8276 if (i.types[this_operand].bitfield.baseindex)
8277 {
8278 as_bad (_("0x%lx out range of signed 32bit displacement"),
8279 (long) exp->X_add_number);
8280 ret = 0;
8281 }
8282 }
8283 }
8284
8285 #if (defined (OBJ_AOUT) || defined (OBJ_MAYBE_AOUT))
8286 else if (exp->X_op != O_constant
8287 && OUTPUT_FLAVOR == bfd_target_aout_flavour
8288 && exp_seg != absolute_section
8289 && exp_seg != text_section
8290 && exp_seg != data_section
8291 && exp_seg != bss_section
8292 && exp_seg != undefined_section
8293 && !bfd_is_com_section (exp_seg))
8294 {
8295 as_bad (_("unimplemented segment %s in operand"), exp_seg->name);
8296 ret = 0;
8297 }
8298 #endif
8299
8300 /* Check if this is a displacement only operand. */
8301 bigdisp = i.types[this_operand];
8302 bigdisp.bitfield.disp8 = 0;
8303 bigdisp.bitfield.disp16 = 0;
8304 bigdisp.bitfield.disp32 = 0;
8305 bigdisp.bitfield.disp32s = 0;
8306 bigdisp.bitfield.disp64 = 0;
8307 if (operand_type_all_zero (&bigdisp))
8308 i.types[this_operand] = operand_type_and (i.types[this_operand],
8309 types);
8310
8311 return ret;
8312 }
8313
8314 /* Make sure the memory operand we've been dealt is valid.
8315 Return 1 on success, 0 on a failure. */
8316
8317 static int
8318 i386_index_check (const char *operand_string)
8319 {
8320 const char *kind = "base/index";
8321 enum flag_code addr_mode;
8322
8323 if (i.prefix[ADDR_PREFIX])
8324 addr_mode = flag_code == CODE_32BIT ? CODE_16BIT : CODE_32BIT;
8325 else
8326 {
8327 addr_mode = flag_code;
8328
8329 #if INFER_ADDR_PREFIX
8330 if (i.mem_operands == 0)
8331 {
8332 /* Infer address prefix from the first memory operand. */
8333 const reg_entry *addr_reg = i.base_reg;
8334
8335 if (addr_reg == NULL)
8336 addr_reg = i.index_reg;
8337
8338 if (addr_reg)
8339 {
8340 if (addr_reg->reg_num == RegEip
8341 || addr_reg->reg_num == RegEiz
8342 || addr_reg->reg_type.bitfield.reg32)
8343 addr_mode = CODE_32BIT;
8344 else if (flag_code != CODE_64BIT
8345 && addr_reg->reg_type.bitfield.reg16)
8346 addr_mode = CODE_16BIT;
8347
8348 if (addr_mode != flag_code)
8349 {
8350 i.prefix[ADDR_PREFIX] = ADDR_PREFIX_OPCODE;
8351 i.prefixes += 1;
8352 /* Change the size of any displacement too. At most one
8353 of Disp16 or Disp32 is set.
8354 FIXME. There doesn't seem to be any real need for
8355 separate Disp16 and Disp32 flags. The same goes for
8356 Imm16 and Imm32. Removing them would probably clean
8357 up the code quite a lot. */
8358 if (flag_code != CODE_64BIT
8359 && (i.types[this_operand].bitfield.disp16
8360 || i.types[this_operand].bitfield.disp32))
8361 i.types[this_operand]
8362 = operand_type_xor (i.types[this_operand], disp16_32);
8363 }
8364 }
8365 }
8366 #endif
8367 }
8368
8369 if (current_templates->start->opcode_modifier.isstring
8370 && !current_templates->start->opcode_modifier.immext
8371 && (current_templates->end[-1].opcode_modifier.isstring
8372 || i.mem_operands))
8373 {
8374 /* Memory operands of string insns are special in that they only allow
8375 a single register (rDI, rSI, or rBX) as their memory address. */
8376 const reg_entry *expected_reg;
8377 static const char *di_si[][2] =
8378 {
8379 { "esi", "edi" },
8380 { "si", "di" },
8381 { "rsi", "rdi" }
8382 };
8383 static const char *bx[] = { "ebx", "bx", "rbx" };
8384
8385 kind = "string address";
8386
8387 if (current_templates->start->opcode_modifier.w)
8388 {
8389 i386_operand_type type = current_templates->end[-1].operand_types[0];
8390
8391 if (!type.bitfield.baseindex
8392 || ((!i.mem_operands != !intel_syntax)
8393 && current_templates->end[-1].operand_types[1]
8394 .bitfield.baseindex))
8395 type = current_templates->end[-1].operand_types[1];
8396 expected_reg = hash_find (reg_hash,
8397 di_si[addr_mode][type.bitfield.esseg]);
8398
8399 }
8400 else
8401 expected_reg = hash_find (reg_hash, bx[addr_mode]);
8402
8403 if (i.base_reg != expected_reg
8404 || i.index_reg
8405 || operand_type_check (i.types[this_operand], disp))
8406 {
8407 /* The second memory operand must have the same size as
8408 the first one. */
8409 if (i.mem_operands
8410 && i.base_reg
8411 && !((addr_mode == CODE_64BIT
8412 && i.base_reg->reg_type.bitfield.reg64)
8413 || (addr_mode == CODE_32BIT
8414 ? i.base_reg->reg_type.bitfield.reg32
8415 : i.base_reg->reg_type.bitfield.reg16)))
8416 goto bad_address;
8417
8418 as_warn (_("`%s' is not valid here (expected `%c%s%s%c')"),
8419 operand_string,
8420 intel_syntax ? '[' : '(',
8421 register_prefix,
8422 expected_reg->reg_name,
8423 intel_syntax ? ']' : ')');
8424 return 1;
8425 }
8426 else
8427 return 1;
8428
8429 bad_address:
8430 as_bad (_("`%s' is not a valid %s expression"),
8431 operand_string, kind);
8432 return 0;
8433 }
8434 else
8435 {
8436 if (addr_mode != CODE_16BIT)
8437 {
8438 /* 32-bit/64-bit checks. */
8439 if ((i.base_reg
8440 && (addr_mode == CODE_64BIT
8441 ? !i.base_reg->reg_type.bitfield.reg64
8442 : !i.base_reg->reg_type.bitfield.reg32)
8443 && (i.index_reg
8444 || (i.base_reg->reg_num
8445 != (addr_mode == CODE_64BIT ? RegRip : RegEip))))
8446 || (i.index_reg
8447 && !i.index_reg->reg_type.bitfield.regxmm
8448 && !i.index_reg->reg_type.bitfield.regymm
8449 && !i.index_reg->reg_type.bitfield.regzmm
8450 && ((addr_mode == CODE_64BIT
8451 ? !(i.index_reg->reg_type.bitfield.reg64
8452 || i.index_reg->reg_num == RegRiz)
8453 : !(i.index_reg->reg_type.bitfield.reg32
8454 || i.index_reg->reg_num == RegEiz))
8455 || !i.index_reg->reg_type.bitfield.baseindex)))
8456 goto bad_address;
8457 }
8458 else
8459 {
8460 /* 16-bit checks. */
8461 if ((i.base_reg
8462 && (!i.base_reg->reg_type.bitfield.reg16
8463 || !i.base_reg->reg_type.bitfield.baseindex))
8464 || (i.index_reg
8465 && (!i.index_reg->reg_type.bitfield.reg16
8466 || !i.index_reg->reg_type.bitfield.baseindex
8467 || !(i.base_reg
8468 && i.base_reg->reg_num < 6
8469 && i.index_reg->reg_num >= 6
8470 && i.log2_scale_factor == 0))))
8471 goto bad_address;
8472 }
8473 }
8474 return 1;
8475 }
8476
8477 /* Handle vector immediates. */
8478
8479 static int
8480 RC_SAE_immediate (const char *imm_start)
8481 {
8482 unsigned int match_found, j;
8483 const char *pstr = imm_start;
8484 expressionS *exp;
8485
8486 if (*pstr != '{')
8487 return 0;
8488
8489 pstr++;
8490 match_found = 0;
8491 for (j = 0; j < ARRAY_SIZE (RC_NamesTable); j++)
8492 {
8493 if (!strncmp (pstr, RC_NamesTable[j].name, RC_NamesTable[j].len))
8494 {
8495 if (!i.rounding)
8496 {
8497 rc_op.type = RC_NamesTable[j].type;
8498 rc_op.operand = this_operand;
8499 i.rounding = &rc_op;
8500 }
8501 else
8502 {
8503 as_bad (_("duplicated `%s'"), imm_start);
8504 return 0;
8505 }
8506 pstr += RC_NamesTable[j].len;
8507 match_found = 1;
8508 break;
8509 }
8510 }
8511 if (!match_found)
8512 return 0;
8513
8514 if (*pstr++ != '}')
8515 {
8516 as_bad (_("Missing '}': '%s'"), imm_start);
8517 return 0;
8518 }
8519 /* RC/SAE immediate string should contain nothing more. */;
8520 if (*pstr != 0)
8521 {
8522 as_bad (_("Junk after '}': '%s'"), imm_start);
8523 return 0;
8524 }
8525
8526 exp = &im_expressions[i.imm_operands++];
8527 i.op[this_operand].imms = exp;
8528
8529 exp->X_op = O_constant;
8530 exp->X_add_number = 0;
8531 exp->X_add_symbol = (symbolS *) 0;
8532 exp->X_op_symbol = (symbolS *) 0;
8533
8534 i.types[this_operand].bitfield.imm8 = 1;
8535 return 1;
8536 }
8537
8538 /* Parse OPERAND_STRING into the i386_insn structure I. Returns zero
8539 on error. */
8540
8541 static int
8542 i386_att_operand (char *operand_string)
8543 {
8544 const reg_entry *r;
8545 char *end_op;
8546 char *op_string = operand_string;
8547
8548 if (is_space_char (*op_string))
8549 ++op_string;
8550
8551 /* We check for an absolute prefix (differentiating,
8552 for example, 'jmp pc_relative_label' from 'jmp *absolute_label'. */
8553 if (*op_string == ABSOLUTE_PREFIX)
8554 {
8555 ++op_string;
8556 if (is_space_char (*op_string))
8557 ++op_string;
8558 i.types[this_operand].bitfield.jumpabsolute = 1;
8559 }
8560
8561 /* Check if operand is a register. */
8562 if ((r = parse_register (op_string, &end_op)) != NULL)
8563 {
8564 i386_operand_type temp;
8565
8566 /* Check for a segment override by searching for ':' after a
8567 segment register. */
8568 op_string = end_op;
8569 if (is_space_char (*op_string))
8570 ++op_string;
8571 if (*op_string == ':'
8572 && (r->reg_type.bitfield.sreg2
8573 || r->reg_type.bitfield.sreg3))
8574 {
8575 switch (r->reg_num)
8576 {
8577 case 0:
8578 i.seg[i.mem_operands] = &es;
8579 break;
8580 case 1:
8581 i.seg[i.mem_operands] = &cs;
8582 break;
8583 case 2:
8584 i.seg[i.mem_operands] = &ss;
8585 break;
8586 case 3:
8587 i.seg[i.mem_operands] = &ds;
8588 break;
8589 case 4:
8590 i.seg[i.mem_operands] = &fs;
8591 break;
8592 case 5:
8593 i.seg[i.mem_operands] = &gs;
8594 break;
8595 }
8596
8597 /* Skip the ':' and whitespace. */
8598 ++op_string;
8599 if (is_space_char (*op_string))
8600 ++op_string;
8601
8602 if (!is_digit_char (*op_string)
8603 && !is_identifier_char (*op_string)
8604 && *op_string != '('
8605 && *op_string != ABSOLUTE_PREFIX)
8606 {
8607 as_bad (_("bad memory operand `%s'"), op_string);
8608 return 0;
8609 }
8610 /* Handle case of %es:*foo. */
8611 if (*op_string == ABSOLUTE_PREFIX)
8612 {
8613 ++op_string;
8614 if (is_space_char (*op_string))
8615 ++op_string;
8616 i.types[this_operand].bitfield.jumpabsolute = 1;
8617 }
8618 goto do_memory_reference;
8619 }
8620
8621 /* Handle vector operations. */
8622 if (*op_string == '{')
8623 {
8624 op_string = check_VecOperations (op_string, NULL);
8625 if (op_string == NULL)
8626 return 0;
8627 }
8628
8629 if (*op_string)
8630 {
8631 as_bad (_("junk `%s' after register"), op_string);
8632 return 0;
8633 }
8634 temp = r->reg_type;
8635 temp.bitfield.baseindex = 0;
8636 i.types[this_operand] = operand_type_or (i.types[this_operand],
8637 temp);
8638 i.types[this_operand].bitfield.unspecified = 0;
8639 i.op[this_operand].regs = r;
8640 i.reg_operands++;
8641 }
8642 else if (*op_string == REGISTER_PREFIX)
8643 {
8644 as_bad (_("bad register name `%s'"), op_string);
8645 return 0;
8646 }
8647 else if (*op_string == IMMEDIATE_PREFIX)
8648 {
8649 ++op_string;
8650 if (i.types[this_operand].bitfield.jumpabsolute)
8651 {
8652 as_bad (_("immediate operand illegal with absolute jump"));
8653 return 0;
8654 }
8655 if (!i386_immediate (op_string))
8656 return 0;
8657 }
8658 else if (RC_SAE_immediate (operand_string))
8659 {
8660 /* If it is a RC or SAE immediate, do nothing. */
8661 ;
8662 }
8663 else if (is_digit_char (*op_string)
8664 || is_identifier_char (*op_string)
8665 || *op_string == '(')
8666 {
8667 /* This is a memory reference of some sort. */
8668 char *base_string;
8669
8670 /* Start and end of displacement string expression (if found). */
8671 char *displacement_string_start;
8672 char *displacement_string_end;
8673 char *vop_start;
8674
8675 do_memory_reference:
8676 if ((i.mem_operands == 1
8677 && !current_templates->start->opcode_modifier.isstring)
8678 || i.mem_operands == 2)
8679 {
8680 as_bad (_("too many memory references for `%s'"),
8681 current_templates->start->name);
8682 return 0;
8683 }
8684
8685 /* Check for base index form. We detect the base index form by
8686 looking for an ')' at the end of the operand, searching
8687 for the '(' matching it, and finding a REGISTER_PREFIX or ','
8688 after the '('. */
8689 base_string = op_string + strlen (op_string);
8690
8691 /* Handle vector operations. */
8692 vop_start = strchr (op_string, '{');
8693 if (vop_start && vop_start < base_string)
8694 {
8695 if (check_VecOperations (vop_start, base_string) == NULL)
8696 return 0;
8697 base_string = vop_start;
8698 }
8699
8700 --base_string;
8701 if (is_space_char (*base_string))
8702 --base_string;
8703
8704 /* If we only have a displacement, set-up for it to be parsed later. */
8705 displacement_string_start = op_string;
8706 displacement_string_end = base_string + 1;
8707
8708 if (*base_string == ')')
8709 {
8710 char *temp_string;
8711 unsigned int parens_balanced = 1;
8712 /* We've already checked that the number of left & right ()'s are
8713 equal, so this loop will not be infinite. */
8714 do
8715 {
8716 base_string--;
8717 if (*base_string == ')')
8718 parens_balanced++;
8719 if (*base_string == '(')
8720 parens_balanced--;
8721 }
8722 while (parens_balanced);
8723
8724 temp_string = base_string;
8725
8726 /* Skip past '(' and whitespace. */
8727 ++base_string;
8728 if (is_space_char (*base_string))
8729 ++base_string;
8730
8731 if (*base_string == ','
8732 || ((i.base_reg = parse_register (base_string, &end_op))
8733 != NULL))
8734 {
8735 displacement_string_end = temp_string;
8736
8737 i.types[this_operand].bitfield.baseindex = 1;
8738
8739 if (i.base_reg)
8740 {
8741 base_string = end_op;
8742 if (is_space_char (*base_string))
8743 ++base_string;
8744 }
8745
8746 /* There may be an index reg or scale factor here. */
8747 if (*base_string == ',')
8748 {
8749 ++base_string;
8750 if (is_space_char (*base_string))
8751 ++base_string;
8752
8753 if ((i.index_reg = parse_register (base_string, &end_op))
8754 != NULL)
8755 {
8756 base_string = end_op;
8757 if (is_space_char (*base_string))
8758 ++base_string;
8759 if (*base_string == ',')
8760 {
8761 ++base_string;
8762 if (is_space_char (*base_string))
8763 ++base_string;
8764 }
8765 else if (*base_string != ')')
8766 {
8767 as_bad (_("expecting `,' or `)' "
8768 "after index register in `%s'"),
8769 operand_string);
8770 return 0;
8771 }
8772 }
8773 else if (*base_string == REGISTER_PREFIX)
8774 {
8775 end_op = strchr (base_string, ',');
8776 if (end_op)
8777 *end_op = '\0';
8778 as_bad (_("bad register name `%s'"), base_string);
8779 return 0;
8780 }
8781
8782 /* Check for scale factor. */
8783 if (*base_string != ')')
8784 {
8785 char *end_scale = i386_scale (base_string);
8786
8787 if (!end_scale)
8788 return 0;
8789
8790 base_string = end_scale;
8791 if (is_space_char (*base_string))
8792 ++base_string;
8793 if (*base_string != ')')
8794 {
8795 as_bad (_("expecting `)' "
8796 "after scale factor in `%s'"),
8797 operand_string);
8798 return 0;
8799 }
8800 }
8801 else if (!i.index_reg)
8802 {
8803 as_bad (_("expecting index register or scale factor "
8804 "after `,'; got '%c'"),
8805 *base_string);
8806 return 0;
8807 }
8808 }
8809 else if (*base_string != ')')
8810 {
8811 as_bad (_("expecting `,' or `)' "
8812 "after base register in `%s'"),
8813 operand_string);
8814 return 0;
8815 }
8816 }
8817 else if (*base_string == REGISTER_PREFIX)
8818 {
8819 end_op = strchr (base_string, ',');
8820 if (end_op)
8821 *end_op = '\0';
8822 as_bad (_("bad register name `%s'"), base_string);
8823 return 0;
8824 }
8825 }
8826
8827 /* If there's an expression beginning the operand, parse it,
8828 assuming displacement_string_start and
8829 displacement_string_end are meaningful. */
8830 if (displacement_string_start != displacement_string_end)
8831 {
8832 if (!i386_displacement (displacement_string_start,
8833 displacement_string_end))
8834 return 0;
8835 }
8836
8837 /* Special case for (%dx) while doing input/output op. */
8838 if (i.base_reg
8839 && operand_type_equal (&i.base_reg->reg_type,
8840 &reg16_inoutportreg)
8841 && i.index_reg == 0
8842 && i.log2_scale_factor == 0
8843 && i.seg[i.mem_operands] == 0
8844 && !operand_type_check (i.types[this_operand], disp))
8845 {
8846 i.types[this_operand] = inoutportreg;
8847 return 1;
8848 }
8849
8850 if (i386_index_check (operand_string) == 0)
8851 return 0;
8852 i.types[this_operand].bitfield.mem = 1;
8853 i.mem_operands++;
8854 }
8855 else
8856 {
8857 /* It's not a memory operand; argh! */
8858 as_bad (_("invalid char %s beginning operand %d `%s'"),
8859 output_invalid (*op_string),
8860 this_operand + 1,
8861 op_string);
8862 return 0;
8863 }
8864 return 1; /* Normal return. */
8865 }
8866 \f
8867 /* Calculate the maximum variable size (i.e., excluding fr_fix)
8868 that an rs_machine_dependent frag may reach. */
8869
8870 unsigned int
8871 i386_frag_max_var (fragS *frag)
8872 {
8873 /* The only relaxable frags are for jumps.
8874 Unconditional jumps can grow by 4 bytes and others by 5 bytes. */
8875 gas_assert (frag->fr_type == rs_machine_dependent);
8876 return TYPE_FROM_RELAX_STATE (frag->fr_subtype) == UNCOND_JUMP ? 4 : 5;
8877 }
8878
8879 /* md_estimate_size_before_relax()
8880
8881 Called just before relax() for rs_machine_dependent frags. The x86
8882 assembler uses these frags to handle variable size jump
8883 instructions.
8884
8885 Any symbol that is now undefined will not become defined.
8886 Return the correct fr_subtype in the frag.
8887 Return the initial "guess for variable size of frag" to caller.
8888 The guess is actually the growth beyond the fixed part. Whatever
8889 we do to grow the fixed or variable part contributes to our
8890 returned value. */
8891
8892 int
8893 md_estimate_size_before_relax (fragS *fragP, segT segment)
8894 {
8895 /* We've already got fragP->fr_subtype right; all we have to do is
8896 check for un-relaxable symbols. On an ELF system, we can't relax
8897 an externally visible symbol, because it may be overridden by a
8898 shared library. */
8899 if (S_GET_SEGMENT (fragP->fr_symbol) != segment
8900 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
8901 || (IS_ELF
8902 && (S_IS_EXTERNAL (fragP->fr_symbol)
8903 || S_IS_WEAK (fragP->fr_symbol)
8904 || ((symbol_get_bfdsym (fragP->fr_symbol)->flags
8905 & BSF_GNU_INDIRECT_FUNCTION))))
8906 #endif
8907 #if defined (OBJ_COFF) && defined (TE_PE)
8908 || (OUTPUT_FLAVOR == bfd_target_coff_flavour
8909 && S_IS_WEAK (fragP->fr_symbol))
8910 #endif
8911 )
8912 {
8913 /* Symbol is undefined in this segment, or we need to keep a
8914 reloc so that weak symbols can be overridden. */
8915 int size = (fragP->fr_subtype & CODE16) ? 2 : 4;
8916 enum bfd_reloc_code_real reloc_type;
8917 unsigned char *opcode;
8918 int old_fr_fix;
8919
8920 if (fragP->fr_var != NO_RELOC)
8921 reloc_type = (enum bfd_reloc_code_real) fragP->fr_var;
8922 else if (size == 2)
8923 reloc_type = BFD_RELOC_16_PCREL;
8924 else
8925 reloc_type = BFD_RELOC_32_PCREL;
8926
8927 old_fr_fix = fragP->fr_fix;
8928 opcode = (unsigned char *) fragP->fr_opcode;
8929
8930 switch (TYPE_FROM_RELAX_STATE (fragP->fr_subtype))
8931 {
8932 case UNCOND_JUMP:
8933 /* Make jmp (0xeb) a (d)word displacement jump. */
8934 opcode[0] = 0xe9;
8935 fragP->fr_fix += size;
8936 fix_new (fragP, old_fr_fix, size,
8937 fragP->fr_symbol,
8938 fragP->fr_offset, 1,
8939 reloc_type);
8940 break;
8941
8942 case COND_JUMP86:
8943 if (size == 2
8944 && (!no_cond_jump_promotion || fragP->fr_var != NO_RELOC))
8945 {
8946 /* Negate the condition, and branch past an
8947 unconditional jump. */
8948 opcode[0] ^= 1;
8949 opcode[1] = 3;
8950 /* Insert an unconditional jump. */
8951 opcode[2] = 0xe9;
8952 /* We added two extra opcode bytes, and have a two byte
8953 offset. */
8954 fragP->fr_fix += 2 + 2;
8955 fix_new (fragP, old_fr_fix + 2, 2,
8956 fragP->fr_symbol,
8957 fragP->fr_offset, 1,
8958 reloc_type);
8959 break;
8960 }
8961 /* Fall through. */
8962
8963 case COND_JUMP:
8964 if (no_cond_jump_promotion && fragP->fr_var == NO_RELOC)
8965 {
8966 fixS *fixP;
8967
8968 fragP->fr_fix += 1;
8969 fixP = fix_new (fragP, old_fr_fix, 1,
8970 fragP->fr_symbol,
8971 fragP->fr_offset, 1,
8972 BFD_RELOC_8_PCREL);
8973 fixP->fx_signed = 1;
8974 break;
8975 }
8976
8977 /* This changes the byte-displacement jump 0x7N
8978 to the (d)word-displacement jump 0x0f,0x8N. */
8979 opcode[1] = opcode[0] + 0x10;
8980 opcode[0] = TWO_BYTE_OPCODE_ESCAPE;
8981 /* We've added an opcode byte. */
8982 fragP->fr_fix += 1 + size;
8983 fix_new (fragP, old_fr_fix + 1, size,
8984 fragP->fr_symbol,
8985 fragP->fr_offset, 1,
8986 reloc_type);
8987 break;
8988
8989 default:
8990 BAD_CASE (fragP->fr_subtype);
8991 break;
8992 }
8993 frag_wane (fragP);
8994 return fragP->fr_fix - old_fr_fix;
8995 }
8996
8997 /* Guess size depending on current relax state. Initially the relax
8998 state will correspond to a short jump and we return 1, because
8999 the variable part of the frag (the branch offset) is one byte
9000 long. However, we can relax a section more than once and in that
9001 case we must either set fr_subtype back to the unrelaxed state,
9002 or return the value for the appropriate branch. */
9003 return md_relax_table[fragP->fr_subtype].rlx_length;
9004 }
9005
9006 /* Called after relax() is finished.
9007
9008 In: Address of frag.
9009 fr_type == rs_machine_dependent.
9010 fr_subtype is what the address relaxed to.
9011
9012 Out: Any fixSs and constants are set up.
9013 Caller will turn frag into a ".space 0". */
9014
9015 void
9016 md_convert_frag (bfd *abfd ATTRIBUTE_UNUSED, segT sec ATTRIBUTE_UNUSED,
9017 fragS *fragP)
9018 {
9019 unsigned char *opcode;
9020 unsigned char *where_to_put_displacement = NULL;
9021 offsetT target_address;
9022 offsetT opcode_address;
9023 unsigned int extension = 0;
9024 offsetT displacement_from_opcode_start;
9025
9026 opcode = (unsigned char *) fragP->fr_opcode;
9027
9028 /* Address we want to reach in file space. */
9029 target_address = S_GET_VALUE (fragP->fr_symbol) + fragP->fr_offset;
9030
9031 /* Address opcode resides at in file space. */
9032 opcode_address = fragP->fr_address + fragP->fr_fix;
9033
9034 /* Displacement from opcode start to fill into instruction. */
9035 displacement_from_opcode_start = target_address - opcode_address;
9036
9037 if ((fragP->fr_subtype & BIG) == 0)
9038 {
9039 /* Don't have to change opcode. */
9040 extension = 1; /* 1 opcode + 1 displacement */
9041 where_to_put_displacement = &opcode[1];
9042 }
9043 else
9044 {
9045 if (no_cond_jump_promotion
9046 && TYPE_FROM_RELAX_STATE (fragP->fr_subtype) != UNCOND_JUMP)
9047 as_warn_where (fragP->fr_file, fragP->fr_line,
9048 _("long jump required"));
9049
9050 switch (fragP->fr_subtype)
9051 {
9052 case ENCODE_RELAX_STATE (UNCOND_JUMP, BIG):
9053 extension = 4; /* 1 opcode + 4 displacement */
9054 opcode[0] = 0xe9;
9055 where_to_put_displacement = &opcode[1];
9056 break;
9057
9058 case ENCODE_RELAX_STATE (UNCOND_JUMP, BIG16):
9059 extension = 2; /* 1 opcode + 2 displacement */
9060 opcode[0] = 0xe9;
9061 where_to_put_displacement = &opcode[1];
9062 break;
9063
9064 case ENCODE_RELAX_STATE (COND_JUMP, BIG):
9065 case ENCODE_RELAX_STATE (COND_JUMP86, BIG):
9066 extension = 5; /* 2 opcode + 4 displacement */
9067 opcode[1] = opcode[0] + 0x10;
9068 opcode[0] = TWO_BYTE_OPCODE_ESCAPE;
9069 where_to_put_displacement = &opcode[2];
9070 break;
9071
9072 case ENCODE_RELAX_STATE (COND_JUMP, BIG16):
9073 extension = 3; /* 2 opcode + 2 displacement */
9074 opcode[1] = opcode[0] + 0x10;
9075 opcode[0] = TWO_BYTE_OPCODE_ESCAPE;
9076 where_to_put_displacement = &opcode[2];
9077 break;
9078
9079 case ENCODE_RELAX_STATE (COND_JUMP86, BIG16):
9080 extension = 4;
9081 opcode[0] ^= 1;
9082 opcode[1] = 3;
9083 opcode[2] = 0xe9;
9084 where_to_put_displacement = &opcode[3];
9085 break;
9086
9087 default:
9088 BAD_CASE (fragP->fr_subtype);
9089 break;
9090 }
9091 }
9092
9093 /* If size if less then four we are sure that the operand fits,
9094 but if it's 4, then it could be that the displacement is larger
9095 then -/+ 2GB. */
9096 if (DISP_SIZE_FROM_RELAX_STATE (fragP->fr_subtype) == 4
9097 && object_64bit
9098 && ((addressT) (displacement_from_opcode_start - extension
9099 + ((addressT) 1 << 31))
9100 > (((addressT) 2 << 31) - 1)))
9101 {
9102 as_bad_where (fragP->fr_file, fragP->fr_line,
9103 _("jump target out of range"));
9104 /* Make us emit 0. */
9105 displacement_from_opcode_start = extension;
9106 }
9107 /* Now put displacement after opcode. */
9108 md_number_to_chars ((char *) where_to_put_displacement,
9109 (valueT) (displacement_from_opcode_start - extension),
9110 DISP_SIZE_FROM_RELAX_STATE (fragP->fr_subtype));
9111 fragP->fr_fix += extension;
9112 }
9113 \f
9114 /* Apply a fixup (fixP) to segment data, once it has been determined
9115 by our caller that we have all the info we need to fix it up.
9116
9117 Parameter valP is the pointer to the value of the bits.
9118
9119 On the 386, immediates, displacements, and data pointers are all in
9120 the same (little-endian) format, so we don't need to care about which
9121 we are handling. */
9122
9123 void
9124 md_apply_fix (fixS *fixP, valueT *valP, segT seg ATTRIBUTE_UNUSED)
9125 {
9126 char *p = fixP->fx_where + fixP->fx_frag->fr_literal;
9127 valueT value = *valP;
9128
9129 #if !defined (TE_Mach)
9130 if (fixP->fx_pcrel)
9131 {
9132 switch (fixP->fx_r_type)
9133 {
9134 default:
9135 break;
9136
9137 case BFD_RELOC_64:
9138 fixP->fx_r_type = BFD_RELOC_64_PCREL;
9139 break;
9140 case BFD_RELOC_32:
9141 case BFD_RELOC_X86_64_32S:
9142 fixP->fx_r_type = BFD_RELOC_32_PCREL;
9143 break;
9144 case BFD_RELOC_16:
9145 fixP->fx_r_type = BFD_RELOC_16_PCREL;
9146 break;
9147 case BFD_RELOC_8:
9148 fixP->fx_r_type = BFD_RELOC_8_PCREL;
9149 break;
9150 }
9151 }
9152
9153 if (fixP->fx_addsy != NULL
9154 && (fixP->fx_r_type == BFD_RELOC_32_PCREL
9155 || fixP->fx_r_type == BFD_RELOC_64_PCREL
9156 || fixP->fx_r_type == BFD_RELOC_16_PCREL
9157 || fixP->fx_r_type == BFD_RELOC_8_PCREL
9158 || fixP->fx_r_type == BFD_RELOC_X86_64_PC32_BND)
9159 && !use_rela_relocations)
9160 {
9161 /* This is a hack. There should be a better way to handle this.
9162 This covers for the fact that bfd_install_relocation will
9163 subtract the current location (for partial_inplace, PC relative
9164 relocations); see more below. */
9165 #ifndef OBJ_AOUT
9166 if (IS_ELF
9167 #ifdef TE_PE
9168 || OUTPUT_FLAVOR == bfd_target_coff_flavour
9169 #endif
9170 )
9171 value += fixP->fx_where + fixP->fx_frag->fr_address;
9172 #endif
9173 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
9174 if (IS_ELF)
9175 {
9176 segT sym_seg = S_GET_SEGMENT (fixP->fx_addsy);
9177
9178 if ((sym_seg == seg
9179 || (symbol_section_p (fixP->fx_addsy)
9180 && sym_seg != absolute_section))
9181 && !generic_force_reloc (fixP))
9182 {
9183 /* Yes, we add the values in twice. This is because
9184 bfd_install_relocation subtracts them out again. I think
9185 bfd_install_relocation is broken, but I don't dare change
9186 it. FIXME. */
9187 value += fixP->fx_where + fixP->fx_frag->fr_address;
9188 }
9189 }
9190 #endif
9191 #if defined (OBJ_COFF) && defined (TE_PE)
9192 /* For some reason, the PE format does not store a
9193 section address offset for a PC relative symbol. */
9194 if (S_GET_SEGMENT (fixP->fx_addsy) != seg
9195 || S_IS_WEAK (fixP->fx_addsy))
9196 value += md_pcrel_from (fixP);
9197 #endif
9198 }
9199 #if defined (OBJ_COFF) && defined (TE_PE)
9200 if (fixP->fx_addsy != NULL
9201 && S_IS_WEAK (fixP->fx_addsy)
9202 /* PR 16858: Do not modify weak function references. */
9203 && ! fixP->fx_pcrel)
9204 {
9205 #if !defined (TE_PEP)
9206 /* For x86 PE weak function symbols are neither PC-relative
9207 nor do they set S_IS_FUNCTION. So the only reliable way
9208 to detect them is to check the flags of their containing
9209 section. */
9210 if (S_GET_SEGMENT (fixP->fx_addsy) != NULL
9211 && S_GET_SEGMENT (fixP->fx_addsy)->flags & SEC_CODE)
9212 ;
9213 else
9214 #endif
9215 value -= S_GET_VALUE (fixP->fx_addsy);
9216 }
9217 #endif
9218
9219 /* Fix a few things - the dynamic linker expects certain values here,
9220 and we must not disappoint it. */
9221 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
9222 if (IS_ELF && fixP->fx_addsy)
9223 switch (fixP->fx_r_type)
9224 {
9225 case BFD_RELOC_386_PLT32:
9226 case BFD_RELOC_X86_64_PLT32:
9227 case BFD_RELOC_X86_64_PLT32_BND:
9228 /* Make the jump instruction point to the address of the operand. At
9229 runtime we merely add the offset to the actual PLT entry. */
9230 value = -4;
9231 break;
9232
9233 case BFD_RELOC_386_TLS_GD:
9234 case BFD_RELOC_386_TLS_LDM:
9235 case BFD_RELOC_386_TLS_IE_32:
9236 case BFD_RELOC_386_TLS_IE:
9237 case BFD_RELOC_386_TLS_GOTIE:
9238 case BFD_RELOC_386_TLS_GOTDESC:
9239 case BFD_RELOC_X86_64_TLSGD:
9240 case BFD_RELOC_X86_64_TLSLD:
9241 case BFD_RELOC_X86_64_GOTTPOFF:
9242 case BFD_RELOC_X86_64_GOTPC32_TLSDESC:
9243 value = 0; /* Fully resolved at runtime. No addend. */
9244 /* Fallthrough */
9245 case BFD_RELOC_386_TLS_LE:
9246 case BFD_RELOC_386_TLS_LDO_32:
9247 case BFD_RELOC_386_TLS_LE_32:
9248 case BFD_RELOC_X86_64_DTPOFF32:
9249 case BFD_RELOC_X86_64_DTPOFF64:
9250 case BFD_RELOC_X86_64_TPOFF32:
9251 case BFD_RELOC_X86_64_TPOFF64:
9252 S_SET_THREAD_LOCAL (fixP->fx_addsy);
9253 break;
9254
9255 case BFD_RELOC_386_TLS_DESC_CALL:
9256 case BFD_RELOC_X86_64_TLSDESC_CALL:
9257 value = 0; /* Fully resolved at runtime. No addend. */
9258 S_SET_THREAD_LOCAL (fixP->fx_addsy);
9259 fixP->fx_done = 0;
9260 return;
9261
9262 case BFD_RELOC_386_GOT32:
9263 case BFD_RELOC_X86_64_GOT32:
9264 value = 0; /* Fully resolved at runtime. No addend. */
9265 break;
9266
9267 case BFD_RELOC_VTABLE_INHERIT:
9268 case BFD_RELOC_VTABLE_ENTRY:
9269 fixP->fx_done = 0;
9270 return;
9271
9272 default:
9273 break;
9274 }
9275 #endif /* defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) */
9276 *valP = value;
9277 #endif /* !defined (TE_Mach) */
9278
9279 /* Are we finished with this relocation now? */
9280 if (fixP->fx_addsy == NULL)
9281 fixP->fx_done = 1;
9282 #if defined (OBJ_COFF) && defined (TE_PE)
9283 else if (fixP->fx_addsy != NULL && S_IS_WEAK (fixP->fx_addsy))
9284 {
9285 fixP->fx_done = 0;
9286 /* Remember value for tc_gen_reloc. */
9287 fixP->fx_addnumber = value;
9288 /* Clear out the frag for now. */
9289 value = 0;
9290 }
9291 #endif
9292 else if (use_rela_relocations)
9293 {
9294 fixP->fx_no_overflow = 1;
9295 /* Remember value for tc_gen_reloc. */
9296 fixP->fx_addnumber = value;
9297 value = 0;
9298 }
9299
9300 md_number_to_chars (p, value, fixP->fx_size);
9301 }
9302 \f
9303 char *
9304 md_atof (int type, char *litP, int *sizeP)
9305 {
9306 /* This outputs the LITTLENUMs in REVERSE order;
9307 in accord with the bigendian 386. */
9308 return ieee_md_atof (type, litP, sizeP, FALSE);
9309 }
9310 \f
9311 static char output_invalid_buf[sizeof (unsigned char) * 2 + 6];
9312
9313 static char *
9314 output_invalid (int c)
9315 {
9316 if (ISPRINT (c))
9317 snprintf (output_invalid_buf, sizeof (output_invalid_buf),
9318 "'%c'", c);
9319 else
9320 snprintf (output_invalid_buf, sizeof (output_invalid_buf),
9321 "(0x%x)", (unsigned char) c);
9322 return output_invalid_buf;
9323 }
9324
9325 /* REG_STRING starts *before* REGISTER_PREFIX. */
9326
9327 static const reg_entry *
9328 parse_real_register (char *reg_string, char **end_op)
9329 {
9330 char *s = reg_string;
9331 char *p;
9332 char reg_name_given[MAX_REG_NAME_SIZE + 1];
9333 const reg_entry *r;
9334
9335 /* Skip possible REGISTER_PREFIX and possible whitespace. */
9336 if (*s == REGISTER_PREFIX)
9337 ++s;
9338
9339 if (is_space_char (*s))
9340 ++s;
9341
9342 p = reg_name_given;
9343 while ((*p++ = register_chars[(unsigned char) *s]) != '\0')
9344 {
9345 if (p >= reg_name_given + MAX_REG_NAME_SIZE)
9346 return (const reg_entry *) NULL;
9347 s++;
9348 }
9349
9350 /* For naked regs, make sure that we are not dealing with an identifier.
9351 This prevents confusing an identifier like `eax_var' with register
9352 `eax'. */
9353 if (allow_naked_reg && identifier_chars[(unsigned char) *s])
9354 return (const reg_entry *) NULL;
9355
9356 *end_op = s;
9357
9358 r = (const reg_entry *) hash_find (reg_hash, reg_name_given);
9359
9360 /* Handle floating point regs, allowing spaces in the (i) part. */
9361 if (r == i386_regtab /* %st is first entry of table */)
9362 {
9363 if (is_space_char (*s))
9364 ++s;
9365 if (*s == '(')
9366 {
9367 ++s;
9368 if (is_space_char (*s))
9369 ++s;
9370 if (*s >= '0' && *s <= '7')
9371 {
9372 int fpr = *s - '0';
9373 ++s;
9374 if (is_space_char (*s))
9375 ++s;
9376 if (*s == ')')
9377 {
9378 *end_op = s + 1;
9379 r = (const reg_entry *) hash_find (reg_hash, "st(0)");
9380 know (r);
9381 return r + fpr;
9382 }
9383 }
9384 /* We have "%st(" then garbage. */
9385 return (const reg_entry *) NULL;
9386 }
9387 }
9388
9389 if (r == NULL || allow_pseudo_reg)
9390 return r;
9391
9392 if (operand_type_all_zero (&r->reg_type))
9393 return (const reg_entry *) NULL;
9394
9395 if ((r->reg_type.bitfield.reg32
9396 || r->reg_type.bitfield.sreg3
9397 || r->reg_type.bitfield.control
9398 || r->reg_type.bitfield.debug
9399 || r->reg_type.bitfield.test)
9400 && !cpu_arch_flags.bitfield.cpui386)
9401 return (const reg_entry *) NULL;
9402
9403 if (r->reg_type.bitfield.floatreg
9404 && !cpu_arch_flags.bitfield.cpu8087
9405 && !cpu_arch_flags.bitfield.cpu287
9406 && !cpu_arch_flags.bitfield.cpu387)
9407 return (const reg_entry *) NULL;
9408
9409 if (r->reg_type.bitfield.regmmx && !cpu_arch_flags.bitfield.cpummx)
9410 return (const reg_entry *) NULL;
9411
9412 if (r->reg_type.bitfield.regxmm && !cpu_arch_flags.bitfield.cpusse)
9413 return (const reg_entry *) NULL;
9414
9415 if (r->reg_type.bitfield.regymm && !cpu_arch_flags.bitfield.cpuavx)
9416 return (const reg_entry *) NULL;
9417
9418 if ((r->reg_type.bitfield.regzmm || r->reg_type.bitfield.regmask)
9419 && !cpu_arch_flags.bitfield.cpuavx512f)
9420 return (const reg_entry *) NULL;
9421
9422 /* Don't allow fake index register unless allow_index_reg isn't 0. */
9423 if (!allow_index_reg
9424 && (r->reg_num == RegEiz || r->reg_num == RegRiz))
9425 return (const reg_entry *) NULL;
9426
9427 /* Upper 16 vector register is only available with VREX in 64bit
9428 mode. */
9429 if ((r->reg_flags & RegVRex))
9430 {
9431 if (!cpu_arch_flags.bitfield.cpuvrex
9432 || flag_code != CODE_64BIT)
9433 return (const reg_entry *) NULL;
9434
9435 i.need_vrex = 1;
9436 }
9437
9438 if (((r->reg_flags & (RegRex64 | RegRex))
9439 || r->reg_type.bitfield.reg64)
9440 && (!cpu_arch_flags.bitfield.cpulm
9441 || !operand_type_equal (&r->reg_type, &control))
9442 && flag_code != CODE_64BIT)
9443 return (const reg_entry *) NULL;
9444
9445 if (r->reg_type.bitfield.sreg3 && r->reg_num == RegFlat && !intel_syntax)
9446 return (const reg_entry *) NULL;
9447
9448 return r;
9449 }
9450
9451 /* REG_STRING starts *before* REGISTER_PREFIX. */
9452
9453 static const reg_entry *
9454 parse_register (char *reg_string, char **end_op)
9455 {
9456 const reg_entry *r;
9457
9458 if (*reg_string == REGISTER_PREFIX || allow_naked_reg)
9459 r = parse_real_register (reg_string, end_op);
9460 else
9461 r = NULL;
9462 if (!r)
9463 {
9464 char *save = input_line_pointer;
9465 char c;
9466 symbolS *symbolP;
9467
9468 input_line_pointer = reg_string;
9469 c = get_symbol_end ();
9470 symbolP = symbol_find (reg_string);
9471 if (symbolP && S_GET_SEGMENT (symbolP) == reg_section)
9472 {
9473 const expressionS *e = symbol_get_value_expression (symbolP);
9474
9475 know (e->X_op == O_register);
9476 know (e->X_add_number >= 0
9477 && (valueT) e->X_add_number < i386_regtab_size);
9478 r = i386_regtab + e->X_add_number;
9479 if ((r->reg_flags & RegVRex))
9480 i.need_vrex = 1;
9481 *end_op = input_line_pointer;
9482 }
9483 *input_line_pointer = c;
9484 input_line_pointer = save;
9485 }
9486 return r;
9487 }
9488
9489 int
9490 i386_parse_name (char *name, expressionS *e, char *nextcharP)
9491 {
9492 const reg_entry *r;
9493 char *end = input_line_pointer;
9494
9495 *end = *nextcharP;
9496 r = parse_register (name, &input_line_pointer);
9497 if (r && end <= input_line_pointer)
9498 {
9499 *nextcharP = *input_line_pointer;
9500 *input_line_pointer = 0;
9501 e->X_op = O_register;
9502 e->X_add_number = r - i386_regtab;
9503 return 1;
9504 }
9505 input_line_pointer = end;
9506 *end = 0;
9507 return intel_syntax ? i386_intel_parse_name (name, e) : 0;
9508 }
9509
9510 void
9511 md_operand (expressionS *e)
9512 {
9513 char *end;
9514 const reg_entry *r;
9515
9516 switch (*input_line_pointer)
9517 {
9518 case REGISTER_PREFIX:
9519 r = parse_real_register (input_line_pointer, &end);
9520 if (r)
9521 {
9522 e->X_op = O_register;
9523 e->X_add_number = r - i386_regtab;
9524 input_line_pointer = end;
9525 }
9526 break;
9527
9528 case '[':
9529 gas_assert (intel_syntax);
9530 end = input_line_pointer++;
9531 expression (e);
9532 if (*input_line_pointer == ']')
9533 {
9534 ++input_line_pointer;
9535 e->X_op_symbol = make_expr_symbol (e);
9536 e->X_add_symbol = NULL;
9537 e->X_add_number = 0;
9538 e->X_op = O_index;
9539 }
9540 else
9541 {
9542 e->X_op = O_absent;
9543 input_line_pointer = end;
9544 }
9545 break;
9546 }
9547 }
9548
9549 \f
9550 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
9551 const char *md_shortopts = "kVQ:sqn";
9552 #else
9553 const char *md_shortopts = "qn";
9554 #endif
9555
9556 #define OPTION_32 (OPTION_MD_BASE + 0)
9557 #define OPTION_64 (OPTION_MD_BASE + 1)
9558 #define OPTION_DIVIDE (OPTION_MD_BASE + 2)
9559 #define OPTION_MARCH (OPTION_MD_BASE + 3)
9560 #define OPTION_MTUNE (OPTION_MD_BASE + 4)
9561 #define OPTION_MMNEMONIC (OPTION_MD_BASE + 5)
9562 #define OPTION_MSYNTAX (OPTION_MD_BASE + 6)
9563 #define OPTION_MINDEX_REG (OPTION_MD_BASE + 7)
9564 #define OPTION_MNAKED_REG (OPTION_MD_BASE + 8)
9565 #define OPTION_MOLD_GCC (OPTION_MD_BASE + 9)
9566 #define OPTION_MSSE2AVX (OPTION_MD_BASE + 10)
9567 #define OPTION_MSSE_CHECK (OPTION_MD_BASE + 11)
9568 #define OPTION_MOPERAND_CHECK (OPTION_MD_BASE + 12)
9569 #define OPTION_MAVXSCALAR (OPTION_MD_BASE + 13)
9570 #define OPTION_X32 (OPTION_MD_BASE + 14)
9571 #define OPTION_MADD_BND_PREFIX (OPTION_MD_BASE + 15)
9572 #define OPTION_MEVEXLIG (OPTION_MD_BASE + 16)
9573 #define OPTION_MEVEXWIG (OPTION_MD_BASE + 17)
9574 #define OPTION_MBIG_OBJ (OPTION_MD_BASE + 18)
9575 #define OPTION_OMIT_LOCK_PREFIX (OPTION_MD_BASE + 19)
9576 #define OPTION_MEVEXRCIG (OPTION_MD_BASE + 20)
9577
9578 struct option md_longopts[] =
9579 {
9580 {"32", no_argument, NULL, OPTION_32},
9581 #if (defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) \
9582 || defined (TE_PE) || defined (TE_PEP) || defined (OBJ_MACH_O))
9583 {"64", no_argument, NULL, OPTION_64},
9584 #endif
9585 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
9586 {"x32", no_argument, NULL, OPTION_X32},
9587 #endif
9588 {"divide", no_argument, NULL, OPTION_DIVIDE},
9589 {"march", required_argument, NULL, OPTION_MARCH},
9590 {"mtune", required_argument, NULL, OPTION_MTUNE},
9591 {"mmnemonic", required_argument, NULL, OPTION_MMNEMONIC},
9592 {"msyntax", required_argument, NULL, OPTION_MSYNTAX},
9593 {"mindex-reg", no_argument, NULL, OPTION_MINDEX_REG},
9594 {"mnaked-reg", no_argument, NULL, OPTION_MNAKED_REG},
9595 {"mold-gcc", no_argument, NULL, OPTION_MOLD_GCC},
9596 {"msse2avx", no_argument, NULL, OPTION_MSSE2AVX},
9597 {"msse-check", required_argument, NULL, OPTION_MSSE_CHECK},
9598 {"moperand-check", required_argument, NULL, OPTION_MOPERAND_CHECK},
9599 {"mavxscalar", required_argument, NULL, OPTION_MAVXSCALAR},
9600 {"madd-bnd-prefix", no_argument, NULL, OPTION_MADD_BND_PREFIX},
9601 {"mevexlig", required_argument, NULL, OPTION_MEVEXLIG},
9602 {"mevexwig", required_argument, NULL, OPTION_MEVEXWIG},
9603 # if defined (TE_PE) || defined (TE_PEP)
9604 {"mbig-obj", no_argument, NULL, OPTION_MBIG_OBJ},
9605 #endif
9606 {"momit-lock-prefix", required_argument, NULL, OPTION_OMIT_LOCK_PREFIX},
9607 {"mevexrcig", required_argument, NULL, OPTION_MEVEXRCIG},
9608 {NULL, no_argument, NULL, 0}
9609 };
9610 size_t md_longopts_size = sizeof (md_longopts);
9611
9612 int
9613 md_parse_option (int c, char *arg)
9614 {
9615 unsigned int j;
9616 char *arch, *next;
9617
9618 switch (c)
9619 {
9620 case 'n':
9621 optimize_align_code = 0;
9622 break;
9623
9624 case 'q':
9625 quiet_warnings = 1;
9626 break;
9627
9628 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
9629 /* -Qy, -Qn: SVR4 arguments controlling whether a .comment section
9630 should be emitted or not. FIXME: Not implemented. */
9631 case 'Q':
9632 break;
9633
9634 /* -V: SVR4 argument to print version ID. */
9635 case 'V':
9636 print_version_id ();
9637 break;
9638
9639 /* -k: Ignore for FreeBSD compatibility. */
9640 case 'k':
9641 break;
9642
9643 case 's':
9644 /* -s: On i386 Solaris, this tells the native assembler to use
9645 .stab instead of .stab.excl. We always use .stab anyhow. */
9646 break;
9647 #endif
9648 #if (defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) \
9649 || defined (TE_PE) || defined (TE_PEP) || defined (OBJ_MACH_O))
9650 case OPTION_64:
9651 {
9652 const char **list, **l;
9653
9654 list = bfd_target_list ();
9655 for (l = list; *l != NULL; l++)
9656 if (CONST_STRNEQ (*l, "elf64-x86-64")
9657 || strcmp (*l, "coff-x86-64") == 0
9658 || strcmp (*l, "pe-x86-64") == 0
9659 || strcmp (*l, "pei-x86-64") == 0
9660 || strcmp (*l, "mach-o-x86-64") == 0)
9661 {
9662 default_arch = "x86_64";
9663 break;
9664 }
9665 if (*l == NULL)
9666 as_fatal (_("no compiled in support for x86_64"));
9667 free (list);
9668 }
9669 break;
9670 #endif
9671
9672 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
9673 case OPTION_X32:
9674 if (IS_ELF)
9675 {
9676 const char **list, **l;
9677
9678 list = bfd_target_list ();
9679 for (l = list; *l != NULL; l++)
9680 if (CONST_STRNEQ (*l, "elf32-x86-64"))
9681 {
9682 default_arch = "x86_64:32";
9683 break;
9684 }
9685 if (*l == NULL)
9686 as_fatal (_("no compiled in support for 32bit x86_64"));
9687 free (list);
9688 }
9689 else
9690 as_fatal (_("32bit x86_64 is only supported for ELF"));
9691 break;
9692 #endif
9693
9694 case OPTION_32:
9695 default_arch = "i386";
9696 break;
9697
9698 case OPTION_DIVIDE:
9699 #ifdef SVR4_COMMENT_CHARS
9700 {
9701 char *n, *t;
9702 const char *s;
9703
9704 n = (char *) xmalloc (strlen (i386_comment_chars) + 1);
9705 t = n;
9706 for (s = i386_comment_chars; *s != '\0'; s++)
9707 if (*s != '/')
9708 *t++ = *s;
9709 *t = '\0';
9710 i386_comment_chars = n;
9711 }
9712 #endif
9713 break;
9714
9715 case OPTION_MARCH:
9716 arch = xstrdup (arg);
9717 do
9718 {
9719 if (*arch == '.')
9720 as_fatal (_("invalid -march= option: `%s'"), arg);
9721 next = strchr (arch, '+');
9722 if (next)
9723 *next++ = '\0';
9724 for (j = 0; j < ARRAY_SIZE (cpu_arch); j++)
9725 {
9726 if (strcmp (arch, cpu_arch [j].name) == 0)
9727 {
9728 /* Processor. */
9729 if (! cpu_arch[j].flags.bitfield.cpui386)
9730 continue;
9731
9732 cpu_arch_name = cpu_arch[j].name;
9733 cpu_sub_arch_name = NULL;
9734 cpu_arch_flags = cpu_arch[j].flags;
9735 cpu_arch_isa = cpu_arch[j].type;
9736 cpu_arch_isa_flags = cpu_arch[j].flags;
9737 if (!cpu_arch_tune_set)
9738 {
9739 cpu_arch_tune = cpu_arch_isa;
9740 cpu_arch_tune_flags = cpu_arch_isa_flags;
9741 }
9742 break;
9743 }
9744 else if (*cpu_arch [j].name == '.'
9745 && strcmp (arch, cpu_arch [j].name + 1) == 0)
9746 {
9747 /* ISA entension. */
9748 i386_cpu_flags flags;
9749
9750 if (!cpu_arch[j].negated)
9751 flags = cpu_flags_or (cpu_arch_flags,
9752 cpu_arch[j].flags);
9753 else
9754 flags = cpu_flags_and_not (cpu_arch_flags,
9755 cpu_arch[j].flags);
9756 if (!cpu_flags_equal (&flags, &cpu_arch_flags))
9757 {
9758 if (cpu_sub_arch_name)
9759 {
9760 char *name = cpu_sub_arch_name;
9761 cpu_sub_arch_name = concat (name,
9762 cpu_arch[j].name,
9763 (const char *) NULL);
9764 free (name);
9765 }
9766 else
9767 cpu_sub_arch_name = xstrdup (cpu_arch[j].name);
9768 cpu_arch_flags = flags;
9769 cpu_arch_isa_flags = flags;
9770 }
9771 break;
9772 }
9773 }
9774
9775 if (j >= ARRAY_SIZE (cpu_arch))
9776 as_fatal (_("invalid -march= option: `%s'"), arg);
9777
9778 arch = next;
9779 }
9780 while (next != NULL );
9781 break;
9782
9783 case OPTION_MTUNE:
9784 if (*arg == '.')
9785 as_fatal (_("invalid -mtune= option: `%s'"), arg);
9786 for (j = 0; j < ARRAY_SIZE (cpu_arch); j++)
9787 {
9788 if (strcmp (arg, cpu_arch [j].name) == 0)
9789 {
9790 cpu_arch_tune_set = 1;
9791 cpu_arch_tune = cpu_arch [j].type;
9792 cpu_arch_tune_flags = cpu_arch[j].flags;
9793 break;
9794 }
9795 }
9796 if (j >= ARRAY_SIZE (cpu_arch))
9797 as_fatal (_("invalid -mtune= option: `%s'"), arg);
9798 break;
9799
9800 case OPTION_MMNEMONIC:
9801 if (strcasecmp (arg, "att") == 0)
9802 intel_mnemonic = 0;
9803 else if (strcasecmp (arg, "intel") == 0)
9804 intel_mnemonic = 1;
9805 else
9806 as_fatal (_("invalid -mmnemonic= option: `%s'"), arg);
9807 break;
9808
9809 case OPTION_MSYNTAX:
9810 if (strcasecmp (arg, "att") == 0)
9811 intel_syntax = 0;
9812 else if (strcasecmp (arg, "intel") == 0)
9813 intel_syntax = 1;
9814 else
9815 as_fatal (_("invalid -msyntax= option: `%s'"), arg);
9816 break;
9817
9818 case OPTION_MINDEX_REG:
9819 allow_index_reg = 1;
9820 break;
9821
9822 case OPTION_MNAKED_REG:
9823 allow_naked_reg = 1;
9824 break;
9825
9826 case OPTION_MOLD_GCC:
9827 old_gcc = 1;
9828 break;
9829
9830 case OPTION_MSSE2AVX:
9831 sse2avx = 1;
9832 break;
9833
9834 case OPTION_MSSE_CHECK:
9835 if (strcasecmp (arg, "error") == 0)
9836 sse_check = check_error;
9837 else if (strcasecmp (arg, "warning") == 0)
9838 sse_check = check_warning;
9839 else if (strcasecmp (arg, "none") == 0)
9840 sse_check = check_none;
9841 else
9842 as_fatal (_("invalid -msse-check= option: `%s'"), arg);
9843 break;
9844
9845 case OPTION_MOPERAND_CHECK:
9846 if (strcasecmp (arg, "error") == 0)
9847 operand_check = check_error;
9848 else if (strcasecmp (arg, "warning") == 0)
9849 operand_check = check_warning;
9850 else if (strcasecmp (arg, "none") == 0)
9851 operand_check = check_none;
9852 else
9853 as_fatal (_("invalid -moperand-check= option: `%s'"), arg);
9854 break;
9855
9856 case OPTION_MAVXSCALAR:
9857 if (strcasecmp (arg, "128") == 0)
9858 avxscalar = vex128;
9859 else if (strcasecmp (arg, "256") == 0)
9860 avxscalar = vex256;
9861 else
9862 as_fatal (_("invalid -mavxscalar= option: `%s'"), arg);
9863 break;
9864
9865 case OPTION_MADD_BND_PREFIX:
9866 add_bnd_prefix = 1;
9867 break;
9868
9869 case OPTION_MEVEXLIG:
9870 if (strcmp (arg, "128") == 0)
9871 evexlig = evexl128;
9872 else if (strcmp (arg, "256") == 0)
9873 evexlig = evexl256;
9874 else if (strcmp (arg, "512") == 0)
9875 evexlig = evexl512;
9876 else
9877 as_fatal (_("invalid -mevexlig= option: `%s'"), arg);
9878 break;
9879
9880 case OPTION_MEVEXRCIG:
9881 if (strcmp (arg, "rne") == 0)
9882 evexrcig = rne;
9883 else if (strcmp (arg, "rd") == 0)
9884 evexrcig = rd;
9885 else if (strcmp (arg, "ru") == 0)
9886 evexrcig = ru;
9887 else if (strcmp (arg, "rz") == 0)
9888 evexrcig = rz;
9889 else
9890 as_fatal (_("invalid -mevexrcig= option: `%s'"), arg);
9891 break;
9892
9893 case OPTION_MEVEXWIG:
9894 if (strcmp (arg, "0") == 0)
9895 evexwig = evexw0;
9896 else if (strcmp (arg, "1") == 0)
9897 evexwig = evexw1;
9898 else
9899 as_fatal (_("invalid -mevexwig= option: `%s'"), arg);
9900 break;
9901
9902 # if defined (TE_PE) || defined (TE_PEP)
9903 case OPTION_MBIG_OBJ:
9904 use_big_obj = 1;
9905 break;
9906 #endif
9907
9908 case OPTION_OMIT_LOCK_PREFIX:
9909 if (strcasecmp (arg, "yes") == 0)
9910 omit_lock_prefix = 1;
9911 else if (strcasecmp (arg, "no") == 0)
9912 omit_lock_prefix = 0;
9913 else
9914 as_fatal (_("invalid -momit-lock-prefix= option: `%s'"), arg);
9915 break;
9916
9917 default:
9918 return 0;
9919 }
9920 return 1;
9921 }
9922
9923 #define MESSAGE_TEMPLATE \
9924 " "
9925
9926 static void
9927 show_arch (FILE *stream, int ext, int check)
9928 {
9929 static char message[] = MESSAGE_TEMPLATE;
9930 char *start = message + 27;
9931 char *p;
9932 int size = sizeof (MESSAGE_TEMPLATE);
9933 int left;
9934 const char *name;
9935 int len;
9936 unsigned int j;
9937
9938 p = start;
9939 left = size - (start - message);
9940 for (j = 0; j < ARRAY_SIZE (cpu_arch); j++)
9941 {
9942 /* Should it be skipped? */
9943 if (cpu_arch [j].skip)
9944 continue;
9945
9946 name = cpu_arch [j].name;
9947 len = cpu_arch [j].len;
9948 if (*name == '.')
9949 {
9950 /* It is an extension. Skip if we aren't asked to show it. */
9951 if (ext)
9952 {
9953 name++;
9954 len--;
9955 }
9956 else
9957 continue;
9958 }
9959 else if (ext)
9960 {
9961 /* It is an processor. Skip if we show only extension. */
9962 continue;
9963 }
9964 else if (check && ! cpu_arch[j].flags.bitfield.cpui386)
9965 {
9966 /* It is an impossible processor - skip. */
9967 continue;
9968 }
9969
9970 /* Reserve 2 spaces for ", " or ",\0" */
9971 left -= len + 2;
9972
9973 /* Check if there is any room. */
9974 if (left >= 0)
9975 {
9976 if (p != start)
9977 {
9978 *p++ = ',';
9979 *p++ = ' ';
9980 }
9981 p = mempcpy (p, name, len);
9982 }
9983 else
9984 {
9985 /* Output the current message now and start a new one. */
9986 *p++ = ',';
9987 *p = '\0';
9988 fprintf (stream, "%s\n", message);
9989 p = start;
9990 left = size - (start - message) - len - 2;
9991
9992 gas_assert (left >= 0);
9993
9994 p = mempcpy (p, name, len);
9995 }
9996 }
9997
9998 *p = '\0';
9999 fprintf (stream, "%s\n", message);
10000 }
10001
10002 void
10003 md_show_usage (FILE *stream)
10004 {
10005 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
10006 fprintf (stream, _("\
10007 -Q ignored\n\
10008 -V print assembler version number\n\
10009 -k ignored\n"));
10010 #endif
10011 fprintf (stream, _("\
10012 -n Do not optimize code alignment\n\
10013 -q quieten some warnings\n"));
10014 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
10015 fprintf (stream, _("\
10016 -s ignored\n"));
10017 #endif
10018 #if (defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) \
10019 || defined (TE_PE) || defined (TE_PEP))
10020 fprintf (stream, _("\
10021 --32/--64/--x32 generate 32bit/64bit/x32 code\n"));
10022 #endif
10023 #ifdef SVR4_COMMENT_CHARS
10024 fprintf (stream, _("\
10025 --divide do not treat `/' as a comment character\n"));
10026 #else
10027 fprintf (stream, _("\
10028 --divide ignored\n"));
10029 #endif
10030 fprintf (stream, _("\
10031 -march=CPU[,+EXTENSION...]\n\
10032 generate code for CPU and EXTENSION, CPU is one of:\n"));
10033 show_arch (stream, 0, 1);
10034 fprintf (stream, _("\
10035 EXTENSION is combination of:\n"));
10036 show_arch (stream, 1, 0);
10037 fprintf (stream, _("\
10038 -mtune=CPU optimize for CPU, CPU is one of:\n"));
10039 show_arch (stream, 0, 0);
10040 fprintf (stream, _("\
10041 -msse2avx encode SSE instructions with VEX prefix\n"));
10042 fprintf (stream, _("\
10043 -msse-check=[none|error|warning]\n\
10044 check SSE instructions\n"));
10045 fprintf (stream, _("\
10046 -moperand-check=[none|error|warning]\n\
10047 check operand combinations for validity\n"));
10048 fprintf (stream, _("\
10049 -mavxscalar=[128|256] encode scalar AVX instructions with specific vector\n\
10050 length\n"));
10051 fprintf (stream, _("\
10052 -mevexlig=[128|256|512] encode scalar EVEX instructions with specific vector\n\
10053 length\n"));
10054 fprintf (stream, _("\
10055 -mevexwig=[0|1] encode EVEX instructions with specific EVEX.W value\n\
10056 for EVEX.W bit ignored instructions\n"));
10057 fprintf (stream, _("\
10058 -mevexrcig=[rne|rd|ru|rz]\n\
10059 encode EVEX instructions with specific EVEX.RC value\n\
10060 for SAE-only ignored instructions\n"));
10061 fprintf (stream, _("\
10062 -mmnemonic=[att|intel] use AT&T/Intel mnemonic\n"));
10063 fprintf (stream, _("\
10064 -msyntax=[att|intel] use AT&T/Intel syntax\n"));
10065 fprintf (stream, _("\
10066 -mindex-reg support pseudo index registers\n"));
10067 fprintf (stream, _("\
10068 -mnaked-reg don't require `%%' prefix for registers\n"));
10069 fprintf (stream, _("\
10070 -mold-gcc support old (<= 2.8.1) versions of gcc\n"));
10071 fprintf (stream, _("\
10072 -madd-bnd-prefix add BND prefix for all valid branches\n"));
10073 # if defined (TE_PE) || defined (TE_PEP)
10074 fprintf (stream, _("\
10075 -mbig-obj generate big object files\n"));
10076 #endif
10077 fprintf (stream, _("\
10078 -momit-lock-prefix=[no|yes]\n\
10079 strip all lock prefixes\n"));
10080 }
10081
10082 #if ((defined (OBJ_MAYBE_COFF) && defined (OBJ_MAYBE_AOUT)) \
10083 || defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF) \
10084 || defined (TE_PE) || defined (TE_PEP) || defined (OBJ_MACH_O))
10085
10086 /* Pick the target format to use. */
10087
10088 const char *
10089 i386_target_format (void)
10090 {
10091 if (!strncmp (default_arch, "x86_64", 6))
10092 {
10093 update_code_flag (CODE_64BIT, 1);
10094 if (default_arch[6] == '\0')
10095 x86_elf_abi = X86_64_ABI;
10096 else
10097 x86_elf_abi = X86_64_X32_ABI;
10098 }
10099 else if (!strcmp (default_arch, "i386"))
10100 update_code_flag (CODE_32BIT, 1);
10101 else
10102 as_fatal (_("unknown architecture"));
10103
10104 if (cpu_flags_all_zero (&cpu_arch_isa_flags))
10105 cpu_arch_isa_flags = cpu_arch[flag_code == CODE_64BIT].flags;
10106 if (cpu_flags_all_zero (&cpu_arch_tune_flags))
10107 cpu_arch_tune_flags = cpu_arch[flag_code == CODE_64BIT].flags;
10108
10109 switch (OUTPUT_FLAVOR)
10110 {
10111 #if defined (OBJ_MAYBE_AOUT) || defined (OBJ_AOUT)
10112 case bfd_target_aout_flavour:
10113 return AOUT_TARGET_FORMAT;
10114 #endif
10115 #if defined (OBJ_MAYBE_COFF) || defined (OBJ_COFF)
10116 # if defined (TE_PE) || defined (TE_PEP)
10117 case bfd_target_coff_flavour:
10118 if (flag_code == CODE_64BIT)
10119 return use_big_obj ? "pe-bigobj-x86-64" : "pe-x86-64";
10120 else
10121 return "pe-i386";
10122 # elif defined (TE_GO32)
10123 case bfd_target_coff_flavour:
10124 return "coff-go32";
10125 # else
10126 case bfd_target_coff_flavour:
10127 return "coff-i386";
10128 # endif
10129 #endif
10130 #if defined (OBJ_MAYBE_ELF) || defined (OBJ_ELF)
10131 case bfd_target_elf_flavour:
10132 {
10133 const char *format;
10134
10135 switch (x86_elf_abi)
10136 {
10137 default:
10138 format = ELF_TARGET_FORMAT;
10139 break;
10140 case X86_64_ABI:
10141 use_rela_relocations = 1;
10142 object_64bit = 1;
10143 format = ELF_TARGET_FORMAT64;
10144 break;
10145 case X86_64_X32_ABI:
10146 use_rela_relocations = 1;
10147 object_64bit = 1;
10148 disallow_64bit_reloc = 1;
10149 format = ELF_TARGET_FORMAT32;
10150 break;
10151 }
10152 if (cpu_arch_isa == PROCESSOR_L1OM)
10153 {
10154 if (x86_elf_abi != X86_64_ABI)
10155 as_fatal (_("Intel L1OM is 64bit only"));
10156 return ELF_TARGET_L1OM_FORMAT;
10157 }
10158 if (cpu_arch_isa == PROCESSOR_K1OM)
10159 {
10160 if (x86_elf_abi != X86_64_ABI)
10161 as_fatal (_("Intel K1OM is 64bit only"));
10162 return ELF_TARGET_K1OM_FORMAT;
10163 }
10164 else
10165 return format;
10166 }
10167 #endif
10168 #if defined (OBJ_MACH_O)
10169 case bfd_target_mach_o_flavour:
10170 if (flag_code == CODE_64BIT)
10171 {
10172 use_rela_relocations = 1;
10173 object_64bit = 1;
10174 return "mach-o-x86-64";
10175 }
10176 else
10177 return "mach-o-i386";
10178 #endif
10179 default:
10180 abort ();
10181 return NULL;
10182 }
10183 }
10184
10185 #endif /* OBJ_MAYBE_ more than one */
10186
10187 #if (defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF))
10188 void
10189 i386_elf_emit_arch_note (void)
10190 {
10191 if (IS_ELF && cpu_arch_name != NULL)
10192 {
10193 char *p;
10194 asection *seg = now_seg;
10195 subsegT subseg = now_subseg;
10196 Elf_Internal_Note i_note;
10197 Elf_External_Note e_note;
10198 asection *note_secp;
10199 int len;
10200
10201 /* Create the .note section. */
10202 note_secp = subseg_new (".note", 0);
10203 bfd_set_section_flags (stdoutput,
10204 note_secp,
10205 SEC_HAS_CONTENTS | SEC_READONLY);
10206
10207 /* Process the arch string. */
10208 len = strlen (cpu_arch_name);
10209
10210 i_note.namesz = len + 1;
10211 i_note.descsz = 0;
10212 i_note.type = NT_ARCH;
10213 p = frag_more (sizeof (e_note.namesz));
10214 md_number_to_chars (p, (valueT) i_note.namesz, sizeof (e_note.namesz));
10215 p = frag_more (sizeof (e_note.descsz));
10216 md_number_to_chars (p, (valueT) i_note.descsz, sizeof (e_note.descsz));
10217 p = frag_more (sizeof (e_note.type));
10218 md_number_to_chars (p, (valueT) i_note.type, sizeof (e_note.type));
10219 p = frag_more (len + 1);
10220 strcpy (p, cpu_arch_name);
10221
10222 frag_align (2, 0, 0);
10223
10224 subseg_set (seg, subseg);
10225 }
10226 }
10227 #endif
10228 \f
10229 symbolS *
10230 md_undefined_symbol (char *name)
10231 {
10232 if (name[0] == GLOBAL_OFFSET_TABLE_NAME[0]
10233 && name[1] == GLOBAL_OFFSET_TABLE_NAME[1]
10234 && name[2] == GLOBAL_OFFSET_TABLE_NAME[2]
10235 && strcmp (name, GLOBAL_OFFSET_TABLE_NAME) == 0)
10236 {
10237 if (!GOT_symbol)
10238 {
10239 if (symbol_find (name))
10240 as_bad (_("GOT already in symbol table"));
10241 GOT_symbol = symbol_new (name, undefined_section,
10242 (valueT) 0, &zero_address_frag);
10243 };
10244 return GOT_symbol;
10245 }
10246 return 0;
10247 }
10248
10249 /* Round up a section size to the appropriate boundary. */
10250
10251 valueT
10252 md_section_align (segT segment ATTRIBUTE_UNUSED, valueT size)
10253 {
10254 #if (defined (OBJ_AOUT) || defined (OBJ_MAYBE_AOUT))
10255 if (OUTPUT_FLAVOR == bfd_target_aout_flavour)
10256 {
10257 /* For a.out, force the section size to be aligned. If we don't do
10258 this, BFD will align it for us, but it will not write out the
10259 final bytes of the section. This may be a bug in BFD, but it is
10260 easier to fix it here since that is how the other a.out targets
10261 work. */
10262 int align;
10263
10264 align = bfd_get_section_alignment (stdoutput, segment);
10265 size = ((size + (1 << align) - 1) & ((valueT) -1 << align));
10266 }
10267 #endif
10268
10269 return size;
10270 }
10271
10272 /* On the i386, PC-relative offsets are relative to the start of the
10273 next instruction. That is, the address of the offset, plus its
10274 size, since the offset is always the last part of the insn. */
10275
10276 long
10277 md_pcrel_from (fixS *fixP)
10278 {
10279 return fixP->fx_size + fixP->fx_where + fixP->fx_frag->fr_address;
10280 }
10281
10282 #ifndef I386COFF
10283
10284 static void
10285 s_bss (int ignore ATTRIBUTE_UNUSED)
10286 {
10287 int temp;
10288
10289 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
10290 if (IS_ELF)
10291 obj_elf_section_change_hook ();
10292 #endif
10293 temp = get_absolute_expression ();
10294 subseg_set (bss_section, (subsegT) temp);
10295 demand_empty_rest_of_line ();
10296 }
10297
10298 #endif
10299
10300 void
10301 i386_validate_fix (fixS *fixp)
10302 {
10303 if (fixp->fx_subsy && fixp->fx_subsy == GOT_symbol)
10304 {
10305 if (fixp->fx_r_type == BFD_RELOC_32_PCREL)
10306 {
10307 if (!object_64bit)
10308 abort ();
10309 fixp->fx_r_type = BFD_RELOC_X86_64_GOTPCREL;
10310 }
10311 else
10312 {
10313 if (!object_64bit)
10314 fixp->fx_r_type = BFD_RELOC_386_GOTOFF;
10315 else
10316 fixp->fx_r_type = BFD_RELOC_X86_64_GOTOFF64;
10317 }
10318 fixp->fx_subsy = 0;
10319 }
10320 }
10321
10322 arelent *
10323 tc_gen_reloc (asection *section ATTRIBUTE_UNUSED, fixS *fixp)
10324 {
10325 arelent *rel;
10326 bfd_reloc_code_real_type code;
10327
10328 switch (fixp->fx_r_type)
10329 {
10330 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
10331 case BFD_RELOC_SIZE32:
10332 case BFD_RELOC_SIZE64:
10333 if (S_IS_DEFINED (fixp->fx_addsy)
10334 && !S_IS_EXTERNAL (fixp->fx_addsy))
10335 {
10336 /* Resolve size relocation against local symbol to size of
10337 the symbol plus addend. */
10338 valueT value = S_GET_SIZE (fixp->fx_addsy) + fixp->fx_offset;
10339 if (fixp->fx_r_type == BFD_RELOC_SIZE32
10340 && !fits_in_unsigned_long (value))
10341 as_bad_where (fixp->fx_file, fixp->fx_line,
10342 _("symbol size computation overflow"));
10343 fixp->fx_addsy = NULL;
10344 fixp->fx_subsy = NULL;
10345 md_apply_fix (fixp, (valueT *) &value, NULL);
10346 return NULL;
10347 }
10348 #endif
10349
10350 case BFD_RELOC_X86_64_PLT32:
10351 case BFD_RELOC_X86_64_PLT32_BND:
10352 case BFD_RELOC_X86_64_GOT32:
10353 case BFD_RELOC_X86_64_GOTPCREL:
10354 case BFD_RELOC_386_PLT32:
10355 case BFD_RELOC_386_GOT32:
10356 case BFD_RELOC_386_GOTOFF:
10357 case BFD_RELOC_386_GOTPC:
10358 case BFD_RELOC_386_TLS_GD:
10359 case BFD_RELOC_386_TLS_LDM:
10360 case BFD_RELOC_386_TLS_LDO_32:
10361 case BFD_RELOC_386_TLS_IE_32:
10362 case BFD_RELOC_386_TLS_IE:
10363 case BFD_RELOC_386_TLS_GOTIE:
10364 case BFD_RELOC_386_TLS_LE_32:
10365 case BFD_RELOC_386_TLS_LE:
10366 case BFD_RELOC_386_TLS_GOTDESC:
10367 case BFD_RELOC_386_TLS_DESC_CALL:
10368 case BFD_RELOC_X86_64_TLSGD:
10369 case BFD_RELOC_X86_64_TLSLD:
10370 case BFD_RELOC_X86_64_DTPOFF32:
10371 case BFD_RELOC_X86_64_DTPOFF64:
10372 case BFD_RELOC_X86_64_GOTTPOFF:
10373 case BFD_RELOC_X86_64_TPOFF32:
10374 case BFD_RELOC_X86_64_TPOFF64:
10375 case BFD_RELOC_X86_64_GOTOFF64:
10376 case BFD_RELOC_X86_64_GOTPC32:
10377 case BFD_RELOC_X86_64_GOT64:
10378 case BFD_RELOC_X86_64_GOTPCREL64:
10379 case BFD_RELOC_X86_64_GOTPC64:
10380 case BFD_RELOC_X86_64_GOTPLT64:
10381 case BFD_RELOC_X86_64_PLTOFF64:
10382 case BFD_RELOC_X86_64_GOTPC32_TLSDESC:
10383 case BFD_RELOC_X86_64_TLSDESC_CALL:
10384 case BFD_RELOC_RVA:
10385 case BFD_RELOC_VTABLE_ENTRY:
10386 case BFD_RELOC_VTABLE_INHERIT:
10387 #ifdef TE_PE
10388 case BFD_RELOC_32_SECREL:
10389 #endif
10390 code = fixp->fx_r_type;
10391 break;
10392 case BFD_RELOC_X86_64_32S:
10393 if (!fixp->fx_pcrel)
10394 {
10395 /* Don't turn BFD_RELOC_X86_64_32S into BFD_RELOC_32. */
10396 code = fixp->fx_r_type;
10397 break;
10398 }
10399 default:
10400 if (fixp->fx_pcrel)
10401 {
10402 switch (fixp->fx_size)
10403 {
10404 default:
10405 as_bad_where (fixp->fx_file, fixp->fx_line,
10406 _("can not do %d byte pc-relative relocation"),
10407 fixp->fx_size);
10408 code = BFD_RELOC_32_PCREL;
10409 break;
10410 case 1: code = BFD_RELOC_8_PCREL; break;
10411 case 2: code = BFD_RELOC_16_PCREL; break;
10412 case 4:
10413 code = (fixp->fx_r_type == BFD_RELOC_X86_64_PC32_BND
10414 ? fixp-> fx_r_type : BFD_RELOC_32_PCREL);
10415 break;
10416 #ifdef BFD64
10417 case 8: code = BFD_RELOC_64_PCREL; break;
10418 #endif
10419 }
10420 }
10421 else
10422 {
10423 switch (fixp->fx_size)
10424 {
10425 default:
10426 as_bad_where (fixp->fx_file, fixp->fx_line,
10427 _("can not do %d byte relocation"),
10428 fixp->fx_size);
10429 code = BFD_RELOC_32;
10430 break;
10431 case 1: code = BFD_RELOC_8; break;
10432 case 2: code = BFD_RELOC_16; break;
10433 case 4: code = BFD_RELOC_32; break;
10434 #ifdef BFD64
10435 case 8: code = BFD_RELOC_64; break;
10436 #endif
10437 }
10438 }
10439 break;
10440 }
10441
10442 if ((code == BFD_RELOC_32
10443 || code == BFD_RELOC_32_PCREL
10444 || code == BFD_RELOC_X86_64_32S)
10445 && GOT_symbol
10446 && fixp->fx_addsy == GOT_symbol)
10447 {
10448 if (!object_64bit)
10449 code = BFD_RELOC_386_GOTPC;
10450 else
10451 code = BFD_RELOC_X86_64_GOTPC32;
10452 }
10453 if ((code == BFD_RELOC_64 || code == BFD_RELOC_64_PCREL)
10454 && GOT_symbol
10455 && fixp->fx_addsy == GOT_symbol)
10456 {
10457 code = BFD_RELOC_X86_64_GOTPC64;
10458 }
10459
10460 rel = (arelent *) xmalloc (sizeof (arelent));
10461 rel->sym_ptr_ptr = (asymbol **) xmalloc (sizeof (asymbol *));
10462 *rel->sym_ptr_ptr = symbol_get_bfdsym (fixp->fx_addsy);
10463
10464 rel->address = fixp->fx_frag->fr_address + fixp->fx_where;
10465
10466 if (!use_rela_relocations)
10467 {
10468 /* HACK: Since i386 ELF uses Rel instead of Rela, encode the
10469 vtable entry to be used in the relocation's section offset. */
10470 if (fixp->fx_r_type == BFD_RELOC_VTABLE_ENTRY)
10471 rel->address = fixp->fx_offset;
10472 #if defined (OBJ_COFF) && defined (TE_PE)
10473 else if (fixp->fx_addsy && S_IS_WEAK (fixp->fx_addsy))
10474 rel->addend = fixp->fx_addnumber - (S_GET_VALUE (fixp->fx_addsy) * 2);
10475 else
10476 #endif
10477 rel->addend = 0;
10478 }
10479 /* Use the rela in 64bit mode. */
10480 else
10481 {
10482 if (disallow_64bit_reloc)
10483 switch (code)
10484 {
10485 case BFD_RELOC_X86_64_DTPOFF64:
10486 case BFD_RELOC_X86_64_TPOFF64:
10487 case BFD_RELOC_64_PCREL:
10488 case BFD_RELOC_X86_64_GOTOFF64:
10489 case BFD_RELOC_X86_64_GOT64:
10490 case BFD_RELOC_X86_64_GOTPCREL64:
10491 case BFD_RELOC_X86_64_GOTPC64:
10492 case BFD_RELOC_X86_64_GOTPLT64:
10493 case BFD_RELOC_X86_64_PLTOFF64:
10494 as_bad_where (fixp->fx_file, fixp->fx_line,
10495 _("cannot represent relocation type %s in x32 mode"),
10496 bfd_get_reloc_code_name (code));
10497 break;
10498 default:
10499 break;
10500 }
10501
10502 if (!fixp->fx_pcrel)
10503 rel->addend = fixp->fx_offset;
10504 else
10505 switch (code)
10506 {
10507 case BFD_RELOC_X86_64_PLT32:
10508 case BFD_RELOC_X86_64_PLT32_BND:
10509 case BFD_RELOC_X86_64_GOT32:
10510 case BFD_RELOC_X86_64_GOTPCREL:
10511 case BFD_RELOC_X86_64_TLSGD:
10512 case BFD_RELOC_X86_64_TLSLD:
10513 case BFD_RELOC_X86_64_GOTTPOFF:
10514 case BFD_RELOC_X86_64_GOTPC32_TLSDESC:
10515 case BFD_RELOC_X86_64_TLSDESC_CALL:
10516 rel->addend = fixp->fx_offset - fixp->fx_size;
10517 break;
10518 default:
10519 rel->addend = (section->vma
10520 - fixp->fx_size
10521 + fixp->fx_addnumber
10522 + md_pcrel_from (fixp));
10523 break;
10524 }
10525 }
10526
10527 rel->howto = bfd_reloc_type_lookup (stdoutput, code);
10528 if (rel->howto == NULL)
10529 {
10530 as_bad_where (fixp->fx_file, fixp->fx_line,
10531 _("cannot represent relocation type %s"),
10532 bfd_get_reloc_code_name (code));
10533 /* Set howto to a garbage value so that we can keep going. */
10534 rel->howto = bfd_reloc_type_lookup (stdoutput, BFD_RELOC_32);
10535 gas_assert (rel->howto != NULL);
10536 }
10537
10538 return rel;
10539 }
10540
10541 #include "tc-i386-intel.c"
10542
10543 void
10544 tc_x86_parse_to_dw2regnum (expressionS *exp)
10545 {
10546 int saved_naked_reg;
10547 char saved_register_dot;
10548
10549 saved_naked_reg = allow_naked_reg;
10550 allow_naked_reg = 1;
10551 saved_register_dot = register_chars['.'];
10552 register_chars['.'] = '.';
10553 allow_pseudo_reg = 1;
10554 expression_and_evaluate (exp);
10555 allow_pseudo_reg = 0;
10556 register_chars['.'] = saved_register_dot;
10557 allow_naked_reg = saved_naked_reg;
10558
10559 if (exp->X_op == O_register && exp->X_add_number >= 0)
10560 {
10561 if ((addressT) exp->X_add_number < i386_regtab_size)
10562 {
10563 exp->X_op = O_constant;
10564 exp->X_add_number = i386_regtab[exp->X_add_number]
10565 .dw2_regnum[flag_code >> 1];
10566 }
10567 else
10568 exp->X_op = O_illegal;
10569 }
10570 }
10571
10572 void
10573 tc_x86_frame_initial_instructions (void)
10574 {
10575 static unsigned int sp_regno[2];
10576
10577 if (!sp_regno[flag_code >> 1])
10578 {
10579 char *saved_input = input_line_pointer;
10580 char sp[][4] = {"esp", "rsp"};
10581 expressionS exp;
10582
10583 input_line_pointer = sp[flag_code >> 1];
10584 tc_x86_parse_to_dw2regnum (&exp);
10585 gas_assert (exp.X_op == O_constant);
10586 sp_regno[flag_code >> 1] = exp.X_add_number;
10587 input_line_pointer = saved_input;
10588 }
10589
10590 cfi_add_CFA_def_cfa (sp_regno[flag_code >> 1], -x86_cie_data_alignment);
10591 cfi_add_CFA_offset (x86_dwarf2_return_column, x86_cie_data_alignment);
10592 }
10593
10594 int
10595 x86_dwarf2_addr_size (void)
10596 {
10597 #if defined (OBJ_MAYBE_ELF) || defined (OBJ_ELF)
10598 if (x86_elf_abi == X86_64_X32_ABI)
10599 return 4;
10600 #endif
10601 return bfd_arch_bits_per_address (stdoutput) / 8;
10602 }
10603
10604 int
10605 i386_elf_section_type (const char *str, size_t len)
10606 {
10607 if (flag_code == CODE_64BIT
10608 && len == sizeof ("unwind") - 1
10609 && strncmp (str, "unwind", 6) == 0)
10610 return SHT_X86_64_UNWIND;
10611
10612 return -1;
10613 }
10614
10615 #ifdef TE_SOLARIS
10616 void
10617 i386_solaris_fix_up_eh_frame (segT sec)
10618 {
10619 if (flag_code == CODE_64BIT)
10620 elf_section_type (sec) = SHT_X86_64_UNWIND;
10621 }
10622 #endif
10623
10624 #ifdef TE_PE
10625 void
10626 tc_pe_dwarf2_emit_offset (symbolS *symbol, unsigned int size)
10627 {
10628 expressionS exp;
10629
10630 exp.X_op = O_secrel;
10631 exp.X_add_symbol = symbol;
10632 exp.X_add_number = 0;
10633 emit_expr (&exp, size);
10634 }
10635 #endif
10636
10637 #if defined (OBJ_ELF) || defined (OBJ_MAYBE_ELF)
10638 /* For ELF on x86-64, add support for SHF_X86_64_LARGE. */
10639
10640 bfd_vma
10641 x86_64_section_letter (int letter, char **ptr_msg)
10642 {
10643 if (flag_code == CODE_64BIT)
10644 {
10645 if (letter == 'l')
10646 return SHF_X86_64_LARGE;
10647
10648 *ptr_msg = _("bad .section directive: want a,l,w,x,M,S,G,T in string");
10649 }
10650 else
10651 *ptr_msg = _("bad .section directive: want a,w,x,M,S,G,T in string");
10652 return -1;
10653 }
10654
10655 bfd_vma
10656 x86_64_section_word (char *str, size_t len)
10657 {
10658 if (len == 5 && flag_code == CODE_64BIT && CONST_STRNEQ (str, "large"))
10659 return SHF_X86_64_LARGE;
10660
10661 return -1;
10662 }
10663
10664 static void
10665 handle_large_common (int small ATTRIBUTE_UNUSED)
10666 {
10667 if (flag_code != CODE_64BIT)
10668 {
10669 s_comm_internal (0, elf_common_parse);
10670 as_warn (_(".largecomm supported only in 64bit mode, producing .comm"));
10671 }
10672 else
10673 {
10674 static segT lbss_section;
10675 asection *saved_com_section_ptr = elf_com_section_ptr;
10676 asection *saved_bss_section = bss_section;
10677
10678 if (lbss_section == NULL)
10679 {
10680 flagword applicable;
10681 segT seg = now_seg;
10682 subsegT subseg = now_subseg;
10683
10684 /* The .lbss section is for local .largecomm symbols. */
10685 lbss_section = subseg_new (".lbss", 0);
10686 applicable = bfd_applicable_section_flags (stdoutput);
10687 bfd_set_section_flags (stdoutput, lbss_section,
10688 applicable & SEC_ALLOC);
10689 seg_info (lbss_section)->bss = 1;
10690
10691 subseg_set (seg, subseg);
10692 }
10693
10694 elf_com_section_ptr = &_bfd_elf_large_com_section;
10695 bss_section = lbss_section;
10696
10697 s_comm_internal (0, elf_common_parse);
10698
10699 elf_com_section_ptr = saved_com_section_ptr;
10700 bss_section = saved_bss_section;
10701 }
10702 }
10703 #endif /* OBJ_ELF || OBJ_MAYBE_ELF */