1 //===-- VOP3Instructions.td - Vector Instruction Defintions ---------------===//
3 // The LLVM Compiler Infrastructure
5 // This file is distributed under the University of Illinois Open Source
6 // License. See LICENSE.TXT for details.
8 //===----------------------------------------------------------------------===//
10 //===----------------------------------------------------------------------===//
12 //===----------------------------------------------------------------------===//
14 class getVOP3ModPat<VOPProfile P, SDPatternOperator node> {
15 dag src0 = !if(P.HasOMod,
16 (VOP3Mods0 P.Src0VT:$src0, i32:$src0_modifiers, i1:$clamp, i32:$omod),
17 (VOP3Mods0 P.Src0VT:$src0, i32:$src0_modifiers, i1:$clamp));
19 list<dag> ret3 = [(set P.DstVT:$vdst,
20 (DivergentFragOrOp<node, P>.ret (P.Src0VT src0),
21 (P.Src1VT (VOP3Mods P.Src1VT:$src1, i32:$src1_modifiers)),
22 (P.Src2VT (VOP3Mods P.Src2VT:$src2, i32:$src2_modifiers))))];
24 list<dag> ret2 = [(set P.DstVT:$vdst,
25 (DivergentFragOrOp<node, P>.ret (P.Src0VT src0),
26 (P.Src1VT (VOP3Mods P.Src1VT:$src1, i32:$src1_modifiers))))];
28 list<dag> ret1 = [(set P.DstVT:$vdst,
29 (DivergentFragOrOp<node, P>.ret (P.Src0VT src0)))];
31 list<dag> ret = !if(!eq(P.NumSrcArgs, 3), ret3,
32 !if(!eq(P.NumSrcArgs, 2), ret2,
36 class getVOP3PModPat<VOPProfile P, SDPatternOperator node> {
37 list<dag> ret3 = [(set P.DstVT:$vdst,
38 (DivergentFragOrOp<node, P>.ret (P.Src0VT !if(P.HasClamp, (VOP3PMods0 P.Src0VT:$src0, i32:$src0_modifiers, i1:$clamp),
39 (VOP3PMods P.Src0VT:$src0, i32:$src0_modifiers))),
40 (P.Src1VT (VOP3PMods P.Src1VT:$src1, i32:$src1_modifiers)),
41 (P.Src2VT (VOP3PMods P.Src2VT:$src2, i32:$src2_modifiers))))];
43 list<dag> ret2 = [(set P.DstVT:$vdst,
44 (DivergentFragOrOp<node, P>.ret !if(P.HasClamp, (P.Src0VT (VOP3PMods0 P.Src0VT:$src0, i32:$src0_modifiers, i1:$clamp)),
45 (P.Src0VT (VOP3PMods P.Src0VT:$src0, i32:$src0_modifiers))),
46 (P.Src1VT (VOP3PMods P.Src1VT:$src1, i32:$src1_modifiers))))];
48 list<dag> ret1 = [(set P.DstVT:$vdst,
49 (DivergentFragOrOp<node, P>.ret (P.Src0VT (VOP3PMods0 P.Src0VT:$src0, i32:$src0_modifiers, i1:$clamp))))];
51 list<dag> ret = !if(!eq(P.NumSrcArgs, 3), ret3,
52 !if(!eq(P.NumSrcArgs, 2), ret2,
56 class getVOP3OpSelPat<VOPProfile P, SDPatternOperator node> {
57 list<dag> ret3 = [(set P.DstVT:$vdst,
58 (DivergentFragOrOp<node, P>.ret (P.Src0VT !if(P.HasClamp, (VOP3OpSel0 P.Src0VT:$src0, i32:$src0_modifiers, i1:$clamp),
59 (VOP3OpSel P.Src0VT:$src0, i32:$src0_modifiers))),
60 (P.Src1VT (VOP3OpSel P.Src1VT:$src1, i32:$src1_modifiers)),
61 (P.Src2VT (VOP3OpSel P.Src2VT:$src2, i32:$src2_modifiers))))];
63 list<dag> ret2 = [(set P.DstVT:$vdst,
64 (DivergentFragOrOp<node, P>.ret !if(P.HasClamp, (P.Src0VT (VOP3OpSel0 P.Src0VT:$src0, i32:$src0_modifiers, i1:$clamp)),
65 (P.Src0VT (VOP3OpSel P.Src0VT:$src0, i32:$src0_modifiers))),
66 (P.Src1VT (VOP3OpSel P.Src1VT:$src1, i32:$src1_modifiers))))];
68 list<dag> ret1 = [(set P.DstVT:$vdst,
69 (DivergentFragOrOp<node, P>.ret (P.Src0VT (VOP3OpSel0 P.Src0VT:$src0, i32:$src0_modifiers, i1:$clamp))))];
71 list<dag> ret = !if(!eq(P.NumSrcArgs, 3), ret3,
72 !if(!eq(P.NumSrcArgs, 2), ret2,
76 class getVOP3OpSelModPat<VOPProfile P, SDPatternOperator node> {
77 list<dag> ret3 = [(set P.DstVT:$vdst,
78 (DivergentFragOrOp<node, P>.ret (P.Src0VT !if(P.HasClamp, (VOP3OpSelMods0 P.Src0VT:$src0, i32:$src0_modifiers, i1:$clamp),
79 (VOP3OpSelMods P.Src0VT:$src0, i32:$src0_modifiers))),
80 (P.Src1VT (VOP3OpSelMods P.Src1VT:$src1, i32:$src1_modifiers)),
81 (P.Src2VT (VOP3OpSelMods P.Src2VT:$src2, i32:$src2_modifiers))))];
83 list<dag> ret2 = [(set P.DstVT:$vdst,
84 (DivergentFragOrOp<node, P>.ret !if(P.HasClamp, (P.Src0VT (VOP3OpSelMods0 P.Src0VT:$src0, i32:$src0_modifiers, i1:$clamp)),
85 (P.Src0VT (VOP3OpSelMods P.Src0VT:$src0, i32:$src0_modifiers))),
86 (P.Src1VT (VOP3OpSelMods P.Src1VT:$src1, i32:$src1_modifiers))))];
88 list<dag> ret1 = [(set P.DstVT:$vdst,
89 (DivergentFragOrOp<node, P>.ret (P.Src0VT (VOP3OpSelMods0 P.Src0VT:$src0, i32:$src0_modifiers, i1:$clamp))))];
91 list<dag> ret = !if(!eq(P.NumSrcArgs, 3), ret3,
92 !if(!eq(P.NumSrcArgs, 2), ret2,
96 class getVOP3Pat<VOPProfile P, SDPatternOperator node> {
97 list<dag> ret3 = [(set P.DstVT:$vdst, (DivergentFragOrOp<node, P>.ret P.Src0VT:$src0, P.Src1VT:$src1, P.Src2VT:$src2))];
98 list<dag> ret2 = [(set P.DstVT:$vdst, (DivergentFragOrOp<node, P>.ret P.Src0VT:$src0, P.Src1VT:$src1))];
99 list<dag> ret1 = [(set P.DstVT:$vdst, (DivergentFragOrOp<node, P>.ret P.Src0VT:$src0))];
100 list<dag> ret = !if(!eq(P.NumSrcArgs, 3), ret3,
101 !if(!eq(P.NumSrcArgs, 2), ret2,
105 class getVOP3ClampPat<VOPProfile P, SDPatternOperator node> {
106 list<dag> ret3 = [(set P.DstVT:$vdst, (node P.Src0VT:$src0, P.Src1VT:$src1, P.Src2VT:$src2, i1:$clamp))];
107 list<dag> ret2 = [(set P.DstVT:$vdst, (node P.Src0VT:$src0, P.Src1VT:$src1, i1:$clamp))];
108 list<dag> ret1 = [(set P.DstVT:$vdst, (node P.Src0VT:$src0, i1:$clamp))];
109 list<dag> ret = !if(!eq(P.NumSrcArgs, 3), ret3,
110 !if(!eq(P.NumSrcArgs, 2), ret2,
114 class VOP3Inst<string OpName, VOPProfile P, SDPatternOperator node = null_frag, bit VOP3Only = 0> :
115 VOP3_Pseudo<OpName, P,
118 getVOP3OpSelModPat<P, node>.ret,
119 getVOP3OpSelPat<P, node>.ret),
121 getVOP3ModPat<P, node>.ret,
123 getVOP3ClampPat<P, node>.ret,
124 getVOP3Pat<P, node>.ret))),
125 VOP3Only, 0, P.HasOpSel> {
127 let IntClamp = P.HasIntClamp;
128 let AsmMatchConverter =
131 !if(!or(P.HasModifiers, !or(P.HasOMod, P.HasIntClamp)),
136 // Special case for v_div_fmas_{f32|f64}, since it seems to be the
137 // only VOP instruction that implicitly reads VCC.
138 let Asm64 = " $vdst, $src0_modifiers, $src1_modifiers, $src2_modifiers$clamp$omod" in {
139 def VOP_F32_F32_F32_F32_VCC : VOPProfile<[f32, f32, f32, f32]> {
140 let Outs64 = (outs DstRC.RegClass:$vdst);
142 def VOP_F64_F64_F64_F64_VCC : VOPProfile<[f64, f64, f64, f64]> {
143 let Outs64 = (outs DstRC.RegClass:$vdst);
147 class getVOP3VCC<VOPProfile P, SDPatternOperator node> {
150 (node (P.Src0VT (VOP3Mods0 P.Src0VT:$src0, i32:$src0_modifiers, i1:$clamp, i32:$omod)),
151 (P.Src1VT (VOP3Mods P.Src1VT:$src1, i32:$src1_modifiers)),
152 (P.Src2VT (VOP3Mods P.Src2VT:$src2, i32:$src2_modifiers)),
156 class VOP3Features<bit Clamp, bit OpSel, bit Packed> {
157 bit HasClamp = Clamp;
158 bit HasOpSel = OpSel;
159 bit IsPacked = Packed;
162 def VOP3_REGULAR : VOP3Features<0, 0, 0>;
163 def VOP3_CLAMP : VOP3Features<1, 0, 0>;
164 def VOP3_OPSEL : VOP3Features<1, 1, 0>;
165 def VOP3_PACKED : VOP3Features<1, 1, 1>;
167 class VOP3_Profile<VOPProfile P, VOP3Features Features = VOP3_REGULAR> : VOPProfile<P.ArgVT> {
169 let HasClamp = !if(Features.HasClamp, 1, P.HasClamp);
170 let HasOpSel = !if(Features.HasOpSel, 1, P.HasOpSel);
171 let IsPacked = !if(Features.IsPacked, 1, P.IsPacked);
173 let HasModifiers = !if(Features.IsPacked, 1, P.HasModifiers);
175 // FIXME: Hack to stop printing _e64
176 let Outs64 = (outs DstRC.RegClass:$vdst);
178 " " # !if(Features.HasOpSel,
179 getAsmVOP3OpSel<NumSrcArgs,
183 HasSrc2FloatMods>.ret,
184 !if(Features.HasClamp,
185 getAsm64<HasDst, NumSrcArgs, HasIntClamp,
186 HasModifiers, HasOMod, DstVT>.ret,
188 let NeedPatGen = P.NeedPatGen;
191 class VOP3b_Profile<ValueType vt> : VOPProfile<[vt, vt, vt, vt]> {
192 // v_div_scale_{f32|f64} do not support input modifiers.
193 let HasModifiers = 0;
195 let Outs64 = (outs DstRC:$vdst, SReg_64:$sdst);
196 let Asm64 = " $vdst, $sdst, $src0, $src1, $src2";
199 def VOP3b_F32_I1_F32_F32_F32 : VOP3b_Profile<f32> {
200 // FIXME: Hack to stop printing _e64
201 let DstRC = RegisterOperand<VGPR_32>;
204 def VOP3b_F64_I1_F64_F64_F64 : VOP3b_Profile<f64> {
205 // FIXME: Hack to stop printing _e64
206 let DstRC = RegisterOperand<VReg_64>;
209 def VOP3b_I64_I1_I32_I32_I64 : VOPProfile<[i64, i32, i32, i64]> {
212 // FIXME: Hack to stop printing _e64
213 let DstRC = RegisterOperand<VReg_64>;
215 let Outs64 = (outs DstRC:$vdst, SReg_64:$sdst);
216 let Asm64 = " $vdst, $sdst, $src0, $src1, $src2$clamp";
219 //===----------------------------------------------------------------------===//
221 //===----------------------------------------------------------------------===//
223 class VOP3Interp<string OpName, VOPProfile P> : VOP3_Pseudo<OpName, P> {
224 let AsmMatchConverter = "cvtVOP3Interp";
227 def VOP3_INTERP : VOPProfile<[f32, f32, i32, untyped]> {
228 let Ins64 = (ins Src0Mod:$src0_modifiers, VRegSrc_32:$src0,
229 Attr:$attr, AttrChan:$attrchan,
230 clampmod:$clamp, omod:$omod);
232 let Asm64 = "$vdst, $src0_modifiers, $attr$attrchan$clamp$omod";
235 def VOP3_INTERP_MOV : VOPProfile<[f32, i32, i32, untyped]> {
236 let Ins64 = (ins InterpSlot:$src0,
237 Attr:$attr, AttrChan:$attrchan,
238 clampmod:$clamp, omod:$omod);
240 let Asm64 = "$vdst, $src0, $attr$attrchan$clamp$omod";
245 class getInterp16Asm <bit HasSrc2, bit HasOMod> {
246 string src2 = !if(HasSrc2, ", $src2_modifiers", "");
247 string omod = !if(HasOMod, "$omod", "");
249 " $vdst, $src0_modifiers, $attr$attrchan"#src2#"$high$clamp"#omod;
252 class getInterp16Ins <bit HasSrc2, bit HasOMod,
253 Operand Src0Mod, Operand Src2Mod> {
254 dag ret = !if(HasSrc2,
256 (ins Src0Mod:$src0_modifiers, VRegSrc_32:$src0,
257 Attr:$attr, AttrChan:$attrchan,
258 Src2Mod:$src2_modifiers, VRegSrc_32:$src2,
259 highmod:$high, clampmod:$clamp, omod:$omod),
260 (ins Src0Mod:$src0_modifiers, VRegSrc_32:$src0,
261 Attr:$attr, AttrChan:$attrchan,
262 Src2Mod:$src2_modifiers, VRegSrc_32:$src2,
263 highmod:$high, clampmod:$clamp)
265 (ins Src0Mod:$src0_modifiers, VRegSrc_32:$src0,
266 Attr:$attr, AttrChan:$attrchan,
267 highmod:$high, clampmod:$clamp, omod:$omod)
271 class VOP3_INTERP16 <list<ValueType> ArgVT> : VOPProfile<ArgVT> {
273 let HasOMod = !if(!eq(DstVT.Value, f16.Value), 0, 1);
276 let Outs64 = (outs VGPR_32:$vdst);
277 let Ins64 = getInterp16Ins<HasSrc2, HasOMod, Src0Mod, Src2Mod>.ret;
278 let Asm64 = getInterp16Asm<HasSrc2, HasOMod>.ret;
281 //===----------------------------------------------------------------------===//
283 //===----------------------------------------------------------------------===//
285 let isCommutable = 1 in {
287 def V_MAD_LEGACY_F32 : VOP3Inst <"v_mad_legacy_f32", VOP3_Profile<VOP_F32_F32_F32_F32>>;
288 def V_MAD_F32 : VOP3Inst <"v_mad_f32", VOP3_Profile<VOP_F32_F32_F32_F32>, fmad>;
289 def V_MAD_I32_I24 : VOP3Inst <"v_mad_i32_i24", VOP3_Profile<VOP_I32_I32_I32_I32, VOP3_CLAMP>>;
290 def V_MAD_U32_U24 : VOP3Inst <"v_mad_u32_u24", VOP3_Profile<VOP_I32_I32_I32_I32, VOP3_CLAMP>>;
291 def V_FMA_F32 : VOP3Inst <"v_fma_f32", VOP3_Profile<VOP_F32_F32_F32_F32>, fma>;
292 def V_LERP_U8 : VOP3Inst <"v_lerp_u8", VOP3_Profile<VOP_I32_I32_I32_I32>, int_amdgcn_lerp>;
294 let SchedRW = [WriteDoubleAdd] in {
295 def V_FMA_F64 : VOP3Inst <"v_fma_f64", VOP3_Profile<VOP_F64_F64_F64_F64>, fma>;
296 def V_ADD_F64 : VOP3Inst <"v_add_f64", VOP3_Profile<VOP_F64_F64_F64>, fadd, 1>;
297 def V_MUL_F64 : VOP3Inst <"v_mul_f64", VOP3_Profile<VOP_F64_F64_F64>, fmul, 1>;
298 def V_MIN_F64 : VOP3Inst <"v_min_f64", VOP3_Profile<VOP_F64_F64_F64>, fminnum, 1>;
299 def V_MAX_F64 : VOP3Inst <"v_max_f64", VOP3_Profile<VOP_F64_F64_F64>, fmaxnum, 1>;
300 } // End SchedRW = [WriteDoubleAdd]
302 let SchedRW = [WriteQuarterRate32] in {
303 def V_MUL_LO_U32 : VOP3Inst <"v_mul_lo_u32", VOP3_Profile<VOP_I32_I32_I32>>;
304 def V_MUL_HI_U32 : VOP3Inst <"v_mul_hi_u32", VOP3_Profile<VOP_I32_I32_I32>, mulhu>;
305 def V_MUL_LO_I32 : VOP3Inst <"v_mul_lo_i32", VOP3_Profile<VOP_I32_I32_I32>>;
306 def V_MUL_HI_I32 : VOP3Inst <"v_mul_hi_i32", VOP3_Profile<VOP_I32_I32_I32>, mulhs>;
307 } // End SchedRW = [WriteQuarterRate32]
309 let Uses = [VCC, EXEC] in {
311 // result = src0 * src1 + src2
315 def V_DIV_FMAS_F32 : VOP3_Pseudo <"v_div_fmas_f32", VOP_F32_F32_F32_F32_VCC,
316 getVOP3VCC<VOP_F32_F32_F32_F32_VCC, AMDGPUdiv_fmas>.ret> {
317 let SchedRW = [WriteFloatFMA];
320 // result = src0 * src1 + src2
324 def V_DIV_FMAS_F64 : VOP3_Pseudo <"v_div_fmas_f64", VOP_F64_F64_F64_F64_VCC,
325 getVOP3VCC<VOP_F64_F64_F64_F64_VCC, AMDGPUdiv_fmas>.ret> {
326 let SchedRW = [WriteDouble];
328 } // End Uses = [VCC, EXEC]
330 } // End isCommutable = 1
332 def V_CUBEID_F32 : VOP3Inst <"v_cubeid_f32", VOP3_Profile<VOP_F32_F32_F32_F32>, int_amdgcn_cubeid>;
333 def V_CUBESC_F32 : VOP3Inst <"v_cubesc_f32", VOP3_Profile<VOP_F32_F32_F32_F32>, int_amdgcn_cubesc>;
334 def V_CUBETC_F32 : VOP3Inst <"v_cubetc_f32", VOP3_Profile<VOP_F32_F32_F32_F32>, int_amdgcn_cubetc>;
335 def V_CUBEMA_F32 : VOP3Inst <"v_cubema_f32", VOP3_Profile<VOP_F32_F32_F32_F32>, int_amdgcn_cubema>;
336 def V_BFE_U32 : VOP3Inst <"v_bfe_u32", VOP3_Profile<VOP_I32_I32_I32_I32>, AMDGPUbfe_u32>;
337 def V_BFE_I32 : VOP3Inst <"v_bfe_i32", VOP3_Profile<VOP_I32_I32_I32_I32>, AMDGPUbfe_i32>;
338 def V_BFI_B32 : VOP3Inst <"v_bfi_b32", VOP3_Profile<VOP_I32_I32_I32_I32>, AMDGPUbfi>;
339 def V_ALIGNBIT_B32 : VOP3Inst <"v_alignbit_b32", VOP3_Profile<VOP_I32_I32_I32_I32>, int_amdgcn_alignbit>;
340 def V_ALIGNBYTE_B32 : VOP3Inst <"v_alignbyte_b32", VOP3_Profile<VOP_I32_I32_I32_I32>, int_amdgcn_alignbyte>;
341 def V_MIN3_F32 : VOP3Inst <"v_min3_f32", VOP3_Profile<VOP_F32_F32_F32_F32>, AMDGPUfmin3>;
342 def V_MIN3_I32 : VOP3Inst <"v_min3_i32", VOP3_Profile<VOP_I32_I32_I32_I32>, AMDGPUsmin3>;
343 def V_MIN3_U32 : VOP3Inst <"v_min3_u32", VOP3_Profile<VOP_I32_I32_I32_I32>, AMDGPUumin3>;
344 def V_MAX3_F32 : VOP3Inst <"v_max3_f32", VOP3_Profile<VOP_F32_F32_F32_F32>, AMDGPUfmax3>;
345 def V_MAX3_I32 : VOP3Inst <"v_max3_i32", VOP3_Profile<VOP_I32_I32_I32_I32>, AMDGPUsmax3>;
346 def V_MAX3_U32 : VOP3Inst <"v_max3_u32", VOP3_Profile<VOP_I32_I32_I32_I32>, AMDGPUumax3>;
347 def V_MED3_F32 : VOP3Inst <"v_med3_f32", VOP3_Profile<VOP_F32_F32_F32_F32>, AMDGPUfmed3>;
348 def V_MED3_I32 : VOP3Inst <"v_med3_i32", VOP3_Profile<VOP_I32_I32_I32_I32>, AMDGPUsmed3>;
349 def V_MED3_U32 : VOP3Inst <"v_med3_u32", VOP3_Profile<VOP_I32_I32_I32_I32>, AMDGPUumed3>;
350 def V_SAD_U8 : VOP3Inst <"v_sad_u8", VOP3_Profile<VOP_I32_I32_I32_I32, VOP3_CLAMP>>;
351 def V_SAD_HI_U8 : VOP3Inst <"v_sad_hi_u8", VOP3_Profile<VOP_I32_I32_I32_I32, VOP3_CLAMP>>;
352 def V_SAD_U16 : VOP3Inst <"v_sad_u16", VOP3_Profile<VOP_I32_I32_I32_I32, VOP3_CLAMP>>;
353 def V_SAD_U32 : VOP3Inst <"v_sad_u32", VOP3_Profile<VOP_I32_I32_I32_I32, VOP3_CLAMP>>;
354 def V_CVT_PK_U8_F32 : VOP3Inst<"v_cvt_pk_u8_f32", VOP3_Profile<VOP_I32_F32_I32_I32>, int_amdgcn_cvt_pk_u8_f32>;
355 def V_DIV_FIXUP_F32 : VOP3Inst <"v_div_fixup_f32", VOP3_Profile<VOP_F32_F32_F32_F32>, AMDGPUdiv_fixup>;
357 let SchedRW = [WriteDoubleAdd] in {
358 def V_DIV_FIXUP_F64 : VOP3Inst <"v_div_fixup_f64", VOP3_Profile<VOP_F64_F64_F64_F64>, AMDGPUdiv_fixup>;
359 def V_LDEXP_F64 : VOP3Inst <"v_ldexp_f64", VOP3_Profile<VOP_F64_F64_I32>, AMDGPUldexp, 1>;
360 } // End SchedRW = [WriteDoubleAdd]
362 def V_DIV_SCALE_F32 : VOP3_Pseudo <"v_div_scale_f32", VOP3b_F32_I1_F32_F32_F32, [], 1> {
363 let SchedRW = [WriteFloatFMA, WriteSALU];
364 let AsmMatchConverter = "";
367 // Double precision division pre-scale.
368 def V_DIV_SCALE_F64 : VOP3_Pseudo <"v_div_scale_f64", VOP3b_F64_I1_F64_F64_F64, [], 1> {
369 let SchedRW = [WriteDouble, WriteSALU];
370 let AsmMatchConverter = "";
373 def V_MSAD_U8 : VOP3Inst <"v_msad_u8", VOP3_Profile<VOP_I32_I32_I32_I32, VOP3_CLAMP>>;
375 let Constraints = "@earlyclobber $vdst" in {
376 def V_MQSAD_PK_U16_U8 : VOP3Inst <"v_mqsad_pk_u16_u8", VOP3_Profile<VOP_I64_I64_I32_I64, VOP3_CLAMP>>;
377 } // End Constraints = "@earlyclobber $vdst"
379 def V_TRIG_PREOP_F64 : VOP3Inst <"v_trig_preop_f64", VOP3_Profile<VOP_F64_F64_I32>, AMDGPUtrig_preop> {
380 let SchedRW = [WriteDouble];
383 let SchedRW = [Write64Bit] in {
384 // These instructions only exist on SI and CI
385 let SubtargetPredicate = isSICI, Predicates = [isSICI] in {
386 def V_LSHL_B64 : VOP3Inst <"v_lshl_b64", VOP3_Profile<VOP_PAT_GEN<VOP_I64_I64_I32>>, shl>;
387 def V_LSHR_B64 : VOP3Inst <"v_lshr_b64", VOP3_Profile<VOP_PAT_GEN<VOP_I64_I64_I32>>, srl>;
388 def V_ASHR_I64 : VOP3Inst <"v_ashr_i64", VOP3_Profile<VOP_PAT_GEN<VOP_I64_I64_I32>>, sra>;
389 def V_MULLIT_F32 : VOP3Inst <"v_mullit_f32", VOP3_Profile<VOP_F32_F32_F32_F32>>;
390 } // End SubtargetPredicate = isSICI, Predicates = [isSICI]
392 let SubtargetPredicate = isVI in {
393 def V_LSHLREV_B64 : VOP3Inst <"v_lshlrev_b64", VOP3_Profile<VOP_I64_I32_I64>>;
394 def V_LSHRREV_B64 : VOP3Inst <"v_lshrrev_b64", VOP3_Profile<VOP_I64_I32_I64>>;
395 def V_ASHRREV_I64 : VOP3Inst <"v_ashrrev_i64", VOP3_Profile<VOP_I64_I32_I64>>;
396 } // End SubtargetPredicate = isVI
397 } // End SchedRW = [Write64Bit]
399 let Predicates = [isVI] in {
401 (getDivergentFrag<shl>.ret i64:$x, i32:$y),
402 (V_LSHLREV_B64 $y, $x)
405 (getDivergentFrag<srl>.ret i64:$x, i32:$y),
406 (V_LSHRREV_B64 $y, $x)
409 (getDivergentFrag<sra>.ret i64:$x, i32:$y),
410 (V_ASHRREV_I64 $y, $x)
415 let SubtargetPredicate = isCIVI in {
417 let Constraints = "@earlyclobber $vdst", SchedRW = [WriteQuarterRate32] in {
418 def V_QSAD_PK_U16_U8 : VOP3Inst <"v_qsad_pk_u16_u8", VOP3_Profile<VOP_I64_I64_I32_I64, VOP3_CLAMP>>;
419 def V_MQSAD_U32_U8 : VOP3Inst <"v_mqsad_u32_u8", VOP3_Profile<VOP_V4I32_I64_I32_V4I32, VOP3_CLAMP>>;
420 } // End Constraints = "@earlyclobber $vdst", SchedRW = [WriteQuarterRate32]
422 let isCommutable = 1 in {
423 let SchedRW = [WriteQuarterRate32, WriteSALU] in {
424 def V_MAD_U64_U32 : VOP3Inst <"v_mad_u64_u32", VOP3b_I64_I1_I32_I32_I64>;
425 def V_MAD_I64_I32 : VOP3Inst <"v_mad_i64_i32", VOP3b_I64_I1_I32_I32_I64>;
426 } // End SchedRW = [WriteDouble, WriteSALU]
427 } // End isCommutable = 1
429 } // End SubtargetPredicate = isCIVI
432 def V_DIV_FIXUP_F16 : VOP3Inst <"v_div_fixup_f16", VOP3_Profile<VOP_F16_F16_F16_F16>, AMDGPUdiv_fixup> {
433 let Predicates = [Has16BitInsts, isVIOnly];
435 def V_DIV_FIXUP_F16_gfx9 : VOP3Inst <"v_div_fixup_f16_gfx9",
436 VOP3_Profile<VOP_F16_F16_F16_F16, VOP3_OPSEL>, AMDGPUdiv_fixup> {
437 let renamedInGFX9 = 1;
438 let Predicates = [Has16BitInsts, isGFX9];
441 let SubtargetPredicate = Has16BitInsts, isCommutable = 1 in {
443 let renamedInGFX9 = 1 in {
444 def V_MAD_F16 : VOP3Inst <"v_mad_f16", VOP3_Profile<VOP_F16_F16_F16_F16>, fmad>;
445 def V_MAD_U16 : VOP3Inst <"v_mad_u16", VOP3_Profile<VOP_I16_I16_I16_I16, VOP3_CLAMP>>;
446 def V_MAD_I16 : VOP3Inst <"v_mad_i16", VOP3_Profile<VOP_I16_I16_I16_I16, VOP3_CLAMP>>;
447 def V_FMA_F16 : VOP3Inst <"v_fma_f16", VOP3_Profile<VOP_F16_F16_F16_F16>, fma>;
448 def V_INTERP_P2_F16 : VOP3Interp <"v_interp_p2_f16", VOP3_INTERP16<[f16, f32, i32, f32]>>;
451 let SubtargetPredicate = isGFX9 in {
452 def V_MAD_F16_gfx9 : VOP3Inst <"v_mad_f16_gfx9", VOP3_Profile<VOP_F16_F16_F16_F16, VOP3_OPSEL>>;
453 def V_MAD_U16_gfx9 : VOP3Inst <"v_mad_u16_gfx9", VOP3_Profile<VOP_I16_I16_I16_I16, VOP3_OPSEL>>;
454 def V_MAD_I16_gfx9 : VOP3Inst <"v_mad_i16_gfx9", VOP3_Profile<VOP_I16_I16_I16_I16, VOP3_OPSEL>>;
455 def V_FMA_F16_gfx9 : VOP3Inst <"v_fma_f16_gfx9", VOP3_Profile<VOP_F16_F16_F16_F16, VOP3_OPSEL>>;
456 def V_INTERP_P2_F16_gfx9 : VOP3Interp <"v_interp_p2_f16_gfx9", VOP3_INTERP16<[f16, f32, i32, f32]>>;
457 } // End SubtargetPredicate = isGFX9
459 def V_INTERP_P1LL_F16 : VOP3Interp <"v_interp_p1ll_f16", VOP3_INTERP16<[f32, f32, i32, untyped]>>;
460 def V_INTERP_P1LV_F16 : VOP3Interp <"v_interp_p1lv_f16", VOP3_INTERP16<[f32, f32, i32, f16]>>;
462 } // End SubtargetPredicate = Has16BitInsts, isCommutable = 1
464 let SubtargetPredicate = isVI in {
465 def V_INTERP_P1_F32_e64 : VOP3Interp <"v_interp_p1_f32", VOP3_INTERP>;
466 def V_INTERP_P2_F32_e64 : VOP3Interp <"v_interp_p2_f32", VOP3_INTERP>;
467 def V_INTERP_MOV_F32_e64 : VOP3Interp <"v_interp_mov_f32", VOP3_INTERP_MOV>;
469 def V_PERM_B32 : VOP3Inst <"v_perm_b32", VOP3_Profile<VOP_I32_I32_I32_I32>, AMDGPUperm>;
470 } // End SubtargetPredicate = isVI
472 let Predicates = [Has16BitInsts] in {
474 multiclass Ternary_i16_Pats <SDPatternOperator op1, SDPatternOperator op2,
475 Instruction inst, SDPatternOperator op3> {
477 (op2 (op1 i16:$src0, i16:$src1), i16:$src2),
478 (inst i16:$src0, i16:$src1, i16:$src2, (i1 0))
483 defm: Ternary_i16_Pats<mul, add, V_MAD_U16, zext>;
484 defm: Ternary_i16_Pats<mul, add, V_MAD_I16, sext>;
486 } // End Predicates = [Has16BitInsts]
488 let SubtargetPredicate = isGFX9 in {
489 def V_PACK_B32_F16 : VOP3Inst <"v_pack_b32_f16", VOP3_Profile<VOP_B32_F16_F16, VOP3_OPSEL>>;
490 def V_LSHL_ADD_U32 : VOP3Inst <"v_lshl_add_u32", VOP3_Profile<VOP_I32_I32_I32_I32>>;
491 def V_ADD_LSHL_U32 : VOP3Inst <"v_add_lshl_u32", VOP3_Profile<VOP_I32_I32_I32_I32>>;
492 def V_ADD3_U32 : VOP3Inst <"v_add3_u32", VOP3_Profile<VOP_I32_I32_I32_I32>>;
493 def V_LSHL_OR_B32 : VOP3Inst <"v_lshl_or_b32", VOP3_Profile<VOP_I32_I32_I32_I32>>;
494 def V_AND_OR_B32 : VOP3Inst <"v_and_or_b32", VOP3_Profile<VOP_I32_I32_I32_I32>>;
495 def V_OR3_B32 : VOP3Inst <"v_or3_b32", VOP3_Profile<VOP_I32_I32_I32_I32>>;
497 def V_XAD_U32 : VOP3Inst <"v_xad_u32", VOP3_Profile<VOP_I32_I32_I32_I32>>;
499 def V_MED3_F16 : VOP3Inst <"v_med3_f16", VOP3_Profile<VOP_F16_F16_F16_F16, VOP3_OPSEL>, AMDGPUfmed3>;
500 def V_MED3_I16 : VOP3Inst <"v_med3_i16", VOP3_Profile<VOP_I16_I16_I16_I16, VOP3_OPSEL>, AMDGPUsmed3>;
501 def V_MED3_U16 : VOP3Inst <"v_med3_u16", VOP3_Profile<VOP_I16_I16_I16_I16, VOP3_OPSEL>, AMDGPUumed3>;
503 def V_MIN3_F16 : VOP3Inst <"v_min3_f16", VOP3_Profile<VOP_F16_F16_F16_F16, VOP3_OPSEL>, AMDGPUfmin3>;
504 def V_MIN3_I16 : VOP3Inst <"v_min3_i16", VOP3_Profile<VOP_I16_I16_I16_I16, VOP3_OPSEL>, AMDGPUsmin3>;
505 def V_MIN3_U16 : VOP3Inst <"v_min3_u16", VOP3_Profile<VOP_I16_I16_I16_I16, VOP3_OPSEL>, AMDGPUumin3>;
507 def V_MAX3_F16 : VOP3Inst <"v_max3_f16", VOP3_Profile<VOP_F16_F16_F16_F16, VOP3_OPSEL>, AMDGPUfmax3>;
508 def V_MAX3_I16 : VOP3Inst <"v_max3_i16", VOP3_Profile<VOP_I16_I16_I16_I16, VOP3_OPSEL>, AMDGPUsmax3>;
509 def V_MAX3_U16 : VOP3Inst <"v_max3_u16", VOP3_Profile<VOP_I16_I16_I16_I16, VOP3_OPSEL>, AMDGPUumax3>;
511 def V_ADD_I16 : VOP3Inst <"v_add_i16", VOP3_Profile<VOP_I16_I16_I16, VOP3_OPSEL>>;
512 def V_SUB_I16 : VOP3Inst <"v_sub_i16", VOP3_Profile<VOP_I16_I16_I16, VOP3_OPSEL>>;
514 def V_MAD_U32_U16 : VOP3Inst <"v_mad_u32_u16", VOP3_Profile<VOP_I32_I16_I16_I32, VOP3_OPSEL>>;
515 def V_MAD_I32_I16 : VOP3Inst <"v_mad_i32_i16", VOP3_Profile<VOP_I32_I16_I16_I32, VOP3_OPSEL>>;
517 def V_CVT_PKNORM_I16_F16 : VOP3Inst <"v_cvt_pknorm_i16_f16", VOP3_Profile<VOP_B32_F16_F16, VOP3_OPSEL>>;
518 def V_CVT_PKNORM_U16_F16 : VOP3Inst <"v_cvt_pknorm_u16_f16", VOP3_Profile<VOP_B32_F16_F16, VOP3_OPSEL>>;
520 def V_ADD_I32_gfx9 : VOP3Inst <"v_add_i32_gfx9", VOP3_Profile<VOP_I32_I32_I32>>;
521 def V_SUB_I32_gfx9 : VOP3Inst <"v_sub_i32_gfx9", VOP3_Profile<VOP_I32_I32_I32>>;
522 } // End SubtargetPredicate = isGFX9
524 //===----------------------------------------------------------------------===//
525 // Integer Clamp Patterns
526 //===----------------------------------------------------------------------===//
528 class getClampPat<VOPProfile P, SDPatternOperator node> {
529 dag ret3 = (P.DstVT (node P.Src0VT:$src0, P.Src1VT:$src1, P.Src2VT:$src2));
530 dag ret2 = (P.DstVT (node P.Src0VT:$src0, P.Src1VT:$src1));
531 dag ret1 = (P.DstVT (node P.Src0VT:$src0));
532 dag ret = !if(!eq(P.NumSrcArgs, 3), ret3,
533 !if(!eq(P.NumSrcArgs, 2), ret2,
537 class getClampRes<VOPProfile P, Instruction inst> {
538 dag ret3 = (inst P.Src0VT:$src0, P.Src1VT:$src1, P.Src2VT:$src2, (i1 0));
539 dag ret2 = (inst P.Src0VT:$src0, P.Src1VT:$src1, (i1 0));
540 dag ret1 = (inst P.Src0VT:$src0, (i1 0));
541 dag ret = !if(!eq(P.NumSrcArgs, 3), ret3,
542 !if(!eq(P.NumSrcArgs, 2), ret2,
546 class IntClampPat<VOP3Inst inst, SDPatternOperator node> : GCNPat<
547 getClampPat<inst.Pfl, node>.ret,
548 getClampRes<inst.Pfl, inst>.ret
551 def : IntClampPat<V_MAD_I32_I24, AMDGPUmad_i24>;
552 def : IntClampPat<V_MAD_U32_U24, AMDGPUmad_u24>;
554 def : IntClampPat<V_SAD_U8, int_amdgcn_sad_u8>;
555 def : IntClampPat<V_SAD_HI_U8, int_amdgcn_sad_hi_u8>;
556 def : IntClampPat<V_SAD_U16, int_amdgcn_sad_u16>;
558 def : IntClampPat<V_MSAD_U8, int_amdgcn_msad_u8>;
559 def : IntClampPat<V_MQSAD_PK_U16_U8, int_amdgcn_mqsad_pk_u16_u8>;
561 def : IntClampPat<V_QSAD_PK_U16_U8, int_amdgcn_qsad_pk_u16_u8>;
562 def : IntClampPat<V_MQSAD_U32_U8, int_amdgcn_mqsad_u32_u8>;
564 //===----------------------------------------------------------------------===//
566 //===----------------------------------------------------------------------===//
568 //===----------------------------------------------------------------------===//
570 //===----------------------------------------------------------------------===//
572 let AssemblerPredicates = [isSICI], DecoderNamespace = "SICI" in {
574 multiclass VOP3_Real_si<bits<9> op> {
575 def _si : VOP3_Real<!cast<VOP3_Pseudo>(NAME), SIEncodingFamily.SI>,
576 VOP3e_si <op, !cast<VOP3_Pseudo>(NAME).Pfl>;
579 multiclass VOP3be_Real_si<bits<9> op> {
580 def _si : VOP3_Real<!cast<VOP3_Pseudo>(NAME), SIEncodingFamily.SI>,
581 VOP3be_si <op, !cast<VOP3_Pseudo>(NAME).Pfl>;
584 } // End AssemblerPredicates = [isSICI], DecoderNamespace = "SICI"
586 defm V_MAD_LEGACY_F32 : VOP3_Real_si <0x140>;
587 defm V_MAD_F32 : VOP3_Real_si <0x141>;
588 defm V_MAD_I32_I24 : VOP3_Real_si <0x142>;
589 defm V_MAD_U32_U24 : VOP3_Real_si <0x143>;
590 defm V_CUBEID_F32 : VOP3_Real_si <0x144>;
591 defm V_CUBESC_F32 : VOP3_Real_si <0x145>;
592 defm V_CUBETC_F32 : VOP3_Real_si <0x146>;
593 defm V_CUBEMA_F32 : VOP3_Real_si <0x147>;
594 defm V_BFE_U32 : VOP3_Real_si <0x148>;
595 defm V_BFE_I32 : VOP3_Real_si <0x149>;
596 defm V_BFI_B32 : VOP3_Real_si <0x14a>;
597 defm V_FMA_F32 : VOP3_Real_si <0x14b>;
598 defm V_FMA_F64 : VOP3_Real_si <0x14c>;
599 defm V_LERP_U8 : VOP3_Real_si <0x14d>;
600 defm V_ALIGNBIT_B32 : VOP3_Real_si <0x14e>;
601 defm V_ALIGNBYTE_B32 : VOP3_Real_si <0x14f>;
602 defm V_MULLIT_F32 : VOP3_Real_si <0x150>;
603 defm V_MIN3_F32 : VOP3_Real_si <0x151>;
604 defm V_MIN3_I32 : VOP3_Real_si <0x152>;
605 defm V_MIN3_U32 : VOP3_Real_si <0x153>;
606 defm V_MAX3_F32 : VOP3_Real_si <0x154>;
607 defm V_MAX3_I32 : VOP3_Real_si <0x155>;
608 defm V_MAX3_U32 : VOP3_Real_si <0x156>;
609 defm V_MED3_F32 : VOP3_Real_si <0x157>;
610 defm V_MED3_I32 : VOP3_Real_si <0x158>;
611 defm V_MED3_U32 : VOP3_Real_si <0x159>;
612 defm V_SAD_U8 : VOP3_Real_si <0x15a>;
613 defm V_SAD_HI_U8 : VOP3_Real_si <0x15b>;
614 defm V_SAD_U16 : VOP3_Real_si <0x15c>;
615 defm V_SAD_U32 : VOP3_Real_si <0x15d>;
616 defm V_CVT_PK_U8_F32 : VOP3_Real_si <0x15e>;
617 defm V_DIV_FIXUP_F32 : VOP3_Real_si <0x15f>;
618 defm V_DIV_FIXUP_F64 : VOP3_Real_si <0x160>;
619 defm V_LSHL_B64 : VOP3_Real_si <0x161>;
620 defm V_LSHR_B64 : VOP3_Real_si <0x162>;
621 defm V_ASHR_I64 : VOP3_Real_si <0x163>;
622 defm V_ADD_F64 : VOP3_Real_si <0x164>;
623 defm V_MUL_F64 : VOP3_Real_si <0x165>;
624 defm V_MIN_F64 : VOP3_Real_si <0x166>;
625 defm V_MAX_F64 : VOP3_Real_si <0x167>;
626 defm V_LDEXP_F64 : VOP3_Real_si <0x168>;
627 defm V_MUL_LO_U32 : VOP3_Real_si <0x169>;
628 defm V_MUL_HI_U32 : VOP3_Real_si <0x16a>;
629 defm V_MUL_LO_I32 : VOP3_Real_si <0x16b>;
630 defm V_MUL_HI_I32 : VOP3_Real_si <0x16c>;
631 defm V_DIV_SCALE_F32 : VOP3be_Real_si <0x16d>;
632 defm V_DIV_SCALE_F64 : VOP3be_Real_si <0x16e>;
633 defm V_DIV_FMAS_F32 : VOP3_Real_si <0x16f>;
634 defm V_DIV_FMAS_F64 : VOP3_Real_si <0x170>;
635 defm V_MSAD_U8 : VOP3_Real_si <0x171>;
636 defm V_MQSAD_PK_U16_U8 : VOP3_Real_si <0x173>;
637 defm V_TRIG_PREOP_F64 : VOP3_Real_si <0x174>;
639 //===----------------------------------------------------------------------===//
641 //===----------------------------------------------------------------------===//
643 multiclass VOP3_Real_ci<bits<9> op> {
644 def _ci : VOP3_Real<!cast<VOP3_Pseudo>(NAME), SIEncodingFamily.SI>,
645 VOP3e_si <op, !cast<VOP3_Pseudo>(NAME).Pfl> {
646 let AssemblerPredicates = [isCIOnly];
647 let DecoderNamespace = "CI";
651 multiclass VOP3be_Real_ci<bits<9> op> {
652 def _ci : VOP3_Real<!cast<VOP3_Pseudo>(NAME), SIEncodingFamily.SI>,
653 VOP3be_si <op, !cast<VOP3_Pseudo>(NAME).Pfl> {
654 let AssemblerPredicates = [isCIOnly];
655 let DecoderNamespace = "CI";
659 defm V_QSAD_PK_U16_U8 : VOP3_Real_ci <0x172>;
660 defm V_MQSAD_U32_U8 : VOP3_Real_ci <0x175>;
661 defm V_MAD_U64_U32 : VOP3be_Real_ci <0x176>;
662 defm V_MAD_I64_I32 : VOP3be_Real_ci <0x177>;
664 //===----------------------------------------------------------------------===//
666 //===----------------------------------------------------------------------===//
668 let AssemblerPredicates = [isVI], DecoderNamespace = "VI" in {
670 multiclass VOP3_Real_vi<bits<10> op> {
671 def _vi : VOP3_Real<!cast<VOP_Pseudo>(NAME), SIEncodingFamily.VI>,
672 VOP3e_vi <op, !cast<VOP_Pseudo>(NAME).Pfl>;
675 multiclass VOP3be_Real_vi<bits<10> op> {
676 def _vi : VOP3_Real<!cast<VOP_Pseudo>(NAME), SIEncodingFamily.VI>,
677 VOP3be_vi <op, !cast<VOP_Pseudo>(NAME).Pfl>;
680 multiclass VOP3OpSel_Real_gfx9<bits<10> op> {
681 def _vi : VOP3_Real<!cast<VOP_Pseudo>(NAME), SIEncodingFamily.VI>,
682 VOP3OpSel_gfx9 <op, !cast<VOP_Pseudo>(NAME).Pfl>;
685 multiclass VOP3Interp_Real_vi<bits<10> op> {
686 def _vi : VOP3_Real<!cast<VOP_Pseudo>(NAME), SIEncodingFamily.VI>,
687 VOP3Interp_vi <op, !cast<VOP_Pseudo>(NAME).Pfl>;
690 } // End AssemblerPredicates = [isVI], DecoderNamespace = "VI"
692 let AssemblerPredicates = [isVIOnly], DecoderNamespace = "VI" in {
694 multiclass VOP3_F16_Real_vi<bits<10> op> {
695 def _vi : VOP3_Real<!cast<VOP3_Pseudo>(NAME), SIEncodingFamily.VI>,
696 VOP3e_vi <op, !cast<VOP3_Pseudo>(NAME).Pfl>;
699 multiclass VOP3Interp_F16_Real_vi<bits<10> op> {
700 def _vi : VOP3_Real<!cast<VOP3_Pseudo>(NAME), SIEncodingFamily.VI>,
701 VOP3Interp_vi <op, !cast<VOP3_Pseudo>(NAME).Pfl>;
704 } // End AssemblerPredicates = [isVIOnly], DecoderNamespace = "VI"
706 let AssemblerPredicates = [isGFX9], DecoderNamespace = "GFX9" in {
708 multiclass VOP3_F16_Real_gfx9<bits<10> op, string OpName, string AsmName> {
709 def _gfx9 : VOP3_Real<!cast<VOP3_Pseudo>(OpName), SIEncodingFamily.GFX9>,
710 VOP3e_vi <op, !cast<VOP3_Pseudo>(OpName).Pfl> {
711 VOP3_Pseudo ps = !cast<VOP3_Pseudo>(OpName);
712 let AsmString = AsmName # ps.AsmOperands;
716 multiclass VOP3OpSel_F16_Real_gfx9<bits<10> op, string AsmName> {
717 def _gfx9 : VOP3_Real<!cast<VOP3_Pseudo>(NAME), SIEncodingFamily.GFX9>,
718 VOP3OpSel_gfx9 <op, !cast<VOP3_Pseudo>(NAME).Pfl> {
719 VOP3_Pseudo ps = !cast<VOP3_Pseudo>(NAME);
720 let AsmString = AsmName # ps.AsmOperands;
724 multiclass VOP3Interp_F16_Real_gfx9<bits<10> op, string OpName, string AsmName> {
725 def _gfx9 : VOP3_Real<!cast<VOP3_Pseudo>(OpName), SIEncodingFamily.GFX9>,
726 VOP3Interp_vi <op, !cast<VOP3_Pseudo>(OpName).Pfl> {
727 VOP3_Pseudo ps = !cast<VOP3_Pseudo>(OpName);
728 let AsmString = AsmName # ps.AsmOperands;
732 multiclass VOP3_Real_gfx9<bits<10> op, string AsmName> {
733 def _gfx9 : VOP3_Real<!cast<VOP_Pseudo>(NAME), SIEncodingFamily.GFX9>,
734 VOP3e_vi <op, !cast<VOP_Pseudo>(NAME).Pfl> {
735 VOP_Pseudo ps = !cast<VOP_Pseudo>(NAME);
736 let AsmString = AsmName # ps.AsmOperands;
740 } // End AssemblerPredicates = [isGFX9], DecoderNamespace = "GFX9"
742 defm V_MAD_U64_U32 : VOP3be_Real_vi <0x1E8>;
743 defm V_MAD_I64_I32 : VOP3be_Real_vi <0x1E9>;
745 defm V_MAD_LEGACY_F32 : VOP3_Real_vi <0x1c0>;
746 defm V_MAD_F32 : VOP3_Real_vi <0x1c1>;
747 defm V_MAD_I32_I24 : VOP3_Real_vi <0x1c2>;
748 defm V_MAD_U32_U24 : VOP3_Real_vi <0x1c3>;
749 defm V_CUBEID_F32 : VOP3_Real_vi <0x1c4>;
750 defm V_CUBESC_F32 : VOP3_Real_vi <0x1c5>;
751 defm V_CUBETC_F32 : VOP3_Real_vi <0x1c6>;
752 defm V_CUBEMA_F32 : VOP3_Real_vi <0x1c7>;
753 defm V_BFE_U32 : VOP3_Real_vi <0x1c8>;
754 defm V_BFE_I32 : VOP3_Real_vi <0x1c9>;
755 defm V_BFI_B32 : VOP3_Real_vi <0x1ca>;
756 defm V_FMA_F32 : VOP3_Real_vi <0x1cb>;
757 defm V_FMA_F64 : VOP3_Real_vi <0x1cc>;
758 defm V_LERP_U8 : VOP3_Real_vi <0x1cd>;
759 defm V_ALIGNBIT_B32 : VOP3_Real_vi <0x1ce>;
760 defm V_ALIGNBYTE_B32 : VOP3_Real_vi <0x1cf>;
761 defm V_MIN3_F32 : VOP3_Real_vi <0x1d0>;
762 defm V_MIN3_I32 : VOP3_Real_vi <0x1d1>;
763 defm V_MIN3_U32 : VOP3_Real_vi <0x1d2>;
764 defm V_MAX3_F32 : VOP3_Real_vi <0x1d3>;
765 defm V_MAX3_I32 : VOP3_Real_vi <0x1d4>;
766 defm V_MAX3_U32 : VOP3_Real_vi <0x1d5>;
767 defm V_MED3_F32 : VOP3_Real_vi <0x1d6>;
768 defm V_MED3_I32 : VOP3_Real_vi <0x1d7>;
769 defm V_MED3_U32 : VOP3_Real_vi <0x1d8>;
770 defm V_SAD_U8 : VOP3_Real_vi <0x1d9>;
771 defm V_SAD_HI_U8 : VOP3_Real_vi <0x1da>;
772 defm V_SAD_U16 : VOP3_Real_vi <0x1db>;
773 defm V_SAD_U32 : VOP3_Real_vi <0x1dc>;
774 defm V_CVT_PK_U8_F32 : VOP3_Real_vi <0x1dd>;
775 defm V_DIV_FIXUP_F32 : VOP3_Real_vi <0x1de>;
776 defm V_DIV_FIXUP_F64 : VOP3_Real_vi <0x1df>;
777 defm V_DIV_SCALE_F32 : VOP3be_Real_vi <0x1e0>;
778 defm V_DIV_SCALE_F64 : VOP3be_Real_vi <0x1e1>;
779 defm V_DIV_FMAS_F32 : VOP3_Real_vi <0x1e2>;
780 defm V_DIV_FMAS_F64 : VOP3_Real_vi <0x1e3>;
781 defm V_MSAD_U8 : VOP3_Real_vi <0x1e4>;
782 defm V_QSAD_PK_U16_U8 : VOP3_Real_vi <0x1e5>;
783 defm V_MQSAD_PK_U16_U8 : VOP3_Real_vi <0x1e6>;
784 defm V_MQSAD_U32_U8 : VOP3_Real_vi <0x1e7>;
786 defm V_PERM_B32 : VOP3_Real_vi <0x1ed>;
788 defm V_MAD_F16 : VOP3_F16_Real_vi <0x1ea>;
789 defm V_MAD_U16 : VOP3_F16_Real_vi <0x1eb>;
790 defm V_MAD_I16 : VOP3_F16_Real_vi <0x1ec>;
791 defm V_FMA_F16 : VOP3_F16_Real_vi <0x1ee>;
792 defm V_DIV_FIXUP_F16 : VOP3_F16_Real_vi <0x1ef>;
793 defm V_INTERP_P2_F16 : VOP3Interp_F16_Real_vi <0x276>;
795 defm V_MAD_LEGACY_F16 : VOP3_F16_Real_gfx9 <0x1ea, "V_MAD_F16", "v_mad_legacy_f16">;
796 defm V_MAD_LEGACY_U16 : VOP3_F16_Real_gfx9 <0x1eb, "V_MAD_U16", "v_mad_legacy_u16">;
797 defm V_MAD_LEGACY_I16 : VOP3_F16_Real_gfx9 <0x1ec, "V_MAD_I16", "v_mad_legacy_i16">;
798 defm V_FMA_LEGACY_F16 : VOP3_F16_Real_gfx9 <0x1ee, "V_FMA_F16", "v_fma_legacy_f16">;
799 defm V_DIV_FIXUP_LEGACY_F16 : VOP3_F16_Real_gfx9 <0x1ef, "V_DIV_FIXUP_F16", "v_div_fixup_legacy_f16">;
800 defm V_INTERP_P2_LEGACY_F16 : VOP3Interp_F16_Real_gfx9 <0x276, "V_INTERP_P2_F16", "v_interp_p2_legacy_f16">;
802 defm V_MAD_F16_gfx9 : VOP3OpSel_F16_Real_gfx9 <0x203, "v_mad_f16">;
803 defm V_MAD_U16_gfx9 : VOP3OpSel_F16_Real_gfx9 <0x204, "v_mad_u16">;
804 defm V_MAD_I16_gfx9 : VOP3OpSel_F16_Real_gfx9 <0x205, "v_mad_i16">;
805 defm V_FMA_F16_gfx9 : VOP3OpSel_F16_Real_gfx9 <0x206, "v_fma_f16">;
806 defm V_DIV_FIXUP_F16_gfx9 : VOP3OpSel_F16_Real_gfx9 <0x207, "v_div_fixup_f16">;
807 defm V_INTERP_P2_F16_gfx9 : VOP3Interp_F16_Real_gfx9 <0x277, "V_INTERP_P2_F16_gfx9", "v_interp_p2_f16">;
809 defm V_ADD_I32_gfx9 : VOP3_Real_gfx9 <0x29c, "v_add_i32">;
810 defm V_SUB_I32_gfx9 : VOP3_Real_gfx9 <0x29d, "v_sub_i32">;
812 defm V_INTERP_P1_F32_e64 : VOP3Interp_Real_vi <0x270>;
813 defm V_INTERP_P2_F32_e64 : VOP3Interp_Real_vi <0x271>;
814 defm V_INTERP_MOV_F32_e64 : VOP3Interp_Real_vi <0x272>;
816 defm V_INTERP_P1LL_F16 : VOP3Interp_Real_vi <0x274>;
817 defm V_INTERP_P1LV_F16 : VOP3Interp_Real_vi <0x275>;
818 defm V_ADD_F64 : VOP3_Real_vi <0x280>;
819 defm V_MUL_F64 : VOP3_Real_vi <0x281>;
820 defm V_MIN_F64 : VOP3_Real_vi <0x282>;
821 defm V_MAX_F64 : VOP3_Real_vi <0x283>;
822 defm V_LDEXP_F64 : VOP3_Real_vi <0x284>;
823 defm V_MUL_LO_U32 : VOP3_Real_vi <0x285>;
825 // removed from VI as identical to V_MUL_LO_U32
826 let isAsmParserOnly = 1 in {
827 defm V_MUL_LO_I32 : VOP3_Real_vi <0x285>;
830 defm V_MUL_HI_U32 : VOP3_Real_vi <0x286>;
831 defm V_MUL_HI_I32 : VOP3_Real_vi <0x287>;
833 defm V_READLANE_B32 : VOP3_Real_vi <0x289>;
834 defm V_WRITELANE_B32 : VOP3_Real_vi <0x28a>;
836 defm V_LSHLREV_B64 : VOP3_Real_vi <0x28f>;
837 defm V_LSHRREV_B64 : VOP3_Real_vi <0x290>;
838 defm V_ASHRREV_I64 : VOP3_Real_vi <0x291>;
839 defm V_TRIG_PREOP_F64 : VOP3_Real_vi <0x292>;
841 defm V_LSHL_ADD_U32 : VOP3_Real_vi <0x1fd>;
842 defm V_ADD_LSHL_U32 : VOP3_Real_vi <0x1fe>;
843 defm V_ADD3_U32 : VOP3_Real_vi <0x1ff>;
844 defm V_LSHL_OR_B32 : VOP3_Real_vi <0x200>;
845 defm V_AND_OR_B32 : VOP3_Real_vi <0x201>;
846 defm V_OR3_B32 : VOP3_Real_vi <0x202>;
847 defm V_PACK_B32_F16 : VOP3OpSel_Real_gfx9 <0x2a0>;
849 defm V_XAD_U32 : VOP3_Real_vi <0x1f3>;
851 defm V_MIN3_F16 : VOP3OpSel_Real_gfx9 <0x1f4>;
852 defm V_MIN3_I16 : VOP3OpSel_Real_gfx9 <0x1f5>;
853 defm V_MIN3_U16 : VOP3OpSel_Real_gfx9 <0x1f6>;
855 defm V_MAX3_F16 : VOP3OpSel_Real_gfx9 <0x1f7>;
856 defm V_MAX3_I16 : VOP3OpSel_Real_gfx9 <0x1f8>;
857 defm V_MAX3_U16 : VOP3OpSel_Real_gfx9 <0x1f9>;
859 defm V_MED3_F16 : VOP3OpSel_Real_gfx9 <0x1fa>;
860 defm V_MED3_I16 : VOP3OpSel_Real_gfx9 <0x1fb>;
861 defm V_MED3_U16 : VOP3OpSel_Real_gfx9 <0x1fc>;
863 defm V_ADD_I16 : VOP3OpSel_Real_gfx9 <0x29e>;
864 defm V_SUB_I16 : VOP3OpSel_Real_gfx9 <0x29f>;
866 defm V_MAD_U32_U16 : VOP3OpSel_Real_gfx9 <0x1f1>;
867 defm V_MAD_I32_I16 : VOP3OpSel_Real_gfx9 <0x1f2>;
869 defm V_CVT_PKNORM_I16_F16 : VOP3OpSel_Real_gfx9 <0x299>;
870 defm V_CVT_PKNORM_U16_F16 : VOP3OpSel_Real_gfx9 <0x29a>;