reference, declaration → definition definition → references, declarations, derived classes, virtual overrides reference to multiple definitions → definitions unreferenced |
549 if (!isUseSafeToFold(TII, *UseMI, UseOp)) 573 if (tryToFoldACImm(TII, UseMI->getOperand(0), RSUseMI, 587 if (tryToFoldACImm(TII, OpToFold, UseMI, UseOpIdx, FoldList)) 590 if (frameIndexMayFold(TII, *UseMI, UseOpIdx, OpToFold)) { 593 MachineOperand *SOff = TII->getNamedOperand(*UseMI, AMDGPU::OpName::soffset); 598 if (TII->getNamedOperand(*UseMI, AMDGPU::OpName::srsrc)->getReg() != 645 TII->isInlineConstant(OpToFold, AMDGPU::OPERAND_REG_INLINE_C_INT32)) { 646 UseMI->setDesc(TII->get(AMDGPU::V_ACCVGPR_WRITE_B32)); 655 unsigned MovOp = TII->getMovOpcode(DestRC); 659 UseMI->setDesc(TII->get(MovOp)); 674 unsigned Size = TII->getOpSize(*UseMI, 1); 688 getRegSeqInit(Defs, UseReg, AMDGPU::OPERAND_REG_INLINE_C_INT32, TII, 693 UseMI->setDesc(TII->get(AMDGPU::REG_SEQUENCE)); 704 TII->isInlineConstant(*Def, AMDGPU::OPERAND_REG_INLINE_C_INT32)) { 709 TII->get(AMDGPU::V_ACCVGPR_WRITE_B32), Tmp).addImm(Imm); 735 BuildMI(MBB, UseMI, DL, TII->get(AMDGPU::COPY), Tmp).add(*Def); 746 BuildMI(MBB, UseMI, DL, TII->get(AMDGPU::COPY), Vgpr).add(*Def); 751 TII->get(AMDGPU::V_ACCVGPR_WRITE_B32), Tmp).addReg(Vgpr); 765 UseMI->setDesc(TII->get(AMDGPU::V_ACCVGPR_WRITE_B32)); 768 UseMI->setDesc(TII->get(AMDGPU::V_ACCVGPR_READ_B32)); 788 UseMI->setDesc(TII->get(AMDGPU::S_MOV_B32)); 811 UseMI->setDesc(TII->get(AMDGPU::COPY)); 831 tryAddToFoldList(FoldList, UseMI, UseOpIdx, &OpToFold, TII); 861 tryAddToFoldList(FoldList, UseMI, UseOpIdx, &ImmOp, TII); 867 tryAddToFoldList(FoldList, UseMI, UseOpIdx, &OpToFold, TII); 1145 if (OpToFold.isImm() && tryConstantFoldOp(*MRI, TII, UseMI, &OpToFold)) { 1172 if (isInlineConstantIfFolded(TII, *UseMI, OpNo, OpToFold)) { 1174 } else if (frameIndexMayFold(TII, *UseMI, OpNo, OpToFold)) { 1218 if (updateOperand(Fold, *TII, *TRI, *ST)) { 1230 tryFoldInst(TII, Fold.UseMI); 1233 TII->commuteInstruction(*Fold.UseMI, false); 1247 if (!TII->getNamedOperand(MI, AMDGPU::OpName::clamp)->getImm()) 1251 const MachineOperand *Src0 = TII->getNamedOperand(MI, AMDGPU::OpName::src0); 1252 const MachineOperand *Src1 = TII->getNamedOperand(MI, AMDGPU::OpName::src1); 1260 if (TII->hasModifiersSet(MI, AMDGPU::OpName::omod)) 1264 = TII->getNamedOperand(MI, AMDGPU::OpName::src0_modifiers)->getImm(); 1266 = TII->getNamedOperand(MI, AMDGPU::OpName::src1_modifiers)->getImm(); 1303 if (TII->getClampMask(*Def) != TII->getClampMask(MI)) 1303 if (TII->getClampMask(*Def) != TII->getClampMask(MI)) 1306 MachineOperand *DefClamp = TII->getNamedOperand(*Def, AMDGPU::OpName::clamp); 1367 const MachineOperand *Src0 = TII->getNamedOperand(MI, AMDGPU::OpName::src0); 1368 const MachineOperand *Src1 = TII->getNamedOperand(MI, AMDGPU::OpName::src1); 1380 TII->hasModifiersSet(MI, AMDGPU::OpName::src0_modifiers) || 1381 TII->hasModifiersSet(MI, AMDGPU::OpName::src1_modifiers) || 1382 TII->hasModifiersSet(MI, AMDGPU::OpName::omod) || 1383 TII->hasModifiersSet(MI, AMDGPU::OpName::clamp)) 1396 const MachineOperand *Src0 = TII->getNamedOperand(MI, AMDGPU::OpName::src0); 1397 const MachineOperand *Src1 = TII->getNamedOperand(MI, AMDGPU::OpName::src1); 1401 !TII->hasModifiersSet(MI, AMDGPU::OpName::src0_modifiers) && 1402 !TII->hasModifiersSet(MI, AMDGPU::OpName::src1_modifiers) && 1403 !TII->hasModifiersSet(MI, AMDGPU::OpName::clamp) && 1404 !TII->hasModifiersSet(MI, AMDGPU::OpName::omod)) 1425 MachineOperand *DefOMod = TII->getNamedOperand(*Def, AMDGPU::OpName::omod); 1431 if (TII->hasModifiersSet(*Def, AMDGPU::OpName::clamp)) 1448 TII = ST->getInstrInfo(); 1449 TRI = &TII->getRegisterInfo(); 1467 tryFoldInst(TII, &MI); 1469 if (!TII->isFoldableCopy(MI)) {