Chris Lattner | a960d95 | 2003-01-13 01:01:59 +0000 | [diff] [blame] | 1 | //===-- PeepholeOptimizer.cpp - X86 Peephole Optimizer --------------------===// |
John Criswell | b576c94 | 2003-10-20 19:43:21 +0000 | [diff] [blame] | 2 | // |
| 3 | // The LLVM Compiler Infrastructure |
| 4 | // |
| 5 | // This file was developed by the LLVM research group and is distributed under |
| 6 | // the University of Illinois Open Source License. See LICENSE.TXT for details. |
| 7 | // |
| 8 | //===----------------------------------------------------------------------===// |
Chris Lattner | a960d95 | 2003-01-13 01:01:59 +0000 | [diff] [blame] | 9 | // |
| 10 | // This file contains a peephole optimizer for the X86. |
| 11 | // |
| 12 | //===----------------------------------------------------------------------===// |
| 13 | |
| 14 | #include "X86.h" |
| 15 | #include "llvm/CodeGen/MachineFunctionPass.h" |
| 16 | #include "llvm/CodeGen/MachineInstrBuilder.h" |
| 17 | |
| 18 | namespace { |
| 19 | struct PH : public MachineFunctionPass { |
| 20 | virtual bool runOnMachineFunction(MachineFunction &MF); |
| 21 | |
| 22 | bool PeepholeOptimize(MachineBasicBlock &MBB, |
| 23 | MachineBasicBlock::iterator &I); |
| 24 | |
| 25 | virtual const char *getPassName() const { return "X86 Peephole Optimizer"; } |
| 26 | }; |
| 27 | } |
| 28 | |
Brian Gaeke | 19df387 | 2003-08-13 18:18:15 +0000 | [diff] [blame] | 29 | FunctionPass *createX86PeepholeOptimizerPass() { return new PH(); } |
Chris Lattner | a960d95 | 2003-01-13 01:01:59 +0000 | [diff] [blame] | 30 | |
| 31 | bool PH::runOnMachineFunction(MachineFunction &MF) { |
| 32 | bool Changed = false; |
| 33 | |
| 34 | for (MachineFunction::iterator BI = MF.begin(), E = MF.end(); BI != E; ++BI) |
Chris Lattner | ee3e435 | 2003-01-16 18:07:13 +0000 | [diff] [blame] | 35 | for (MachineBasicBlock::iterator I = BI->begin(); I != BI->end(); ) |
Chris Lattner | a960d95 | 2003-01-13 01:01:59 +0000 | [diff] [blame] | 36 | if (PeepholeOptimize(*BI, I)) |
| 37 | Changed = true; |
| 38 | else |
| 39 | ++I; |
| 40 | |
| 41 | return Changed; |
| 42 | } |
| 43 | |
| 44 | |
| 45 | bool PH::PeepholeOptimize(MachineBasicBlock &MBB, |
| 46 | MachineBasicBlock::iterator &I) { |
| 47 | MachineInstr *MI = *I; |
| 48 | MachineInstr *Next = (I+1 != MBB.end()) ? *(I+1) : 0; |
| 49 | unsigned Size = 0; |
| 50 | switch (MI->getOpcode()) { |
| 51 | case X86::MOVrr8: |
| 52 | case X86::MOVrr16: |
| 53 | case X86::MOVrr32: // Destroy X = X copies... |
| 54 | if (MI->getOperand(0).getReg() == MI->getOperand(1).getReg()) { |
| 55 | I = MBB.erase(I); |
| 56 | delete MI; |
| 57 | return true; |
| 58 | } |
| 59 | return false; |
| 60 | |
Chris Lattner | 43a5ff8 | 2003-10-20 05:53:31 +0000 | [diff] [blame] | 61 | // A large number of X86 instructions have forms which take an 8-bit |
| 62 | // immediate despite the fact that the operands are 16 or 32 bits. Because |
| 63 | // this can save three bytes of code size (and icache space), we want to |
| 64 | // shrink them if possible. |
| 65 | case X86::ADDri16: case X86::ADDri32: |
| 66 | case X86::SUBri16: case X86::SUBri32: |
| 67 | case X86::IMULri16: case X86::IMULri32: |
| 68 | case X86::ANDri16: case X86::ANDri32: |
| 69 | case X86::ORri16: case X86::ORri32: |
| 70 | case X86::XORri16: case X86::XORri32: |
| 71 | assert(MI->getNumOperands() == 3 && "These should all have 3 operands!"); |
| 72 | if (MI->getOperand(2).isImmediate()) { |
| 73 | int Val = MI->getOperand(2).getImmedValue(); |
| 74 | // If the value is the same when signed extended from 8 bits... |
| 75 | if (Val == (signed int)(signed char)Val) { |
| 76 | unsigned Opcode; |
| 77 | switch (MI->getOpcode()) { |
| 78 | default: assert(0 && "Unknown opcode value!"); |
| 79 | case X86::ADDri16: Opcode = X86::ADDri16b; break; |
| 80 | case X86::ADDri32: Opcode = X86::ADDri32b; break; |
| 81 | case X86::SUBri16: Opcode = X86::SUBri16b; break; |
| 82 | case X86::SUBri32: Opcode = X86::SUBri32b; break; |
| 83 | case X86::IMULri16: Opcode = X86::IMULri16b; break; |
| 84 | case X86::IMULri32: Opcode = X86::IMULri32b; break; |
| 85 | case X86::ANDri16: Opcode = X86::ANDri16b; break; |
| 86 | case X86::ANDri32: Opcode = X86::ANDri32b; break; |
| 87 | case X86::ORri16: Opcode = X86::ORri16b; break; |
| 88 | case X86::ORri32: Opcode = X86::ORri32b; break; |
| 89 | case X86::XORri16: Opcode = X86::XORri16b; break; |
| 90 | case X86::XORri32: Opcode = X86::XORri32b; break; |
| 91 | } |
| 92 | unsigned R0 = MI->getOperand(0).getReg(); |
| 93 | unsigned R1 = MI->getOperand(1).getReg(); |
| 94 | *I = BuildMI(Opcode, 2, R0).addReg(R1).addZImm((char)Val); |
| 95 | delete MI; |
| 96 | return true; |
| 97 | } |
| 98 | } |
| 99 | return false; |
| 100 | |
Chris Lattner | a960d95 | 2003-01-13 01:01:59 +0000 | [diff] [blame] | 101 | #if 0 |
| 102 | case X86::MOVir32: Size++; |
| 103 | case X86::MOVir16: Size++; |
| 104 | case X86::MOVir8: |
| 105 | // FIXME: We can only do this transformation if we know that flags are not |
| 106 | // used here, because XOR clobbers the flags! |
| 107 | if (MI->getOperand(1).isImmediate()) { // avoid mov EAX, <value> |
| 108 | int Val = MI->getOperand(1).getImmedValue(); |
| 109 | if (Val == 0) { // mov EAX, 0 -> xor EAX, EAX |
| 110 | static const unsigned Opcode[] ={X86::XORrr8,X86::XORrr16,X86::XORrr32}; |
| 111 | unsigned Reg = MI->getOperand(0).getReg(); |
| 112 | *I = BuildMI(Opcode[Size], 2, Reg).addReg(Reg).addReg(Reg); |
| 113 | delete MI; |
| 114 | return true; |
| 115 | } else if (Val == -1) { // mov EAX, -1 -> or EAX, -1 |
| 116 | // TODO: 'or Reg, -1' has a smaller encoding than 'mov Reg, -1' |
| 117 | } |
| 118 | } |
| 119 | return false; |
| 120 | #endif |
| 121 | case X86::BSWAPr32: // Change bswap EAX, bswap EAX into nothing |
| 122 | if (Next->getOpcode() == X86::BSWAPr32 && |
| 123 | MI->getOperand(0).getReg() == Next->getOperand(0).getReg()) { |
| 124 | I = MBB.erase(MBB.erase(I)); |
| 125 | delete MI; |
| 126 | delete Next; |
| 127 | return true; |
| 128 | } |
| 129 | return false; |
| 130 | default: |
| 131 | return false; |
| 132 | } |
| 133 | } |