|  | Date: Fri, 1 Jun 2001 17:08:44 -0500 (CDT) | 
|  | From: Chris Lattner <sabre@nondot.org> | 
|  | To: Vikram S. Adve <vadve@cs.uiuc.edu> | 
|  | Subject: RE: Interesting: GCC passes | 
|  |  | 
|  | > That is very interesting.  I agree that some of these could be done on LLVM | 
|  | > at link-time, but it is the extra time required that concerns me.  Link-time | 
|  | > optimization is severely time-constrained. | 
|  |  | 
|  | If we were to reimplement any of these optimizations, I assume that we | 
|  | could do them a translation unit at a time, just as GCC does now.  This | 
|  | would lead to a pipeline like this: | 
|  |  | 
|  | Static optimizations, xlation unit at a time: | 
|  | .c --GCC--> .llvm --llvmopt--> .llvm | 
|  |  | 
|  | Link time optimizations: | 
|  | .llvm --llvm-ld--> .llvm --llvm-link-opt--> .llvm | 
|  |  | 
|  | Of course, many optimizations could be shared between llvmopt and | 
|  | llvm-link-opt, but the wouldn't need to be shared...  Thus compile time | 
|  | could be faster, because we are using a "smarter" IR (SSA based). | 
|  |  | 
|  | > BTW, about SGI, "borrowing" SSA-based optimizations from one compiler and | 
|  | > putting it into another is not necessarily easier than re-doing it. | 
|  | > Optimization code is usually heavily tied in to the specific IR they use. | 
|  |  | 
|  | Understood.  The only reason that I brought this up is because SGI's IR is | 
|  | more similar to LLVM than it is different in many respects (SSA based, | 
|  | relatively low level, etc), and could be easily adapted.  Also their | 
|  | optimizations are written in C++ and are actually somewhat | 
|  | structured... of course it would be no walk in the park, but it would be | 
|  | much less time consuming to adapt, say, SSA-PRE than to rewrite it. | 
|  |  | 
|  | > But your larger point is valid that adding SSA based optimizations is | 
|  | > feasible and should be fun.  (Again, link time cost is the issue.) | 
|  |  | 
|  | Assuming linktime cost wasn't an issue, the question is: | 
|  | Does using GCC's backend buy us anything? | 
|  |  | 
|  | > It also occurs to me that GCC is probably doing quite a bit of back-end | 
|  | > optimization (step 16 in your list).  Do you have a breakdown of that? | 
|  |  | 
|  | Not really.  The irritating part of GCC is that it mixes it all up and | 
|  | doesn't have a clean separation of concerns.  A lot of the "back end | 
|  | optimization" happens right along with other data optimizations (ie, CSE | 
|  | of machine specific things). | 
|  |  | 
|  | As far as REAL back end optimizations go, it looks something like this: | 
|  |  | 
|  | 1. Instruction combination: try to make CISCy instructions, if available | 
|  | 2. Register movement: try to get registers in the right places for the | 
|  | architecture to avoid register to register moves.  For example, try to get | 
|  | the first argument of a function to naturally land in %o0 for sparc. | 
|  | 3. Instruction scheduling: 'nuff said :) | 
|  | 4. Register class preferencing: ?? | 
|  | 5. Local register allocation | 
|  | 6. global register allocation | 
|  | 7. Spilling | 
|  | 8. Local regalloc | 
|  | 9. Jump optimization | 
|  | 10. Delay slot scheduling | 
|  | 11. Branch shorting for CISC machines | 
|  | 12. Instruction selection & peephole optimization | 
|  | 13. Debug info output | 
|  |  | 
|  | But none of this would be usable for LLVM anyways, unless we were using | 
|  | GCC as a static compiler. | 
|  |  | 
|  | -Chris | 
|  |  |