blob: c83a2a4a091203b901efa2682a532beccc767afc [file] [log] [blame]
buzbee67bf8852011-08-17 17:51:35 -07001/*
2 * Copyright (C) 2011 The Android Open Source Project
3 *
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at
7 *
8 * http://www.apache.org/licenses/LICENSE-2.0
9 *
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
15 */
16
buzbee67bf8852011-08-17 17:51:35 -070017static const RegLocation badLoc = {kLocDalvikFrame, 0, 0, INVALID_REG,
18 INVALID_REG, INVALID_SREG, 0,
19 kLocDalvikFrame, INVALID_REG, INVALID_REG,
20 INVALID_OFFSET};
21static const RegLocation retLoc = LOC_DALVIK_RETURN_VAL;
22static const RegLocation retLocWide = LOC_DALVIK_RETURN_VAL_WIDE;
23
buzbeedfd3d702011-08-28 12:56:51 -070024/*
25 * Let helper function take care of everything. Will call
26 * Array::AllocFromCode(type_idx, method, count);
27 * Note: AllocFromCode will handle checks for errNegativeArraySize.
28 */
buzbee67bf8852011-08-17 17:51:35 -070029static void genNewArray(CompilationUnit* cUnit, MIR* mir, RegLocation rlDest,
30 RegLocation rlSrc)
31{
buzbeedfd3d702011-08-28 12:56:51 -070032 oatFlushAllRegs(cUnit); /* Everything to home location */
33 loadWordDisp(cUnit, rSELF,
34 OFFSETOF_MEMBER(Thread, pAllocFromCode), rLR);
35 loadCurrMethodDirect(cUnit, r1); // arg1 <- Method*
36 loadConstant(cUnit, r0, mir->dalvikInsn.vC); // arg0 <- type_id
37 loadValueDirectFixed(cUnit, rlSrc, r2); // arg2 <- count
38 opReg(cUnit, kOpBlx, rLR);
39 oatClobberCallRegs(cUnit);
40 RegLocation rlResult = oatGetReturn(cUnit);
41 storeValue(cUnit, rlDest, rlResult);
buzbee67bf8852011-08-17 17:51:35 -070042}
43
44/*
45 * Similar to genNewArray, but with post-allocation initialization.
46 * Verifier guarantees we're dealing with an array class. Current
47 * code throws runtime exception "bad Filled array req" for 'D' and 'J'.
48 * Current code also throws internal unimp if not 'L', '[' or 'I'.
49 */
50static void genFilledNewArray(CompilationUnit* cUnit, MIR* mir, bool isRange)
51{
52 DecodedInstruction* dInsn = &mir->dalvikInsn;
53 int elems;
buzbeedfd3d702011-08-28 12:56:51 -070054 int typeId;
buzbee67bf8852011-08-17 17:51:35 -070055 if (isRange) {
56 elems = dInsn->vA;
buzbeedfd3d702011-08-28 12:56:51 -070057 typeId = dInsn->vB;
buzbee67bf8852011-08-17 17:51:35 -070058 } else {
59 elems = dInsn->vB;
buzbeedfd3d702011-08-28 12:56:51 -070060 typeId = dInsn->vC;
buzbee67bf8852011-08-17 17:51:35 -070061 }
buzbeedfd3d702011-08-28 12:56:51 -070062 oatFlushAllRegs(cUnit); /* Everything to home location */
63 // TODO: Alloc variant that checks types (see header comment) */
64 UNIMPLEMENTED(WARNING) << "Need AllocFromCode variant w/ extra checks";
65 loadWordDisp(cUnit, rSELF,
66 OFFSETOF_MEMBER(Thread, pAllocFromCode), rLR);
67 loadCurrMethodDirect(cUnit, r1); // arg1 <- Method*
68 loadConstant(cUnit, r0, typeId); // arg0 <- type_id
69 loadConstant(cUnit, r2, elems); // arg2 <- count
70 opReg(cUnit, kOpBlx, rLR);
buzbee67bf8852011-08-17 17:51:35 -070071 /*
buzbeedfd3d702011-08-28 12:56:51 -070072 * NOTE: the implicit target for OP_FILLED_NEW_ARRAY is the
73 * return region. Because AllocFromCode placed the new array
74 * in r0, we'll just lock it into place. When debugger support is
75 * added, it may be necessary to additionally copy all return
76 * values to a home location in thread-local storage
buzbee67bf8852011-08-17 17:51:35 -070077 */
buzbee67bf8852011-08-17 17:51:35 -070078 oatLockTemp(cUnit, r0);
buzbeedfd3d702011-08-28 12:56:51 -070079
buzbee67bf8852011-08-17 17:51:35 -070080 // Having a range of 0 is legal
81 if (isRange && (dInsn->vA > 0)) {
82 /*
83 * Bit of ugliness here. We're going generate a mem copy loop
84 * on the register range, but it is possible that some regs
85 * in the range have been promoted. This is unlikely, but
86 * before generating the copy, we'll just force a flush
87 * of any regs in the source range that have been promoted to
88 * home location.
89 */
90 for (unsigned int i = 0; i < dInsn->vA; i++) {
91 RegLocation loc = oatUpdateLoc(cUnit,
92 oatGetSrc(cUnit, mir, i));
93 if (loc.location == kLocPhysReg) {
94 storeBaseDisp(cUnit, rSP, loc.spOffset, loc.lowReg, kWord);
95 }
96 }
97 /*
98 * TUNING note: generated code here could be much improved, but
99 * this is an uncommon operation and isn't especially performance
100 * critical.
101 */
102 int rSrc = oatAllocTemp(cUnit);
103 int rDst = oatAllocTemp(cUnit);
104 int rIdx = oatAllocTemp(cUnit);
105 int rVal = rLR; // Using a lot of temps, rLR is known free here
106 // Set up source pointer
107 RegLocation rlFirst = oatGetSrc(cUnit, mir, 0);
108 opRegRegImm(cUnit, kOpAdd, rSrc, rSP, rlFirst.spOffset);
109 // Set up the target pointer
110 opRegRegImm(cUnit, kOpAdd, rDst, r0,
buzbeec143c552011-08-20 17:38:58 -0700111 Array::DataOffset().Int32Value());
buzbee67bf8852011-08-17 17:51:35 -0700112 // Set up the loop counter (known to be > 0)
113 loadConstant(cUnit, rIdx, dInsn->vA);
114 // Generate the copy loop. Going backwards for convenience
115 ArmLIR* target = newLIR0(cUnit, kArmPseudoTargetLabel);
116 target->defMask = ENCODE_ALL;
117 // Copy next element
118 loadBaseIndexed(cUnit, rSrc, rIdx, rVal, 2, kWord);
119 storeBaseIndexed(cUnit, rDst, rIdx, rVal, 2, kWord);
120 // Use setflags encoding here
121 newLIR3(cUnit, kThumb2SubsRRI12, rIdx, rIdx, 1);
122 ArmLIR* branch = opCondBranch(cUnit, kArmCondNe);
123 branch->generic.target = (LIR*)target;
124 } else if (!isRange) {
125 // TUNING: interleave
126 for (unsigned int i = 0; i < dInsn->vA; i++) {
127 RegLocation rlArg = loadValue(cUnit,
128 oatGetSrc(cUnit, mir, i), kCoreReg);
buzbeec143c552011-08-20 17:38:58 -0700129 storeBaseDisp(cUnit, r0,
130 Array::DataOffset().Int32Value() +
buzbee67bf8852011-08-17 17:51:35 -0700131 i * 4, rlArg.lowReg, kWord);
132 // If the loadValue caused a temp to be allocated, free it
133 if (oatIsTemp(cUnit, rlArg.lowReg)) {
134 oatFreeTemp(cUnit, rlArg.lowReg);
135 }
136 }
137 }
138}
139
140static void genSput(CompilationUnit* cUnit, MIR* mir, RegLocation rlSrc)
141{
buzbeec143c552011-08-20 17:38:58 -0700142 UNIMPLEMENTED(FATAL) << "Must update for new world";
143#if 0
buzbee67bf8852011-08-17 17:51:35 -0700144 int valOffset = OFFSETOF_MEMBER(StaticField, value);
145 int tReg = oatAllocTemp(cUnit);
146 int objHead;
147 bool isVolatile;
148 bool isSputObject;
149 const Method *method = (mir->OptimizationFlags & MIR_CALLEE) ?
150 mir->meta.calleeMethod : cUnit->method;
151 void* fieldPtr = (void*)
152 (method->clazz->pDvmDex->pResFields[mir->dalvikInsn.vB]);
153 Opcode opcode = mir->dalvikInsn.opcode;
154
155 if (fieldPtr == NULL) {
156 // FIXME: need to handle this case for oat();
157 UNIMPLEMENTED(FATAL);
158 }
159
160#if ANDROID_SMP != 0
161 isVolatile = (opcode == OP_SPUT_VOLATILE) ||
162 (opcode == OP_SPUT_VOLATILE_JUMBO) ||
163 (opcode == OP_SPUT_OBJECT_VOLATILE) ||
164 (opcode == OP_SPUT_OBJECT_VOLATILE_JUMBO);
buzbeec143c552011-08-20 17:38:58 -0700165 assert(isVolatile == artIsVolatileField((Field *) fieldPtr));
buzbee67bf8852011-08-17 17:51:35 -0700166#else
buzbeec143c552011-08-20 17:38:58 -0700167 isVolatile = artIsVolatileField((Field *) fieldPtr);
buzbee67bf8852011-08-17 17:51:35 -0700168#endif
169
170 isSputObject = (opcode == OP_SPUT_OBJECT) ||
171 (opcode == OP_SPUT_OBJECT_VOLATILE);
172
173 rlSrc = oatGetSrc(cUnit, mir, 0);
174 rlSrc = loadValue(cUnit, rlSrc, kAnyReg);
175 loadConstant(cUnit, tReg, (int) fieldPtr);
176 if (isSputObject) {
177 objHead = oatAllocTemp(cUnit);
178 loadWordDisp(cUnit, tReg, OFFSETOF_MEMBER(Field, clazz), objHead);
179 }
180 storeWordDisp(cUnit, tReg, valOffset ,rlSrc.lowReg);
181 oatFreeTemp(cUnit, tReg);
182 if (isVolatile) {
183 oatGenMemBarrier(cUnit, kSY);
184 }
185 if (isSputObject) {
186 /* NOTE: marking card based sfield->clazz */
187 markGCCard(cUnit, rlSrc.lowReg, objHead);
188 oatFreeTemp(cUnit, objHead);
189 }
buzbeec143c552011-08-20 17:38:58 -0700190#endif
buzbee67bf8852011-08-17 17:51:35 -0700191}
192
193static void genSputWide(CompilationUnit* cUnit, MIR* mir, RegLocation rlSrc)
194{
buzbeec143c552011-08-20 17:38:58 -0700195 UNIMPLEMENTED(FATAL) << "Must update for new world";
196#if 0
buzbee67bf8852011-08-17 17:51:35 -0700197 int tReg = oatAllocTemp(cUnit);
198 int valOffset = OFFSETOF_MEMBER(StaticField, value);
199 const Method *method = (mir->OptimizationFlags & MIR_CALLEE) ?
200 mir->meta.calleeMethod : cUnit->method;
201 void* fieldPtr = (void*)
202 (method->clazz->pDvmDex->pResFields[mir->dalvikInsn.vB]);
203
204 if (fieldPtr == NULL) {
205 // FIXME: need to handle this case for oat();
206 UNIMPLEMENTED(FATAL);
207 }
208
209 rlSrc = oatGetSrcWide(cUnit, mir, 0, 1);
210 rlSrc = loadValueWide(cUnit, rlSrc, kAnyReg);
211 loadConstant(cUnit, tReg, (int) fieldPtr + valOffset);
212
213 storePair(cUnit, tReg, rlSrc.lowReg, rlSrc.highReg);
buzbeec143c552011-08-20 17:38:58 -0700214#endif
buzbee67bf8852011-08-17 17:51:35 -0700215}
216
217
218
219static void genSgetWide(CompilationUnit* cUnit, MIR* mir,
220 RegLocation rlResult, RegLocation rlDest)
221{
buzbeec143c552011-08-20 17:38:58 -0700222 UNIMPLEMENTED(FATAL) << "Must update for new world";
223#if 0
buzbee67bf8852011-08-17 17:51:35 -0700224 int valOffset = OFFSETOF_MEMBER(StaticField, value);
225 const Method *method = (mir->OptimizationFlags & MIR_CALLEE) ?
226 mir->meta.calleeMethod : cUnit->method;
227 void* fieldPtr = (void*)
228 (method->clazz->pDvmDex->pResFields[mir->dalvikInsn.vB]);
229
230 if (fieldPtr == NULL) {
231 // FIXME: need to handle this case for oat();
232 UNIMPLEMENTED(FATAL);
233 }
234
235 int tReg = oatAllocTemp(cUnit);
236 rlDest = oatGetDestWide(cUnit, mir, 0, 1);
237 rlResult = oatEvalLoc(cUnit, rlDest, kAnyReg, true);
238 loadConstant(cUnit, tReg, (int) fieldPtr + valOffset);
239
240 loadPair(cUnit, tReg, rlResult.lowReg, rlResult.highReg);
241
242 storeValueWide(cUnit, rlDest, rlResult);
buzbeec143c552011-08-20 17:38:58 -0700243#endif
buzbee67bf8852011-08-17 17:51:35 -0700244}
245
246static void genSget(CompilationUnit* cUnit, MIR* mir,
247 RegLocation rlResult, RegLocation rlDest)
248{
buzbeec143c552011-08-20 17:38:58 -0700249 UNIMPLEMENTED(FATAL) << "Must update for new world";
250#if 0
buzbee67bf8852011-08-17 17:51:35 -0700251 int valOffset = OFFSETOF_MEMBER(StaticField, value);
252 int tReg = oatAllocTemp(cUnit);
253 bool isVolatile;
254 const Method *method = cUnit->method;
255 void* fieldPtr = (void*)
256 (method->clazz->pDvmDex->pResFields[mir->dalvikInsn.vB]);
257
258 if (fieldPtr == NULL) {
259 // FIXME: need to handle this case for oat();
260 UNIMPLEMENTED(FATAL);
261 }
262
263 /*
264 * On SMP systems, Dalvik opcodes found to be referencing
265 * volatile fields are rewritten to their _VOLATILE variant.
266 * However, this does not happen on non-SMP systems. The compiler
267 * still needs to know about volatility to avoid unsafe
268 * optimizations so we determine volatility based on either
269 * the opcode or the field access flags.
270 */
271#if ANDROID_SMP != 0
272 Opcode opcode = mir->dalvikInsn.opcode;
273 isVolatile = (opcode == OP_SGET_VOLATILE) ||
274 (opcode == OP_SGET_OBJECT_VOLATILE);
buzbeec143c552011-08-20 17:38:58 -0700275 assert(isVolatile == artIsVolatileField((Field *) fieldPtr));
buzbee67bf8852011-08-17 17:51:35 -0700276#else
buzbeec143c552011-08-20 17:38:58 -0700277 isVolatile = artIsVolatileField((Field *) fieldPtr);
buzbee67bf8852011-08-17 17:51:35 -0700278#endif
279
280 rlDest = oatGetDest(cUnit, mir, 0);
281 rlResult = oatEvalLoc(cUnit, rlDest, kAnyReg, true);
282 loadConstant(cUnit, tReg, (int) fieldPtr + valOffset);
283
284 if (isVolatile) {
285 oatGenMemBarrier(cUnit, kSY);
286 }
287 loadWordDisp(cUnit, tReg, 0, rlResult.lowReg);
288
289 storeValue(cUnit, rlDest, rlResult);
buzbeec143c552011-08-20 17:38:58 -0700290#endif
buzbee67bf8852011-08-17 17:51:35 -0700291}
292
293typedef int (*NextCallInsn)(CompilationUnit*, MIR*, DecodedInstruction*, int);
294
295/*
296 * Bit of a hack here - in leiu of a real scheduling pass,
297 * emit the next instruction in static & direct invoke sequences.
298 */
299static int nextSDCallInsn(CompilationUnit* cUnit, MIR* mir,
300 DecodedInstruction* dInsn, int state)
301{
buzbeec143c552011-08-20 17:38:58 -0700302 UNIMPLEMENTED(FATAL) << "Update with new cache model";
303#if 0
buzbee67bf8852011-08-17 17:51:35 -0700304 switch(state) {
305 case 0: // Get the current Method* [sets r0]
buzbeedfd3d702011-08-28 12:56:51 -0700306 loadCurrMethodDirect(cUnit, r0);
buzbee67bf8852011-08-17 17:51:35 -0700307 break;
308 case 1: // Get the pResMethods pointer [uses r0, sets r0]
buzbeec143c552011-08-20 17:38:58 -0700309 UNIMPLEMENTED(FATAL) << "Update with new cache";
buzbee67bf8852011-08-17 17:51:35 -0700310 loadBaseDisp(cUnit, mir, r0, OFFSETOF_MEMBER(Method, pResMethods),
311 r0, kWord, INVALID_SREG);
312 break;
313 case 2: // Get the target Method* [uses r0, sets r0]
314 loadBaseDisp(cUnit, mir, r0, dInsn->vB * 4, r0,
315 kWord, INVALID_SREG);
316 break;
317 case 3: // Get the target compiled code address [uses r0, sets rLR]
318 loadBaseDisp(cUnit, mir, r0,
319 OFFSETOF_MEMBER(Method, compiledInsns), rLR,
320 kWord, INVALID_SREG);
321 break;
322 default:
323 return -1;
324 }
buzbeec143c552011-08-20 17:38:58 -0700325#endif
buzbee67bf8852011-08-17 17:51:35 -0700326 return state + 1;
327}
328
buzbeec5ef0462011-08-25 18:44:49 -0700329// Slow path static & direct invoke launch sequence
330static int nextSDCallInsnSP(CompilationUnit* cUnit, MIR* mir,
331 DecodedInstruction* dInsn, int state)
332{
333 switch(state) {
334 case 0: // Get the current Method* [sets r0]
buzbeedfd3d702011-08-28 12:56:51 -0700335 loadCurrMethodDirect(cUnit, r0);
buzbeec5ef0462011-08-25 18:44:49 -0700336 break;
337 case 1: // Get the current Method->DeclaringClass() [sets r0]
338 loadBaseDisp(cUnit, mir, r0,
339 OFFSETOF_MEMBER(art::Method, declaring_class_),
340 r0, kWord, INVALID_SREG);
341 break;
342 case 2: // Method->DeclaringClass()->GetDexCache() [sets r0]
343 loadBaseDisp(cUnit, mir, r0,
344 OFFSETOF_MEMBER(art::Class, dex_cache_), r0, kWord,
345 INVALID_SREG);
346 break;
347 case 3: // Method->DeclaringClass()->GetDexCache()->methodsObjectArr
buzbee5cd21802011-08-26 10:40:14 -0700348 loadBaseDisp(cUnit, mir, r0,
349 art::DexCache::MethodsOffset().Int32Value(), r0,
350 kWord, INVALID_SREG);
buzbeec5ef0462011-08-25 18:44:49 -0700351 break;
352 case 4: // Skip past the object header
353 opRegImm(cUnit, kOpAdd, r0, art::Array::DataOffset().Int32Value());
354 break;
355 case 5: // Get the target Method* [uses r0, sets r0]
356 loadBaseDisp(cUnit, mir, r0, dInsn->vB * 4, r0,
357 kWord, INVALID_SREG);
358 break;
359 case 6: // Get the target compiled code address [uses r0, sets rLR]
360 loadBaseDisp(cUnit, mir, r0, art::Method::GetCodeOffset(), rLR,
361 kWord, INVALID_SREG);
362 break;
363 default:
364 return -1;
365 }
366 return state + 1;
367}
368
buzbee67bf8852011-08-17 17:51:35 -0700369/*
370 * Bit of a hack here - in leiu of a real scheduling pass,
371 * emit the next instruction in a virtual invoke sequence.
372 * We can use rLR as a temp prior to target address loading
373 * Note also that we'll load the first argument ("this") into
374 * r1 here rather than the standard loadArgRegs.
375 */
376static int nextVCallInsn(CompilationUnit* cUnit, MIR* mir,
377 DecodedInstruction* dInsn, int state)
378{
buzbeec143c552011-08-20 17:38:58 -0700379 UNIMPLEMENTED(FATAL) << "Update with new cache model";
380#if 0
buzbee67bf8852011-08-17 17:51:35 -0700381 RegLocation rlArg;
382 switch(state) {
383 case 0: // Get the current Method* [set r0]
buzbeedfd3d702011-08-28 12:56:51 -0700384 loadCurrMethodDirect(cUnit, r0);
buzbee67bf8852011-08-17 17:51:35 -0700385 // Load "this" [set r1]
386 rlArg = oatGetSrc(cUnit, mir, 0);
387 loadValueDirectFixed(cUnit, rlArg, r1);
388 break;
389 case 1: // Get the pResMethods pointer [use r0, set r12]
390 loadBaseDisp(cUnit, mir, r0, OFFSETOF_MEMBER(Method, pResMethods),
391 r12, kWord, INVALID_SREG);
392 // Is "this" null? [use r1]
393 genNullCheck(cUnit, oatSSASrc(mir,0), r1,
394 mir->offset, NULL);
395 break;
396 case 2: // Get the base Method* [use r12, set r0]
397 loadBaseDisp(cUnit, mir, r12, dInsn->vB * 4, r0,
398 kWord, INVALID_SREG);
399 // get this->clazz [use r1, set rLR]
400 loadBaseDisp(cUnit, mir, r1, OFFSETOF_MEMBER(Object, clazz), rLR,
401 kWord, INVALID_SREG);
402 break;
403 case 3: // Get the method index [use r0, set r12]
404 loadBaseDisp(cUnit, mir, r0, OFFSETOF_MEMBER(Method, methodIndex),
405 r12, kUnsignedHalf, INVALID_SREG);
406 // get this->clazz->vtable [use rLR, set rLR]
407 loadBaseDisp(cUnit, mir, rLR,
buzbeec143c552011-08-20 17:38:58 -0700408 OFFSETOF_MEMBER(Class, vtable), rLR, kWord,
buzbee67bf8852011-08-17 17:51:35 -0700409 INVALID_SREG);
410 break;
411 case 4: // get target Method* [use rLR, use r12, set r0]
412 loadBaseIndexed(cUnit, rLR, r12, r0, 2, kWord);
413 break;
414 case 5: // Get the target compiled code address [use r0, set rLR]
buzbeec143c552011-08-20 17:38:58 -0700415 UNIMPLEMENTED(FATAL) << "Update with new cache";
buzbee67bf8852011-08-17 17:51:35 -0700416 loadBaseDisp(cUnit, mir, r0, OFFSETOF_MEMBER(Method, compiledInsns),
417 rLR, kWord, INVALID_SREG);
418 break;
419 default:
420 return -1;
421 }
buzbeec143c552011-08-20 17:38:58 -0700422#endif
buzbee67bf8852011-08-17 17:51:35 -0700423 return state + 1;
424}
425
buzbee7b1b86d2011-08-26 18:59:10 -0700426// Slow path sequence for virtual calls
427static int nextVCallInsnSP(CompilationUnit* cUnit, MIR* mir,
428 DecodedInstruction* dInsn, int state)
429{
430 RegLocation rlArg;
431 switch(state) {
432 case 0: // Get the current Method* [sets r0]
buzbeedfd3d702011-08-28 12:56:51 -0700433 loadCurrMethodDirect(cUnit, r0);
buzbee7b1b86d2011-08-26 18:59:10 -0700434 break;
435 case 1: // Get the current Method->DeclaringClass() [uses/sets r0]
436 loadBaseDisp(cUnit, mir, r0,
437 OFFSETOF_MEMBER(art::Method, declaring_class_),
438 r0, kWord, INVALID_SREG);
439 break;
440 case 2: // Method->DeclaringClass()->GetDexCache() [uses/sets r0]
441 loadBaseDisp(cUnit, mir, r0,
442 OFFSETOF_MEMBER(art::Class, dex_cache_), r0, kWord,
443 INVALID_SREG);
444 break;
445 case 3: // ...()->GetDexCache()->methodsObjectArr [uses/sets r0]
446 loadBaseDisp(cUnit, mir, r0,
447 art::DexCache::MethodsOffset().Int32Value(), r0,
448 kWord, INVALID_SREG);
449 // Load "this" [set r1]
450 rlArg = oatGetSrc(cUnit, mir, 0);
451 loadValueDirectFixed(cUnit, rlArg, r1);
452 // Skip past the object header
453 opRegImm(cUnit, kOpAdd, r0, art::Array::DataOffset().Int32Value());
454 break;
455 case 4:
456 // Is "this" null? [use r1]
457 genNullCheck(cUnit, oatSSASrc(mir,0), r1, mir->offset, NULL);
458 // get this->clazz [use r1, set rLR]
459 loadBaseDisp(cUnit, mir, r1, OFFSETOF_MEMBER(Object, klass_), rLR,
460 kWord, INVALID_SREG);
461 // Get the base Method* [uses r0, sets r0]
462 loadBaseDisp(cUnit, mir, r0, dInsn->vB * 4, r0,
463 kWord, INVALID_SREG);
464 // get this->clazz->vtable [use rLR, set rLR]
465 loadBaseDisp(cUnit, mir, rLR,
466 OFFSETOF_MEMBER(Class, vtable_), rLR, kWord,
467 INVALID_SREG);
468 // Get the method index [use r0, set r12]
469 loadBaseDisp(cUnit, mir, r0, OFFSETOF_MEMBER(Method, method_index_),
470 r12, kUnsignedHalf, INVALID_SREG);
471 // Skip past the object header
472 opRegImm(cUnit, kOpAdd, rLR, art::Array::DataOffset().Int32Value());
473 // Get target Method*
474 loadBaseIndexed(cUnit, rLR, r12, r0, 2, kWord);
475 break;
476 case 5: // Get the target compiled code address [uses r0, sets rLR]
477 loadBaseDisp(cUnit, mir, r0, art::Method::GetCodeOffset(), rLR,
478 kWord, INVALID_SREG);
479 break;
480 default:
481 return -1;
482 }
483 return state + 1;
484}
485
buzbee67bf8852011-08-17 17:51:35 -0700486/* Load up to 3 arguments in r1..r3 */
487static int loadArgRegs(CompilationUnit* cUnit, MIR* mir,
488 DecodedInstruction* dInsn, int callState,
489 int *args, NextCallInsn nextCallInsn)
490{
491 for (int i = 0; i < 3; i++) {
492 if (args[i] != INVALID_REG) {
493 RegLocation rlArg = oatGetSrc(cUnit, mir, i);
494 loadValueDirectFixed(cUnit, rlArg, r1 + i);
495 callState = nextCallInsn(cUnit, mir, dInsn, callState);
496 }
497 }
498 return callState;
499}
500
501/*
502 * Interleave launch code for INVOKE_INTERFACE. The target is
503 * identified using artFindInterfaceMethodInCache(class, ref, method, dex)
504 * Note that we'll have to reload "this" following the helper call.
505 *
506 * FIXME: do we need to have artFindInterfaceMethodInCache return
507 * a NULL if not found so we can throw exception here? Otherwise,
508 * may need to pass some additional info to allow the helper function
509 * to throw on its own.
510 */
511static int nextInterfaceCallInsn(CompilationUnit* cUnit, MIR* mir,
512 DecodedInstruction* dInsn, int state)
513{
buzbeec143c552011-08-20 17:38:58 -0700514 UNIMPLEMENTED(FATAL) << "Update with new cache model";
515#if 0
buzbee67bf8852011-08-17 17:51:35 -0700516 RegLocation rlArg;
517 switch(state) {
518 case 0:
519 // Load "this" [set r12]
520 rlArg = oatGetSrc(cUnit, mir, 0);
521 loadValueDirectFixed(cUnit, rlArg, r12);
522 // Get the current Method* [set arg2]
buzbeedfd3d702011-08-28 12:56:51 -0700523 loadCurrMethodDirect(cUnit, r2);
buzbee67bf8852011-08-17 17:51:35 -0700524 // Is "this" null? [use r12]
525 genNullCheck(cUnit, oatSSASrc(mir,0), r12,
526 mir->offset, NULL);
527 // Get curMethod->clazz [set arg3]
528 loadBaseDisp(cUnit, mir, r2, OFFSETOF_MEMBER(Method, clazz),
529 r3, kWord, INVALID_SREG);
530 // Load this->class [usr r12, set arg0]
buzbeec143c552011-08-20 17:38:58 -0700531 loadBaseDisp(cUnit, mir, r12, OFFSETOF_MEMBER(Class, clazz),
buzbee67bf8852011-08-17 17:51:35 -0700532 r3, kWord, INVALID_SREG);
533 // Load address of helper function
534 loadBaseDisp(cUnit, mir, rSELF,
535 OFFSETOF_MEMBER(Thread, pArtFindInterfaceMethodInCache),
536 rLR, kWord, INVALID_SREG);
537 // Get dvmDex
buzbeec143c552011-08-20 17:38:58 -0700538 loadBaseDisp(cUnit, mir, r3, OFFSETOF_MEMBER(Class, pDvmDex),
buzbee67bf8852011-08-17 17:51:35 -0700539 r3, kWord, INVALID_SREG);
540 // Load ref [set arg1]
541 loadConstant(cUnit, r1, dInsn->vB);
542 // Call out to helper, target Method returned in ret0
543 newLIR1(cUnit, kThumbBlxR, rLR);
544 break;
545 case 1: // Get the target compiled code address [use r0, set rLR]
546 loadBaseDisp(cUnit, mir, r0, OFFSETOF_MEMBER(Method, compiledInsns),
547 rLR, kWord, INVALID_SREG);
548 default:
549 return -1;
550 }
buzbeec143c552011-08-20 17:38:58 -0700551#endif
buzbee67bf8852011-08-17 17:51:35 -0700552 return state + 1;
553}
554
555
556/*
557 * Interleave launch code for INVOKE_SUPER. See comments
558 * for nextVCallIns.
559 */
560static int nextSuperCallInsn(CompilationUnit* cUnit, MIR* mir,
561 DecodedInstruction* dInsn, int state)
562{
buzbeec143c552011-08-20 17:38:58 -0700563 UNIMPLEMENTED(FATAL) << "Update with new cache model";
564#if 0
buzbee67bf8852011-08-17 17:51:35 -0700565 RegLocation rlArg;
566 switch(state) {
567 case 0:
568 // Get the current Method* [set r0]
buzbeedfd3d702011-08-28 12:56:51 -0700569 loadCurrMethodDirect(cUnit, r0);
buzbee67bf8852011-08-17 17:51:35 -0700570 // Load "this" [set r1]
571 rlArg = oatGetSrc(cUnit, mir, 0);
572 loadValueDirectFixed(cUnit, rlArg, r1);
573 // Get method->clazz [use r0, set r12]
574 loadBaseDisp(cUnit, mir, r0, OFFSETOF_MEMBER(Method, clazz),
575 r12, kWord, INVALID_SREG);
576 // Get pResmethods [use r0, set rLR]
577 loadBaseDisp(cUnit, mir, r0, OFFSETOF_MEMBER(Method, pResMethods),
578 rLR, kWord, INVALID_SREG);
579 // Get clazz->super [use r12, set r12]
buzbeec143c552011-08-20 17:38:58 -0700580 loadBaseDisp(cUnit, mir, r12, OFFSETOF_MEMBER(Class, super),
buzbee67bf8852011-08-17 17:51:35 -0700581 r12, kWord, INVALID_SREG);
582 // Get base method [use rLR, set r0]
583 loadBaseDisp(cUnit, mir, rLR, dInsn->vB * 4, r0,
584 kWord, INVALID_SREG);
585 // Is "this" null? [use r1]
586 genNullCheck(cUnit, oatSSASrc(mir,0), r1,
587 mir->offset, NULL);
588 // Get methodIndex [use r0, set rLR]
589 loadBaseDisp(cUnit, mir, r0, OFFSETOF_MEMBER(Method, methodIndex),
590 rLR, kUnsignedHalf, INVALID_SREG);
591 // Get vtableCount [use r12, set r0]
592 loadBaseDisp(cUnit, mir, r12,
buzbeec143c552011-08-20 17:38:58 -0700593 OFFSETOF_MEMBER(Class, vtableCount),
buzbee67bf8852011-08-17 17:51:35 -0700594 r0, kWord, INVALID_SREG);
595 // Compare method index w/ vtable count [use r12, use rLR]
596 genRegRegCheck(cUnit, kArmCondGe, rLR, r0, mir->offset, NULL);
597 // get target Method* [use rLR, use r12, set r0]
598 loadBaseIndexed(cUnit, r0, r12, rLR, 2, kWord);
599 case 1: // Get the target compiled code address [use r0, set rLR]
600 loadBaseDisp(cUnit, mir, r0, OFFSETOF_MEMBER(Method, compiledInsns),
601 rLR, kWord, INVALID_SREG);
602 default:
603 return -1;
604 }
buzbeec143c552011-08-20 17:38:58 -0700605#endif
buzbee67bf8852011-08-17 17:51:35 -0700606 return state + 1;
607}
608
609/*
610 * Load up to 5 arguments, the first three of which will be in
611 * r1 .. r3. On entry r0 contains the current method pointer,
612 * and as part of the load sequence, it must be replaced with
613 * the target method pointer. Note, this may also be called
614 * for "range" variants if the number of arguments is 5 or fewer.
615 */
616static int genDalvikArgsNoRange(CompilationUnit* cUnit, MIR* mir,
617 DecodedInstruction* dInsn, int callState,
618 ArmLIR** pcrLabel, bool isRange,
619 NextCallInsn nextCallInsn)
620{
621 RegLocation rlArg;
622 int registerArgs[3];
623
624 /* If no arguments, just return */
625 if (dInsn->vA == 0)
626 return callState;
627
628 oatLockAllTemps(cUnit);
629 callState = nextCallInsn(cUnit, mir, dInsn, callState);
630
631 /*
632 * Load frame arguments arg4 & arg5 first. Coded a little odd to
633 * pre-schedule the method pointer target.
634 */
635 for (unsigned int i=3; i < dInsn->vA; i++) {
636 int reg;
637 int arg = (isRange) ? dInsn->vC + i : i;
638 rlArg = oatUpdateLoc(cUnit, oatGetSrc(cUnit, mir, arg));
639 if (rlArg.location == kLocPhysReg) {
640 reg = rlArg.lowReg;
641 } else {
642 reg = r1;
643 loadValueDirectFixed(cUnit, rlArg, r1);
644 callState = nextCallInsn(cUnit, mir, dInsn, callState);
645 }
646 storeBaseDisp(cUnit, rSP, (i + 1) * 4, reg, kWord);
647 callState = nextCallInsn(cUnit, mir, dInsn, callState);
648 }
649
650 /* Load register arguments r1..r3 */
651 for (unsigned int i = 0; i < 3; i++) {
652 if (i < dInsn->vA)
653 registerArgs[i] = (isRange) ? dInsn->vC + i : i;
654 else
655 registerArgs[i] = INVALID_REG;
656 }
657 callState = loadArgRegs(cUnit, mir, dInsn, callState, registerArgs,
658 nextCallInsn);
659
660 // Load direct & need a "this" null check?
661 if (pcrLabel) {
662 *pcrLabel = genNullCheck(cUnit, oatSSASrc(mir,0), r1,
663 mir->offset, NULL);
664 }
665 return callState;
666}
667
668/*
669 * May have 0+ arguments (also used for jumbo). Note that
670 * source virtual registers may be in physical registers, so may
671 * need to be flushed to home location before copying. This
672 * applies to arg3 and above (see below).
673 *
674 * Two general strategies:
675 * If < 20 arguments
676 * Pass args 3-18 using vldm/vstm block copy
677 * Pass arg0, arg1 & arg2 in r1-r3
678 * If 20+ arguments
679 * Pass args arg19+ using memcpy block copy
680 * Pass arg0, arg1 & arg2 in r1-r3
681 *
682 */
683static int genDalvikArgsRange(CompilationUnit* cUnit, MIR* mir,
684 DecodedInstruction* dInsn, int callState,
685 ArmLIR** pcrLabel, NextCallInsn nextCallInsn)
686{
687 int firstArg = dInsn->vC;
688 int numArgs = dInsn->vA;
689
690 // If we can treat it as non-range (Jumbo ops will use range form)
691 if (numArgs <= 5)
692 return genDalvikArgsNoRange(cUnit, mir, dInsn, callState, pcrLabel,
693 true, nextCallInsn);
694 /*
695 * Make sure range list doesn't span the break between in normal
696 * Dalvik vRegs and the ins.
697 */
698 int highestVreg = oatGetSrc(cUnit, mir, numArgs-1).sRegLow;
buzbeec143c552011-08-20 17:38:58 -0700699 if (highestVreg >= cUnit->method->num_registers_ -
700 cUnit->method->num_ins_) {
buzbee67bf8852011-08-17 17:51:35 -0700701 LOG(FATAL) << "Wide argument spanned locals & args";
702 }
703
704 /*
705 * First load the non-register arguments. Both forms expect all
706 * of the source arguments to be in their home frame location, so
707 * scan the sReg names and flush any that have been promoted to
708 * frame backing storage.
709 */
710 // Scan the rest of the args - if in physReg flush to memory
711 for (int i = 4; i < numArgs; i++) {
712 RegLocation loc = oatUpdateLoc(cUnit,
713 oatGetSrc(cUnit, mir, i));
714 if (loc.location == kLocPhysReg) { // TUNING: if dirty?
715 storeBaseDisp(cUnit, rSP, loc.spOffset, loc.lowReg, kWord);
716 callState = nextCallInsn(cUnit, mir, dInsn, callState);
717 }
718 }
719
720 int startOffset = cUnit->regLocation[mir->ssaRep->uses[3]].spOffset;
721 int outsOffset = 4 /* Method* */ + (3 * 4);
722 if (numArgs >= 20) {
723 // Generate memcpy, but first make sure all of
724 opRegRegImm(cUnit, kOpAdd, r0, rSP, startOffset);
725 opRegRegImm(cUnit, kOpAdd, r1, rSP, outsOffset);
726 loadWordDisp(cUnit, rSELF, OFFSETOF_MEMBER(Thread, pMemcpy), rLR);
727 loadConstant(cUnit, r2, (numArgs - 3) * 4);
728 newLIR1(cUnit, kThumbBlxR, rLR);
729 } else {
730 // Use vldm/vstm pair using r3 as a temp
buzbeec143c552011-08-20 17:38:58 -0700731 int regsLeft = std::min(numArgs - 3, 16);
buzbee67bf8852011-08-17 17:51:35 -0700732 callState = nextCallInsn(cUnit, mir, dInsn, callState);
733 opRegRegImm(cUnit, kOpAdd, r3, rSP, startOffset);
734 newLIR3(cUnit, kThumb2Vldms, r3, fr0 & FP_REG_MASK, regsLeft);
735 callState = nextCallInsn(cUnit, mir, dInsn, callState);
736 opRegRegImm(cUnit, kOpAdd, r3, rSP, 4 /* Method* */ + (3 * 4));
737 callState = nextCallInsn(cUnit, mir, dInsn, callState);
738 newLIR3(cUnit, kThumb2Vstms, r3, fr0 & FP_REG_MASK, regsLeft);
739 callState = nextCallInsn(cUnit, mir, dInsn, callState);
740 }
741
742 // Handle the 1st 3 in r1, r2 & r3
743 for (unsigned int i = 0; i < dInsn->vA && i < 3; i++) {
744 RegLocation loc = oatGetSrc(cUnit, mir, firstArg + i);
745 loadValueDirectFixed(cUnit, loc, r1 + i);
746 callState = nextCallInsn(cUnit, mir, dInsn, callState);
747 }
748
749 // Finally, deal with the register arguments
750 // We'll be using fixed registers here
751 oatLockAllTemps(cUnit);
752 callState = nextCallInsn(cUnit, mir, dInsn, callState);
753 return callState;
754}
755
756static void genInvokeStatic(CompilationUnit* cUnit, MIR* mir)
757{
758 DecodedInstruction* dInsn = &mir->dalvikInsn;
759 int callState = 0;
buzbeec5ef0462011-08-25 18:44:49 -0700760 int fastPath = false; // TODO: set based on resolution results
761
762 NextCallInsn nextCallInsn = fastPath ? nextSDCallInsn : nextSDCallInsnSP;
763
buzbee67bf8852011-08-17 17:51:35 -0700764 if (mir->dalvikInsn.opcode == OP_INVOKE_STATIC) {
765 callState = genDalvikArgsNoRange(cUnit, mir, dInsn, callState, NULL,
buzbeec5ef0462011-08-25 18:44:49 -0700766 false, nextCallInsn);
buzbee67bf8852011-08-17 17:51:35 -0700767 } else {
768 callState = genDalvikArgsRange(cUnit, mir, dInsn, callState, NULL,
buzbeec5ef0462011-08-25 18:44:49 -0700769 nextCallInsn);
buzbee67bf8852011-08-17 17:51:35 -0700770 }
771 // Finish up any of the call sequence not interleaved in arg loading
772 while (callState >= 0) {
buzbeec5ef0462011-08-25 18:44:49 -0700773 callState = nextCallInsn(cUnit, mir, dInsn, callState);
buzbee67bf8852011-08-17 17:51:35 -0700774 }
775 newLIR1(cUnit, kThumbBlxR, rLR);
776}
777
778static void genInvokeDirect(CompilationUnit* cUnit, MIR* mir)
779{
780 DecodedInstruction* dInsn = &mir->dalvikInsn;
781 int callState = 0;
782 ArmLIR* nullCk;
buzbee7b1b86d2011-08-26 18:59:10 -0700783 int fastPath = false; // TODO: set based on resolution results
784
785 NextCallInsn nextCallInsn = fastPath ? nextSDCallInsn : nextSDCallInsnSP;
buzbee67bf8852011-08-17 17:51:35 -0700786 if (mir->dalvikInsn.opcode == OP_INVOKE_DIRECT)
787 callState = genDalvikArgsNoRange(cUnit, mir, dInsn, callState, &nullCk,
buzbee7b1b86d2011-08-26 18:59:10 -0700788 false, nextCallInsn);
buzbee67bf8852011-08-17 17:51:35 -0700789 else
790 callState = genDalvikArgsRange(cUnit, mir, dInsn, callState, &nullCk,
buzbee7b1b86d2011-08-26 18:59:10 -0700791 nextCallInsn);
buzbee67bf8852011-08-17 17:51:35 -0700792 // Finish up any of the call sequence not interleaved in arg loading
793 while (callState >= 0) {
buzbee7b1b86d2011-08-26 18:59:10 -0700794 callState = nextCallInsn(cUnit, mir, dInsn, callState);
buzbee67bf8852011-08-17 17:51:35 -0700795 }
796 newLIR1(cUnit, kThumbBlxR, rLR);
797}
798
799static void genInvokeInterface(CompilationUnit* cUnit, MIR* mir)
800{
801 DecodedInstruction* dInsn = &mir->dalvikInsn;
802 int callState = 0;
803 ArmLIR* nullCk;
804 /* Note: must call nextInterfaceCallInsn() prior to 1st argument load */
805 callState = nextInterfaceCallInsn(cUnit, mir, dInsn, callState);
806 if (mir->dalvikInsn.opcode == OP_INVOKE_INTERFACE)
807 callState = genDalvikArgsNoRange(cUnit, mir, dInsn, callState, &nullCk,
808 false, nextInterfaceCallInsn);
809 else
810 callState = genDalvikArgsRange(cUnit, mir, dInsn, callState, &nullCk,
811 nextInterfaceCallInsn);
812 // Finish up any of the call sequence not interleaved in arg loading
813 while (callState >= 0) {
814 callState = nextInterfaceCallInsn(cUnit, mir, dInsn, callState);
815 }
816 newLIR1(cUnit, kThumbBlxR, rLR);
817}
818
819static void genInvokeSuper(CompilationUnit* cUnit, MIR* mir)
820{
821 DecodedInstruction* dInsn = &mir->dalvikInsn;
822 int callState = 0;
823 ArmLIR* nullCk;
824// FIXME - redundantly loading arg0/r1 ("this")
825 if (mir->dalvikInsn.opcode == OP_INVOKE_SUPER)
826 callState = genDalvikArgsNoRange(cUnit, mir, dInsn, callState, &nullCk,
827 false, nextSuperCallInsn);
828 else
829 callState = genDalvikArgsRange(cUnit, mir, dInsn, callState, &nullCk,
830 nextSuperCallInsn);
831 // Finish up any of the call sequence not interleaved in arg loading
832 while (callState >= 0) {
833 callState = nextSuperCallInsn(cUnit, mir, dInsn, callState);
834 }
835 newLIR1(cUnit, kThumbBlxR, rLR);
836}
837
838static void genInvokeVirtual(CompilationUnit* cUnit, MIR* mir)
839{
840 DecodedInstruction* dInsn = &mir->dalvikInsn;
841 int callState = 0;
842 ArmLIR* nullCk;
buzbee7b1b86d2011-08-26 18:59:10 -0700843 int fastPath = false; // TODO: set based on resolution results
844
845 NextCallInsn nextCallInsn = fastPath ? nextVCallInsn : nextVCallInsnSP;
846 // TODO - redundantly loading arg0/r1 ("this")
buzbee67bf8852011-08-17 17:51:35 -0700847 if (mir->dalvikInsn.opcode == OP_INVOKE_VIRTUAL)
848 callState = genDalvikArgsNoRange(cUnit, mir, dInsn, callState, &nullCk,
buzbee7b1b86d2011-08-26 18:59:10 -0700849 false, nextCallInsn);
buzbee67bf8852011-08-17 17:51:35 -0700850 else
851 callState = genDalvikArgsRange(cUnit, mir, dInsn, callState, &nullCk,
buzbee7b1b86d2011-08-26 18:59:10 -0700852 nextCallInsn);
buzbee67bf8852011-08-17 17:51:35 -0700853 // Finish up any of the call sequence not interleaved in arg loading
854 while (callState >= 0) {
buzbee7b1b86d2011-08-26 18:59:10 -0700855 callState = nextCallInsn(cUnit, mir, dInsn, callState);
buzbee67bf8852011-08-17 17:51:35 -0700856 }
857 newLIR1(cUnit, kThumbBlxR, rLR);
858}
859
860// TODO: break out the case handlers. Might make it easier to support x86
861static bool compileDalvikInstruction(CompilationUnit* cUnit, MIR* mir,
862 BasicBlock* bb, ArmLIR* labelList)
863{
864 bool res = false; // Assume success
865 RegLocation rlSrc[3];
866 RegLocation rlDest = badLoc;
867 RegLocation rlResult = badLoc;
868 Opcode opcode = mir->dalvikInsn.opcode;
869
870 /* Prep Src and Dest locations */
871 int nextSreg = 0;
872 int nextLoc = 0;
873 int attrs = oatDataFlowAttributes[opcode];
874 rlSrc[0] = rlSrc[1] = rlSrc[2] = badLoc;
875 if (attrs & DF_UA) {
876 rlSrc[nextLoc++] = oatGetSrc(cUnit, mir, nextSreg);
877 nextSreg++;
878 } else if (attrs & DF_UA_WIDE) {
879 rlSrc[nextLoc++] = oatGetSrcWide(cUnit, mir, nextSreg,
880 nextSreg + 1);
881 nextSreg+= 2;
882 }
883 if (attrs & DF_UB) {
884 rlSrc[nextLoc++] = oatGetSrc(cUnit, mir, nextSreg);
885 nextSreg++;
886 } else if (attrs & DF_UB_WIDE) {
887 rlSrc[nextLoc++] = oatGetSrcWide(cUnit, mir, nextSreg,
888 nextSreg + 1);
889 nextSreg+= 2;
890 }
891 if (attrs & DF_UC) {
892 rlSrc[nextLoc++] = oatGetSrc(cUnit, mir, nextSreg);
893 } else if (attrs & DF_UC_WIDE) {
894 rlSrc[nextLoc++] = oatGetSrcWide(cUnit, mir, nextSreg,
895 nextSreg + 1);
896 }
897 if (attrs & DF_DA) {
898 rlDest = oatGetDest(cUnit, mir, 0);
899 } else if (attrs & DF_DA_WIDE) {
900 rlDest = oatGetDestWide(cUnit, mir, 0, 1);
901 }
902
903 switch(opcode) {
904 case OP_NOP:
905 break;
906
907 case OP_MOVE_EXCEPTION:
908 int exOffset;
909 int resetReg;
buzbeec143c552011-08-20 17:38:58 -0700910 exOffset = Thread::ExceptionOffset().Int32Value();
buzbee67bf8852011-08-17 17:51:35 -0700911 resetReg = oatAllocTemp(cUnit);
912 rlResult = oatEvalLoc(cUnit, rlDest, kCoreReg, true);
913 loadWordDisp(cUnit, rSELF, exOffset, rlResult.lowReg);
914 loadConstant(cUnit, resetReg, 0);
915 storeWordDisp(cUnit, rSELF, exOffset, resetReg);
916 storeValue(cUnit, rlDest, rlResult);
917 break;
918
919 case OP_RETURN_VOID:
920 break;
921
922 case OP_RETURN:
923 case OP_RETURN_OBJECT:
924 storeValue(cUnit, retLoc, rlSrc[0]);
925 break;
926
927 case OP_RETURN_WIDE:
928 rlDest = retLocWide;
929 rlDest.fp = rlSrc[0].fp;
930 storeValueWide(cUnit, rlDest, rlSrc[0]);
931 break;
932
933 case OP_MOVE_RESULT_WIDE:
934 if (mir->OptimizationFlags & MIR_INLINED)
935 break; // Nop - combined w/ previous invoke
936 /*
937 * Somewhat hacky here. Because we're now passing
938 * return values in registers, we have to let the
939 * register allocation utilities know that the return
940 * registers are live and may not be used for address
941 * formation in storeValueWide.
942 */
943 assert(retLocWide.lowReg == r0);
944 assert(retLocWide.lowReg == r1);
945 oatLockTemp(cUnit, retLocWide.lowReg);
946 oatLockTemp(cUnit, retLocWide.highReg);
947 storeValueWide(cUnit, rlDest, retLocWide);
948 oatFreeTemp(cUnit, retLocWide.lowReg);
949 oatFreeTemp(cUnit, retLocWide.highReg);
950 break;
951
952 case OP_MOVE_RESULT:
953 case OP_MOVE_RESULT_OBJECT:
954 if (mir->OptimizationFlags & MIR_INLINED)
955 break; // Nop - combined w/ previous invoke
956 /* See comment for OP_MOVE_RESULT_WIDE */
957 assert(retLoc.lowReg == r0);
958 oatLockTemp(cUnit, retLoc.lowReg);
959 storeValue(cUnit, rlDest, retLoc);
960 oatFreeTemp(cUnit, retLoc.lowReg);
961 break;
962
963 case OP_MOVE:
964 case OP_MOVE_OBJECT:
965 case OP_MOVE_16:
966 case OP_MOVE_OBJECT_16:
967 case OP_MOVE_FROM16:
968 case OP_MOVE_OBJECT_FROM16:
969 storeValue(cUnit, rlDest, rlSrc[0]);
970 break;
971
972 case OP_MOVE_WIDE:
973 case OP_MOVE_WIDE_16:
974 case OP_MOVE_WIDE_FROM16:
975 storeValueWide(cUnit, rlDest, rlSrc[0]);
976 break;
977
978 case OP_CONST:
979 case OP_CONST_4:
980 case OP_CONST_16:
981 rlResult = oatEvalLoc(cUnit, rlDest, kAnyReg, true);
982 loadConstantNoClobber(cUnit, rlResult.lowReg, mir->dalvikInsn.vB);
983 storeValue(cUnit, rlDest, rlResult);
984 break;
985
986 case OP_CONST_HIGH16:
987 rlResult = oatEvalLoc(cUnit, rlDest, kAnyReg, true);
988 loadConstantNoClobber(cUnit, rlResult.lowReg,
989 mir->dalvikInsn.vB << 16);
990 storeValue(cUnit, rlDest, rlResult);
991 break;
992
993 case OP_CONST_WIDE_16:
994 case OP_CONST_WIDE_32:
995 rlResult = oatEvalLoc(cUnit, rlDest, kCoreReg, true);
996 loadConstantNoClobber(cUnit, rlResult.lowReg, mir->dalvikInsn.vB);
997 //TUNING: do high separately to avoid load dependency
998 opRegRegImm(cUnit, kOpAsr, rlResult.highReg, rlResult.lowReg, 31);
999 storeValueWide(cUnit, rlDest, rlResult);
1000 break;
1001
1002 case OP_CONST_WIDE:
1003 rlResult = oatEvalLoc(cUnit, rlDest, kAnyReg, true);
1004 loadConstantValueWide(cUnit, rlResult.lowReg, rlResult.highReg,
buzbee54330722011-08-23 16:46:55 -07001005 mir->dalvikInsn.vB_wide & 0xffffffff,
1006 (mir->dalvikInsn.vB_wide >> 32) & 0xffffffff);
buzbee3ea4ec52011-08-22 17:37:19 -07001007 storeValueWide(cUnit, rlDest, rlResult);
buzbee67bf8852011-08-17 17:51:35 -07001008 break;
1009
1010 case OP_CONST_WIDE_HIGH16:
1011 rlResult = oatEvalLoc(cUnit, rlDest, kAnyReg, true);
1012 loadConstantValueWide(cUnit, rlResult.lowReg, rlResult.highReg,
1013 0, mir->dalvikInsn.vB << 16);
buzbee7b1b86d2011-08-26 18:59:10 -07001014 storeValueWide(cUnit, rlDest, rlResult);
buzbee67bf8852011-08-17 17:51:35 -07001015 break;
1016
1017 case OP_MONITOR_ENTER:
1018 genMonitorEnter(cUnit, mir, rlSrc[0]);
1019 break;
1020
1021 case OP_MONITOR_EXIT:
1022 genMonitorExit(cUnit, mir, rlSrc[0]);
1023 break;
1024
1025 case OP_CHECK_CAST:
1026 genCheckCast(cUnit, mir, rlSrc[0]);
1027 break;
1028
1029 case OP_INSTANCE_OF:
1030 genInstanceof(cUnit, mir, rlDest, rlSrc[0]);
1031 break;
1032
1033 case OP_NEW_INSTANCE:
1034 genNewInstance(cUnit, mir, rlDest);
1035 break;
1036
1037 case OP_THROW:
1038 genThrow(cUnit, mir, rlSrc[0]);
1039 break;
1040
1041 case OP_ARRAY_LENGTH:
1042 int lenOffset;
buzbeec143c552011-08-20 17:38:58 -07001043 lenOffset = Array::LengthOffset().Int32Value();
buzbee7b1b86d2011-08-26 18:59:10 -07001044 rlSrc[0] = loadValue(cUnit, rlSrc[0], kCoreReg);
buzbee67bf8852011-08-17 17:51:35 -07001045 genNullCheck(cUnit, rlSrc[0].sRegLow, rlSrc[0].lowReg,
1046 mir->offset, NULL);
1047 rlResult = oatEvalLoc(cUnit, rlDest, kCoreReg, true);
1048 loadWordDisp(cUnit, rlSrc[0].lowReg, lenOffset,
1049 rlResult.lowReg);
1050 storeValue(cUnit, rlDest, rlResult);
1051 break;
1052
1053 case OP_CONST_STRING:
1054 case OP_CONST_STRING_JUMBO:
1055 genConstString(cUnit, mir, rlDest, rlSrc[0]);
1056 break;
1057
1058 case OP_CONST_CLASS:
1059 genConstClass(cUnit, mir, rlDest, rlSrc[0]);
1060 break;
1061
1062 case OP_FILL_ARRAY_DATA:
1063 genFillArrayData(cUnit, mir, rlSrc[0]);
1064 break;
1065
1066 case OP_FILLED_NEW_ARRAY:
1067 genFilledNewArray(cUnit, mir, false /* not range */);
1068 break;
1069
1070 case OP_FILLED_NEW_ARRAY_RANGE:
1071 genFilledNewArray(cUnit, mir, true /* range */);
1072 break;
1073
1074 case OP_NEW_ARRAY:
1075 genNewArray(cUnit, mir, rlDest, rlSrc[0]);
1076 break;
1077
1078 case OP_GOTO:
1079 case OP_GOTO_16:
1080 case OP_GOTO_32:
1081 // TUNING: add MIR flag to disable when unnecessary
1082 bool backwardBranch;
1083 backwardBranch = (bb->taken->startOffset <= mir->offset);
1084 if (backwardBranch) {
1085 genSuspendPoll(cUnit, mir);
1086 }
1087 genUnconditionalBranch(cUnit, &labelList[bb->taken->id]);
1088 break;
1089
1090 case OP_PACKED_SWITCH:
1091 genPackedSwitch(cUnit, mir, rlSrc[0]);
1092 break;
1093
1094 case OP_SPARSE_SWITCH:
1095 genSparseSwitch(cUnit, mir, rlSrc[0]);
1096 break;
1097
1098 case OP_CMPL_FLOAT:
1099 case OP_CMPG_FLOAT:
1100 case OP_CMPL_DOUBLE:
1101 case OP_CMPG_DOUBLE:
1102 res = genCmpFP(cUnit, mir, rlDest, rlSrc[0], rlSrc[1]);
1103 break;
1104
1105 case OP_CMP_LONG:
1106 genCmpLong(cUnit, mir, rlDest, rlSrc[0], rlSrc[1]);
1107 break;
1108
1109 case OP_IF_EQ:
1110 case OP_IF_NE:
1111 case OP_IF_LT:
1112 case OP_IF_GE:
1113 case OP_IF_GT:
1114 case OP_IF_LE: {
1115 bool backwardBranch;
1116 ArmConditionCode cond;
1117 backwardBranch = (bb->taken->startOffset <= mir->offset);
1118 if (backwardBranch) {
1119 genSuspendPoll(cUnit, mir);
1120 }
1121 rlSrc[0] = loadValue(cUnit, rlSrc[0], kCoreReg);
1122 rlSrc[1] = loadValue(cUnit, rlSrc[1], kCoreReg);
1123 opRegReg(cUnit, kOpCmp, rlSrc[0].lowReg, rlSrc[1].lowReg);
1124 switch(opcode) {
1125 case OP_IF_EQ:
1126 cond = kArmCondEq;
1127 break;
1128 case OP_IF_NE:
1129 cond = kArmCondNe;
1130 break;
1131 case OP_IF_LT:
1132 cond = kArmCondLt;
1133 break;
1134 case OP_IF_GE:
1135 cond = kArmCondGe;
1136 break;
1137 case OP_IF_GT:
1138 cond = kArmCondGt;
1139 break;
1140 case OP_IF_LE:
1141 cond = kArmCondLe;
1142 break;
1143 default:
1144 cond = (ArmConditionCode)0;
1145 LOG(FATAL) << "Unexpected opcode " << (int)opcode;
1146 }
1147 genConditionalBranch(cUnit, cond, &labelList[bb->taken->id]);
1148 genUnconditionalBranch(cUnit, &labelList[bb->fallThrough->id]);
1149 break;
1150 }
1151
1152 case OP_IF_EQZ:
1153 case OP_IF_NEZ:
1154 case OP_IF_LTZ:
1155 case OP_IF_GEZ:
1156 case OP_IF_GTZ:
1157 case OP_IF_LEZ: {
1158 bool backwardBranch;
1159 ArmConditionCode cond;
1160 backwardBranch = (bb->taken->startOffset <= mir->offset);
1161 if (backwardBranch) {
1162 genSuspendPoll(cUnit, mir);
1163 }
1164 rlSrc[0] = loadValue(cUnit, rlSrc[0], kCoreReg);
1165 opRegImm(cUnit, kOpCmp, rlSrc[0].lowReg, 0);
1166 switch(opcode) {
1167 case OP_IF_EQZ:
1168 cond = kArmCondEq;
1169 break;
1170 case OP_IF_NEZ:
1171 cond = kArmCondNe;
1172 break;
1173 case OP_IF_LTZ:
1174 cond = kArmCondLt;
1175 break;
1176 case OP_IF_GEZ:
1177 cond = kArmCondGe;
1178 break;
1179 case OP_IF_GTZ:
1180 cond = kArmCondGt;
1181 break;
1182 case OP_IF_LEZ:
1183 cond = kArmCondLe;
1184 break;
1185 default:
1186 cond = (ArmConditionCode)0;
1187 LOG(FATAL) << "Unexpected opcode " << (int)opcode;
1188 }
1189 genConditionalBranch(cUnit, cond, &labelList[bb->taken->id]);
1190 genUnconditionalBranch(cUnit, &labelList[bb->fallThrough->id]);
1191 break;
1192 }
1193
1194 case OP_AGET_WIDE:
1195 genArrayGet(cUnit, mir, kLong, rlSrc[0], rlSrc[1], rlDest, 3);
1196 break;
1197 case OP_AGET:
1198 case OP_AGET_OBJECT:
1199 genArrayGet(cUnit, mir, kWord, rlSrc[0], rlSrc[1], rlDest, 2);
1200 break;
1201 case OP_AGET_BOOLEAN:
1202 genArrayGet(cUnit, mir, kUnsignedByte, rlSrc[0], rlSrc[1],
1203 rlDest, 0);
1204 break;
1205 case OP_AGET_BYTE:
1206 genArrayGet(cUnit, mir, kSignedByte, rlSrc[0], rlSrc[1], rlDest, 0);
1207 break;
1208 case OP_AGET_CHAR:
1209 genArrayGet(cUnit, mir, kUnsignedHalf, rlSrc[0], rlSrc[1],
1210 rlDest, 1);
1211 break;
1212 case OP_AGET_SHORT:
1213 genArrayGet(cUnit, mir, kSignedHalf, rlSrc[0], rlSrc[1], rlDest, 1);
1214 break;
1215 case OP_APUT_WIDE:
1216 genArrayPut(cUnit, mir, kLong, rlSrc[1], rlSrc[2], rlSrc[0], 3);
1217 break;
1218 case OP_APUT:
1219 genArrayPut(cUnit, mir, kWord, rlSrc[1], rlSrc[2], rlSrc[0], 2);
1220 break;
1221 case OP_APUT_OBJECT:
buzbeec143c552011-08-20 17:38:58 -07001222 genArrayPut(cUnit, mir, rlSrc[1], rlSrc[2], rlSrc[0], 2);
buzbee67bf8852011-08-17 17:51:35 -07001223 break;
1224 case OP_APUT_SHORT:
1225 case OP_APUT_CHAR:
1226 genArrayPut(cUnit, mir, kUnsignedHalf, rlSrc[1], rlSrc[2],
1227 rlSrc[0], 1);
1228 break;
1229 case OP_APUT_BYTE:
1230 case OP_APUT_BOOLEAN:
1231 genArrayPut(cUnit, mir, kUnsignedByte, rlSrc[1], rlSrc[2],
1232 rlSrc[0], 0);
1233 break;
1234
1235 case OP_IGET_WIDE:
1236 case OP_IGET_WIDE_VOLATILE:
1237 genIGetWideX(cUnit, mir, rlDest, rlSrc[0]);
1238 break;
1239
1240 case OP_IGET:
1241 case OP_IGET_VOLATILE:
1242 case OP_IGET_OBJECT:
1243 case OP_IGET_OBJECT_VOLATILE:
1244 genIGetX(cUnit, mir, kWord, rlDest, rlSrc[0]);
1245 break;
1246
1247 case OP_IGET_BOOLEAN:
1248 case OP_IGET_BYTE:
1249 genIGetX(cUnit, mir, kUnsignedByte, rlDest, rlSrc[0]);
1250 break;
1251
1252 case OP_IGET_CHAR:
1253 genIGetX(cUnit, mir, kUnsignedHalf, rlDest, rlSrc[0]);
1254 break;
1255
1256 case OP_IGET_SHORT:
1257 genIGetX(cUnit, mir, kSignedHalf, rlDest, rlSrc[0]);
1258 break;
1259
1260 case OP_IPUT_WIDE:
1261 case OP_IPUT_WIDE_VOLATILE:
1262 genIPutWideX(cUnit, mir, rlSrc[0], rlSrc[1]);
1263 break;
1264
1265 case OP_IPUT_OBJECT:
1266 case OP_IPUT_OBJECT_VOLATILE:
1267 genIPutX(cUnit, mir, kWord, rlSrc[0], rlSrc[1], true);
1268 break;
1269
1270 case OP_IPUT:
1271 case OP_IPUT_VOLATILE:
1272 genIPutX(cUnit, mir, kWord, rlSrc[0], rlSrc[1], false);
1273 break;
1274
1275 case OP_IPUT_BOOLEAN:
1276 case OP_IPUT_BYTE:
1277 genIPutX(cUnit, mir, kUnsignedByte, rlSrc[0], rlSrc[1], false);
1278 break;
1279
1280 case OP_IPUT_CHAR:
1281 genIPutX(cUnit, mir, kUnsignedHalf, rlSrc[0], rlSrc[1], false);
1282 break;
1283
1284 case OP_IPUT_SHORT:
1285 genIPutX(cUnit, mir, kSignedHalf, rlSrc[0], rlSrc[1], false);
1286 break;
1287
1288 case OP_SGET:
1289 case OP_SGET_OBJECT:
1290 case OP_SGET_BOOLEAN:
1291 case OP_SGET_BYTE:
1292 case OP_SGET_CHAR:
1293 case OP_SGET_SHORT:
1294 genSget(cUnit, mir, rlResult, rlDest);
1295 break;
1296
1297 case OP_SGET_WIDE:
1298 genSgetWide(cUnit, mir, rlResult, rlDest);
1299 break;
1300
1301 case OP_SPUT:
1302 case OP_SPUT_OBJECT:
1303 case OP_SPUT_BOOLEAN:
1304 case OP_SPUT_BYTE:
1305 case OP_SPUT_CHAR:
1306 case OP_SPUT_SHORT:
1307 genSput(cUnit, mir, rlSrc[0]);
1308 break;
1309
1310 case OP_SPUT_WIDE:
1311 genSputWide(cUnit, mir, rlSrc[0]);
1312 break;
1313
1314 case OP_INVOKE_STATIC_RANGE:
1315 case OP_INVOKE_STATIC:
1316 genInvokeStatic(cUnit, mir);
1317 break;
1318
1319 case OP_INVOKE_DIRECT:
1320 case OP_INVOKE_DIRECT_RANGE:
1321 genInvokeDirect(cUnit, mir);
1322 break;
1323
1324 case OP_INVOKE_VIRTUAL:
1325 case OP_INVOKE_VIRTUAL_RANGE:
1326 genInvokeVirtual(cUnit, mir);
1327 break;
1328
1329 case OP_INVOKE_SUPER:
1330 case OP_INVOKE_SUPER_RANGE:
1331 genInvokeSuper(cUnit, mir);
1332 break;
1333
1334 case OP_INVOKE_INTERFACE:
1335 case OP_INVOKE_INTERFACE_RANGE:
1336 genInvokeInterface(cUnit, mir);
1337 break;
1338
1339 case OP_NEG_INT:
1340 case OP_NOT_INT:
1341 res = genArithOpInt(cUnit, mir, rlDest, rlSrc[0], rlSrc[0]);
1342 break;
1343
1344 case OP_NEG_LONG:
1345 case OP_NOT_LONG:
1346 res = genArithOpLong(cUnit, mir, rlDest, rlSrc[0], rlSrc[0]);
1347 break;
1348
1349 case OP_NEG_FLOAT:
1350 res = genArithOpFloat(cUnit, mir, rlDest, rlSrc[0], rlSrc[0]);
1351 break;
1352
1353 case OP_NEG_DOUBLE:
1354 res = genArithOpDouble(cUnit, mir, rlDest, rlSrc[0], rlSrc[0]);
1355 break;
1356
1357 case OP_INT_TO_LONG:
1358 rlResult = oatEvalLoc(cUnit, rlDest, kCoreReg, true);
1359 if (rlSrc[0].location == kLocPhysReg) {
1360 genRegCopy(cUnit, rlResult.lowReg, rlSrc[0].lowReg);
1361 } else {
1362 loadValueDirect(cUnit, rlSrc[0], rlResult.lowReg);
1363 }
1364 opRegRegImm(cUnit, kOpAsr, rlResult.highReg,
1365 rlResult.lowReg, 31);
1366 storeValueWide(cUnit, rlDest, rlResult);
1367 break;
1368
1369 case OP_LONG_TO_INT:
1370 rlSrc[0] = oatUpdateLocWide(cUnit, rlSrc[0]);
1371 rlSrc[0] = oatWideToNarrow(cUnit, rlSrc[0]);
1372 storeValue(cUnit, rlDest, rlSrc[0]);
1373 break;
1374
1375 case OP_INT_TO_BYTE:
1376 rlSrc[0] = loadValue(cUnit, rlSrc[0], kCoreReg);
1377 rlResult = oatEvalLoc(cUnit, rlDest, kCoreReg, true);
1378 opRegReg(cUnit, kOp2Byte, rlResult.lowReg, rlSrc[0].lowReg);
1379 storeValue(cUnit, rlDest, rlResult);
1380 break;
1381
1382 case OP_INT_TO_SHORT:
1383 rlSrc[0] = loadValue(cUnit, rlSrc[0], kCoreReg);
1384 rlResult = oatEvalLoc(cUnit, rlDest, kCoreReg, true);
1385 opRegReg(cUnit, kOp2Short, rlResult.lowReg, rlSrc[0].lowReg);
1386 storeValue(cUnit, rlDest, rlResult);
1387 break;
1388
1389 case OP_INT_TO_CHAR:
1390 rlSrc[0] = loadValue(cUnit, rlSrc[0], kCoreReg);
1391 rlResult = oatEvalLoc(cUnit, rlDest, kCoreReg, true);
1392 opRegReg(cUnit, kOp2Char, rlResult.lowReg, rlSrc[0].lowReg);
1393 storeValue(cUnit, rlDest, rlResult);
1394 break;
1395
1396 case OP_INT_TO_FLOAT:
1397 case OP_INT_TO_DOUBLE:
1398 case OP_LONG_TO_FLOAT:
1399 case OP_LONG_TO_DOUBLE:
1400 case OP_FLOAT_TO_INT:
1401 case OP_FLOAT_TO_LONG:
1402 case OP_FLOAT_TO_DOUBLE:
1403 case OP_DOUBLE_TO_INT:
1404 case OP_DOUBLE_TO_LONG:
1405 case OP_DOUBLE_TO_FLOAT:
1406 genConversion(cUnit, mir);
1407 break;
1408
1409 case OP_ADD_INT:
1410 case OP_SUB_INT:
1411 case OP_MUL_INT:
1412 case OP_DIV_INT:
1413 case OP_REM_INT:
1414 case OP_AND_INT:
1415 case OP_OR_INT:
1416 case OP_XOR_INT:
1417 case OP_SHL_INT:
1418 case OP_SHR_INT:
1419 case OP_USHR_INT:
1420 case OP_ADD_INT_2ADDR:
1421 case OP_SUB_INT_2ADDR:
1422 case OP_MUL_INT_2ADDR:
1423 case OP_DIV_INT_2ADDR:
1424 case OP_REM_INT_2ADDR:
1425 case OP_AND_INT_2ADDR:
1426 case OP_OR_INT_2ADDR:
1427 case OP_XOR_INT_2ADDR:
1428 case OP_SHL_INT_2ADDR:
1429 case OP_SHR_INT_2ADDR:
1430 case OP_USHR_INT_2ADDR:
1431 genArithOpInt(cUnit, mir, rlDest, rlSrc[0], rlSrc[1]);
1432 break;
1433
1434 case OP_ADD_LONG:
1435 case OP_SUB_LONG:
1436 case OP_MUL_LONG:
1437 case OP_DIV_LONG:
1438 case OP_REM_LONG:
1439 case OP_AND_LONG:
1440 case OP_OR_LONG:
1441 case OP_XOR_LONG:
1442 case OP_ADD_LONG_2ADDR:
1443 case OP_SUB_LONG_2ADDR:
1444 case OP_MUL_LONG_2ADDR:
1445 case OP_DIV_LONG_2ADDR:
1446 case OP_REM_LONG_2ADDR:
1447 case OP_AND_LONG_2ADDR:
1448 case OP_OR_LONG_2ADDR:
1449 case OP_XOR_LONG_2ADDR:
1450 genArithOpLong(cUnit, mir, rlDest, rlSrc[0], rlSrc[1]);
1451 break;
1452
buzbee67bf8852011-08-17 17:51:35 -07001453 case OP_SHL_LONG:
1454 case OP_SHR_LONG:
1455 case OP_USHR_LONG:
buzbeee6d61962011-08-27 11:58:19 -07001456 case OP_SHL_LONG_2ADDR:
1457 case OP_SHR_LONG_2ADDR:
1458 case OP_USHR_LONG_2ADDR:
buzbee67bf8852011-08-17 17:51:35 -07001459 genShiftOpLong(cUnit,mir, rlDest, rlSrc[0], rlSrc[1]);
1460 break;
1461
1462 case OP_ADD_FLOAT:
1463 case OP_SUB_FLOAT:
1464 case OP_MUL_FLOAT:
1465 case OP_DIV_FLOAT:
1466 case OP_REM_FLOAT:
1467 case OP_ADD_FLOAT_2ADDR:
1468 case OP_SUB_FLOAT_2ADDR:
1469 case OP_MUL_FLOAT_2ADDR:
1470 case OP_DIV_FLOAT_2ADDR:
1471 case OP_REM_FLOAT_2ADDR:
1472 genArithOpFloat(cUnit, mir, rlDest, rlSrc[0], rlSrc[1]);
1473 break;
1474
1475 case OP_ADD_DOUBLE:
1476 case OP_SUB_DOUBLE:
1477 case OP_MUL_DOUBLE:
1478 case OP_DIV_DOUBLE:
1479 case OP_REM_DOUBLE:
1480 case OP_ADD_DOUBLE_2ADDR:
1481 case OP_SUB_DOUBLE_2ADDR:
1482 case OP_MUL_DOUBLE_2ADDR:
1483 case OP_DIV_DOUBLE_2ADDR:
1484 case OP_REM_DOUBLE_2ADDR:
1485 genArithOpDouble(cUnit, mir, rlDest, rlSrc[0], rlSrc[1]);
1486 break;
1487
1488 case OP_RSUB_INT:
1489 case OP_ADD_INT_LIT16:
1490 case OP_MUL_INT_LIT16:
1491 case OP_DIV_INT_LIT16:
1492 case OP_REM_INT_LIT16:
1493 case OP_AND_INT_LIT16:
1494 case OP_OR_INT_LIT16:
1495 case OP_XOR_INT_LIT16:
1496 case OP_ADD_INT_LIT8:
1497 case OP_RSUB_INT_LIT8:
1498 case OP_MUL_INT_LIT8:
1499 case OP_DIV_INT_LIT8:
1500 case OP_REM_INT_LIT8:
1501 case OP_AND_INT_LIT8:
1502 case OP_OR_INT_LIT8:
1503 case OP_XOR_INT_LIT8:
1504 case OP_SHL_INT_LIT8:
1505 case OP_SHR_INT_LIT8:
1506 case OP_USHR_INT_LIT8:
1507 genArithOpIntLit(cUnit, mir, rlDest, rlSrc[0], mir->dalvikInsn.vC);
1508 break;
1509
1510 default:
1511 res = true;
1512 }
1513 return res;
1514}
1515
1516static const char *extendedMIROpNames[kMirOpLast - kMirOpFirst] = {
1517 "kMirOpPhi",
1518 "kMirOpNullNRangeUpCheck",
1519 "kMirOpNullNRangeDownCheck",
1520 "kMirOpLowerBound",
1521 "kMirOpPunt",
1522 "kMirOpCheckInlinePrediction",
1523};
1524
1525/* Extended MIR instructions like PHI */
1526static void handleExtendedMethodMIR(CompilationUnit* cUnit, MIR* mir)
1527{
1528 int opOffset = mir->dalvikInsn.opcode - kMirOpFirst;
1529 char* msg = (char*)oatNew(strlen(extendedMIROpNames[opOffset]) + 1, false);
1530 strcpy(msg, extendedMIROpNames[opOffset]);
1531 ArmLIR* op = newLIR1(cUnit, kArmPseudoExtended, (int) msg);
1532
1533 switch ((ExtendedMIROpcode)mir->dalvikInsn.opcode) {
1534 case kMirOpPhi: {
1535 char* ssaString = oatGetSSAString(cUnit, mir->ssaRep);
1536 op->flags.isNop = true;
1537 newLIR1(cUnit, kArmPseudoSSARep, (int) ssaString);
1538 break;
1539 }
1540 default:
1541 break;
1542 }
1543}
1544
1545/* If there are any ins passed in registers that have not been promoted
1546 * to a callee-save register, flush them to the frame.
buzbeedfd3d702011-08-28 12:56:51 -07001547 * Note: at this pointCopy any ins that are passed in register to their
1548 * home location */
buzbee67bf8852011-08-17 17:51:35 -07001549static void flushIns(CompilationUnit* cUnit)
1550{
buzbeec143c552011-08-20 17:38:58 -07001551 if (cUnit->method->num_ins_ == 0)
buzbee67bf8852011-08-17 17:51:35 -07001552 return;
buzbeec143c552011-08-20 17:38:58 -07001553 int inRegs = (cUnit->method->num_ins_ > 2) ? 3 : cUnit->method->num_ins_;
buzbee67bf8852011-08-17 17:51:35 -07001554 int startReg = r1;
buzbeec143c552011-08-20 17:38:58 -07001555 int startLoc = cUnit->method->num_registers_ - cUnit->method->num_ins_;
buzbee67bf8852011-08-17 17:51:35 -07001556 for (int i = 0; i < inRegs; i++) {
1557 RegLocation loc = cUnit->regLocation[startLoc + i];
buzbeedfd3d702011-08-28 12:56:51 -07001558 //TUNING: be smarter about flushing ins to frame
1559 storeBaseDisp(cUnit, rSP, loc.spOffset, startReg + i, kWord);
buzbee67bf8852011-08-17 17:51:35 -07001560 if (loc.location == kLocPhysReg) {
1561 genRegCopy(cUnit, loc.lowReg, startReg + i);
buzbee67bf8852011-08-17 17:51:35 -07001562 }
1563 }
1564
1565 // Handle special case of wide argument half in regs, half in frame
1566 if (inRegs == 3) {
1567 RegLocation loc = cUnit->regLocation[startLoc + 2];
1568 if (loc.wide && loc.location == kLocPhysReg) {
1569 // Load the other half of the arg into the promoted pair
1570 loadBaseDisp(cUnit, NULL, rSP, loc.spOffset+4,
1571 loc.highReg, kWord, INVALID_SREG);
1572 inRegs++;
1573 }
1574 }
1575
1576 // Now, do initial assignment of all promoted arguments passed in frame
buzbeec143c552011-08-20 17:38:58 -07001577 for (int i = inRegs; i < cUnit->method->num_ins_;) {
buzbee67bf8852011-08-17 17:51:35 -07001578 RegLocation loc = cUnit->regLocation[startLoc + i];
1579 if (loc.fpLocation == kLocPhysReg) {
1580 loc.location = kLocPhysReg;
1581 loc.fp = true;
1582 loc.lowReg = loc.fpLowReg;
1583 loc.highReg = loc.fpHighReg;
1584 }
1585 if (loc.location == kLocPhysReg) {
1586 if (loc.wide) {
1587 loadBaseDispWide(cUnit, NULL, rSP, loc.spOffset,
1588 loc.lowReg, loc.highReg, INVALID_SREG);
1589 i++;
1590 } else {
1591 loadBaseDisp(cUnit, NULL, rSP, loc.spOffset,
1592 loc.lowReg, kWord, INVALID_SREG);
1593 }
1594 }
1595 i++;
1596 }
1597}
1598
1599/* Handle the content in each basic block */
1600static bool methodBlockCodeGen(CompilationUnit* cUnit, BasicBlock* bb)
1601{
1602 MIR* mir;
1603 ArmLIR* labelList = (ArmLIR*) cUnit->blockLabelList;
1604 int blockId = bb->id;
1605
1606 cUnit->curBlock = bb;
1607 labelList[blockId].operands[0] = bb->startOffset;
1608
1609 /* Insert the block label */
1610 labelList[blockId].opcode = kArmPseudoNormalBlockLabel;
1611 oatAppendLIR(cUnit, (LIR*) &labelList[blockId]);
1612
1613 oatClobberAllRegs(cUnit);
1614 oatResetNullCheck(cUnit);
1615
1616 ArmLIR* headLIR = NULL;
1617
1618 if (bb->blockType == kEntryBlock) {
1619 /*
1620 * On entry, r0, r1, r2 & r3 are live. Let the register allocation
1621 * mechanism know so it doesn't try to use any of them when
1622 * expanding the frame or flushing. This leaves the utility
1623 * code with a single temp: r12. This should be enough.
1624 */
1625 oatLockTemp(cUnit, r0);
1626 oatLockTemp(cUnit, r1);
1627 oatLockTemp(cUnit, r2);
1628 oatLockTemp(cUnit, r3);
1629 newLIR0(cUnit, kArmPseudoMethodEntry);
1630 /* Spill core callee saves */
1631 newLIR1(cUnit, kThumb2Push, cUnit->coreSpillMask);
1632 /* Need to spill any FP regs? */
1633 if (cUnit->numFPSpills) {
1634 newLIR1(cUnit, kThumb2VPushCS, cUnit->numFPSpills);
1635 }
1636 opRegImm(cUnit, kOpSub, rSP, cUnit->frameSize - (cUnit->numSpills * 4));
1637 storeBaseDisp(cUnit, rSP, 0, r0, kWord);
1638 flushIns(cUnit);
1639 oatFreeTemp(cUnit, r0);
1640 oatFreeTemp(cUnit, r1);
1641 oatFreeTemp(cUnit, r2);
1642 oatFreeTemp(cUnit, r3);
1643 } else if (bb->blockType == kExitBlock) {
1644 newLIR0(cUnit, kArmPseudoMethodExit);
1645 opRegImm(cUnit, kOpAdd, rSP, cUnit->frameSize - (cUnit->numSpills * 4));
1646 /* Need to restore any FP callee saves? */
1647 if (cUnit->numFPSpills) {
1648 newLIR1(cUnit, kThumb2VPopCS, cUnit->numFPSpills);
1649 }
1650 if (cUnit->coreSpillMask & (1 << rLR)) {
1651 /* Unspill rLR to rPC */
1652 cUnit->coreSpillMask &= ~(1 << rLR);
1653 cUnit->coreSpillMask |= (1 << rPC);
1654 }
1655 newLIR1(cUnit, kThumb2Pop, cUnit->coreSpillMask);
1656 if (!(cUnit->coreSpillMask & (1 << rPC))) {
1657 /* We didn't pop to rPC, so must do a bv rLR */
1658 newLIR1(cUnit, kThumbBx, rLR);
1659 }
1660 }
1661
1662 for (mir = bb->firstMIRInsn; mir; mir = mir->next) {
1663
1664 oatResetRegPool(cUnit);
1665 if (cUnit->disableOpt & (1 << kTrackLiveTemps)) {
1666 oatClobberAllRegs(cUnit);
1667 }
1668
1669 if (cUnit->disableOpt & (1 << kSuppressLoads)) {
1670 oatResetDefTracking(cUnit);
1671 }
1672
1673 if ((int)mir->dalvikInsn.opcode >= (int)kMirOpFirst) {
1674 handleExtendedMethodMIR(cUnit, mir);
1675 continue;
1676 }
1677
1678 cUnit->currentDalvikOffset = mir->offset;
1679
1680 Opcode dalvikOpcode = mir->dalvikInsn.opcode;
1681 InstructionFormat dalvikFormat =
1682 dexGetFormatFromOpcode(dalvikOpcode);
1683
1684 ArmLIR* boundaryLIR;
1685
1686 /* Mark the beginning of a Dalvik instruction for line tracking */
1687 boundaryLIR = newLIR1(cUnit, kArmPseudoDalvikByteCodeBoundary,
1688 (int) oatGetDalvikDisassembly(
1689 &mir->dalvikInsn, ""));
1690 /* Remember the first LIR for this block */
1691 if (headLIR == NULL) {
1692 headLIR = boundaryLIR;
1693 /* Set the first boundaryLIR as a scheduling barrier */
1694 headLIR->defMask = ENCODE_ALL;
1695 }
1696
1697 /* Don't generate the SSA annotation unless verbose mode is on */
1698 if (cUnit->printMe && mir->ssaRep) {
1699 char *ssaString = oatGetSSAString(cUnit, mir->ssaRep);
1700 newLIR1(cUnit, kArmPseudoSSARep, (int) ssaString);
1701 }
1702
1703 bool notHandled = compileDalvikInstruction(cUnit, mir, bb, labelList);
1704
1705 if (notHandled) {
1706 char buf[100];
1707 snprintf(buf, 100, "%#06x: Opcode %#x (%s) / Fmt %d not handled",
1708 mir->offset,
1709 dalvikOpcode, dexGetOpcodeName(dalvikOpcode),
1710 dalvikFormat);
1711 LOG(FATAL) << buf;
1712 }
1713 }
1714
1715 if (headLIR) {
1716 /*
1717 * Eliminate redundant loads/stores and delay stores into later
1718 * slots
1719 */
1720 oatApplyLocalOptimizations(cUnit, (LIR*) headLIR,
1721 cUnit->lastLIRInsn);
1722
1723 /*
1724 * Generate an unconditional branch to the fallthrough block.
1725 */
1726 if (bb->fallThrough) {
1727 genUnconditionalBranch(cUnit,
1728 &labelList[bb->fallThrough->id]);
1729 }
1730 }
1731 return false;
1732}
1733
1734/*
1735 * Nop any unconditional branches that go to the next instruction.
1736 * Note: new redundant branches may be inserted later, and we'll
1737 * use a check in final instruction assembly to nop those out.
1738 */
1739void removeRedundantBranches(CompilationUnit* cUnit)
1740{
1741 ArmLIR* thisLIR;
1742
1743 for (thisLIR = (ArmLIR*) cUnit->firstLIRInsn;
1744 thisLIR != (ArmLIR*) cUnit->lastLIRInsn;
1745 thisLIR = NEXT_LIR(thisLIR)) {
1746
1747 /* Branch to the next instruction */
1748 if ((thisLIR->opcode == kThumbBUncond) ||
1749 (thisLIR->opcode == kThumb2BUncond)) {
1750 ArmLIR* nextLIR = thisLIR;
1751
1752 while (true) {
1753 nextLIR = NEXT_LIR(nextLIR);
1754
1755 /*
1756 * Is the branch target the next instruction?
1757 */
1758 if (nextLIR == (ArmLIR*) thisLIR->generic.target) {
1759 thisLIR->flags.isNop = true;
1760 break;
1761 }
1762
1763 /*
1764 * Found real useful stuff between the branch and the target.
1765 * Need to explicitly check the lastLIRInsn here because it
1766 * might be the last real instruction.
1767 */
1768 if (!isPseudoOpcode(nextLIR->opcode) ||
1769 (nextLIR = (ArmLIR*) cUnit->lastLIRInsn))
1770 break;
1771 }
1772 }
1773 }
1774}
1775
1776void oatMethodMIR2LIR(CompilationUnit* cUnit)
1777{
1778 /* Used to hold the labels of each block */
1779 cUnit->blockLabelList =
1780 (void *) oatNew(sizeof(ArmLIR) * cUnit->numBlocks, true);
1781
1782 oatDataFlowAnalysisDispatcher(cUnit, methodBlockCodeGen,
1783 kPreOrderDFSTraversal, false /* Iterative */);
1784 removeRedundantBranches(cUnit);
1785}
1786
1787/* Common initialization routine for an architecture family */
1788bool oatArchInit()
1789{
1790 int i;
1791
1792 for (i = 0; i < kArmLast; i++) {
1793 if (EncodingMap[i].opcode != i) {
1794 LOG(FATAL) << "Encoding order for " << EncodingMap[i].name <<
1795 " is wrong: expecting " << i << ", seeing " <<
1796 (int)EncodingMap[i].opcode;
1797 }
1798 }
1799
1800 return oatArchVariantInit();
1801}
1802
1803/* Needed by the Assembler */
1804void oatSetupResourceMasks(ArmLIR* lir)
1805{
1806 setupResourceMasks(lir);
1807}
1808
1809/* Needed by the ld/st optmizatons */
1810ArmLIR* oatRegCopyNoInsert(CompilationUnit* cUnit, int rDest, int rSrc)
1811{
1812 return genRegCopyNoInsert(cUnit, rDest, rSrc);
1813}
1814
1815/* Needed by the register allocator */
1816ArmLIR* oatRegCopy(CompilationUnit* cUnit, int rDest, int rSrc)
1817{
1818 return genRegCopy(cUnit, rDest, rSrc);
1819}
1820
1821/* Needed by the register allocator */
1822void oatRegCopyWide(CompilationUnit* cUnit, int destLo, int destHi,
1823 int srcLo, int srcHi)
1824{
1825 genRegCopyWide(cUnit, destLo, destHi, srcLo, srcHi);
1826}
1827
1828void oatFlushRegImpl(CompilationUnit* cUnit, int rBase,
1829 int displacement, int rSrc, OpSize size)
1830{
1831 storeBaseDisp(cUnit, rBase, displacement, rSrc, size);
1832}
1833
1834void oatFlushRegWideImpl(CompilationUnit* cUnit, int rBase,
1835 int displacement, int rSrcLo, int rSrcHi)
1836{
1837 storeBaseDispWide(cUnit, rBase, displacement, rSrcLo, rSrcHi);
1838}