30 #if defined(V8_TARGET_ARCH_ARM)
38 #define __ ACCESS_MASM(masm)
61 ASSERT(!masm->has_frame());
62 masm->set_has_frame(
true);
69 masm->set_has_frame(
false);
77 MacroAssembler* masm) {
88 __ RecordWriteField(
r2,
100 MacroAssembler* masm, Label* fail) {
109 Label loop, entry, convert_hole, gc_required, only_change_map, done;
115 __ CompareRoot(
r4, Heap::kEmptyFixedArrayRootIndex);
116 __ b(
eq, &only_change_map);
129 __ LoadRoot(
r9, Heap::kFixedDoubleArrayMapRootIndex);
135 __ RecordWriteField(
r2,
146 __ RecordWriteField(
r2,
166 if (!vfp3_supported)
__ Push(
r1,
r0);
170 __ bind(&only_change_map);
172 __ RecordWriteField(
r2,
183 __ bind(&gc_required);
191 __ UntagAndJumpIfNotSmi(
r9,
r9, &convert_hole);
194 if (vfp3_supported) {
195 CpuFeatures::Scope scope(
VFP3);
199 __ add(
r7,
r7, Operand(8));
214 __ bind(&convert_hole);
215 if (FLAG_debug_code) {
218 __ orr(
r9,
r9, Operand(1));
219 __ CompareRoot(
r9, Heap::kTheHoleValueRootIndex);
220 __ Assert(
eq,
"object found in smi-only array");
228 if (!vfp3_supported)
__ Pop(
r1,
r0);
235 MacroAssembler* masm, Label* fail) {
244 Label entry, loop, convert_hole, gc_required, only_change_map;
249 __ CompareRoot(
r4, Heap::kEmptyFixedArrayRootIndex);
250 __ b(
eq, &only_change_map);
264 __ LoadRoot(
r9, Heap::kFixedArrayMapRootIndex);
273 __ LoadRoot(
r7, Heap::kTheHoleValueRootIndex);
274 __ LoadRoot(
r9, Heap::kHeapNumberMapRootIndex);
285 __ bind(&gc_required);
295 __ b(
eq, &convert_hole);
298 __ AllocateHeapNumber(
r2,
r0,
lr,
r9, &gc_required);
314 __ bind(&convert_hole);
324 __ RecordWriteField(
r2,
334 __ bind(&only_change_map);
337 __ RecordWriteField(
r2,
352 Label* call_runtime) {
358 Label check_sequential;
360 __ b(
eq, &check_sequential);
365 __ b(
eq, &cons_string);
368 Label indirect_string_loaded;
372 __ jmp(&indirect_string_loaded);
379 __ bind(&cons_string);
381 __ CompareRoot(result, Heap::kEmptyStringRootIndex);
382 __ b(
ne, call_runtime);
386 __ bind(&indirect_string_loaded);
393 Label external_string, check_encoding;
394 __ bind(&check_sequential);
397 __ b(
ne, &external_string);
404 __ jmp(&check_encoding);
407 __ bind(&external_string);
408 if (FLAG_debug_code) {
412 __ Assert(
eq,
"external string expected, but not found");
417 __ b(
ne, call_runtime);
421 __ bind(&check_encoding);
438 #endif // V8_TARGET_ARCH_ARM
static const int kResourceDataOffset
STATIC_CHECK((kStringRepresentationMask|kStringEncodingMask)==Internals::kFullStringRepresentationMask)
const uint32_t kTwoByteStringTag
static bool IsSupported(CpuFeature f)
#define ASSERT(condition)
virtual void AfterCall(MacroAssembler *masm) const
double(* UnaryMathFunction)(double x)
const uint32_t kStringRepresentationMask
const uint32_t kShortExternalStringMask
static const int kFirstOffset
static const int kParentOffset
STATIC_ASSERT((FixedDoubleArray::kHeaderSize &kDoubleAlignmentMask)==0)
static const int kValueOffset
const uint32_t kHoleNanUpper32
const uint32_t kIsIndirectStringMask
const uint32_t kHoleNanLower32
UnaryMathFunction CreateSqrtFunction()
static void GenerateSmiToDouble(MacroAssembler *masm, Label *fail)
const uint32_t kShortExternalStringTag
static void GenerateMapChangeElementsTransition(MacroAssembler *masm)
static void Generate(MacroAssembler *masm, Register string, Register index, Register result, Label *call_runtime)
static const int kHeaderSize
static const int kElementsOffset
static const int kOffsetOffset
static const int kHeaderSize
static const int kMapOffset
static void GenerateDoubleToObject(MacroAssembler *masm, Label *fail)
const uint32_t kSlicedNotConsMask
static const int kLengthOffset
static void ConvertIntToDouble(MacroAssembler *masm, Register int_scratch, Destination destination, DwVfpRegister double_dst, Register dst1, Register dst2, Register scratch2, SwVfpRegister single_scratch)
static const int kSecondOffset
MemOperand FieldMemOperand(Register object, int offset)
activate correct semantics for inheriting readonliness enable harmony semantics for typeof enable harmony enable harmony proxies enable all harmony harmony_scoping harmony_proxies harmony_scoping tracks arrays with only smi values automatically unbox arrays of doubles use crankshaft use hydrogen range analysis use hydrogen global value numbering use function inlining maximum number of AST nodes considered for a single inlining loop invariant code motion print statistics for hydrogen trace generated IR for specified phases trace register allocator trace range analysis trace representation types environment for every instruction put a break point before deoptimizing polymorphic inlining perform array bounds checks elimination trace on stack replacement optimize closures functions with arguments object optimize functions containing for in loops profiler considers IC stability primitive functions trigger their own optimization re try self optimization if it failed insert an interrupt check at function exit execution budget before interrupt is triggered call count before self optimization self_optimization count_based_interrupts weighted_back_edges trace_opt emit comments in code disassembly enable use of SSE3 instructions if available enable use of CMOV instruction if available enable use of SAHF instruction if enable use of VFP3 instructions if available this implies enabling ARMv7 enable use of ARMv7 instructions if enable use of MIPS FPU instructions if NULL
virtual void BeforeCall(MacroAssembler *masm) const
UnaryMathFunction CreateTranscendentalFunction(TranscendentalCache::Type type)
const uint32_t kStringEncodingMask
static const int kInstanceTypeOffset