| // Copyright 2006-2009 the V8 project authors. All rights reserved. |
| // Redistribution and use in source and binary forms, with or without |
| // modification, are permitted provided that the following conditions are |
| // met: |
| // |
| // * Redistributions of source code must retain the above copyright |
| // notice, this list of conditions and the following disclaimer. |
| // * Redistributions in binary form must reproduce the above |
| // copyright notice, this list of conditions and the following |
| // disclaimer in the documentation and/or other materials provided |
| // with the distribution. |
| // * Neither the name of Google Inc. nor the names of its |
| // contributors may be used to endorse or promote products derived |
| // from this software without specific prior written permission. |
| // |
| // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS |
| // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT |
| // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR |
| // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT |
| // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, |
| // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT |
| // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
| // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
| // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
| // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
| // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| |
| #include "v8.h" |
| |
| #include "ic-inl.h" |
| #include "codegen-inl.h" |
| #include "stub-cache.h" |
| |
| namespace v8 { |
| namespace internal { |
| |
| #define __ ACCESS_MASM(masm) |
| |
| |
| static void ProbeTable(MacroAssembler* masm, |
| Code::Flags flags, |
| StubCache::Table table, |
| Register name, |
| Register offset) { |
| ExternalReference key_offset(SCTableReference::keyReference(table)); |
| ExternalReference value_offset(SCTableReference::valueReference(table)); |
| |
| Label miss; |
| |
| // Save the offset on the stack. |
| __ push(offset); |
| |
| // Check that the key in the entry matches the name. |
| __ mov(ip, Operand(key_offset)); |
| __ ldr(ip, MemOperand(ip, offset, LSL, 1)); |
| __ cmp(name, Operand(ip)); |
| __ b(ne, &miss); |
| |
| // Get the code entry from the cache. |
| __ mov(ip, Operand(value_offset)); |
| __ ldr(offset, MemOperand(ip, offset, LSL, 1)); |
| |
| // Check that the flags match what we're looking for. |
| __ ldr(offset, FieldMemOperand(offset, Code::kFlagsOffset)); |
| __ and_(offset, offset, Operand(~Code::kFlagsNotUsedInLookup)); |
| __ cmp(offset, Operand(flags)); |
| __ b(ne, &miss); |
| |
| // Restore offset and re-load code entry from cache. |
| __ pop(offset); |
| __ mov(ip, Operand(value_offset)); |
| __ ldr(offset, MemOperand(ip, offset, LSL, 1)); |
| |
| // Jump to the first instruction in the code stub. |
| __ add(offset, offset, Operand(Code::kHeaderSize - kHeapObjectTag)); |
| __ Jump(offset); |
| |
| // Miss: Restore offset and fall through. |
| __ bind(&miss); |
| __ pop(offset); |
| } |
| |
| |
| void StubCache::GenerateProbe(MacroAssembler* masm, |
| Code::Flags flags, |
| Register receiver, |
| Register name, |
| Register scratch, |
| Register extra) { |
| Label miss; |
| |
| // Make sure that code is valid. The shifting code relies on the |
| // entry size being 8. |
| ASSERT(sizeof(Entry) == 8); |
| |
| // Make sure the flags does not name a specific type. |
| ASSERT(Code::ExtractTypeFromFlags(flags) == 0); |
| |
| // Make sure that there are no register conflicts. |
| ASSERT(!scratch.is(receiver)); |
| ASSERT(!scratch.is(name)); |
| |
| // Check that the receiver isn't a smi. |
| __ tst(receiver, Operand(kSmiTagMask)); |
| __ b(eq, &miss); |
| |
| // Get the map of the receiver and compute the hash. |
| __ ldr(scratch, FieldMemOperand(name, String::kHashFieldOffset)); |
| __ ldr(ip, FieldMemOperand(receiver, HeapObject::kMapOffset)); |
| __ add(scratch, scratch, Operand(ip)); |
| __ eor(scratch, scratch, Operand(flags)); |
| __ and_(scratch, |
| scratch, |
| Operand((kPrimaryTableSize - 1) << kHeapObjectTagSize)); |
| |
| // Probe the primary table. |
| ProbeTable(masm, flags, kPrimary, name, scratch); |
| |
| // Primary miss: Compute hash for secondary probe. |
| __ sub(scratch, scratch, Operand(name)); |
| __ add(scratch, scratch, Operand(flags)); |
| __ and_(scratch, |
| scratch, |
| Operand((kSecondaryTableSize - 1) << kHeapObjectTagSize)); |
| |
| // Probe the secondary table. |
| ProbeTable(masm, flags, kSecondary, name, scratch); |
| |
| // Cache miss: Fall-through and let caller handle the miss by |
| // entering the runtime system. |
| __ bind(&miss); |
| } |
| |
| |
| void StubCompiler::GenerateLoadGlobalFunctionPrototype(MacroAssembler* masm, |
| int index, |
| Register prototype) { |
| // Load the global or builtins object from the current context. |
| __ ldr(prototype, MemOperand(cp, Context::SlotOffset(Context::GLOBAL_INDEX))); |
| // Load the global context from the global or builtins object. |
| __ ldr(prototype, |
| FieldMemOperand(prototype, GlobalObject::kGlobalContextOffset)); |
| // Load the function from the global context. |
| __ ldr(prototype, MemOperand(prototype, Context::SlotOffset(index))); |
| // Load the initial map. The global functions all have initial maps. |
| __ ldr(prototype, |
| FieldMemOperand(prototype, JSFunction::kPrototypeOrInitialMapOffset)); |
| // Load the prototype from the initial map. |
| __ ldr(prototype, FieldMemOperand(prototype, Map::kPrototypeOffset)); |
| } |
| |
| |
| // Load a fast property out of a holder object (src). In-object properties |
| // are loaded directly otherwise the property is loaded from the properties |
| // fixed array. |
| void StubCompiler::GenerateFastPropertyLoad(MacroAssembler* masm, |
| Register dst, Register src, |
| JSObject* holder, int index) { |
| // Adjust for the number of properties stored in the holder. |
| index -= holder->map()->inobject_properties(); |
| if (index < 0) { |
| // Get the property straight out of the holder. |
| int offset = holder->map()->instance_size() + (index * kPointerSize); |
| __ ldr(dst, FieldMemOperand(src, offset)); |
| } else { |
| // Calculate the offset into the properties array. |
| int offset = index * kPointerSize + FixedArray::kHeaderSize; |
| __ ldr(dst, FieldMemOperand(src, JSObject::kPropertiesOffset)); |
| __ ldr(dst, FieldMemOperand(dst, offset)); |
| } |
| } |
| |
| |
| void StubCompiler::GenerateLoadArrayLength(MacroAssembler* masm, |
| Register receiver, |
| Register scratch, |
| Label* miss_label) { |
| // Check that the receiver isn't a smi. |
| __ tst(receiver, Operand(kSmiTagMask)); |
| __ b(eq, miss_label); |
| |
| // Check that the object is a JS array. |
| __ CompareObjectType(receiver, scratch, scratch, JS_ARRAY_TYPE); |
| __ b(ne, miss_label); |
| |
| // Load length directly from the JS array. |
| __ ldr(r0, FieldMemOperand(receiver, JSArray::kLengthOffset)); |
| __ Ret(); |
| } |
| |
| |
| // Generate code to check if an object is a string. If the object is a |
| // heap object, its map's instance type is left in the scratch1 register. |
| // If this is not needed, scratch1 and scratch2 may be the same register. |
| static void GenerateStringCheck(MacroAssembler* masm, |
| Register receiver, |
| Register scratch1, |
| Register scratch2, |
| Label* smi, |
| Label* non_string_object) { |
| // Check that the receiver isn't a smi. |
| __ tst(receiver, Operand(kSmiTagMask)); |
| __ b(eq, smi); |
| |
| // Check that the object is a string. |
| __ ldr(scratch1, FieldMemOperand(receiver, HeapObject::kMapOffset)); |
| __ ldrb(scratch1, FieldMemOperand(scratch1, Map::kInstanceTypeOffset)); |
| __ and_(scratch2, scratch1, Operand(kIsNotStringMask)); |
| // The cast is to resolve the overload for the argument of 0x0. |
| __ cmp(scratch2, Operand(static_cast<int32_t>(kStringTag))); |
| __ b(ne, non_string_object); |
| } |
| |
| |
| // Generate code to load the length from a string object and return the length. |
| // If the receiver object is not a string or a wrapped string object the |
| // execution continues at the miss label. The register containing the |
| // receiver is potentially clobbered. |
| void StubCompiler::GenerateLoadStringLength(MacroAssembler* masm, |
| Register receiver, |
| Register scratch1, |
| Register scratch2, |
| Label* miss) { |
| Label check_wrapper; |
| |
| // Check if the object is a string leaving the instance type in the |
| // scratch1 register. |
| GenerateStringCheck(masm, receiver, scratch1, scratch2, miss, &check_wrapper); |
| |
| // Load length directly from the string. |
| __ ldr(r0, FieldMemOperand(receiver, String::kLengthOffset)); |
| __ mov(r0, Operand(r0, LSL, kSmiTagSize)); |
| __ Ret(); |
| |
| // Check if the object is a JSValue wrapper. |
| __ bind(&check_wrapper); |
| __ cmp(scratch1, Operand(JS_VALUE_TYPE)); |
| __ b(ne, miss); |
| |
| // Unwrap the value and check if the wrapped value is a string. |
| __ ldr(scratch1, FieldMemOperand(receiver, JSValue::kValueOffset)); |
| GenerateStringCheck(masm, scratch1, scratch2, scratch2, miss, miss); |
| __ ldr(r0, FieldMemOperand(scratch1, String::kLengthOffset)); |
| __ mov(r0, Operand(r0, LSL, kSmiTagSize)); |
| __ Ret(); |
| } |
| |
| |
| void StubCompiler::GenerateLoadFunctionPrototype(MacroAssembler* masm, |
| Register receiver, |
| Register scratch1, |
| Register scratch2, |
| Label* miss_label) { |
| __ TryGetFunctionPrototype(receiver, scratch1, scratch2, miss_label); |
| __ mov(r0, scratch1); |
| __ Ret(); |
| } |
| |
| |
| // Generate StoreField code, value is passed in r0 register. |
| // When leaving generated code after success, the receiver_reg and name_reg |
| // may be clobbered. Upon branch to miss_label, the receiver and name |
| // registers have their original values. |
| void StubCompiler::GenerateStoreField(MacroAssembler* masm, |
| JSObject* object, |
| int index, |
| Map* transition, |
| Register receiver_reg, |
| Register name_reg, |
| Register scratch, |
| Label* miss_label) { |
| // r0 : value |
| Label exit; |
| |
| // Check that the receiver isn't a smi. |
| __ tst(receiver_reg, Operand(kSmiTagMask)); |
| __ b(eq, miss_label); |
| |
| // Check that the map of the receiver hasn't changed. |
| __ ldr(scratch, FieldMemOperand(receiver_reg, HeapObject::kMapOffset)); |
| __ cmp(scratch, Operand(Handle<Map>(object->map()))); |
| __ b(ne, miss_label); |
| |
| // Perform global security token check if needed. |
| if (object->IsJSGlobalProxy()) { |
| __ CheckAccessGlobalProxy(receiver_reg, scratch, miss_label); |
| } |
| |
| // Stub never generated for non-global objects that require access |
| // checks. |
| ASSERT(object->IsJSGlobalProxy() || !object->IsAccessCheckNeeded()); |
| |
| // Perform map transition for the receiver if necessary. |
| if ((transition != NULL) && (object->map()->unused_property_fields() == 0)) { |
| // The properties must be extended before we can store the value. |
| // We jump to a runtime call that extends the properties array. |
| __ push(receiver_reg); |
| __ mov(r2, Operand(Handle<Map>(transition))); |
| __ stm(db_w, sp, r2.bit() | r0.bit()); |
| __ TailCallRuntime( |
| ExternalReference(IC_Utility(IC::kSharedStoreIC_ExtendStorage)), |
| 3, 1); |
| return; |
| } |
| |
| if (transition != NULL) { |
| // Update the map of the object; no write barrier updating is |
| // needed because the map is never in new space. |
| __ mov(ip, Operand(Handle<Map>(transition))); |
| __ str(ip, FieldMemOperand(receiver_reg, HeapObject::kMapOffset)); |
| } |
| |
| // Adjust for the number of properties stored in the object. Even in the |
| // face of a transition we can use the old map here because the size of the |
| // object and the number of in-object properties is not going to change. |
| index -= object->map()->inobject_properties(); |
| |
| if (index < 0) { |
| // Set the property straight into the object. |
| int offset = object->map()->instance_size() + (index * kPointerSize); |
| __ str(r0, FieldMemOperand(receiver_reg, offset)); |
| |
| // Skip updating write barrier if storing a smi. |
| __ tst(r0, Operand(kSmiTagMask)); |
| __ b(eq, &exit); |
| |
| // Update the write barrier for the array address. |
| // Pass the value being stored in the now unused name_reg. |
| __ mov(name_reg, Operand(offset)); |
| __ RecordWrite(receiver_reg, name_reg, scratch); |
| } else { |
| // Write to the properties array. |
| int offset = index * kPointerSize + FixedArray::kHeaderSize; |
| // Get the properties array |
| __ ldr(scratch, FieldMemOperand(receiver_reg, JSObject::kPropertiesOffset)); |
| __ str(r0, FieldMemOperand(scratch, offset)); |
| |
| // Skip updating write barrier if storing a smi. |
| __ tst(r0, Operand(kSmiTagMask)); |
| __ b(eq, &exit); |
| |
| // Update the write barrier for the array address. |
| // Ok to clobber receiver_reg and name_reg, since we return. |
| __ mov(name_reg, Operand(offset)); |
| __ RecordWrite(scratch, name_reg, receiver_reg); |
| } |
| |
| // Return the value (register r0). |
| __ bind(&exit); |
| __ Ret(); |
| } |
| |
| |
| void StubCompiler::GenerateLoadMiss(MacroAssembler* masm, Code::Kind kind) { |
| ASSERT(kind == Code::LOAD_IC || kind == Code::KEYED_LOAD_IC); |
| Code* code = NULL; |
| if (kind == Code::LOAD_IC) { |
| code = Builtins::builtin(Builtins::LoadIC_Miss); |
| } else { |
| code = Builtins::builtin(Builtins::KeyedLoadIC_Miss); |
| } |
| |
| Handle<Code> ic(code); |
| __ Jump(ic, RelocInfo::CODE_TARGET); |
| } |
| |
| |
| static void GenerateCallFunction(MacroAssembler* masm, |
| Object* object, |
| const ParameterCount& arguments, |
| Label* miss) { |
| // ----------- S t a t e ------------- |
| // -- r0: receiver |
| // -- r1: function to call |
| // ----------------------------------- |
| |
| // Check that the function really is a function. |
| __ BranchOnSmi(r1, miss); |
| __ CompareObjectType(r1, r3, r3, JS_FUNCTION_TYPE); |
| __ b(ne, miss); |
| |
| // Patch the receiver on the stack with the global proxy if |
| // necessary. |
| if (object->IsGlobalObject()) { |
| __ ldr(r3, FieldMemOperand(r0, GlobalObject::kGlobalReceiverOffset)); |
| __ str(r3, MemOperand(sp, arguments.immediate() * kPointerSize)); |
| } |
| |
| // Invoke the function. |
| __ InvokeFunction(r1, arguments, JUMP_FUNCTION); |
| } |
| |
| |
| static void PushInterceptorArguments(MacroAssembler* masm, |
| Register receiver, |
| Register holder, |
| Register name, |
| JSObject* holder_obj) { |
| __ push(receiver); |
| __ push(holder); |
| __ push(name); |
| InterceptorInfo* interceptor = holder_obj->GetNamedInterceptor(); |
| ASSERT(!Heap::InNewSpace(interceptor)); |
| |
| Register scratch = receiver; |
| __ mov(scratch, Operand(Handle<Object>(interceptor))); |
| __ push(scratch); |
| __ ldr(scratch, FieldMemOperand(scratch, InterceptorInfo::kDataOffset)); |
| __ push(scratch); |
| } |
| |
| |
| static void CompileCallLoadPropertyWithInterceptor(MacroAssembler* masm, |
| Register receiver, |
| Register holder, |
| Register name, |
| JSObject* holder_obj) { |
| PushInterceptorArguments(masm, receiver, holder, name, holder_obj); |
| |
| ExternalReference ref = |
| ExternalReference(IC_Utility(IC::kLoadPropertyWithInterceptorOnly)); |
| __ mov(r0, Operand(5)); |
| __ mov(r1, Operand(ref)); |
| |
| CEntryStub stub(1); |
| __ CallStub(&stub); |
| } |
| |
| |
| class LoadInterceptorCompiler BASE_EMBEDDED { |
| public: |
| explicit LoadInterceptorCompiler(Register name) : name_(name) {} |
| |
| void CompileCacheable(MacroAssembler* masm, |
| StubCompiler* stub_compiler, |
| Register receiver, |
| Register holder, |
| Register scratch1, |
| Register scratch2, |
| JSObject* holder_obj, |
| LookupResult* lookup, |
| String* name, |
| Label* miss_label) { |
| AccessorInfo* callback = NULL; |
| bool optimize = false; |
| // So far the most popular follow ups for interceptor loads are FIELD |
| // and CALLBACKS, so inline only them, other cases may be added |
| // later. |
| if (lookup->type() == FIELD) { |
| optimize = true; |
| } else if (lookup->type() == CALLBACKS) { |
| Object* callback_object = lookup->GetCallbackObject(); |
| if (callback_object->IsAccessorInfo()) { |
| callback = AccessorInfo::cast(callback_object); |
| optimize = callback->getter() != NULL; |
| } |
| } |
| |
| if (!optimize) { |
| CompileRegular(masm, receiver, holder, scratch2, holder_obj, miss_label); |
| return; |
| } |
| |
| // Note: starting a frame here makes GC aware of pointers pushed below. |
| __ EnterInternalFrame(); |
| |
| __ push(receiver); |
| __ push(holder); |
| __ push(name_); |
| |
| CompileCallLoadPropertyWithInterceptor(masm, |
| receiver, |
| holder, |
| name_, |
| holder_obj); |
| |
| Label interceptor_failed; |
| // Compare with no_interceptor_result_sentinel. |
| __ LoadRoot(scratch1, Heap::kNoInterceptorResultSentinelRootIndex); |
| __ cmp(r0, scratch1); |
| __ b(eq, &interceptor_failed); |
| __ LeaveInternalFrame(); |
| __ Ret(); |
| |
| __ bind(&interceptor_failed); |
| __ pop(name_); |
| __ pop(holder); |
| __ pop(receiver); |
| |
| __ LeaveInternalFrame(); |
| |
| if (lookup->type() == FIELD) { |
| holder = stub_compiler->CheckPrototypes(holder_obj, |
| holder, |
| lookup->holder(), |
| scratch1, |
| scratch2, |
| name, |
| miss_label); |
| stub_compiler->GenerateFastPropertyLoad(masm, |
| r0, |
| holder, |
| lookup->holder(), |
| lookup->GetFieldIndex()); |
| __ Ret(); |
| } else { |
| ASSERT(lookup->type() == CALLBACKS); |
| ASSERT(lookup->GetCallbackObject()->IsAccessorInfo()); |
| ASSERT(callback != NULL); |
| ASSERT(callback->getter() != NULL); |
| |
| Label cleanup; |
| __ pop(scratch2); |
| __ push(receiver); |
| __ push(scratch2); |
| |
| holder = stub_compiler->CheckPrototypes(holder_obj, holder, |
| lookup->holder(), scratch1, |
| scratch2, |
| name, |
| &cleanup); |
| |
| __ push(holder); |
| __ Move(holder, Handle<AccessorInfo>(callback)); |
| __ push(holder); |
| __ ldr(scratch1, FieldMemOperand(holder, AccessorInfo::kDataOffset)); |
| __ push(scratch1); |
| __ push(name_); |
| |
| ExternalReference ref = |
| ExternalReference(IC_Utility(IC::kLoadCallbackProperty)); |
| __ TailCallRuntime(ref, 5, 1); |
| |
| __ bind(&cleanup); |
| __ pop(scratch1); |
| __ pop(scratch2); |
| __ push(scratch1); |
| } |
| } |
| |
| |
| void CompileRegular(MacroAssembler* masm, |
| Register receiver, |
| Register holder, |
| Register scratch, |
| JSObject* holder_obj, |
| Label* miss_label) { |
| PushInterceptorArguments(masm, receiver, holder, name_, holder_obj); |
| |
| ExternalReference ref = ExternalReference( |
| IC_Utility(IC::kLoadPropertyWithInterceptorForLoad)); |
| __ TailCallRuntime(ref, 5, 1); |
| } |
| |
| private: |
| Register name_; |
| }; |
| |
| |
| static void CompileLoadInterceptor(LoadInterceptorCompiler* compiler, |
| StubCompiler* stub_compiler, |
| MacroAssembler* masm, |
| JSObject* object, |
| JSObject* holder, |
| String* name, |
| LookupResult* lookup, |
| Register receiver, |
| Register scratch1, |
| Register scratch2, |
| Label* miss) { |
| ASSERT(holder->HasNamedInterceptor()); |
| ASSERT(!holder->GetNamedInterceptor()->getter()->IsUndefined()); |
| |
| // Check that the receiver isn't a smi. |
| __ BranchOnSmi(receiver, miss); |
| |
| // Check that the maps haven't changed. |
| Register reg = |
| stub_compiler->CheckPrototypes(object, receiver, holder, |
| scratch1, scratch2, name, miss); |
| |
| if (lookup->IsProperty() && lookup->IsCacheable()) { |
| compiler->CompileCacheable(masm, |
| stub_compiler, |
| receiver, |
| reg, |
| scratch1, |
| scratch2, |
| holder, |
| lookup, |
| name, |
| miss); |
| } else { |
| compiler->CompileRegular(masm, |
| receiver, |
| reg, |
| scratch2, |
| holder, |
| miss); |
| } |
| } |
| |
| |
| #undef __ |
| #define __ ACCESS_MASM(masm()) |
| |
| |
| Register StubCompiler::CheckPrototypes(JSObject* object, |
| Register object_reg, |
| JSObject* holder, |
| Register holder_reg, |
| Register scratch, |
| String* name, |
| int save_at_depth, |
| Label* miss) { |
| // TODO(602): support object saving. |
| ASSERT(save_at_depth == kInvalidProtoDepth); |
| |
| // Check that the maps haven't changed. |
| Register result = |
| masm()->CheckMaps(object, object_reg, holder, holder_reg, scratch, miss); |
| |
| // If we've skipped any global objects, it's not enough to verify |
| // that their maps haven't changed. |
| while (object != holder) { |
| if (object->IsGlobalObject()) { |
| GlobalObject* global = GlobalObject::cast(object); |
| Object* probe = global->EnsurePropertyCell(name); |
| if (probe->IsFailure()) { |
| set_failure(Failure::cast(probe)); |
| return result; |
| } |
| JSGlobalPropertyCell* cell = JSGlobalPropertyCell::cast(probe); |
| ASSERT(cell->value()->IsTheHole()); |
| __ mov(scratch, Operand(Handle<Object>(cell))); |
| __ ldr(scratch, |
| FieldMemOperand(scratch, JSGlobalPropertyCell::kValueOffset)); |
| __ LoadRoot(ip, Heap::kTheHoleValueRootIndex); |
| __ cmp(scratch, ip); |
| __ b(ne, miss); |
| } |
| object = JSObject::cast(object->GetPrototype()); |
| } |
| |
| // Return the register containing the holder. |
| return result; |
| } |
| |
| |
| void StubCompiler::GenerateLoadField(JSObject* object, |
| JSObject* holder, |
| Register receiver, |
| Register scratch1, |
| Register scratch2, |
| int index, |
| String* name, |
| Label* miss) { |
| // Check that the receiver isn't a smi. |
| __ tst(receiver, Operand(kSmiTagMask)); |
| __ b(eq, miss); |
| |
| // Check that the maps haven't changed. |
| Register reg = |
| CheckPrototypes(object, receiver, holder, scratch1, scratch2, name, miss); |
| GenerateFastPropertyLoad(masm(), r0, reg, holder, index); |
| __ Ret(); |
| } |
| |
| |
| void StubCompiler::GenerateLoadConstant(JSObject* object, |
| JSObject* holder, |
| Register receiver, |
| Register scratch1, |
| Register scratch2, |
| Object* value, |
| String* name, |
| Label* miss) { |
| // Check that the receiver isn't a smi. |
| __ tst(receiver, Operand(kSmiTagMask)); |
| __ b(eq, miss); |
| |
| // Check that the maps haven't changed. |
| Register reg = |
| CheckPrototypes(object, receiver, holder, scratch1, scratch2, name, miss); |
| |
| // Return the constant value. |
| __ mov(r0, Operand(Handle<Object>(value))); |
| __ Ret(); |
| } |
| |
| |
| bool StubCompiler::GenerateLoadCallback(JSObject* object, |
| JSObject* holder, |
| Register receiver, |
| Register name_reg, |
| Register scratch1, |
| Register scratch2, |
| AccessorInfo* callback, |
| String* name, |
| Label* miss, |
| Failure** failure) { |
| // Check that the receiver isn't a smi. |
| __ tst(receiver, Operand(kSmiTagMask)); |
| __ b(eq, miss); |
| |
| // Check that the maps haven't changed. |
| Register reg = |
| CheckPrototypes(object, receiver, holder, scratch1, scratch2, name, miss); |
| |
| // Push the arguments on the JS stack of the caller. |
| __ push(receiver); // receiver |
| __ push(reg); // holder |
| __ mov(ip, Operand(Handle<AccessorInfo>(callback))); // callback data |
| __ push(ip); |
| __ ldr(reg, FieldMemOperand(ip, AccessorInfo::kDataOffset)); |
| __ push(reg); |
| __ push(name_reg); // name |
| |
| // Do tail-call to the runtime system. |
| ExternalReference load_callback_property = |
| ExternalReference(IC_Utility(IC::kLoadCallbackProperty)); |
| __ TailCallRuntime(load_callback_property, 5, 1); |
| |
| return true; |
| } |
| |
| |
| void StubCompiler::GenerateLoadInterceptor(JSObject* object, |
| JSObject* holder, |
| LookupResult* lookup, |
| Register receiver, |
| Register name_reg, |
| Register scratch1, |
| Register scratch2, |
| String* name, |
| Label* miss) { |
| LoadInterceptorCompiler compiler(name_reg); |
| CompileLoadInterceptor(&compiler, |
| this, |
| masm(), |
| object, |
| holder, |
| name, |
| lookup, |
| receiver, |
| scratch1, |
| scratch2, |
| miss); |
| } |
| |
| |
| Object* StubCompiler::CompileLazyCompile(Code::Flags flags) { |
| // ----------- S t a t e ------------- |
| // -- r1: function |
| // -- lr: return address |
| // ----------------------------------- |
| |
| // Enter an internal frame. |
| __ EnterInternalFrame(); |
| |
| // Preserve the function. |
| __ push(r1); |
| |
| // Push the function on the stack as the argument to the runtime function. |
| __ push(r1); |
| __ CallRuntime(Runtime::kLazyCompile, 1); |
| |
| // Calculate the entry point. |
| __ add(r2, r0, Operand(Code::kHeaderSize - kHeapObjectTag)); |
| |
| // Restore saved function. |
| __ pop(r1); |
| |
| // Tear down temporary frame. |
| __ LeaveInternalFrame(); |
| |
| // Do a tail-call of the compiled function. |
| __ Jump(r2); |
| |
| return GetCodeWithFlags(flags, "LazyCompileStub"); |
| } |
| |
| |
| Object* CallStubCompiler::CompileCallField(JSObject* object, |
| JSObject* holder, |
| int index, |
| String* name) { |
| // ----------- S t a t e ------------- |
| // -- r2 : name |
| // -- lr : return address |
| // ----------------------------------- |
| Label miss; |
| |
| const int argc = arguments().immediate(); |
| |
| // Get the receiver of the function from the stack into r0. |
| __ ldr(r0, MemOperand(sp, argc * kPointerSize)); |
| // Check that the receiver isn't a smi. |
| __ tst(r0, Operand(kSmiTagMask)); |
| __ b(eq, &miss); |
| |
| // Do the right check and compute the holder register. |
| Register reg = CheckPrototypes(object, r0, holder, r1, r3, name, &miss); |
| GenerateFastPropertyLoad(masm(), r1, reg, holder, index); |
| |
| GenerateCallFunction(masm(), object, arguments(), &miss); |
| |
| // Handle call cache miss. |
| __ bind(&miss); |
| Handle<Code> ic = ComputeCallMiss(arguments().immediate()); |
| __ Jump(ic, RelocInfo::CODE_TARGET); |
| |
| // Return the generated code. |
| return GetCode(FIELD, name); |
| } |
| |
| |
| Object* CallStubCompiler::CompileCallConstant(Object* object, |
| JSObject* holder, |
| JSFunction* function, |
| String* name, |
| CheckType check) { |
| // ----------- S t a t e ------------- |
| // -- r2 : name |
| // -- lr : return address |
| // ----------------------------------- |
| Label miss; |
| |
| // Get the receiver from the stack |
| const int argc = arguments().immediate(); |
| __ ldr(r1, MemOperand(sp, argc * kPointerSize)); |
| |
| // Check that the receiver isn't a smi. |
| if (check != NUMBER_CHECK) { |
| __ tst(r1, Operand(kSmiTagMask)); |
| __ b(eq, &miss); |
| } |
| |
| // Make sure that it's okay not to patch the on stack receiver |
| // unless we're doing a receiver map check. |
| ASSERT(!object->IsGlobalObject() || check == RECEIVER_MAP_CHECK); |
| |
| switch (check) { |
| case RECEIVER_MAP_CHECK: |
| // Check that the maps haven't changed. |
| CheckPrototypes(JSObject::cast(object), r1, holder, r3, r0, name, &miss); |
| |
| // Patch the receiver on the stack with the global proxy if |
| // necessary. |
| if (object->IsGlobalObject()) { |
| __ ldr(r3, FieldMemOperand(r1, GlobalObject::kGlobalReceiverOffset)); |
| __ str(r3, MemOperand(sp, argc * kPointerSize)); |
| } |
| break; |
| |
| case STRING_CHECK: |
| if (!function->IsBuiltin()) { |
| // Calling non-builtins with a value as receiver requires boxing. |
| __ jmp(&miss); |
| } else { |
| // Check that the object is a two-byte string or a symbol. |
| __ CompareObjectType(r1, r3, r3, FIRST_NONSTRING_TYPE); |
| __ b(hs, &miss); |
| // Check that the maps starting from the prototype haven't changed. |
| GenerateLoadGlobalFunctionPrototype(masm(), |
| Context::STRING_FUNCTION_INDEX, |
| r0); |
| CheckPrototypes(JSObject::cast(object->GetPrototype()), r0, holder, r3, |
| r1, name, &miss); |
| } |
| break; |
| |
| case NUMBER_CHECK: { |
| if (!function->IsBuiltin()) { |
| // Calling non-builtins with a value as receiver requires boxing. |
| __ jmp(&miss); |
| } else { |
| Label fast; |
| // Check that the object is a smi or a heap number. |
| __ tst(r1, Operand(kSmiTagMask)); |
| __ b(eq, &fast); |
| __ CompareObjectType(r1, r0, r0, HEAP_NUMBER_TYPE); |
| __ b(ne, &miss); |
| __ bind(&fast); |
| // Check that the maps starting from the prototype haven't changed. |
| GenerateLoadGlobalFunctionPrototype(masm(), |
| Context::NUMBER_FUNCTION_INDEX, |
| r0); |
| CheckPrototypes(JSObject::cast(object->GetPrototype()), r0, holder, r3, |
| r1, name, &miss); |
| } |
| break; |
| } |
| |
| case BOOLEAN_CHECK: { |
| if (!function->IsBuiltin()) { |
| // Calling non-builtins with a value as receiver requires boxing. |
| __ jmp(&miss); |
| } else { |
| Label fast; |
| // Check that the object is a boolean. |
| __ LoadRoot(ip, Heap::kTrueValueRootIndex); |
| __ cmp(r1, ip); |
| __ b(eq, &fast); |
| __ LoadRoot(ip, Heap::kFalseValueRootIndex); |
| __ cmp(r1, ip); |
| __ b(ne, &miss); |
| __ bind(&fast); |
| // Check that the maps starting from the prototype haven't changed. |
| GenerateLoadGlobalFunctionPrototype(masm(), |
| Context::BOOLEAN_FUNCTION_INDEX, |
| r0); |
| CheckPrototypes(JSObject::cast(object->GetPrototype()), r0, holder, r3, |
| r1, name, &miss); |
| } |
| break; |
| } |
| |
| case JSARRAY_HAS_FAST_ELEMENTS_CHECK: |
| CheckPrototypes(JSObject::cast(object), r1, holder, r3, r0, name, &miss); |
| // Make sure object->HasFastElements(). |
| // Get the elements array of the object. |
| __ ldr(r3, FieldMemOperand(r1, JSObject::kElementsOffset)); |
| // Check that the object is in fast mode (not dictionary). |
| __ ldr(r0, FieldMemOperand(r3, HeapObject::kMapOffset)); |
| __ LoadRoot(ip, Heap::kFixedArrayMapRootIndex); |
| __ cmp(r0, ip); |
| __ b(ne, &miss); |
| break; |
| |
| default: |
| UNREACHABLE(); |
| } |
| |
| __ InvokeFunction(function, arguments(), JUMP_FUNCTION); |
| |
| // Handle call cache miss. |
| __ bind(&miss); |
| Handle<Code> ic = ComputeCallMiss(arguments().immediate()); |
| __ Jump(ic, RelocInfo::CODE_TARGET); |
| |
| // Return the generated code. |
| String* function_name = NULL; |
| if (function->shared()->name()->IsString()) { |
| function_name = String::cast(function->shared()->name()); |
| } |
| return GetCode(CONSTANT_FUNCTION, function_name); |
| } |
| |
| |
| Object* CallStubCompiler::CompileCallInterceptor(JSObject* object, |
| JSObject* holder, |
| String* name) { |
| // ----------- S t a t e ------------- |
| // -- r2 : name |
| // -- lr : return address |
| // ----------------------------------- |
| ASSERT(holder->HasNamedInterceptor()); |
| ASSERT(!holder->GetNamedInterceptor()->getter()->IsUndefined()); |
| Label miss; |
| |
| const Register receiver = r0; |
| const Register holder_reg = r1; |
| const Register name_reg = r2; |
| const Register scratch = r3; |
| |
| // Get the number of arguments. |
| const int argc = arguments().immediate(); |
| |
| LookupResult lookup; |
| LookupPostInterceptor(holder, name, &lookup); |
| |
| // Get the receiver from the stack into r0. |
| __ ldr(r0, MemOperand(sp, argc * kPointerSize)); |
| |
| // Check that the receiver isn't a smi. |
| __ BranchOnSmi(receiver, &miss); |
| |
| // Check that the maps haven't changed. |
| Register reg = CheckPrototypes(object, receiver, holder, holder_reg, |
| scratch, name, &miss); |
| if (!reg.is(holder_reg)) { |
| __ mov(holder_reg, reg); |
| } |
| |
| // If we call a constant function when the interceptor returns |
| // the no-result sentinel, generate code that optimizes this case. |
| if (lookup.IsProperty() && |
| lookup.IsCacheable() && |
| lookup.type() == CONSTANT_FUNCTION && |
| lookup.GetConstantFunction()->is_compiled() && |
| !holder->IsJSArray()) { |
| // Constant functions cannot sit on global object. |
| ASSERT(!lookup.holder()->IsGlobalObject()); |
| |
| // Call the interceptor. |
| __ EnterInternalFrame(); |
| __ push(holder_reg); |
| __ push(name_reg); |
| CompileCallLoadPropertyWithInterceptor(masm(), |
| receiver, |
| holder_reg, |
| name_reg, |
| holder); |
| __ pop(name_reg); |
| __ pop(holder_reg); |
| __ LeaveInternalFrame(); |
| // r0 no longer contains the receiver. |
| |
| // If interceptor returns no-result sentinal, call the constant function. |
| __ LoadRoot(scratch, Heap::kNoInterceptorResultSentinelRootIndex); |
| __ cmp(r0, scratch); |
| Label invoke; |
| __ b(ne, &invoke); |
| // Check the prototypes between the interceptor's holder and the |
| // constant function's holder. |
| CheckPrototypes(holder, holder_reg, |
| lookup.holder(), r0, |
| scratch, |
| name, |
| &miss); |
| |
| __ InvokeFunction(lookup.GetConstantFunction(), |
| arguments(), |
| JUMP_FUNCTION); |
| |
| __ bind(&invoke); |
| |
| } else { |
| // Call a runtime function to load the interceptor property. |
| __ EnterInternalFrame(); |
| __ push(name_reg); |
| |
| PushInterceptorArguments(masm(), receiver, holder_reg, name_reg, holder); |
| |
| __ CallExternalReference( |
| ExternalReference(IC_Utility(IC::kLoadPropertyWithInterceptorForCall)), |
| 5); |
| |
| __ pop(name_reg); |
| __ LeaveInternalFrame(); |
| } |
| |
| // Move returned value, the function to call, to r1. |
| __ mov(r1, r0); |
| // Restore receiver. |
| __ ldr(receiver, MemOperand(sp, argc * kPointerSize)); |
| |
| GenerateCallFunction(masm(), object, arguments(), &miss); |
| |
| // Handle call cache miss. |
| __ bind(&miss); |
| Handle<Code> ic = ComputeCallMiss(arguments().immediate()); |
| __ Jump(ic, RelocInfo::CODE_TARGET); |
| |
| // Return the generated code. |
| return GetCode(INTERCEPTOR, name); |
| } |
| |
| |
| Object* CallStubCompiler::CompileCallGlobal(JSObject* object, |
| GlobalObject* holder, |
| JSGlobalPropertyCell* cell, |
| JSFunction* function, |
| String* name) { |
| // ----------- S t a t e ------------- |
| // -- r2 : name |
| // -- lr : return address |
| // ----------------------------------- |
| Label miss; |
| |
| // Get the number of arguments. |
| const int argc = arguments().immediate(); |
| |
| // Get the receiver from the stack. |
| __ ldr(r0, MemOperand(sp, argc * kPointerSize)); |
| |
| // If the object is the holder then we know that it's a global |
| // object which can only happen for contextual calls. In this case, |
| // the receiver cannot be a smi. |
| if (object != holder) { |
| __ tst(r0, Operand(kSmiTagMask)); |
| __ b(eq, &miss); |
| } |
| |
| // Check that the maps haven't changed. |
| CheckPrototypes(object, r0, holder, r3, r1, name, &miss); |
| |
| // Get the value from the cell. |
| __ mov(r3, Operand(Handle<JSGlobalPropertyCell>(cell))); |
| __ ldr(r1, FieldMemOperand(r3, JSGlobalPropertyCell::kValueOffset)); |
| |
| // Check that the cell contains the same function. |
| if (Heap::InNewSpace(function)) { |
| // We can't embed a pointer to a function in new space so we have |
| // to verify that the shared function info is unchanged. This has |
| // the nice side effect that multiple closures based on the same |
| // function can all use this call IC. Before we load through the |
| // function, we have to verify that it still is a function. |
| __ tst(r1, Operand(kSmiTagMask)); |
| __ b(eq, &miss); |
| __ CompareObjectType(r1, r3, r3, JS_FUNCTION_TYPE); |
| __ b(ne, &miss); |
| |
| // Check the shared function info. Make sure it hasn't changed. |
| __ mov(r3, Operand(Handle<SharedFunctionInfo>(function->shared()))); |
| __ ldr(r4, FieldMemOperand(r1, JSFunction::kSharedFunctionInfoOffset)); |
| __ cmp(r4, r3); |
| __ b(ne, &miss); |
| } else { |
| __ cmp(r1, Operand(Handle<JSFunction>(function))); |
| __ b(ne, &miss); |
| } |
| |
| // Patch the receiver on the stack with the global proxy if |
| // necessary. |
| if (object->IsGlobalObject()) { |
| __ ldr(r3, FieldMemOperand(r0, GlobalObject::kGlobalReceiverOffset)); |
| __ str(r3, MemOperand(sp, argc * kPointerSize)); |
| } |
| |
| // Setup the context (function already in r1). |
| __ ldr(cp, FieldMemOperand(r1, JSFunction::kContextOffset)); |
| |
| // Jump to the cached code (tail call). |
| __ IncrementCounter(&Counters::call_global_inline, 1, r1, r3); |
| ASSERT(function->is_compiled()); |
| Handle<Code> code(function->code()); |
| ParameterCount expected(function->shared()->formal_parameter_count()); |
| __ InvokeCode(code, expected, arguments(), |
| RelocInfo::CODE_TARGET, JUMP_FUNCTION); |
| |
| // Handle call cache miss. |
| __ bind(&miss); |
| __ IncrementCounter(&Counters::call_global_inline_miss, 1, r1, r3); |
| Handle<Code> ic = ComputeCallMiss(arguments().immediate()); |
| __ Jump(ic, RelocInfo::CODE_TARGET); |
| |
| // Return the generated code. |
| return GetCode(NORMAL, name); |
| } |
| |
| |
| Object* StoreStubCompiler::CompileStoreField(JSObject* object, |
| int index, |
| Map* transition, |
| String* name) { |
| // ----------- S t a t e ------------- |
| // -- r0 : value |
| // -- r1 : receiver |
| // -- r2 : name |
| // -- lr : return address |
| // ----------------------------------- |
| Label miss; |
| |
| GenerateStoreField(masm(), |
| object, |
| index, |
| transition, |
| r1, r2, r3, |
| &miss); |
| __ bind(&miss); |
| Handle<Code> ic(Builtins::builtin(Builtins::StoreIC_Miss)); |
| __ Jump(ic, RelocInfo::CODE_TARGET); |
| |
| // Return the generated code. |
| return GetCode(transition == NULL ? FIELD : MAP_TRANSITION, name); |
| } |
| |
| |
| Object* StoreStubCompiler::CompileStoreCallback(JSObject* object, |
| AccessorInfo* callback, |
| String* name) { |
| // ----------- S t a t e ------------- |
| // -- r0 : value |
| // -- r1 : receiver |
| // -- r2 : name |
| // -- lr : return address |
| // ----------------------------------- |
| Label miss; |
| |
| // Check that the object isn't a smi. |
| __ tst(r1, Operand(kSmiTagMask)); |
| __ b(eq, &miss); |
| |
| // Check that the map of the object hasn't changed. |
| __ ldr(r3, FieldMemOperand(r1, HeapObject::kMapOffset)); |
| __ cmp(r3, Operand(Handle<Map>(object->map()))); |
| __ b(ne, &miss); |
| |
| // Perform global security token check if needed. |
| if (object->IsJSGlobalProxy()) { |
| __ CheckAccessGlobalProxy(r1, r3, &miss); |
| } |
| |
| // Stub never generated for non-global objects that require access |
| // checks. |
| ASSERT(object->IsJSGlobalProxy() || !object->IsAccessCheckNeeded()); |
| |
| __ push(r1); // receiver |
| __ mov(ip, Operand(Handle<AccessorInfo>(callback))); // callback info |
| __ stm(db_w, sp, ip.bit() | r2.bit() | r0.bit()); |
| |
| // Do tail-call to the runtime system. |
| ExternalReference store_callback_property = |
| ExternalReference(IC_Utility(IC::kStoreCallbackProperty)); |
| __ TailCallRuntime(store_callback_property, 4, 1); |
| |
| // Handle store cache miss. |
| __ bind(&miss); |
| Handle<Code> ic(Builtins::builtin(Builtins::StoreIC_Miss)); |
| __ Jump(ic, RelocInfo::CODE_TARGET); |
| |
| // Return the generated code. |
| return GetCode(CALLBACKS, name); |
| } |
| |
| |
| Object* StoreStubCompiler::CompileStoreInterceptor(JSObject* receiver, |
| String* name) { |
| // ----------- S t a t e ------------- |
| // -- r0 : value |
| // -- r1 : receiver |
| // -- r2 : name |
| // -- lr : return address |
| // ----------------------------------- |
| Label miss; |
| |
| // Check that the object isn't a smi. |
| __ tst(r1, Operand(kSmiTagMask)); |
| __ b(eq, &miss); |
| |
| // Check that the map of the object hasn't changed. |
| __ ldr(r3, FieldMemOperand(r1, HeapObject::kMapOffset)); |
| __ cmp(r3, Operand(Handle<Map>(receiver->map()))); |
| __ b(ne, &miss); |
| |
| // Perform global security token check if needed. |
| if (receiver->IsJSGlobalProxy()) { |
| __ CheckAccessGlobalProxy(r1, r3, &miss); |
| } |
| |
| // Stub is never generated for non-global objects that require access |
| // checks. |
| ASSERT(receiver->IsJSGlobalProxy() || !receiver->IsAccessCheckNeeded()); |
| |
| __ push(r1); // receiver. |
| __ push(r2); // name. |
| __ push(r0); // value. |
| |
| // Do tail-call to the runtime system. |
| ExternalReference store_ic_property = |
| ExternalReference(IC_Utility(IC::kStoreInterceptorProperty)); |
| __ TailCallRuntime(store_ic_property, 3, 1); |
| |
| // Handle store cache miss. |
| __ bind(&miss); |
| Handle<Code> ic(Builtins::builtin(Builtins::StoreIC_Miss)); |
| __ Jump(ic, RelocInfo::CODE_TARGET); |
| |
| // Return the generated code. |
| return GetCode(INTERCEPTOR, name); |
| } |
| |
| |
| Object* StoreStubCompiler::CompileStoreGlobal(GlobalObject* object, |
| JSGlobalPropertyCell* cell, |
| String* name) { |
| // ----------- S t a t e ------------- |
| // -- r0 : value |
| // -- r1 : receiver |
| // -- r2 : name |
| // -- lr : return address |
| // ----------------------------------- |
| Label miss; |
| |
| // Check that the map of the global has not changed. |
| __ ldr(r3, FieldMemOperand(r1, HeapObject::kMapOffset)); |
| __ cmp(r3, Operand(Handle<Map>(object->map()))); |
| __ b(ne, &miss); |
| |
| // Store the value in the cell. |
| __ mov(r2, Operand(Handle<JSGlobalPropertyCell>(cell))); |
| __ str(r0, FieldMemOperand(r2, JSGlobalPropertyCell::kValueOffset)); |
| |
| __ IncrementCounter(&Counters::named_store_global_inline, 1, r4, r3); |
| __ Ret(); |
| |
| // Handle store cache miss. |
| __ bind(&miss); |
| __ IncrementCounter(&Counters::named_store_global_inline_miss, 1, r4, r3); |
| Handle<Code> ic(Builtins::builtin(Builtins::StoreIC_Miss)); |
| __ Jump(ic, RelocInfo::CODE_TARGET); |
| |
| // Return the generated code. |
| return GetCode(NORMAL, name); |
| } |
| |
| |
| Object* LoadStubCompiler::CompileLoadField(JSObject* object, |
| JSObject* holder, |
| int index, |
| String* name) { |
| // ----------- S t a t e ------------- |
| // -- r2 : name |
| // -- lr : return address |
| // -- [sp] : receiver |
| // ----------------------------------- |
| Label miss; |
| |
| __ ldr(r0, MemOperand(sp, 0)); |
| |
| GenerateLoadField(object, holder, r0, r3, r1, index, name, &miss); |
| __ bind(&miss); |
| GenerateLoadMiss(masm(), Code::LOAD_IC); |
| |
| // Return the generated code. |
| return GetCode(FIELD, name); |
| } |
| |
| |
| Object* LoadStubCompiler::CompileLoadCallback(String* name, |
| JSObject* object, |
| JSObject* holder, |
| AccessorInfo* callback) { |
| // ----------- S t a t e ------------- |
| // -- r2 : name |
| // -- lr : return address |
| // -- [sp] : receiver |
| // ----------------------------------- |
| Label miss; |
| |
| __ ldr(r0, MemOperand(sp, 0)); |
| Failure* failure = Failure::InternalError(); |
| bool success = GenerateLoadCallback(object, holder, r0, r2, r3, r1, |
| callback, name, &miss, &failure); |
| if (!success) return failure; |
| |
| __ bind(&miss); |
| GenerateLoadMiss(masm(), Code::LOAD_IC); |
| |
| // Return the generated code. |
| return GetCode(CALLBACKS, name); |
| } |
| |
| |
| Object* LoadStubCompiler::CompileLoadConstant(JSObject* object, |
| JSObject* holder, |
| Object* value, |
| String* name) { |
| // ----------- S t a t e ------------- |
| // -- r2 : name |
| // -- lr : return address |
| // -- [sp] : receiver |
| // ----------------------------------- |
| Label miss; |
| |
| __ ldr(r0, MemOperand(sp, 0)); |
| |
| GenerateLoadConstant(object, holder, r0, r3, r1, value, name, &miss); |
| __ bind(&miss); |
| GenerateLoadMiss(masm(), Code::LOAD_IC); |
| |
| // Return the generated code. |
| return GetCode(CONSTANT_FUNCTION, name); |
| } |
| |
| |
| Object* LoadStubCompiler::CompileLoadInterceptor(JSObject* object, |
| JSObject* holder, |
| String* name) { |
| // ----------- S t a t e ------------- |
| // -- r2 : name |
| // -- lr : return address |
| // -- [sp] : receiver |
| // ----------------------------------- |
| Label miss; |
| |
| __ ldr(r0, MemOperand(sp, 0)); |
| |
| LookupResult lookup; |
| LookupPostInterceptor(holder, name, &lookup); |
| GenerateLoadInterceptor(object, |
| holder, |
| &lookup, |
| r0, |
| r2, |
| r3, |
| r1, |
| name, |
| &miss); |
| __ bind(&miss); |
| GenerateLoadMiss(masm(), Code::LOAD_IC); |
| |
| // Return the generated code. |
| return GetCode(INTERCEPTOR, name); |
| } |
| |
| |
| Object* LoadStubCompiler::CompileLoadGlobal(JSObject* object, |
| GlobalObject* holder, |
| JSGlobalPropertyCell* cell, |
| String* name, |
| bool is_dont_delete) { |
| // ----------- S t a t e ------------- |
| // -- r2 : name |
| // -- lr : return address |
| // -- [sp] : receiver |
| // ----------------------------------- |
| Label miss; |
| |
| // Get the receiver from the stack. |
| __ ldr(r1, MemOperand(sp, 0 * kPointerSize)); |
| |
| // If the object is the holder then we know that it's a global |
| // object which can only happen for contextual calls. In this case, |
| // the receiver cannot be a smi. |
| if (object != holder) { |
| __ tst(r1, Operand(kSmiTagMask)); |
| __ b(eq, &miss); |
| } |
| |
| // Check that the map of the global has not changed. |
| CheckPrototypes(object, r1, holder, r3, r0, name, &miss); |
| |
| // Get the value from the cell. |
| __ mov(r3, Operand(Handle<JSGlobalPropertyCell>(cell))); |
| __ ldr(r0, FieldMemOperand(r3, JSGlobalPropertyCell::kValueOffset)); |
| |
| // Check for deleted property if property can actually be deleted. |
| if (!is_dont_delete) { |
| __ LoadRoot(ip, Heap::kTheHoleValueRootIndex); |
| __ cmp(r0, ip); |
| __ b(eq, &miss); |
| } |
| |
| __ IncrementCounter(&Counters::named_load_global_inline, 1, r1, r3); |
| __ Ret(); |
| |
| __ bind(&miss); |
| __ IncrementCounter(&Counters::named_load_global_inline_miss, 1, r1, r3); |
| GenerateLoadMiss(masm(), Code::LOAD_IC); |
| |
| // Return the generated code. |
| return GetCode(NORMAL, name); |
| } |
| |
| |
| Object* KeyedLoadStubCompiler::CompileLoadField(String* name, |
| JSObject* receiver, |
| JSObject* holder, |
| int index) { |
| // ----------- S t a t e ------------- |
| // -- lr : return address |
| // -- sp[0] : key |
| // -- sp[4] : receiver |
| // ----------------------------------- |
| Label miss; |
| |
| __ ldr(r2, MemOperand(sp, 0)); |
| __ ldr(r0, MemOperand(sp, kPointerSize)); |
| |
| __ cmp(r2, Operand(Handle<String>(name))); |
| __ b(ne, &miss); |
| |
| GenerateLoadField(receiver, holder, r0, r3, r1, index, name, &miss); |
| __ bind(&miss); |
| GenerateLoadMiss(masm(), Code::KEYED_LOAD_IC); |
| |
| return GetCode(FIELD, name); |
| } |
| |
| |
| Object* KeyedLoadStubCompiler::CompileLoadCallback(String* name, |
| JSObject* receiver, |
| JSObject* holder, |
| AccessorInfo* callback) { |
| // ----------- S t a t e ------------- |
| // -- lr : return address |
| // -- sp[0] : key |
| // -- sp[4] : receiver |
| // ----------------------------------- |
| Label miss; |
| |
| __ ldr(r2, MemOperand(sp, 0)); |
| __ ldr(r0, MemOperand(sp, kPointerSize)); |
| |
| __ cmp(r2, Operand(Handle<String>(name))); |
| __ b(ne, &miss); |
| |
| Failure* failure = Failure::InternalError(); |
| bool success = GenerateLoadCallback(receiver, holder, r0, r2, r3, r1, |
| callback, name, &miss, &failure); |
| if (!success) return failure; |
| |
| __ bind(&miss); |
| GenerateLoadMiss(masm(), Code::KEYED_LOAD_IC); |
| |
| return GetCode(CALLBACKS, name); |
| } |
| |
| |
| Object* KeyedLoadStubCompiler::CompileLoadConstant(String* name, |
| JSObject* receiver, |
| JSObject* holder, |
| Object* value) { |
| // ----------- S t a t e ------------- |
| // -- lr : return address |
| // -- sp[0] : key |
| // -- sp[4] : receiver |
| // ----------------------------------- |
| Label miss; |
| |
| // Check the key is the cached one |
| __ ldr(r2, MemOperand(sp, 0)); |
| __ ldr(r0, MemOperand(sp, kPointerSize)); |
| |
| __ cmp(r2, Operand(Handle<String>(name))); |
| __ b(ne, &miss); |
| |
| GenerateLoadConstant(receiver, holder, r0, r3, r1, value, name, &miss); |
| __ bind(&miss); |
| GenerateLoadMiss(masm(), Code::KEYED_LOAD_IC); |
| |
| // Return the generated code. |
| return GetCode(CONSTANT_FUNCTION, name); |
| } |
| |
| |
| Object* KeyedLoadStubCompiler::CompileLoadInterceptor(JSObject* receiver, |
| JSObject* holder, |
| String* name) { |
| // ----------- S t a t e ------------- |
| // -- lr : return address |
| // -- sp[0] : key |
| // -- sp[4] : receiver |
| // ----------------------------------- |
| Label miss; |
| |
| // Check the key is the cached one |
| __ ldr(r2, MemOperand(sp, 0)); |
| __ ldr(r0, MemOperand(sp, kPointerSize)); |
| |
| __ cmp(r2, Operand(Handle<String>(name))); |
| __ b(ne, &miss); |
| |
| LookupResult lookup; |
| LookupPostInterceptor(holder, name, &lookup); |
| GenerateLoadInterceptor(receiver, |
| holder, |
| &lookup, |
| r0, |
| r2, |
| r3, |
| r1, |
| name, |
| &miss); |
| __ bind(&miss); |
| GenerateLoadMiss(masm(), Code::KEYED_LOAD_IC); |
| |
| return GetCode(INTERCEPTOR, name); |
| } |
| |
| |
| Object* KeyedLoadStubCompiler::CompileLoadArrayLength(String* name) { |
| // ----------- S t a t e ------------- |
| // -- lr : return address |
| // -- sp[0] : key |
| // -- sp[4] : receiver |
| // ----------------------------------- |
| Label miss; |
| |
| // Check the key is the cached one |
| __ ldr(r2, MemOperand(sp, 0)); |
| __ ldr(r0, MemOperand(sp, kPointerSize)); |
| |
| __ cmp(r2, Operand(Handle<String>(name))); |
| __ b(ne, &miss); |
| |
| GenerateLoadArrayLength(masm(), r0, r3, &miss); |
| __ bind(&miss); |
| GenerateLoadMiss(masm(), Code::KEYED_LOAD_IC); |
| |
| return GetCode(CALLBACKS, name); |
| } |
| |
| |
| Object* KeyedLoadStubCompiler::CompileLoadStringLength(String* name) { |
| // ----------- S t a t e ------------- |
| // -- lr : return address |
| // -- sp[0] : key |
| // -- sp[4] : receiver |
| // ----------------------------------- |
| Label miss; |
| __ IncrementCounter(&Counters::keyed_load_string_length, 1, r1, r3); |
| |
| __ ldr(r2, MemOperand(sp)); |
| __ ldr(r0, MemOperand(sp, kPointerSize)); // receiver |
| |
| __ cmp(r2, Operand(Handle<String>(name))); |
| __ b(ne, &miss); |
| |
| GenerateLoadStringLength(masm(), r0, r1, r3, &miss); |
| __ bind(&miss); |
| __ DecrementCounter(&Counters::keyed_load_string_length, 1, r1, r3); |
| |
| GenerateLoadMiss(masm(), Code::KEYED_LOAD_IC); |
| |
| return GetCode(CALLBACKS, name); |
| } |
| |
| |
| // TODO(1224671): implement the fast case. |
| Object* KeyedLoadStubCompiler::CompileLoadFunctionPrototype(String* name) { |
| // ----------- S t a t e ------------- |
| // -- lr : return address |
| // -- sp[0] : key |
| // -- sp[4] : receiver |
| // ----------------------------------- |
| GenerateLoadMiss(masm(), Code::KEYED_LOAD_IC); |
| |
| return GetCode(CALLBACKS, name); |
| } |
| |
| |
| Object* KeyedStoreStubCompiler::CompileStoreField(JSObject* object, |
| int index, |
| Map* transition, |
| String* name) { |
| // ----------- S t a t e ------------- |
| // -- r0 : value |
| // -- r2 : name |
| // -- lr : return address |
| // -- [sp] : receiver |
| // ----------------------------------- |
| Label miss; |
| |
| __ IncrementCounter(&Counters::keyed_store_field, 1, r1, r3); |
| |
| // Check that the name has not changed. |
| __ cmp(r2, Operand(Handle<String>(name))); |
| __ b(ne, &miss); |
| |
| // Load receiver from the stack. |
| __ ldr(r3, MemOperand(sp)); |
| // r1 is used as scratch register, r3 and r2 might be clobbered. |
| GenerateStoreField(masm(), |
| object, |
| index, |
| transition, |
| r3, r2, r1, |
| &miss); |
| __ bind(&miss); |
| |
| __ DecrementCounter(&Counters::keyed_store_field, 1, r1, r3); |
| __ mov(r2, Operand(Handle<String>(name))); // restore name register. |
| Handle<Code> ic(Builtins::builtin(Builtins::KeyedStoreIC_Miss)); |
| __ Jump(ic, RelocInfo::CODE_TARGET); |
| |
| // Return the generated code. |
| return GetCode(transition == NULL ? FIELD : MAP_TRANSITION, name); |
| } |
| |
| |
| Object* ConstructStubCompiler::CompileConstructStub( |
| SharedFunctionInfo* shared) { |
| // ----------- S t a t e ------------- |
| // -- r0 : argc |
| // -- r1 : constructor |
| // -- lr : return address |
| // -- [sp] : last argument |
| // ----------------------------------- |
| Label generic_stub_call; |
| |
| // Use r7 for holding undefined which is used in several places below. |
| __ LoadRoot(r7, Heap::kUndefinedValueRootIndex); |
| |
| #ifdef ENABLE_DEBUGGER_SUPPORT |
| // Check to see whether there are any break points in the function code. If |
| // there are jump to the generic constructor stub which calls the actual |
| // code for the function thereby hitting the break points. |
| __ ldr(r2, FieldMemOperand(r1, JSFunction::kSharedFunctionInfoOffset)); |
| __ ldr(r2, FieldMemOperand(r2, SharedFunctionInfo::kDebugInfoOffset)); |
| __ cmp(r2, r7); |
| __ b(ne, &generic_stub_call); |
| #endif |
| |
| // Load the initial map and verify that it is in fact a map. |
| // r1: constructor function |
| // r7: undefined |
| __ ldr(r2, FieldMemOperand(r1, JSFunction::kPrototypeOrInitialMapOffset)); |
| __ tst(r2, Operand(kSmiTagMask)); |
| __ b(eq, &generic_stub_call); |
| __ CompareObjectType(r2, r3, r4, MAP_TYPE); |
| __ b(ne, &generic_stub_call); |
| |
| #ifdef DEBUG |
| // Cannot construct functions this way. |
| // r0: argc |
| // r1: constructor function |
| // r2: initial map |
| // r7: undefined |
| __ CompareInstanceType(r2, r3, JS_FUNCTION_TYPE); |
| __ Check(ne, "Function constructed by construct stub."); |
| #endif |
| |
| // Now allocate the JSObject in new space. |
| // r0: argc |
| // r1: constructor function |
| // r2: initial map |
| // r7: undefined |
| __ ldrb(r3, FieldMemOperand(r2, Map::kInstanceSizeOffset)); |
| __ AllocateInNewSpace(r3, |
| r4, |
| r5, |
| r6, |
| &generic_stub_call, |
| NO_ALLOCATION_FLAGS); |
| |
| // Allocated the JSObject, now initialize the fields. Map is set to initial |
| // map and properties and elements are set to empty fixed array. |
| // r0: argc |
| // r1: constructor function |
| // r2: initial map |
| // r3: object size (in words) |
| // r4: JSObject (not tagged) |
| // r7: undefined |
| __ LoadRoot(r6, Heap::kEmptyFixedArrayRootIndex); |
| __ mov(r5, r4); |
| ASSERT_EQ(0 * kPointerSize, JSObject::kMapOffset); |
| __ str(r2, MemOperand(r5, kPointerSize, PostIndex)); |
| ASSERT_EQ(1 * kPointerSize, JSObject::kPropertiesOffset); |
| __ str(r6, MemOperand(r5, kPointerSize, PostIndex)); |
| ASSERT_EQ(2 * kPointerSize, JSObject::kElementsOffset); |
| __ str(r6, MemOperand(r5, kPointerSize, PostIndex)); |
| |
| // Calculate the location of the first argument. The stack contains only the |
| // argc arguments. |
| __ add(r1, sp, Operand(r0, LSL, kPointerSizeLog2)); |
| |
| // Fill all the in-object properties with undefined. |
| // r0: argc |
| // r1: first argument |
| // r3: object size (in words) |
| // r4: JSObject (not tagged) |
| // r5: First in-object property of JSObject (not tagged) |
| // r7: undefined |
| // Fill the initialized properties with a constant value or a passed argument |
| // depending on the this.x = ...; assignment in the function. |
| for (int i = 0; i < shared->this_property_assignments_count(); i++) { |
| if (shared->IsThisPropertyAssignmentArgument(i)) { |
| Label not_passed, next; |
| // Check if the argument assigned to the property is actually passed. |
| int arg_number = shared->GetThisPropertyAssignmentArgument(i); |
| __ cmp(r0, Operand(arg_number)); |
| __ b(le, ¬_passed); |
| // Argument passed - find it on the stack. |
| __ ldr(r2, MemOperand(r1, (arg_number + 1) * -kPointerSize)); |
| __ str(r2, MemOperand(r5, kPointerSize, PostIndex)); |
| __ b(&next); |
| __ bind(¬_passed); |
| // Set the property to undefined. |
| __ str(r7, MemOperand(r5, kPointerSize, PostIndex)); |
| __ bind(&next); |
| } else { |
| // Set the property to the constant value. |
| Handle<Object> constant(shared->GetThisPropertyAssignmentConstant(i)); |
| __ mov(r2, Operand(constant)); |
| __ str(r2, MemOperand(r5, kPointerSize, PostIndex)); |
| } |
| } |
| |
| // Fill the unused in-object property fields with undefined. |
| for (int i = shared->this_property_assignments_count(); |
| i < shared->CalculateInObjectProperties(); |
| i++) { |
| __ str(r7, MemOperand(r5, kPointerSize, PostIndex)); |
| } |
| |
| // r0: argc |
| // r4: JSObject (not tagged) |
| // Move argc to r1 and the JSObject to return to r0 and tag it. |
| __ mov(r1, r0); |
| __ mov(r0, r4); |
| __ orr(r0, r0, Operand(kHeapObjectTag)); |
| |
| // r0: JSObject |
| // r1: argc |
| // Remove caller arguments and receiver from the stack and return. |
| __ add(sp, sp, Operand(r1, LSL, kPointerSizeLog2)); |
| __ add(sp, sp, Operand(kPointerSize)); |
| __ IncrementCounter(&Counters::constructed_objects, 1, r1, r2); |
| __ IncrementCounter(&Counters::constructed_objects_stub, 1, r1, r2); |
| __ Jump(lr); |
| |
| // Jump to the generic stub in case the specialized code cannot handle the |
| // construction. |
| __ bind(&generic_stub_call); |
| Code* code = Builtins::builtin(Builtins::JSConstructStubGeneric); |
| Handle<Code> generic_construct_stub(code); |
| __ Jump(generic_construct_stub, RelocInfo::CODE_TARGET); |
| |
| // Return the generated code. |
| return GetCode(); |
| } |
| |
| |
| #undef __ |
| |
| } } // namespace v8::internal |