| // Copyright 2006-2008 the V8 project authors. All rights reserved. |
| // Redistribution and use in source and binary forms, with or without |
| // modification, are permitted provided that the following conditions are |
| // met: |
| // |
| // * Redistributions of source code must retain the above copyright |
| // notice, this list of conditions and the following disclaimer. |
| // * Redistributions in binary form must reproduce the above |
| // copyright notice, this list of conditions and the following |
| // disclaimer in the documentation and/or other materials provided |
| // with the distribution. |
| // * Neither the name of Google Inc. nor the names of its |
| // contributors may be used to endorse or promote products derived |
| // from this software without specific prior written permission. |
| // |
| // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS |
| // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT |
| // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR |
| // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT |
| // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, |
| // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT |
| // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
| // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
| // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
| // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
| // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| |
| #ifndef V8_STUB_CACHE_H_ |
| #define V8_STUB_CACHE_H_ |
| |
| #include "macro-assembler.h" |
| |
| namespace v8 { |
| namespace internal { |
| |
| |
| // The stub cache is used for megamorphic calls and property accesses. |
| // It maps (map, name, type)->Code* |
| |
| // The design of the table uses the inline cache stubs used for |
| // mono-morphic calls. The beauty of this, we do not have to |
| // invalidate the cache whenever a prototype map is changed. The stub |
| // validates the map chain as in the mono-morphic case. |
| |
| class SCTableReference; |
| |
| class StubCache : public AllStatic { |
| public: |
| struct Entry { |
| String* key; |
| Code* value; |
| }; |
| |
| |
| static void Initialize(bool create_heap_objects); |
| |
| // Computes the right stub matching. Inserts the result in the |
| // cache before returning. This might compile a stub if needed. |
| static Object* ComputeLoadField(String* name, |
| JSObject* receiver, |
| JSObject* holder, |
| int field_index); |
| |
| static Object* ComputeLoadCallback(String* name, |
| JSObject* receiver, |
| JSObject* holder, |
| AccessorInfo* callback); |
| |
| static Object* ComputeLoadConstant(String* name, |
| JSObject* receiver, |
| JSObject* holder, |
| Object* value); |
| |
| static Object* ComputeLoadInterceptor(String* name, |
| JSObject* receiver, |
| JSObject* holder); |
| |
| static Object* ComputeLoadNormal(String* name, JSObject* receiver); |
| |
| |
| static Object* ComputeLoadGlobal(String* name, |
| JSObject* receiver, |
| GlobalObject* holder, |
| JSGlobalPropertyCell* cell, |
| bool is_dont_delete); |
| |
| |
| // --- |
| |
| static Object* ComputeKeyedLoadField(String* name, |
| JSObject* receiver, |
| JSObject* holder, |
| int field_index); |
| |
| static Object* ComputeKeyedLoadCallback(String* name, |
| JSObject* receiver, |
| JSObject* holder, |
| AccessorInfo* callback); |
| |
| static Object* ComputeKeyedLoadConstant(String* name, JSObject* receiver, |
| JSObject* holder, Object* value); |
| |
| static Object* ComputeKeyedLoadInterceptor(String* name, |
| JSObject* receiver, |
| JSObject* holder); |
| |
| static Object* ComputeKeyedLoadArrayLength(String* name, JSArray* receiver); |
| |
| static Object* ComputeKeyedLoadStringLength(String* name, |
| String* receiver); |
| |
| static Object* ComputeKeyedLoadFunctionPrototype(String* name, |
| JSFunction* receiver); |
| |
| // --- |
| |
| static Object* ComputeStoreField(String* name, |
| JSObject* receiver, |
| int field_index, |
| Map* transition = NULL); |
| |
| static Object* ComputeStoreGlobal(String* name, |
| GlobalObject* receiver, |
| JSGlobalPropertyCell* cell); |
| |
| static Object* ComputeStoreCallback(String* name, |
| JSObject* receiver, |
| AccessorInfo* callback); |
| |
| static Object* ComputeStoreInterceptor(String* name, JSObject* receiver); |
| |
| // --- |
| |
| static Object* ComputeKeyedStoreField(String* name, |
| JSObject* receiver, |
| int field_index, |
| Map* transition = NULL); |
| |
| // --- |
| |
| static Object* ComputeCallField(int argc, |
| InLoopFlag in_loop, |
| String* name, |
| Object* object, |
| JSObject* holder, |
| int index); |
| |
| static Object* ComputeCallConstant(int argc, |
| InLoopFlag in_loop, |
| String* name, |
| Object* object, |
| JSObject* holder, |
| JSFunction* function); |
| |
| static Object* ComputeCallNormal(int argc, |
| InLoopFlag in_loop, |
| String* name, |
| JSObject* receiver); |
| |
| static Object* ComputeCallInterceptor(int argc, |
| String* name, |
| Object* object, |
| JSObject* holder); |
| |
| static Object* ComputeCallGlobal(int argc, |
| InLoopFlag in_loop, |
| String* name, |
| JSObject* receiver, |
| GlobalObject* holder, |
| JSGlobalPropertyCell* cell, |
| JSFunction* function); |
| |
| // --- |
| |
| static Object* ComputeCallInitialize(int argc, InLoopFlag in_loop); |
| static Object* ComputeCallPreMonomorphic(int argc, InLoopFlag in_loop); |
| static Object* ComputeCallNormal(int argc, InLoopFlag in_loop); |
| static Object* ComputeCallMegamorphic(int argc, InLoopFlag in_loop); |
| static Object* ComputeCallMiss(int argc); |
| |
| // Finds the Code object stored in the Heap::non_monomorphic_cache(). |
| static Code* FindCallInitialize(int argc, InLoopFlag in_loop); |
| |
| #ifdef ENABLE_DEBUGGER_SUPPORT |
| static Object* ComputeCallDebugBreak(int argc); |
| static Object* ComputeCallDebugPrepareStepIn(int argc); |
| #endif |
| |
| static Object* ComputeLazyCompile(int argc); |
| |
| |
| // Update cache for entry hash(name, map). |
| static Code* Set(String* name, Map* map, Code* code); |
| |
| // Clear the lookup table (@ mark compact collection). |
| static void Clear(); |
| |
| // Functions for generating stubs at startup. |
| static void GenerateMiss(MacroAssembler* masm); |
| |
| // Generate code for probing the stub cache table. |
| // If extra != no_reg it might be used as am extra scratch register. |
| static void GenerateProbe(MacroAssembler* masm, |
| Code::Flags flags, |
| Register receiver, |
| Register name, |
| Register scratch, |
| Register extra); |
| |
| enum Table { |
| kPrimary, |
| kSecondary |
| }; |
| |
| private: |
| friend class SCTableReference; |
| static const int kPrimaryTableSize = 2048; |
| static const int kSecondaryTableSize = 512; |
| static Entry primary_[]; |
| static Entry secondary_[]; |
| |
| // Computes the hashed offsets for primary and secondary caches. |
| static int PrimaryOffset(String* name, Code::Flags flags, Map* map) { |
| // This works well because the heap object tag size and the hash |
| // shift are equal. Shifting down the length field to get the |
| // hash code would effectively throw away two bits of the hash |
| // code. |
| ASSERT(kHeapObjectTagSize == String::kHashShift); |
| // Compute the hash of the name (use entire length field). |
| ASSERT(name->HasHashCode()); |
| uint32_t field = name->length_field(); |
| // Using only the low bits in 64-bit mode is unlikely to increase the |
| // risk of collision even if the heap is spread over an area larger than |
| // 4Gb (and not at all if it isn't). |
| uint32_t map_low32bits = |
| static_cast<uint32_t>(reinterpret_cast<uintptr_t>(map)); |
| // We always set the in_loop bit to zero when generating the lookup code |
| // so do it here too so the hash codes match. |
| uint32_t iflags = |
| (static_cast<uint32_t>(flags) & ~Code::kFlagsNotUsedInLookup); |
| // Base the offset on a simple combination of name, flags, and map. |
| uint32_t key = (map_low32bits + field) ^ iflags; |
| return key & ((kPrimaryTableSize - 1) << kHeapObjectTagSize); |
| } |
| |
| static int SecondaryOffset(String* name, Code::Flags flags, int seed) { |
| // Use the seed from the primary cache in the secondary cache. |
| uint32_t string_low32bits = |
| static_cast<uint32_t>(reinterpret_cast<uintptr_t>(name)); |
| // We always set the in_loop bit to zero when generating the lookup code |
| // so do it here too so the hash codes match. |
| uint32_t iflags = |
| (static_cast<uint32_t>(flags) & ~Code::kFlagsICInLoopMask); |
| uint32_t key = seed - string_low32bits + iflags; |
| return key & ((kSecondaryTableSize - 1) << kHeapObjectTagSize); |
| } |
| |
| // Compute the entry for a given offset in exactly the same way as |
| // we do in generated code. We generate an hash code that already |
| // ends in String::kHashShift 0s. Then we shift it so it is a multiple |
| // of sizeof(Entry). This makes it easier to avoid making mistakes |
| // in the hashed offset computations. |
| static Entry* entry(Entry* table, int offset) { |
| const int shift_amount = kPointerSizeLog2 + 1 - String::kHashShift; |
| return reinterpret_cast<Entry*>( |
| reinterpret_cast<Address>(table) + (offset << shift_amount)); |
| } |
| }; |
| |
| |
| class SCTableReference { |
| public: |
| static SCTableReference keyReference(StubCache::Table table) { |
| return SCTableReference( |
| reinterpret_cast<Address>(&first_entry(table)->key)); |
| } |
| |
| |
| static SCTableReference valueReference(StubCache::Table table) { |
| return SCTableReference( |
| reinterpret_cast<Address>(&first_entry(table)->value)); |
| } |
| |
| Address address() const { return address_; } |
| |
| private: |
| explicit SCTableReference(Address address) : address_(address) {} |
| |
| static StubCache::Entry* first_entry(StubCache::Table table) { |
| switch (table) { |
| case StubCache::kPrimary: return StubCache::primary_; |
| case StubCache::kSecondary: return StubCache::secondary_; |
| } |
| UNREACHABLE(); |
| return NULL; |
| } |
| |
| Address address_; |
| }; |
| |
| // ------------------------------------------------------------------------ |
| |
| |
| // Support functions for IC stubs for callbacks. |
| Object* LoadCallbackProperty(Arguments args); |
| Object* StoreCallbackProperty(Arguments args); |
| |
| |
| // Support functions for IC stubs for interceptors. |
| Object* LoadPropertyWithInterceptorOnly(Arguments args); |
| Object* LoadPropertyWithInterceptorForLoad(Arguments args); |
| Object* LoadPropertyWithInterceptorForCall(Arguments args); |
| Object* StoreInterceptorProperty(Arguments args); |
| Object* CallInterceptorProperty(Arguments args); |
| |
| |
| // Support function for computing call IC miss stubs. |
| Handle<Code> ComputeCallMiss(int argc); |
| |
| |
| // The stub compiler compiles stubs for the stub cache. |
| class StubCompiler BASE_EMBEDDED { |
| public: |
| enum CheckType { |
| RECEIVER_MAP_CHECK, |
| STRING_CHECK, |
| NUMBER_CHECK, |
| BOOLEAN_CHECK, |
| JSARRAY_HAS_FAST_ELEMENTS_CHECK |
| }; |
| |
| StubCompiler() : scope_(), masm_(NULL, 256), failure_(NULL) { } |
| |
| Object* CompileCallInitialize(Code::Flags flags); |
| Object* CompileCallPreMonomorphic(Code::Flags flags); |
| Object* CompileCallNormal(Code::Flags flags); |
| Object* CompileCallMegamorphic(Code::Flags flags); |
| Object* CompileCallMiss(Code::Flags flags); |
| #ifdef ENABLE_DEBUGGER_SUPPORT |
| Object* CompileCallDebugBreak(Code::Flags flags); |
| Object* CompileCallDebugPrepareStepIn(Code::Flags flags); |
| #endif |
| Object* CompileLazyCompile(Code::Flags flags); |
| |
| // Static functions for generating parts of stubs. |
| static void GenerateLoadGlobalFunctionPrototype(MacroAssembler* masm, |
| int index, |
| Register prototype); |
| static void GenerateFastPropertyLoad(MacroAssembler* masm, |
| Register dst, Register src, |
| JSObject* holder, int index); |
| |
| static void GenerateLoadArrayLength(MacroAssembler* masm, |
| Register receiver, |
| Register scratch, |
| Label* miss_label); |
| static void GenerateLoadStringLength(MacroAssembler* masm, |
| Register receiver, |
| Register scratch, |
| Label* miss_label); |
| static void GenerateLoadStringLength2(MacroAssembler* masm, |
| Register receiver, |
| Register scratch1, |
| Register scratch2, |
| Label* miss_label); |
| static void GenerateLoadFunctionPrototype(MacroAssembler* masm, |
| Register receiver, |
| Register scratch1, |
| Register scratch2, |
| Label* miss_label); |
| static void GenerateStoreField(MacroAssembler* masm, |
| Builtins::Name storage_extend, |
| JSObject* object, |
| int index, |
| Map* transition, |
| Register receiver_reg, |
| Register name_reg, |
| Register scratch, |
| Label* miss_label); |
| static void GenerateLoadMiss(MacroAssembler* masm, Code::Kind kind); |
| |
| // Check the integrity of the prototype chain to make sure that the |
| // current IC is still valid. |
| Register CheckPrototypes(JSObject* object, |
| Register object_reg, |
| JSObject* holder, |
| Register holder_reg, |
| Register scratch, |
| String* name, |
| Label* miss); |
| |
| protected: |
| Object* GetCodeWithFlags(Code::Flags flags, const char* name); |
| Object* GetCodeWithFlags(Code::Flags flags, String* name); |
| |
| MacroAssembler* masm() { return &masm_; } |
| void set_failure(Failure* failure) { failure_ = failure; } |
| |
| void GenerateLoadField(JSObject* object, |
| JSObject* holder, |
| Register receiver, |
| Register scratch1, |
| Register scratch2, |
| int index, |
| String* name, |
| Label* miss); |
| |
| void GenerateLoadCallback(JSObject* object, |
| JSObject* holder, |
| Register receiver, |
| Register name_reg, |
| Register scratch1, |
| Register scratch2, |
| AccessorInfo* callback, |
| String* name, |
| Label* miss); |
| |
| void GenerateLoadConstant(JSObject* object, |
| JSObject* holder, |
| Register receiver, |
| Register scratch1, |
| Register scratch2, |
| Object* value, |
| String* name, |
| Label* miss); |
| |
| void GenerateLoadInterceptor(JSObject* object, |
| JSObject* holder, |
| LookupResult* lookup, |
| Register receiver, |
| Register name_reg, |
| Register scratch1, |
| Register scratch2, |
| String* name, |
| Label* miss); |
| |
| private: |
| HandleScope scope_; |
| MacroAssembler masm_; |
| Failure* failure_; |
| }; |
| |
| |
| class LoadStubCompiler: public StubCompiler { |
| public: |
| Object* CompileLoadField(JSObject* object, |
| JSObject* holder, |
| int index, |
| String* name); |
| Object* CompileLoadCallback(JSObject* object, |
| JSObject* holder, |
| AccessorInfo* callback, |
| String* name); |
| Object* CompileLoadConstant(JSObject* object, |
| JSObject* holder, |
| Object* value, |
| String* name); |
| Object* CompileLoadInterceptor(JSObject* object, |
| JSObject* holder, |
| String* name); |
| |
| Object* CompileLoadGlobal(JSObject* object, |
| GlobalObject* holder, |
| JSGlobalPropertyCell* cell, |
| String* name, |
| bool is_dont_delete); |
| |
| private: |
| Object* GetCode(PropertyType type, String* name); |
| }; |
| |
| |
| class KeyedLoadStubCompiler: public StubCompiler { |
| public: |
| Object* CompileLoadField(String* name, |
| JSObject* object, |
| JSObject* holder, |
| int index); |
| Object* CompileLoadCallback(String* name, |
| JSObject* object, |
| JSObject* holder, |
| AccessorInfo* callback); |
| Object* CompileLoadConstant(String* name, |
| JSObject* object, |
| JSObject* holder, |
| Object* value); |
| Object* CompileLoadInterceptor(JSObject* object, |
| JSObject* holder, |
| String* name); |
| Object* CompileLoadArrayLength(String* name); |
| Object* CompileLoadStringLength(String* name); |
| Object* CompileLoadFunctionPrototype(String* name); |
| |
| private: |
| Object* GetCode(PropertyType type, String* name); |
| }; |
| |
| |
| class StoreStubCompiler: public StubCompiler { |
| public: |
| Object* CompileStoreField(JSObject* object, |
| int index, |
| Map* transition, |
| String* name); |
| Object* CompileStoreCallback(JSObject* object, |
| AccessorInfo* callbacks, |
| String* name); |
| Object* CompileStoreInterceptor(JSObject* object, String* name); |
| Object* CompileStoreGlobal(GlobalObject* object, |
| JSGlobalPropertyCell* holder, |
| String* name); |
| |
| |
| private: |
| Object* GetCode(PropertyType type, String* name); |
| }; |
| |
| |
| class KeyedStoreStubCompiler: public StubCompiler { |
| public: |
| Object* CompileStoreField(JSObject* object, |
| int index, |
| Map* transition, |
| String* name); |
| |
| private: |
| Object* GetCode(PropertyType type, String* name); |
| }; |
| |
| |
| class CallStubCompiler: public StubCompiler { |
| public: |
| explicit CallStubCompiler(int argc, InLoopFlag in_loop) |
| : arguments_(argc), in_loop_(in_loop) { } |
| |
| Object* CompileCallField(Object* object, |
| JSObject* holder, |
| int index, |
| String* name); |
| Object* CompileCallConstant(Object* object, |
| JSObject* holder, |
| JSFunction* function, |
| String* name, |
| CheckType check); |
| Object* CompileCallInterceptor(Object* object, |
| JSObject* holder, |
| String* name); |
| Object* CompileCallGlobal(JSObject* object, |
| GlobalObject* holder, |
| JSGlobalPropertyCell* cell, |
| JSFunction* function, |
| String* name); |
| |
| private: |
| const ParameterCount arguments_; |
| const InLoopFlag in_loop_; |
| |
| const ParameterCount& arguments() { return arguments_; } |
| |
| Object* GetCode(PropertyType type, String* name); |
| }; |
| |
| |
| class ConstructStubCompiler: public StubCompiler { |
| public: |
| explicit ConstructStubCompiler() {} |
| |
| Object* CompileConstructStub(SharedFunctionInfo* shared); |
| |
| private: |
| Object* GetCode(); |
| }; |
| |
| |
| } } // namespace v8::internal |
| |
| #endif // V8_STUB_CACHE_H_ |