1 /* 2 * Copyright (C) 2011 The Android Open Source Project 3 * 4 * Licensed under the Apache License, Version 2.0 (the "License"); 5 * you may not use this file except in compliance with the License. 6 * You may obtain a copy of the License at 7 * 8 * http://www.apache.org/licenses/LICENSE-2.0 9 * 10 * Unless required by applicable law or agreed to in writing, software 11 * distributed under the License is distributed on an "AS IS" BASIS, 12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13 * See the License for the specific language governing permissions and 14 * limitations under the License. 15 */ 16 17 #include "intern_table.h" 18 19 #include <memory> 20 21 #include "dex/utf.h" 22 #include "gc/collector/garbage_collector.h" 23 #include "gc/space/image_space.h" 24 #include "gc/weak_root_state.h" 25 #include "gc_root-inl.h" 26 #include "handle_scope-inl.h" 27 #include "image-inl.h" 28 #include "mirror/dex_cache-inl.h" 29 #include "mirror/object-inl.h" 30 #include "mirror/object_array-inl.h" 31 #include "mirror/string-inl.h" 32 #include "object_callbacks.h" 33 #include "scoped_thread_state_change-inl.h" 34 #include "thread.h" 35 36 namespace art { 37 InternTable()38 InternTable::InternTable() 39 : log_new_roots_(false), 40 weak_intern_condition_("New intern condition", *Locks::intern_table_lock_), 41 weak_root_state_(gc::kWeakRootStateNormal) { 42 } 43 Size() const44 size_t InternTable::Size() const { 45 MutexLock mu(Thread::Current(), *Locks::intern_table_lock_); 46 return strong_interns_.Size() + weak_interns_.Size(); 47 } 48 StrongSize() const49 size_t InternTable::StrongSize() const { 50 MutexLock mu(Thread::Current(), *Locks::intern_table_lock_); 51 return strong_interns_.Size(); 52 } 53 WeakSize() const54 size_t InternTable::WeakSize() const { 55 MutexLock mu(Thread::Current(), *Locks::intern_table_lock_); 56 return weak_interns_.Size(); 57 } 58 DumpForSigQuit(std::ostream & os) const59 void InternTable::DumpForSigQuit(std::ostream& os) const { 60 os << "Intern table: " << StrongSize() << " strong; " << WeakSize() << " weak\n"; 61 } 62 VisitRoots(RootVisitor * visitor,VisitRootFlags flags)63 void InternTable::VisitRoots(RootVisitor* visitor, VisitRootFlags flags) { 64 MutexLock mu(Thread::Current(), *Locks::intern_table_lock_); 65 if ((flags & kVisitRootFlagAllRoots) != 0) { 66 strong_interns_.VisitRoots(visitor); 67 } else if ((flags & kVisitRootFlagNewRoots) != 0) { 68 for (auto& root : new_strong_intern_roots_) { 69 ObjPtr<mirror::String> old_ref = root.Read<kWithoutReadBarrier>(); 70 root.VisitRoot(visitor, RootInfo(kRootInternedString)); 71 ObjPtr<mirror::String> new_ref = root.Read<kWithoutReadBarrier>(); 72 if (new_ref != old_ref) { 73 // The GC moved a root in the log. Need to search the strong interns and update the 74 // corresponding object. This is slow, but luckily for us, this may only happen with a 75 // concurrent moving GC. 76 strong_interns_.Remove(old_ref); 77 strong_interns_.Insert(new_ref); 78 } 79 } 80 } 81 if ((flags & kVisitRootFlagClearRootLog) != 0) { 82 new_strong_intern_roots_.clear(); 83 } 84 if ((flags & kVisitRootFlagStartLoggingNewRoots) != 0) { 85 log_new_roots_ = true; 86 } else if ((flags & kVisitRootFlagStopLoggingNewRoots) != 0) { 87 log_new_roots_ = false; 88 } 89 // Note: we deliberately don't visit the weak_interns_ table and the immutable image roots. 90 } 91 LookupWeak(Thread * self,ObjPtr<mirror::String> s)92 ObjPtr<mirror::String> InternTable::LookupWeak(Thread* self, ObjPtr<mirror::String> s) { 93 MutexLock mu(self, *Locks::intern_table_lock_); 94 return LookupWeakLocked(s); 95 } 96 LookupStrong(Thread * self,ObjPtr<mirror::String> s)97 ObjPtr<mirror::String> InternTable::LookupStrong(Thread* self, ObjPtr<mirror::String> s) { 98 MutexLock mu(self, *Locks::intern_table_lock_); 99 return LookupStrongLocked(s); 100 } 101 LookupStrong(Thread * self,uint32_t utf16_length,const char * utf8_data)102 ObjPtr<mirror::String> InternTable::LookupStrong(Thread* self, 103 uint32_t utf16_length, 104 const char* utf8_data) { 105 DCHECK_EQ(utf16_length, CountModifiedUtf8Chars(utf8_data)); 106 Utf8String string(utf16_length, 107 utf8_data, 108 ComputeUtf16HashFromModifiedUtf8(utf8_data, utf16_length)); 109 MutexLock mu(self, *Locks::intern_table_lock_); 110 return strong_interns_.Find(string); 111 } 112 LookupWeakLocked(ObjPtr<mirror::String> s)113 ObjPtr<mirror::String> InternTable::LookupWeakLocked(ObjPtr<mirror::String> s) { 114 return weak_interns_.Find(s); 115 } 116 LookupStrongLocked(ObjPtr<mirror::String> s)117 ObjPtr<mirror::String> InternTable::LookupStrongLocked(ObjPtr<mirror::String> s) { 118 return strong_interns_.Find(s); 119 } 120 AddNewTable()121 void InternTable::AddNewTable() { 122 MutexLock mu(Thread::Current(), *Locks::intern_table_lock_); 123 weak_interns_.AddNewTable(); 124 strong_interns_.AddNewTable(); 125 } 126 InsertStrong(ObjPtr<mirror::String> s)127 ObjPtr<mirror::String> InternTable::InsertStrong(ObjPtr<mirror::String> s) { 128 Runtime* runtime = Runtime::Current(); 129 if (runtime->IsActiveTransaction()) { 130 runtime->RecordStrongStringInsertion(s); 131 } 132 if (log_new_roots_) { 133 new_strong_intern_roots_.push_back(GcRoot<mirror::String>(s)); 134 } 135 strong_interns_.Insert(s); 136 return s; 137 } 138 InsertWeak(ObjPtr<mirror::String> s)139 ObjPtr<mirror::String> InternTable::InsertWeak(ObjPtr<mirror::String> s) { 140 Runtime* runtime = Runtime::Current(); 141 if (runtime->IsActiveTransaction()) { 142 runtime->RecordWeakStringInsertion(s); 143 } 144 weak_interns_.Insert(s); 145 return s; 146 } 147 RemoveStrong(ObjPtr<mirror::String> s)148 void InternTable::RemoveStrong(ObjPtr<mirror::String> s) { 149 strong_interns_.Remove(s); 150 } 151 RemoveWeak(ObjPtr<mirror::String> s)152 void InternTable::RemoveWeak(ObjPtr<mirror::String> s) { 153 Runtime* runtime = Runtime::Current(); 154 if (runtime->IsActiveTransaction()) { 155 runtime->RecordWeakStringRemoval(s); 156 } 157 weak_interns_.Remove(s); 158 } 159 160 // Insert/remove methods used to undo changes made during an aborted transaction. InsertStrongFromTransaction(ObjPtr<mirror::String> s)161 ObjPtr<mirror::String> InternTable::InsertStrongFromTransaction(ObjPtr<mirror::String> s) { 162 DCHECK(!Runtime::Current()->IsActiveTransaction()); 163 return InsertStrong(s); 164 } 165 InsertWeakFromTransaction(ObjPtr<mirror::String> s)166 ObjPtr<mirror::String> InternTable::InsertWeakFromTransaction(ObjPtr<mirror::String> s) { 167 DCHECK(!Runtime::Current()->IsActiveTransaction()); 168 return InsertWeak(s); 169 } 170 RemoveStrongFromTransaction(ObjPtr<mirror::String> s)171 void InternTable::RemoveStrongFromTransaction(ObjPtr<mirror::String> s) { 172 DCHECK(!Runtime::Current()->IsActiveTransaction()); 173 RemoveStrong(s); 174 } 175 RemoveWeakFromTransaction(ObjPtr<mirror::String> s)176 void InternTable::RemoveWeakFromTransaction(ObjPtr<mirror::String> s) { 177 DCHECK(!Runtime::Current()->IsActiveTransaction()); 178 RemoveWeak(s); 179 } 180 BroadcastForNewInterns()181 void InternTable::BroadcastForNewInterns() { 182 Thread* self = Thread::Current(); 183 MutexLock mu(self, *Locks::intern_table_lock_); 184 weak_intern_condition_.Broadcast(self); 185 } 186 WaitUntilAccessible(Thread * self)187 void InternTable::WaitUntilAccessible(Thread* self) { 188 Locks::intern_table_lock_->ExclusiveUnlock(self); 189 { 190 ScopedThreadSuspension sts(self, kWaitingWeakGcRootRead); 191 MutexLock mu(self, *Locks::intern_table_lock_); 192 while ((!kUseReadBarrier && weak_root_state_ == gc::kWeakRootStateNoReadsOrWrites) || 193 (kUseReadBarrier && !self->GetWeakRefAccessEnabled())) { 194 weak_intern_condition_.Wait(self); 195 } 196 } 197 Locks::intern_table_lock_->ExclusiveLock(self); 198 } 199 Insert(ObjPtr<mirror::String> s,bool is_strong,bool holding_locks)200 ObjPtr<mirror::String> InternTable::Insert(ObjPtr<mirror::String> s, 201 bool is_strong, 202 bool holding_locks) { 203 if (s == nullptr) { 204 return nullptr; 205 } 206 Thread* const self = Thread::Current(); 207 MutexLock mu(self, *Locks::intern_table_lock_); 208 if (kDebugLocking && !holding_locks) { 209 Locks::mutator_lock_->AssertSharedHeld(self); 210 CHECK_EQ(2u, self->NumberOfHeldMutexes()) << "may only safely hold the mutator lock"; 211 } 212 while (true) { 213 if (holding_locks) { 214 if (!kUseReadBarrier) { 215 CHECK_EQ(weak_root_state_, gc::kWeakRootStateNormal); 216 } else { 217 CHECK(self->GetWeakRefAccessEnabled()); 218 } 219 } 220 // Check the strong table for a match. 221 ObjPtr<mirror::String> strong = LookupStrongLocked(s); 222 if (strong != nullptr) { 223 return strong; 224 } 225 if ((!kUseReadBarrier && weak_root_state_ != gc::kWeakRootStateNoReadsOrWrites) || 226 (kUseReadBarrier && self->GetWeakRefAccessEnabled())) { 227 break; 228 } 229 // weak_root_state_ is set to gc::kWeakRootStateNoReadsOrWrites in the GC pause but is only 230 // cleared after SweepSystemWeaks has completed. This is why we need to wait until it is 231 // cleared. 232 CHECK(!holding_locks); 233 StackHandleScope<1> hs(self); 234 auto h = hs.NewHandleWrapper(&s); 235 WaitUntilAccessible(self); 236 } 237 if (!kUseReadBarrier) { 238 CHECK_EQ(weak_root_state_, gc::kWeakRootStateNormal); 239 } else { 240 CHECK(self->GetWeakRefAccessEnabled()); 241 } 242 // There is no match in the strong table, check the weak table. 243 ObjPtr<mirror::String> weak = LookupWeakLocked(s); 244 if (weak != nullptr) { 245 if (is_strong) { 246 // A match was found in the weak table. Promote to the strong table. 247 RemoveWeak(weak); 248 return InsertStrong(weak); 249 } 250 return weak; 251 } 252 // No match in the strong table or the weak table. Insert into the strong / weak table. 253 return is_strong ? InsertStrong(s) : InsertWeak(s); 254 } 255 InternStrong(int32_t utf16_length,const char * utf8_data)256 ObjPtr<mirror::String> InternTable::InternStrong(int32_t utf16_length, const char* utf8_data) { 257 DCHECK(utf8_data != nullptr); 258 Thread* self = Thread::Current(); 259 // Try to avoid allocation. 260 ObjPtr<mirror::String> s = LookupStrong(self, utf16_length, utf8_data); 261 if (s != nullptr) { 262 return s; 263 } 264 return InternStrong(mirror::String::AllocFromModifiedUtf8( 265 self, utf16_length, utf8_data)); 266 } 267 InternStrong(const char * utf8_data)268 ObjPtr<mirror::String> InternTable::InternStrong(const char* utf8_data) { 269 DCHECK(utf8_data != nullptr); 270 return InternStrong(mirror::String::AllocFromModifiedUtf8(Thread::Current(), utf8_data)); 271 } 272 InternStrongImageString(ObjPtr<mirror::String> s)273 ObjPtr<mirror::String> InternTable::InternStrongImageString(ObjPtr<mirror::String> s) { 274 // May be holding the heap bitmap lock. 275 return Insert(s, true, true); 276 } 277 InternStrong(ObjPtr<mirror::String> s)278 ObjPtr<mirror::String> InternTable::InternStrong(ObjPtr<mirror::String> s) { 279 return Insert(s, true, false); 280 } 281 InternWeak(ObjPtr<mirror::String> s)282 ObjPtr<mirror::String> InternTable::InternWeak(ObjPtr<mirror::String> s) { 283 return Insert(s, false, false); 284 } 285 ContainsWeak(ObjPtr<mirror::String> s)286 bool InternTable::ContainsWeak(ObjPtr<mirror::String> s) { 287 return LookupWeak(Thread::Current(), s) == s; 288 } 289 SweepInternTableWeaks(IsMarkedVisitor * visitor)290 void InternTable::SweepInternTableWeaks(IsMarkedVisitor* visitor) { 291 MutexLock mu(Thread::Current(), *Locks::intern_table_lock_); 292 weak_interns_.SweepWeaks(visitor); 293 } 294 WriteToMemory(uint8_t * ptr)295 size_t InternTable::WriteToMemory(uint8_t* ptr) { 296 MutexLock mu(Thread::Current(), *Locks::intern_table_lock_); 297 return strong_interns_.WriteToMemory(ptr); 298 } 299 operator ()(const GcRoot<mirror::String> & root) const300 std::size_t InternTable::StringHashEquals::operator()(const GcRoot<mirror::String>& root) const { 301 if (kIsDebugBuild) { 302 Locks::mutator_lock_->AssertSharedHeld(Thread::Current()); 303 } 304 // An additional cast to prevent undesired sign extension. 305 return static_cast<size_t>( 306 static_cast<uint32_t>(root.Read<kWithoutReadBarrier>()->GetHashCode())); 307 } 308 operator ()(const GcRoot<mirror::String> & a,const GcRoot<mirror::String> & b) const309 bool InternTable::StringHashEquals::operator()(const GcRoot<mirror::String>& a, 310 const GcRoot<mirror::String>& b) const { 311 if (kIsDebugBuild) { 312 Locks::mutator_lock_->AssertSharedHeld(Thread::Current()); 313 } 314 return a.Read<kWithoutReadBarrier>()->Equals(b.Read<kWithoutReadBarrier>()); 315 } 316 operator ()(const GcRoot<mirror::String> & a,const Utf8String & b) const317 bool InternTable::StringHashEquals::operator()(const GcRoot<mirror::String>& a, 318 const Utf8String& b) const { 319 if (kIsDebugBuild) { 320 Locks::mutator_lock_->AssertSharedHeld(Thread::Current()); 321 } 322 ObjPtr<mirror::String> a_string = a.Read<kWithoutReadBarrier>(); 323 uint32_t a_length = static_cast<uint32_t>(a_string->GetLength()); 324 if (a_length != b.GetUtf16Length()) { 325 return false; 326 } 327 if (a_string->IsCompressed()) { 328 size_t b_byte_count = strlen(b.GetUtf8Data()); 329 size_t b_utf8_length = CountModifiedUtf8Chars(b.GetUtf8Data(), b_byte_count); 330 // Modified UTF-8 single byte character range is 0x01 .. 0x7f 331 // The string compression occurs on regular ASCII with same exact range, 332 // not on extended ASCII which up to 0xff 333 const bool is_b_regular_ascii = (b_byte_count == b_utf8_length); 334 if (is_b_regular_ascii) { 335 return memcmp(b.GetUtf8Data(), 336 a_string->GetValueCompressed(), a_length * sizeof(uint8_t)) == 0; 337 } else { 338 return false; 339 } 340 } else { 341 const uint16_t* a_value = a_string->GetValue(); 342 return CompareModifiedUtf8ToUtf16AsCodePointValues(b.GetUtf8Data(), a_value, a_length) == 0; 343 } 344 } 345 WriteToMemory(uint8_t * ptr)346 size_t InternTable::Table::WriteToMemory(uint8_t* ptr) { 347 if (tables_.empty()) { 348 return 0; 349 } 350 UnorderedSet* table_to_write; 351 UnorderedSet combined; 352 if (tables_.size() > 1) { 353 table_to_write = &combined; 354 for (InternalTable& table : tables_) { 355 for (GcRoot<mirror::String>& string : table.set_) { 356 combined.insert(string); 357 } 358 } 359 } else { 360 table_to_write = &tables_.back().set_; 361 } 362 return table_to_write->WriteToMemory(ptr); 363 } 364 Remove(ObjPtr<mirror::String> s)365 void InternTable::Table::Remove(ObjPtr<mirror::String> s) { 366 for (InternalTable& table : tables_) { 367 auto it = table.set_.find(GcRoot<mirror::String>(s)); 368 if (it != table.set_.end()) { 369 table.set_.erase(it); 370 return; 371 } 372 } 373 LOG(FATAL) << "Attempting to remove non-interned string " << s->ToModifiedUtf8(); 374 } 375 Find(ObjPtr<mirror::String> s)376 ObjPtr<mirror::String> InternTable::Table::Find(ObjPtr<mirror::String> s) { 377 Locks::intern_table_lock_->AssertHeld(Thread::Current()); 378 for (InternalTable& table : tables_) { 379 auto it = table.set_.find(GcRoot<mirror::String>(s)); 380 if (it != table.set_.end()) { 381 return it->Read(); 382 } 383 } 384 return nullptr; 385 } 386 Find(const Utf8String & string)387 ObjPtr<mirror::String> InternTable::Table::Find(const Utf8String& string) { 388 Locks::intern_table_lock_->AssertHeld(Thread::Current()); 389 for (InternalTable& table : tables_) { 390 auto it = table.set_.find(string); 391 if (it != table.set_.end()) { 392 return it->Read(); 393 } 394 } 395 return nullptr; 396 } 397 AddNewTable()398 void InternTable::Table::AddNewTable() { 399 tables_.push_back(InternalTable()); 400 } 401 Insert(ObjPtr<mirror::String> s)402 void InternTable::Table::Insert(ObjPtr<mirror::String> s) { 403 // Always insert the last table, the image tables are before and we avoid inserting into these 404 // to prevent dirty pages. 405 DCHECK(!tables_.empty()); 406 tables_.back().set_.insert(GcRoot<mirror::String>(s)); 407 } 408 VisitRoots(RootVisitor * visitor)409 void InternTable::Table::VisitRoots(RootVisitor* visitor) { 410 BufferedRootVisitor<kDefaultBufferedRootCount> buffered_visitor( 411 visitor, RootInfo(kRootInternedString)); 412 for (InternalTable& table : tables_) { 413 for (auto& intern : table.set_) { 414 buffered_visitor.VisitRoot(intern); 415 } 416 } 417 } 418 SweepWeaks(IsMarkedVisitor * visitor)419 void InternTable::Table::SweepWeaks(IsMarkedVisitor* visitor) { 420 for (InternalTable& table : tables_) { 421 SweepWeaks(&table.set_, visitor); 422 } 423 } 424 SweepWeaks(UnorderedSet * set,IsMarkedVisitor * visitor)425 void InternTable::Table::SweepWeaks(UnorderedSet* set, IsMarkedVisitor* visitor) { 426 for (auto it = set->begin(), end = set->end(); it != end;) { 427 // This does not need a read barrier because this is called by GC. 428 mirror::Object* object = it->Read<kWithoutReadBarrier>(); 429 mirror::Object* new_object = visitor->IsMarked(object); 430 if (new_object == nullptr) { 431 it = set->erase(it); 432 } else { 433 *it = GcRoot<mirror::String>(new_object->AsString()); 434 ++it; 435 } 436 } 437 } 438 Size() const439 size_t InternTable::Table::Size() const { 440 return std::accumulate(tables_.begin(), 441 tables_.end(), 442 0U, 443 [](size_t sum, const InternalTable& table) { 444 return sum + table.Size(); 445 }); 446 } 447 ChangeWeakRootState(gc::WeakRootState new_state)448 void InternTable::ChangeWeakRootState(gc::WeakRootState new_state) { 449 MutexLock mu(Thread::Current(), *Locks::intern_table_lock_); 450 ChangeWeakRootStateLocked(new_state); 451 } 452 ChangeWeakRootStateLocked(gc::WeakRootState new_state)453 void InternTable::ChangeWeakRootStateLocked(gc::WeakRootState new_state) { 454 CHECK(!kUseReadBarrier); 455 weak_root_state_ = new_state; 456 if (new_state != gc::kWeakRootStateNoReadsOrWrites) { 457 weak_intern_condition_.Broadcast(Thread::Current()); 458 } 459 } 460 Table()461 InternTable::Table::Table() { 462 Runtime* const runtime = Runtime::Current(); 463 InternalTable initial_table; 464 initial_table.set_.SetLoadFactor(runtime->GetHashTableMinLoadFactor(), 465 runtime->GetHashTableMaxLoadFactor()); 466 tables_.push_back(std::move(initial_table)); 467 } 468 469 } // namespace art 470