mirror of
https://github.com/dart-lang/sdk
synced 2024-11-02 12:24:24 +00:00
10e9606861
'is' tests against record types are split into series of 'is' tests of record fields. This is more efficient as record instances cannot be added to subtype test caches (because type of a record instance depends on types of its fields). This change also adds canonicalization and constant evaluation of LoadFieldInstr for record fields. Performance on a trivial micro-benchmark (on x64): JIT (RunTime): 4519104.5 -> 20031.5 AOT (RunTime): 4352583.0 -> 26281.6 TEST=ci Issue: https://github.com/dart-lang/sdk/issues/49719 Change-Id: I2ed464cd3b31f365b17805f4e7debe1d6d1051fa Reviewed-on: https://dart-review.googlesource.com/c/sdk/+/268080 Reviewed-by: Ryan Macnak <rmacnak@google.com> Reviewed-by: Slava Egorov <vegorov@google.com> Commit-Queue: Alexander Markov <alexmarkov@google.com>
2795 lines
98 KiB
C++
2795 lines
98 KiB
C++
// Copyright (c) 2016, the Dart project authors. Please see the AUTHORS file
|
|
// for details. All rights reserved. Use of this source code is governed by a
|
|
// BSD-style license that can be found in the LICENSE file.
|
|
|
|
#include "vm/isolate_reload.h"
|
|
|
|
#include <memory>
|
|
|
|
#include "vm/bit_vector.h"
|
|
#include "vm/compiler/jit/compiler.h"
|
|
#include "vm/dart_api_impl.h"
|
|
#if !defined(PRODUCT) && !defined(DART_PRECOMPILED_RUNTIME)
|
|
#include "vm/hash.h"
|
|
#endif
|
|
#include "vm/hash_table.h"
|
|
#include "vm/heap/become.h"
|
|
#include "vm/heap/safepoint.h"
|
|
#include "vm/isolate.h"
|
|
#include "vm/kernel_isolate.h"
|
|
#include "vm/kernel_loader.h"
|
|
#include "vm/log.h"
|
|
#include "vm/longjump.h"
|
|
#include "vm/object.h"
|
|
#include "vm/object_store.h"
|
|
#include "vm/parser.h"
|
|
#include "vm/runtime_entry.h"
|
|
#include "vm/service_event.h"
|
|
#include "vm/stack_frame.h"
|
|
#include "vm/thread.h"
|
|
#include "vm/timeline.h"
|
|
#include "vm/type_testing_stubs.h"
|
|
#include "vm/visitor.h"
|
|
|
|
namespace dart {
|
|
|
|
DEFINE_FLAG(int, reload_every, 0, "Reload every N stack overflow checks.");
|
|
DEFINE_FLAG(bool, trace_reload, false, "Trace isolate reloading");
|
|
|
|
#if !defined(PRODUCT) && !defined(DART_PRECOMPILED_RUNTIME)
|
|
DEFINE_FLAG(bool,
|
|
trace_reload_verbose,
|
|
false,
|
|
"trace isolate reloading verbose");
|
|
DEFINE_FLAG(bool, identity_reload, false, "Enable checks for identity reload.");
|
|
DEFINE_FLAG(bool, reload_every_optimized, true, "Only from optimized code.");
|
|
DEFINE_FLAG(bool,
|
|
reload_every_back_off,
|
|
false,
|
|
"Double the --reload-every value after each reload.");
|
|
DEFINE_FLAG(bool,
|
|
reload_force_rollback,
|
|
false,
|
|
"Force all reloads to fail and rollback.");
|
|
DEFINE_FLAG(bool,
|
|
check_reloaded,
|
|
false,
|
|
"Assert that an isolate has reloaded at least once.")
|
|
DEFINE_FLAG(bool, gc_during_reload, false, "Cause explicit GC during reload.");
|
|
|
|
DECLARE_FLAG(bool, trace_deoptimization);
|
|
|
|
#define IG (isolate_group())
|
|
#define Z zone_
|
|
|
|
#define TIMELINE_SCOPE(name) \
|
|
TimelineBeginEndScope tbes##name(Thread::Current(), \
|
|
Timeline::GetIsolateStream(), #name)
|
|
|
|
// The ObjectLocator is used for collecting instances that
|
|
// needs to be morphed.
|
|
class ObjectLocator : public ObjectVisitor {
|
|
public:
|
|
explicit ObjectLocator(IsolateGroupReloadContext* context)
|
|
: context_(context), count_(0) {}
|
|
|
|
void VisitObject(ObjectPtr obj) {
|
|
InstanceMorpher* morpher =
|
|
context_->instance_morpher_by_cid_.LookupValue(obj->GetClassId());
|
|
if (morpher != NULL) {
|
|
morpher->AddObject(obj);
|
|
count_++;
|
|
}
|
|
}
|
|
|
|
// Return the number of located objects for morphing.
|
|
intptr_t count() { return count_; }
|
|
|
|
private:
|
|
IsolateGroupReloadContext* context_;
|
|
intptr_t count_;
|
|
};
|
|
|
|
static bool HasNoTasks(Heap* heap) {
|
|
MonitorLocker ml(heap->old_space()->tasks_lock());
|
|
return heap->old_space()->tasks() == 0;
|
|
}
|
|
|
|
InstanceMorpher* InstanceMorpher::CreateFromClassDescriptors(
|
|
Zone* zone,
|
|
ClassTable* class_table,
|
|
const Class& from,
|
|
const Class& to) {
|
|
auto mapping = new (zone) FieldMappingArray();
|
|
auto new_fields_offsets = new (zone) FieldOffsetArray();
|
|
|
|
if (from.NumTypeArguments() > 0) {
|
|
// Add copying of the optional type argument field.
|
|
intptr_t from_offset = from.host_type_arguments_field_offset();
|
|
ASSERT(from_offset != Class::kNoTypeArguments);
|
|
intptr_t to_offset = to.host_type_arguments_field_offset();
|
|
ASSERT(to_offset != Class::kNoTypeArguments);
|
|
mapping->Add({from_offset, kIllegalCid});
|
|
mapping->Add({to_offset, kIllegalCid});
|
|
}
|
|
|
|
// Add copying of the instance fields if matching by name.
|
|
// Note: currently the type of the fields are ignored.
|
|
const Array& from_fields = Array::Handle(
|
|
from.OffsetToFieldMap(IsolateGroup::Current()->heap_walk_class_table()));
|
|
const Array& to_fields = Array::Handle(to.OffsetToFieldMap());
|
|
Field& from_field = Field::Handle();
|
|
Field& to_field = Field::Handle();
|
|
String& from_name = String::Handle();
|
|
String& to_name = String::Handle();
|
|
|
|
auto ensure_boxed_and_guarded = [&](const Field& field) {
|
|
field.set_needs_load_guard(true);
|
|
if (field.is_unboxed()) {
|
|
to.MarkFieldBoxedDuringReload(class_table, field);
|
|
}
|
|
};
|
|
|
|
// Scan across all the fields in the new class definition.
|
|
for (intptr_t i = 0; i < to_fields.Length(); i++) {
|
|
if (to_fields.At(i) == Field::null()) {
|
|
continue; // Ignore non-fields.
|
|
}
|
|
|
|
// Grab the field's name.
|
|
to_field = Field::RawCast(to_fields.At(i));
|
|
ASSERT(to_field.is_instance());
|
|
to_name = to_field.name();
|
|
|
|
// Did this field not exist in the old class definition?
|
|
bool new_field = true;
|
|
|
|
// Find this field in the old class.
|
|
for (intptr_t j = 0; j < from_fields.Length(); j++) {
|
|
if (from_fields.At(j) == Field::null()) {
|
|
continue; // Ignore non-fields.
|
|
}
|
|
from_field = Field::RawCast(from_fields.At(j));
|
|
ASSERT(from_field.is_instance());
|
|
from_name = from_field.name();
|
|
if (from_name.Equals(to_name)) {
|
|
intptr_t from_box_cid = kIllegalCid;
|
|
intptr_t to_box_cid = kIllegalCid;
|
|
|
|
// Check if either of the fields are unboxed.
|
|
if ((from_field.is_unboxed() && from_field.type() != to_field.type()) ||
|
|
(from_field.is_unboxed() != to_field.is_unboxed())) {
|
|
// For simplicity we just migrate to boxed fields if such
|
|
// situation occurs.
|
|
ensure_boxed_and_guarded(to_field);
|
|
}
|
|
|
|
if (from_field.is_unboxed()) {
|
|
const auto field_cid = from_field.guarded_cid();
|
|
switch (field_cid) {
|
|
case kDoubleCid:
|
|
case kFloat32x4Cid:
|
|
case kFloat64x2Cid:
|
|
from_box_cid = field_cid;
|
|
break;
|
|
default:
|
|
from_box_cid = kIntegerCid;
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (to_field.is_unboxed()) {
|
|
const auto field_cid = to_field.guarded_cid();
|
|
switch (field_cid) {
|
|
case kDoubleCid:
|
|
case kFloat32x4Cid:
|
|
case kFloat64x2Cid:
|
|
to_box_cid = field_cid;
|
|
break;
|
|
default:
|
|
to_box_cid = kIntegerCid;
|
|
break;
|
|
}
|
|
}
|
|
|
|
// Field can't become unboxed if it was boxed.
|
|
ASSERT(from_box_cid != kIllegalCid || to_box_cid == kIllegalCid);
|
|
|
|
// Success
|
|
mapping->Add({from_field.HostOffset(), from_box_cid});
|
|
mapping->Add({to_field.HostOffset(), to_box_cid});
|
|
|
|
// Field did exist in old class deifnition.
|
|
new_field = false;
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (new_field) {
|
|
ensure_boxed_and_guarded(to_field);
|
|
new_fields_offsets->Add(to_field.HostOffset());
|
|
}
|
|
}
|
|
|
|
ASSERT(from.id() == to.id());
|
|
return new (zone)
|
|
InstanceMorpher(zone, to.id(), class_table, mapping, new_fields_offsets);
|
|
}
|
|
|
|
InstanceMorpher::InstanceMorpher(Zone* zone,
|
|
classid_t cid,
|
|
ClassTable* class_table,
|
|
FieldMappingArray* mapping,
|
|
FieldOffsetArray* new_fields_offsets)
|
|
: zone_(zone),
|
|
cid_(cid),
|
|
class_table_(class_table),
|
|
mapping_(mapping),
|
|
new_fields_offsets_(new_fields_offsets),
|
|
before_(zone, 16) {}
|
|
|
|
void InstanceMorpher::AddObject(ObjectPtr object) {
|
|
ASSERT(object->GetClassId() == cid_);
|
|
const Instance& instance = Instance::Cast(Object::Handle(Z, object));
|
|
before_.Add(&instance);
|
|
}
|
|
|
|
void InstanceMorpher::CreateMorphedCopies(Become* become) {
|
|
Instance& after = Instance::Handle(Z);
|
|
Object& value = Object::Handle(Z);
|
|
for (intptr_t i = 0; i < before_.length(); i++) {
|
|
const Instance& before = *before_.At(i);
|
|
|
|
// Code can reference constants / canonical objects either directly in the
|
|
// instruction stream (ia32) or via an object pool.
|
|
//
|
|
// We have the following invariants:
|
|
//
|
|
// a) Those canonical objects don't change state (i.e. are not mutable):
|
|
// our optimizer can e.g. execute loads of such constants at
|
|
// compile-time.
|
|
//
|
|
// => We ensure that const-classes with live constants cannot be
|
|
// reloaded to become non-const classes (see Class::CheckReload).
|
|
//
|
|
// b) Those canonical objects live in old space: e.g. on ia32 the
|
|
// scavenger does not make the RX pages writable and therefore cannot
|
|
// update pointers embedded in the instruction stream.
|
|
//
|
|
// In order to maintain these invariants we ensure to always morph canonical
|
|
// objects to old space.
|
|
const bool is_canonical = before.IsCanonical();
|
|
const Heap::Space space = is_canonical ? Heap::kOld : Heap::kNew;
|
|
after = Instance::NewFromCidAndSize(class_table_, cid_, space);
|
|
|
|
// We preserve the canonical bit of the object, since this object is present
|
|
// in the class's constants.
|
|
if (is_canonical) {
|
|
after.SetCanonical();
|
|
}
|
|
#if defined(HASH_IN_OBJECT_HEADER)
|
|
const uint32_t hash = Object::GetCachedHash(before.ptr());
|
|
Object::SetCachedHashIfNotSet(after.ptr(), hash);
|
|
#endif
|
|
|
|
// Morph the context from [before] to [after] using mapping_.
|
|
for (intptr_t i = 0; i < mapping_->length(); i += 2) {
|
|
const auto& from = mapping_->At(i);
|
|
const auto& to = mapping_->At(i + 1);
|
|
ASSERT(from.offset > 0);
|
|
ASSERT(to.offset > 0);
|
|
if (from.box_cid == kIllegalCid) {
|
|
// Boxed to boxed field migration.
|
|
ASSERT(to.box_cid == kIllegalCid);
|
|
value = before.RawGetFieldAtOffset(from.offset);
|
|
after.RawSetFieldAtOffset(to.offset, value);
|
|
} else if (to.box_cid == kIllegalCid) {
|
|
// Unboxed to boxed field migration.
|
|
switch (from.box_cid) {
|
|
case kDoubleCid: {
|
|
const auto unboxed_value =
|
|
before.RawGetUnboxedFieldAtOffset<double>(from.offset);
|
|
value = Double::New(unboxed_value);
|
|
break;
|
|
}
|
|
case kFloat32x4Cid: {
|
|
const auto unboxed_value =
|
|
before.RawGetUnboxedFieldAtOffset<simd128_value_t>(from.offset);
|
|
value = Float32x4::New(unboxed_value);
|
|
break;
|
|
}
|
|
case kFloat64x2Cid: {
|
|
const auto unboxed_value =
|
|
before.RawGetUnboxedFieldAtOffset<simd128_value_t>(from.offset);
|
|
value = Float64x2::New(unboxed_value);
|
|
break;
|
|
}
|
|
case kIntegerCid: {
|
|
const auto unboxed_value =
|
|
before.RawGetUnboxedFieldAtOffset<int64_t>(from.offset);
|
|
value = Integer::New(unboxed_value);
|
|
break;
|
|
}
|
|
}
|
|
if (is_canonical) {
|
|
value = Instance::Cast(value).Canonicalize(Thread::Current());
|
|
}
|
|
after.RawSetFieldAtOffset(to.offset, value);
|
|
} else {
|
|
// Unboxed to unboxed field migration.
|
|
ASSERT(to.box_cid == from.box_cid);
|
|
switch (from.box_cid) {
|
|
case kDoubleCid: {
|
|
const auto unboxed_value =
|
|
before.RawGetUnboxedFieldAtOffset<double>(from.offset);
|
|
after.RawSetUnboxedFieldAtOffset<double>(to.offset, unboxed_value);
|
|
break;
|
|
}
|
|
case kFloat32x4Cid:
|
|
case kFloat64x2Cid: {
|
|
const auto unboxed_value =
|
|
before.RawGetUnboxedFieldAtOffset<simd128_value_t>(from.offset);
|
|
after.RawSetUnboxedFieldAtOffset<simd128_value_t>(to.offset,
|
|
unboxed_value);
|
|
break;
|
|
}
|
|
case kIntegerCid: {
|
|
const auto unboxed_value =
|
|
before.RawGetUnboxedFieldAtOffset<int64_t>(from.offset);
|
|
after.RawSetUnboxedFieldAtOffset<int64_t>(to.offset, unboxed_value);
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
for (intptr_t i = 0; i < new_fields_offsets_->length(); i++) {
|
|
const auto& field_offset = new_fields_offsets_->At(i);
|
|
after.RawSetFieldAtOffset(field_offset, Object::sentinel());
|
|
}
|
|
|
|
// Convert the old instance into a filler object. We will switch to the new
|
|
// class table before the next heap walk, so there must be no instances of
|
|
// any class with the old size.
|
|
Become::MakeDummyObject(before);
|
|
|
|
become->Add(before, after);
|
|
}
|
|
}
|
|
|
|
static const char* BoxCidToCString(intptr_t box_cid) {
|
|
switch (box_cid) {
|
|
case kDoubleCid:
|
|
return "double";
|
|
case kFloat32x4Cid:
|
|
return "float32x4";
|
|
case kFloat64x2Cid:
|
|
return "float64x2";
|
|
case kIntegerCid:
|
|
return "int64";
|
|
}
|
|
return "?";
|
|
}
|
|
|
|
void InstanceMorpher::Dump() const {
|
|
LogBlock blocker;
|
|
THR_Print("Morphing objects with cid: %d via this mapping: ", cid_);
|
|
for (int i = 0; i < mapping_->length(); i += 2) {
|
|
const auto& from = mapping_->At(i);
|
|
const auto& to = mapping_->At(i + 1);
|
|
THR_Print(" %" Pd "->%" Pd "", from.offset, to.offset);
|
|
THR_Print(" (%" Pd " -> %" Pd ")", from.box_cid, to.box_cid);
|
|
if (to.box_cid == kIllegalCid && from.box_cid != kIllegalCid) {
|
|
THR_Print("[box %s]", BoxCidToCString(from.box_cid));
|
|
} else if (to.box_cid != kIllegalCid) {
|
|
THR_Print("[%s]", BoxCidToCString(from.box_cid));
|
|
}
|
|
}
|
|
THR_Print("\n");
|
|
}
|
|
|
|
void InstanceMorpher::AppendTo(JSONArray* array) {
|
|
JSONObject jsobj(array);
|
|
jsobj.AddProperty("type", "ShapeChangeMapping");
|
|
jsobj.AddProperty64("class-id", cid_);
|
|
jsobj.AddProperty("instanceCount", before_.length());
|
|
JSONArray map(&jsobj, "fieldOffsetMappings");
|
|
for (int i = 0; i < mapping_->length(); i += 2) {
|
|
const auto& from = mapping_->At(i);
|
|
const auto& to = mapping_->At(i + 1);
|
|
|
|
JSONArray pair(&map);
|
|
pair.AddValue(from.offset);
|
|
pair.AddValue(to.offset);
|
|
if (to.box_cid == kIllegalCid && from.box_cid != kIllegalCid) {
|
|
pair.AddValueF("box %s", BoxCidToCString(from.box_cid));
|
|
} else if (to.box_cid != kIllegalCid) {
|
|
pair.AddValueF("%s", BoxCidToCString(from.box_cid));
|
|
}
|
|
}
|
|
}
|
|
|
|
void ReasonForCancelling::Report(IsolateGroupReloadContext* context) {
|
|
const Error& error = Error::Handle(ToError());
|
|
context->ReportError(error);
|
|
}
|
|
|
|
ErrorPtr ReasonForCancelling::ToError() {
|
|
// By default create the error returned from ToString.
|
|
const String& message = String::Handle(ToString());
|
|
return LanguageError::New(message);
|
|
}
|
|
|
|
StringPtr ReasonForCancelling::ToString() {
|
|
UNREACHABLE();
|
|
return NULL;
|
|
}
|
|
|
|
void ReasonForCancelling::AppendTo(JSONArray* array) {
|
|
JSONObject jsobj(array);
|
|
jsobj.AddProperty("type", "ReasonForCancelling");
|
|
const String& message = String::Handle(ToString());
|
|
jsobj.AddProperty("message", message.ToCString());
|
|
}
|
|
|
|
ClassReasonForCancelling::ClassReasonForCancelling(Zone* zone,
|
|
const Class& from,
|
|
const Class& to)
|
|
: ReasonForCancelling(zone),
|
|
from_(Class::ZoneHandle(zone, from.ptr())),
|
|
to_(Class::ZoneHandle(zone, to.ptr())) {}
|
|
|
|
void ClassReasonForCancelling::AppendTo(JSONArray* array) {
|
|
JSONObject jsobj(array);
|
|
jsobj.AddProperty("type", "ReasonForCancelling");
|
|
jsobj.AddProperty("class", from_);
|
|
const String& message = String::Handle(ToString());
|
|
jsobj.AddProperty("message", message.ToCString());
|
|
}
|
|
|
|
ErrorPtr IsolateGroupReloadContext::error() const {
|
|
ASSERT(!reasons_to_cancel_reload_.is_empty());
|
|
// Report the first error to the surroundings.
|
|
return reasons_to_cancel_reload_.At(0)->ToError();
|
|
}
|
|
|
|
class ScriptUrlSetTraits {
|
|
public:
|
|
static bool ReportStats() { return false; }
|
|
static const char* Name() { return "ScriptUrlSetTraits"; }
|
|
|
|
static bool IsMatch(const Object& a, const Object& b) {
|
|
if (!a.IsString() || !b.IsString()) {
|
|
return false;
|
|
}
|
|
|
|
return String::Cast(a).Equals(String::Cast(b));
|
|
}
|
|
|
|
static uword Hash(const Object& obj) { return String::Cast(obj).Hash(); }
|
|
};
|
|
|
|
class ClassMapTraits {
|
|
public:
|
|
static bool ReportStats() { return false; }
|
|
static const char* Name() { return "ClassMapTraits"; }
|
|
|
|
static bool IsMatch(const Object& a, const Object& b) {
|
|
if (!a.IsClass() || !b.IsClass()) {
|
|
return false;
|
|
}
|
|
return ProgramReloadContext::IsSameClass(Class::Cast(a), Class::Cast(b));
|
|
}
|
|
|
|
static uword Hash(const Object& obj) {
|
|
uword class_name_hash = String::HashRawSymbol(Class::Cast(obj).Name());
|
|
LibraryPtr raw_library = Class::Cast(obj).library();
|
|
if (raw_library == Library::null()) {
|
|
return class_name_hash;
|
|
}
|
|
return FinalizeHash(
|
|
CombineHashes(class_name_hash,
|
|
String::Hash(Library::Handle(raw_library).private_key())),
|
|
/* hashbits= */ 30);
|
|
}
|
|
};
|
|
|
|
class LibraryMapTraits {
|
|
public:
|
|
static bool ReportStats() { return false; }
|
|
static const char* Name() { return "LibraryMapTraits"; }
|
|
|
|
static bool IsMatch(const Object& a, const Object& b) {
|
|
if (!a.IsLibrary() || !b.IsLibrary()) {
|
|
return false;
|
|
}
|
|
return ProgramReloadContext::IsSameLibrary(Library::Cast(a),
|
|
Library::Cast(b));
|
|
}
|
|
|
|
static uword Hash(const Object& obj) { return Library::Cast(obj).UrlHash(); }
|
|
};
|
|
|
|
bool ProgramReloadContext::IsSameClass(const Class& a, const Class& b) {
|
|
// TODO(turnidge): We need to look at generic type arguments for
|
|
// synthetic mixin classes. Their names are not necessarily unique
|
|
// currently.
|
|
const String& a_name = String::Handle(a.Name());
|
|
const String& b_name = String::Handle(b.Name());
|
|
|
|
if (!a_name.Equals(b_name)) {
|
|
return false;
|
|
}
|
|
|
|
const Library& a_lib = Library::Handle(a.library());
|
|
const Library& b_lib = Library::Handle(b.library());
|
|
|
|
if (a_lib.IsNull() || b_lib.IsNull()) {
|
|
return a_lib.ptr() == b_lib.ptr();
|
|
}
|
|
return (a_lib.private_key() == b_lib.private_key());
|
|
}
|
|
|
|
bool ProgramReloadContext::IsSameLibrary(const Library& a_lib,
|
|
const Library& b_lib) {
|
|
const String& a_lib_url =
|
|
String::Handle(a_lib.IsNull() ? String::null() : a_lib.url());
|
|
const String& b_lib_url =
|
|
String::Handle(b_lib.IsNull() ? String::null() : b_lib.url());
|
|
return a_lib_url.Equals(b_lib_url);
|
|
}
|
|
|
|
IsolateGroupReloadContext::IsolateGroupReloadContext(
|
|
IsolateGroup* isolate_group,
|
|
ClassTable* class_table,
|
|
JSONStream* js)
|
|
: zone_(Thread::Current()->zone()),
|
|
isolate_group_(isolate_group),
|
|
class_table_(class_table),
|
|
start_time_micros_(OS::GetCurrentMonotonicMicros()),
|
|
reload_timestamp_(OS::GetCurrentTimeMillis()),
|
|
js_(js),
|
|
instance_morphers_(zone_, 0),
|
|
reasons_to_cancel_reload_(zone_, 0),
|
|
instance_morpher_by_cid_(zone_),
|
|
root_lib_url_(String::Handle(Z, String::null())),
|
|
root_url_prefix_(String::null()),
|
|
old_root_url_prefix_(String::null()) {}
|
|
IsolateGroupReloadContext::~IsolateGroupReloadContext() {}
|
|
|
|
ProgramReloadContext::ProgramReloadContext(
|
|
std::shared_ptr<IsolateGroupReloadContext> group_reload_context,
|
|
IsolateGroup* isolate_group)
|
|
: zone_(Thread::Current()->zone()),
|
|
group_reload_context_(group_reload_context),
|
|
isolate_group_(isolate_group),
|
|
old_classes_set_storage_(Array::null()),
|
|
class_map_storage_(Array::null()),
|
|
removed_class_set_storage_(Array::null()),
|
|
old_libraries_set_storage_(Array::null()),
|
|
library_map_storage_(Array::null()),
|
|
saved_root_library_(Library::null()),
|
|
saved_libraries_(GrowableObjectArray::null()) {
|
|
// NOTE: DO NOT ALLOCATE ANY RAW OBJECTS HERE. The ProgramReloadContext is not
|
|
// associated with the isolate yet and if a GC is triggered here the raw
|
|
// objects will not be properly accounted for.
|
|
ASSERT(zone_ != NULL);
|
|
}
|
|
|
|
ProgramReloadContext::~ProgramReloadContext() {
|
|
ASSERT(zone_ == Thread::Current()->zone());
|
|
ASSERT(IG->class_table() == IG->heap_walk_class_table());
|
|
}
|
|
|
|
void IsolateGroupReloadContext::ReportError(const Error& error) {
|
|
IsolateGroup* isolate_group = IsolateGroup::Current();
|
|
if (IsolateGroup::IsSystemIsolateGroup(isolate_group)) {
|
|
return;
|
|
}
|
|
TIR_Print("ISO-RELOAD: Error: %s\n", error.ToErrorCString());
|
|
ServiceEvent service_event(isolate_group, ServiceEvent::kIsolateReload);
|
|
service_event.set_reload_error(&error);
|
|
Service::HandleEvent(&service_event);
|
|
}
|
|
|
|
void IsolateGroupReloadContext::ReportSuccess() {
|
|
IsolateGroup* isolate_group = IsolateGroup::Current();
|
|
if (IsolateGroup::IsSystemIsolateGroup(isolate_group)) {
|
|
return;
|
|
}
|
|
ServiceEvent service_event(isolate_group, ServiceEvent::kIsolateReload);
|
|
Service::HandleEvent(&service_event);
|
|
}
|
|
|
|
class Aborted : public ReasonForCancelling {
|
|
public:
|
|
Aborted(Zone* zone, const Error& error)
|
|
: ReasonForCancelling(zone),
|
|
error_(Error::ZoneHandle(zone, error.ptr())) {}
|
|
|
|
private:
|
|
const Error& error_;
|
|
|
|
ErrorPtr ToError() { return error_.ptr(); }
|
|
StringPtr ToString() {
|
|
return String::NewFormatted("%s", error_.ToErrorCString());
|
|
}
|
|
};
|
|
|
|
static intptr_t CommonSuffixLength(const char* a, const char* b) {
|
|
const intptr_t a_length = strlen(a);
|
|
const intptr_t b_length = strlen(b);
|
|
intptr_t a_cursor = a_length;
|
|
intptr_t b_cursor = b_length;
|
|
|
|
while ((a_cursor >= 0) && (b_cursor >= 0)) {
|
|
if (a[a_cursor] != b[b_cursor]) {
|
|
break;
|
|
}
|
|
a_cursor--;
|
|
b_cursor--;
|
|
}
|
|
|
|
ASSERT((a_length - a_cursor) == (b_length - b_cursor));
|
|
return (a_length - a_cursor);
|
|
}
|
|
|
|
static ObjectPtr AcceptCompilation(Thread* thread) {
|
|
TransitionVMToNative transition(thread);
|
|
Dart_KernelCompilationResult result = KernelIsolate::AcceptCompilation();
|
|
if (result.status != Dart_KernelCompilationStatus_Ok) {
|
|
if (result.status != Dart_KernelCompilationStatus_MsgFailed) {
|
|
FATAL1(
|
|
"An error occurred while accepting the most recent"
|
|
" compilation results: %s",
|
|
result.error);
|
|
}
|
|
TIR_Print(
|
|
"An error occurred while accepting the most recent"
|
|
" compilation results: %s",
|
|
result.error);
|
|
Zone* zone = thread->zone();
|
|
const auto& error_str = String::Handle(zone, String::New(result.error));
|
|
free(result.error);
|
|
return ApiError::New(error_str);
|
|
}
|
|
return Object::null();
|
|
}
|
|
|
|
// If [root_script_url] is null, attempt to load from [kernel_buffer].
|
|
bool IsolateGroupReloadContext::Reload(bool force_reload,
|
|
const char* root_script_url,
|
|
const char* packages_url,
|
|
const uint8_t* kernel_buffer,
|
|
intptr_t kernel_buffer_size) {
|
|
TIMELINE_SCOPE(Reload);
|
|
|
|
Thread* thread = Thread::Current();
|
|
|
|
Heap* heap = IG->heap();
|
|
num_old_libs_ =
|
|
GrowableObjectArray::Handle(Z, IG->object_store()->libraries()).Length();
|
|
|
|
// Grab root library before calling CheckpointBeforeReload.
|
|
GetRootLibUrl(root_script_url);
|
|
|
|
std::unique_ptr<kernel::Program> kernel_program;
|
|
|
|
// Reset stats.
|
|
num_received_libs_ = 0;
|
|
bytes_received_libs_ = 0;
|
|
num_received_classes_ = 0;
|
|
num_received_procedures_ = 0;
|
|
|
|
bool did_kernel_compilation = false;
|
|
bool skip_reload = false;
|
|
{
|
|
// Load the kernel program and figure out the modified libraries.
|
|
intptr_t* p_num_received_classes = nullptr;
|
|
intptr_t* p_num_received_procedures = nullptr;
|
|
|
|
// ReadKernelFromFile checks to see if the file at
|
|
// root_script_url is a valid .dill file. If that's the case, a Program*
|
|
// is returned. Otherwise, this is likely a source file that needs to be
|
|
// compiled, so ReadKernelFromFile returns NULL.
|
|
kernel_program = kernel::Program::ReadFromFile(root_script_url);
|
|
if (kernel_program != nullptr) {
|
|
num_received_libs_ = kernel_program->library_count();
|
|
bytes_received_libs_ = kernel_program->kernel_data_size();
|
|
p_num_received_classes = &num_received_classes_;
|
|
p_num_received_procedures = &num_received_procedures_;
|
|
} else {
|
|
if (kernel_buffer == NULL || kernel_buffer_size == 0) {
|
|
char* error = CompileToKernel(force_reload, packages_url,
|
|
&kernel_buffer, &kernel_buffer_size);
|
|
did_kernel_compilation = true;
|
|
if (error != nullptr) {
|
|
TIR_Print("---- LOAD FAILED, ABORTING RELOAD\n");
|
|
const auto& error_str = String::Handle(Z, String::New(error));
|
|
free(error);
|
|
const ApiError& error = ApiError::Handle(Z, ApiError::New(error_str));
|
|
AddReasonForCancelling(new Aborted(Z, error));
|
|
ReportReasonsForCancelling();
|
|
CommonFinalizeTail(num_old_libs_);
|
|
return false;
|
|
}
|
|
}
|
|
const auto& typed_data = ExternalTypedData::Handle(
|
|
Z, ExternalTypedData::NewFinalizeWithFree(
|
|
const_cast<uint8_t*>(kernel_buffer), kernel_buffer_size));
|
|
kernel_program = kernel::Program::ReadFromTypedData(typed_data);
|
|
}
|
|
|
|
NoActiveIsolateScope no_active_isolate_scope;
|
|
|
|
ExternalTypedData& external_typed_data =
|
|
ExternalTypedData::Handle(Z, kernel_program->typed_data()->ptr());
|
|
IsolateGroupSource* source = IsolateGroup::Current()->source();
|
|
source->add_loaded_blob(Z, external_typed_data);
|
|
|
|
modified_libs_ = new (Z) BitVector(Z, num_old_libs_);
|
|
kernel::KernelLoader::FindModifiedLibraries(
|
|
kernel_program.get(), IG, modified_libs_, force_reload, &skip_reload,
|
|
p_num_received_classes, p_num_received_procedures);
|
|
modified_libs_transitive_ = new (Z) BitVector(Z, num_old_libs_);
|
|
BuildModifiedLibrariesClosure(modified_libs_);
|
|
|
|
ASSERT(num_saved_libs_ == -1);
|
|
num_saved_libs_ = 0;
|
|
for (intptr_t i = 0; i < modified_libs_->length(); i++) {
|
|
if (!modified_libs_->Contains(i)) {
|
|
num_saved_libs_++;
|
|
}
|
|
}
|
|
}
|
|
|
|
NoActiveIsolateScope no_active_isolate_scope;
|
|
|
|
if (skip_reload) {
|
|
ASSERT(modified_libs_->IsEmpty());
|
|
reload_skipped_ = true;
|
|
ReportOnJSON(js_, num_old_libs_);
|
|
|
|
// If we use the CFE and performed a compilation, we need to notify that
|
|
// we have accepted the compilation to clear some state in the incremental
|
|
// compiler.
|
|
if (did_kernel_compilation) {
|
|
const auto& result = Object::Handle(Z, AcceptCompilation(thread));
|
|
if (result.IsError()) {
|
|
const auto& error = Error::Cast(result);
|
|
AddReasonForCancelling(new Aborted(Z, error));
|
|
ReportReasonsForCancelling();
|
|
CommonFinalizeTail(num_old_libs_);
|
|
return false;
|
|
}
|
|
}
|
|
TIR_Print("---- SKIPPING RELOAD (No libraries were modified)\n");
|
|
return false;
|
|
}
|
|
|
|
TIR_Print("---- STARTING RELOAD\n");
|
|
|
|
intptr_t number_of_isolates = 0;
|
|
isolate_group_->ForEachIsolate(
|
|
[&](Isolate* isolate) { number_of_isolates++; });
|
|
|
|
// Disable the background compiler while we are performing the reload.
|
|
NoBackgroundCompilerScope stop_bg_compiler(thread);
|
|
|
|
// Wait for any concurrent marking tasks to finish and turn off the
|
|
// concurrent marker during reload as we might be allocating new instances
|
|
// (constants) when loading the new kernel file and this could cause
|
|
// inconsistency between the saved class table and the new class table.
|
|
const bool old_concurrent_mark_flag =
|
|
heap->old_space()->enable_concurrent_mark();
|
|
if (old_concurrent_mark_flag) {
|
|
heap->WaitForMarkerTasks(thread);
|
|
heap->old_space()->set_enable_concurrent_mark(false);
|
|
}
|
|
|
|
// Ensure all functions on the stack have unoptimized code.
|
|
// Deoptimize all code that had optimizing decisions that are dependent on
|
|
// assumptions from field guards or CHA or deferred library prefixes.
|
|
// TODO(johnmccutchan): Deoptimizing dependent code here (before the reload)
|
|
// is paranoid. This likely can be moved to the commit phase.
|
|
IG->program_reload_context()->EnsuredUnoptimizedCodeForStack();
|
|
IG->program_reload_context()->DeoptimizeDependentCode();
|
|
IG->program_reload_context()->ReloadPhase1AllocateStorageMapsAndCheckpoint();
|
|
|
|
// Renumbering the libraries has invalidated this.
|
|
modified_libs_ = nullptr;
|
|
modified_libs_transitive_ = nullptr;
|
|
|
|
if (FLAG_gc_during_reload) {
|
|
// We force the GC to compact, which is more likely to discover untracked
|
|
// pointers (and other issues, like incorrect class table).
|
|
heap->CollectAllGarbage(GCReason::kDebugging, /*compact=*/ true);
|
|
}
|
|
|
|
// Clone the class table.
|
|
{
|
|
TIMELINE_SCOPE(CheckpointClasses);
|
|
IG->program_reload_context()->CheckpointClasses();
|
|
}
|
|
|
|
if (FLAG_gc_during_reload) {
|
|
// We force the GC to compact, which is more likely to discover untracked
|
|
// pointers (and other issues, like incorrect class table).
|
|
heap->CollectAllGarbage(GCReason::kDebugging, /*compact=*/ true);
|
|
}
|
|
|
|
// We synchronously load the hot-reload kernel diff (which includes changed
|
|
// libraries and any libraries transitively depending on them).
|
|
//
|
|
// If loading the hot-reload diff succeeded we'll finalize the loading, which
|
|
// will either commit or reject the reload request.
|
|
const auto& result =
|
|
Object::Handle(Z, IG->program_reload_context()->ReloadPhase2LoadKernel(
|
|
kernel_program.get(), root_lib_url_));
|
|
|
|
if (result.IsError()) {
|
|
TIR_Print("---- LOAD FAILED, ABORTING RELOAD\n");
|
|
|
|
const auto& error = Error::Cast(result);
|
|
AddReasonForCancelling(new Aborted(Z, error));
|
|
|
|
IG->program_reload_context()->ReloadPhase4Rollback();
|
|
CommonFinalizeTail(num_old_libs_);
|
|
} else {
|
|
ASSERT(!reload_skipped_ && !reload_finalized_);
|
|
TIR_Print("---- LOAD SUCCEEDED\n");
|
|
|
|
IG->program_reload_context()->ReloadPhase3FinalizeLoading();
|
|
|
|
if (FLAG_gc_during_reload) {
|
|
// We force the GC to compact, which is more likely to discover untracked
|
|
// pointers (and other issues, like incorrect class table).
|
|
heap->CollectAllGarbage(GCReason::kDebugging, /*compact=*/ true);
|
|
}
|
|
|
|
// If we use the CFE and performed a compilation, we need to notify that
|
|
// we have accepted the compilation to clear some state in the incremental
|
|
// compiler.
|
|
if (did_kernel_compilation) {
|
|
const auto& result = Object::Handle(Z, AcceptCompilation(thread));
|
|
if (result.IsError()) {
|
|
const auto& error = Error::Cast(result);
|
|
AddReasonForCancelling(new Aborted(Z, error));
|
|
}
|
|
}
|
|
|
|
if (!FLAG_reload_force_rollback && !HasReasonsForCancelling()) {
|
|
TIR_Print("---- COMMITTING RELOAD\n");
|
|
isolate_group_->program_reload_context()->ReloadPhase4CommitPrepare();
|
|
bool discard_class_tables = true;
|
|
if (HasInstanceMorphers()) {
|
|
// Find all objects that need to be morphed (reallocated to a new
|
|
// layout).
|
|
ObjectLocator locator(this);
|
|
{
|
|
HeapIterationScope iteration(Thread::Current());
|
|
iteration.IterateObjects(&locator);
|
|
}
|
|
|
|
// We are still using the old class table at this point.
|
|
if (FLAG_gc_during_reload) {
|
|
// We force the GC to compact, which is more likely to discover
|
|
// untracked pointers (and other issues, like incorrect class table).
|
|
heap->CollectAllGarbage(GCReason::kDebugging, /*compact=*/ true);
|
|
}
|
|
const intptr_t count = locator.count();
|
|
if (count > 0) {
|
|
TIMELINE_SCOPE(MorphInstances);
|
|
|
|
// While we are reallocating instances to their new layout, the heap
|
|
// will contain a mix of instances with the old and new layouts that
|
|
// have the same cid. This makes the heap unwalkable until the
|
|
// "become" operation below replaces all the instances of the old
|
|
// layout with forwarding corpses. Force heap growth to prevent layout
|
|
// confusion during this period.
|
|
ForceGrowthScope force_growth(thread);
|
|
// The HeapIterationScope above ensures no other GC tasks can be
|
|
// active.
|
|
ASSERT(HasNoTasks(heap));
|
|
|
|
MorphInstancesPhase1Allocate(&locator, IG->become());
|
|
{
|
|
// Apply the new class table before "become". Become will replace
|
|
// all the instances of the old layout with forwarding corpses, then
|
|
// perform a heap walk to fix references to the forwarding corpses.
|
|
// During this heap walk, it will encounter instances of the new
|
|
// layout, so it requires the new class table.
|
|
ASSERT(HasNoTasks(heap));
|
|
|
|
// We accepted the hot-reload and morphed instances. So now we can
|
|
// commit to the changed class table and deleted the saved one.
|
|
IG->DropOriginalClassTable();
|
|
}
|
|
MorphInstancesPhase2Become(IG->become());
|
|
|
|
discard_class_tables = false;
|
|
}
|
|
// We are using the new class table now.
|
|
if (FLAG_gc_during_reload) {
|
|
// We force the GC to compact, which is more likely to discover
|
|
// untracked pointers (and other issues, like incorrect class table).
|
|
heap->CollectAllGarbage(GCReason::kDebugging, /*compact=*/ true);
|
|
}
|
|
}
|
|
if (FLAG_identity_reload) {
|
|
if (!discard_class_tables) {
|
|
TIR_Print("Identity reload failed! Some instances were morphed\n");
|
|
}
|
|
if (IG->heap_walk_class_table()->NumCids() !=
|
|
IG->class_table()->NumCids()) {
|
|
TIR_Print("Identity reload failed! B#C=%" Pd " A#C=%" Pd "\n",
|
|
IG->heap_walk_class_table()->NumCids(),
|
|
IG->class_table()->NumCids());
|
|
}
|
|
if (IG->heap_walk_class_table()->NumTopLevelCids() !=
|
|
IG->class_table()->NumTopLevelCids()) {
|
|
TIR_Print("Identity reload failed! B#TLC=%" Pd " A#TLC=%" Pd "\n",
|
|
IG->heap_walk_class_table()->NumTopLevelCids(),
|
|
IG->class_table()->NumTopLevelCids());
|
|
}
|
|
}
|
|
if (discard_class_tables) {
|
|
IG->DropOriginalClassTable();
|
|
}
|
|
isolate_group_->program_reload_context()->ReloadPhase4CommitFinish();
|
|
TIR_Print("---- DONE COMMIT\n");
|
|
isolate_group_->set_last_reload_timestamp(reload_timestamp_);
|
|
} else {
|
|
TIR_Print("---- ROLLING BACK");
|
|
isolate_group_->program_reload_context()->ReloadPhase4Rollback();
|
|
}
|
|
|
|
// ValidateReload mutates the direct subclass information and does
|
|
// not remove dead subclasses. Rebuild the direct subclass
|
|
// information from scratch.
|
|
{
|
|
SafepointWriteRwLocker ml(thread, IG->program_lock());
|
|
IG->program_reload_context()->RebuildDirectSubclasses();
|
|
}
|
|
const intptr_t final_library_count =
|
|
GrowableObjectArray::Handle(Z, IG->object_store()->libraries())
|
|
.Length();
|
|
CommonFinalizeTail(final_library_count);
|
|
}
|
|
|
|
// Reenable concurrent marking if it was initially on.
|
|
if (old_concurrent_mark_flag) {
|
|
heap->old_space()->set_enable_concurrent_mark(true);
|
|
}
|
|
|
|
bool success;
|
|
if (!result.IsError() || HasReasonsForCancelling()) {
|
|
ReportSuccess();
|
|
success = true;
|
|
} else {
|
|
ReportReasonsForCancelling();
|
|
success = false;
|
|
}
|
|
|
|
Array& null_array = Array::Handle(Z);
|
|
// Invalidate the URI mapping caches.
|
|
IG->object_store()->set_uri_to_resolved_uri_map(null_array);
|
|
IG->object_store()->set_resolved_uri_to_uri_map(null_array);
|
|
|
|
// Re-queue any shutdown requests so they can inform each isolate's own thread
|
|
// to shut down.
|
|
if (result.IsUnwindError()) {
|
|
const auto& error = UnwindError::Cast(result);
|
|
ForEachIsolate([&](Isolate* isolate) {
|
|
Isolate::KillIfExists(isolate, error.is_user_initiated()
|
|
? Isolate::kKillMsg
|
|
: Isolate::kInternalKillMsg);
|
|
});
|
|
}
|
|
|
|
return success;
|
|
}
|
|
|
|
/// Copied in from https://dart-review.googlesource.com/c/sdk/+/77722.
|
|
static void PropagateLibraryModified(
|
|
const ZoneGrowableArray<ZoneGrowableArray<intptr_t>*>* imported_by,
|
|
intptr_t lib_index,
|
|
BitVector* modified_libs) {
|
|
ZoneGrowableArray<intptr_t>* dep_libs = (*imported_by)[lib_index];
|
|
for (intptr_t i = 0; i < dep_libs->length(); i++) {
|
|
intptr_t dep_lib_index = (*dep_libs)[i];
|
|
if (!modified_libs->Contains(dep_lib_index)) {
|
|
modified_libs->Add(dep_lib_index);
|
|
PropagateLibraryModified(imported_by, dep_lib_index, modified_libs);
|
|
}
|
|
}
|
|
}
|
|
|
|
/// Copied in from https://dart-review.googlesource.com/c/sdk/+/77722.
|
|
void IsolateGroupReloadContext::BuildModifiedLibrariesClosure(
|
|
BitVector* modified_libs) {
|
|
const GrowableObjectArray& libs =
|
|
GrowableObjectArray::Handle(IG->object_store()->libraries());
|
|
Library& lib = Library::Handle();
|
|
intptr_t num_libs = libs.Length();
|
|
|
|
// Construct the imported-by graph.
|
|
ZoneGrowableArray<ZoneGrowableArray<intptr_t>*>* imported_by = new (zone_)
|
|
ZoneGrowableArray<ZoneGrowableArray<intptr_t>*>(zone_, num_libs);
|
|
imported_by->SetLength(num_libs);
|
|
for (intptr_t i = 0; i < num_libs; i++) {
|
|
(*imported_by)[i] = new (zone_) ZoneGrowableArray<intptr_t>(zone_, 0);
|
|
}
|
|
Array& ports = Array::Handle();
|
|
Namespace& ns = Namespace::Handle();
|
|
Library& target = Library::Handle();
|
|
String& target_url = String::Handle();
|
|
|
|
for (intptr_t lib_idx = 0; lib_idx < num_libs; lib_idx++) {
|
|
lib ^= libs.At(lib_idx);
|
|
ASSERT(lib_idx == lib.index());
|
|
if (lib.is_dart_scheme()) {
|
|
// We don't care about imports among dart scheme libraries.
|
|
continue;
|
|
}
|
|
|
|
// Add imports to the import-by graph.
|
|
ports = lib.imports();
|
|
for (intptr_t import_idx = 0; import_idx < ports.Length(); import_idx++) {
|
|
ns ^= ports.At(import_idx);
|
|
if (!ns.IsNull()) {
|
|
target = ns.target();
|
|
target_url = target.url();
|
|
(*imported_by)[target.index()]->Add(lib.index());
|
|
}
|
|
}
|
|
|
|
// Add exports to the import-by graph.
|
|
ports = lib.exports();
|
|
for (intptr_t export_idx = 0; export_idx < ports.Length(); export_idx++) {
|
|
ns ^= ports.At(export_idx);
|
|
if (!ns.IsNull()) {
|
|
target = ns.target();
|
|
(*imported_by)[target.index()]->Add(lib.index());
|
|
}
|
|
}
|
|
|
|
// Add prefixed imports to the import-by graph.
|
|
DictionaryIterator entries(lib);
|
|
Object& entry = Object::Handle();
|
|
LibraryPrefix& prefix = LibraryPrefix::Handle();
|
|
while (entries.HasNext()) {
|
|
entry = entries.GetNext();
|
|
if (entry.IsLibraryPrefix()) {
|
|
prefix ^= entry.ptr();
|
|
ports = prefix.imports();
|
|
for (intptr_t import_idx = 0; import_idx < ports.Length();
|
|
import_idx++) {
|
|
ns ^= ports.At(import_idx);
|
|
if (!ns.IsNull()) {
|
|
target = ns.target();
|
|
(*imported_by)[target.index()]->Add(lib.index());
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
for (intptr_t lib_idx = 0; lib_idx < num_libs; lib_idx++) {
|
|
lib ^= libs.At(lib_idx);
|
|
if (lib.is_dart_scheme() || modified_libs_transitive_->Contains(lib_idx)) {
|
|
// We don't consider dart scheme libraries during reload. If
|
|
// the modified libs set already contains this library, then we
|
|
// have already visited it.
|
|
continue;
|
|
}
|
|
if (modified_libs->Contains(lib_idx)) {
|
|
modified_libs_transitive_->Add(lib_idx);
|
|
PropagateLibraryModified(imported_by, lib_idx, modified_libs_transitive_);
|
|
}
|
|
}
|
|
}
|
|
|
|
void IsolateGroupReloadContext::GetRootLibUrl(const char* root_script_url) {
|
|
const auto& old_root_lib =
|
|
Library::Handle(IG->object_store()->root_library());
|
|
ASSERT(!old_root_lib.IsNull());
|
|
const auto& old_root_lib_url = String::Handle(old_root_lib.url());
|
|
|
|
// Root library url.
|
|
if (root_script_url != nullptr) {
|
|
root_lib_url_ = String::New(root_script_url);
|
|
} else {
|
|
root_lib_url_ = old_root_lib_url.ptr();
|
|
}
|
|
|
|
// Check to see if the base url of the loaded libraries has moved.
|
|
if (!old_root_lib_url.Equals(root_lib_url_)) {
|
|
const char* old_root_library_url_c = old_root_lib_url.ToCString();
|
|
const char* root_library_url_c = root_lib_url_.ToCString();
|
|
const intptr_t common_suffix_length =
|
|
CommonSuffixLength(root_library_url_c, old_root_library_url_c);
|
|
root_url_prefix_ = String::SubString(
|
|
root_lib_url_, 0, root_lib_url_.Length() - common_suffix_length + 1);
|
|
old_root_url_prefix_ =
|
|
String::SubString(old_root_lib_url, 0,
|
|
old_root_lib_url.Length() - common_suffix_length + 1);
|
|
}
|
|
}
|
|
|
|
char* IsolateGroupReloadContext::CompileToKernel(bool force_reload,
|
|
const char* packages_url,
|
|
const uint8_t** kernel_buffer,
|
|
intptr_t* kernel_buffer_size) {
|
|
Dart_SourceFile* modified_scripts = nullptr;
|
|
intptr_t modified_scripts_count = 0;
|
|
FindModifiedSources(force_reload, &modified_scripts, &modified_scripts_count,
|
|
packages_url);
|
|
|
|
Dart_KernelCompilationResult retval = {};
|
|
{
|
|
const char* root_lib_url = root_lib_url_.ToCString();
|
|
TransitionVMToNative transition(Thread::Current());
|
|
retval = KernelIsolate::CompileToKernel(
|
|
root_lib_url, nullptr, 0, modified_scripts_count, modified_scripts,
|
|
/*incremental_compile=*/true,
|
|
/*snapshot_compile=*/false,
|
|
/*package_config=*/nullptr,
|
|
/*multiroot_filepaths=*/nullptr,
|
|
/*multiroot_scheme=*/nullptr, FLAG_sound_null_safety);
|
|
}
|
|
if (retval.status != Dart_KernelCompilationStatus_Ok) {
|
|
if (retval.kernel != nullptr) {
|
|
free(retval.kernel);
|
|
}
|
|
return retval.error;
|
|
}
|
|
*kernel_buffer = retval.kernel;
|
|
*kernel_buffer_size = retval.kernel_size;
|
|
return nullptr;
|
|
}
|
|
|
|
void ProgramReloadContext::ReloadPhase1AllocateStorageMapsAndCheckpoint() {
|
|
// Preallocate storage for maps.
|
|
old_classes_set_storage_ =
|
|
HashTables::New<UnorderedHashSet<ClassMapTraits> >(4);
|
|
class_map_storage_ = HashTables::New<UnorderedHashMap<ClassMapTraits> >(4);
|
|
removed_class_set_storage_ =
|
|
HashTables::New<UnorderedHashSet<ClassMapTraits> >(4);
|
|
old_libraries_set_storage_ =
|
|
HashTables::New<UnorderedHashSet<LibraryMapTraits> >(4);
|
|
library_map_storage_ =
|
|
HashTables::New<UnorderedHashMap<LibraryMapTraits> >(4);
|
|
|
|
// While reloading everything we do must be reversible so that we can abort
|
|
// safely if the reload fails. This function stashes things to the side and
|
|
// prepares the isolate for the reload attempt.
|
|
{
|
|
TIMELINE_SCOPE(Checkpoint);
|
|
CheckpointLibraries();
|
|
}
|
|
}
|
|
|
|
ObjectPtr ProgramReloadContext::ReloadPhase2LoadKernel(
|
|
kernel::Program* program,
|
|
const String& root_lib_url) {
|
|
Thread* thread = Thread::Current();
|
|
|
|
LongJumpScope jump;
|
|
if (setjmp(*jump.Set()) == 0) {
|
|
const Object& tmp = kernel::KernelLoader::LoadEntireProgram(program);
|
|
if (tmp.IsError()) {
|
|
return tmp.ptr();
|
|
}
|
|
|
|
// If main method disappeared or were not there to begin with then
|
|
// KernelLoader will return null. In this case lookup library by
|
|
// URL.
|
|
auto& lib = Library::Handle(Library::RawCast(tmp.ptr()));
|
|
if (lib.IsNull()) {
|
|
lib = Library::LookupLibrary(thread, root_lib_url);
|
|
}
|
|
IG->object_store()->set_root_library(lib);
|
|
return Object::null();
|
|
} else {
|
|
return thread->StealStickyError();
|
|
}
|
|
}
|
|
|
|
void ProgramReloadContext::ReloadPhase3FinalizeLoading() {
|
|
BuildLibraryMapping();
|
|
BuildRemovedClassesSet();
|
|
ValidateReload();
|
|
}
|
|
|
|
void ProgramReloadContext::ReloadPhase4CommitPrepare() {
|
|
CommitBeforeInstanceMorphing();
|
|
}
|
|
|
|
void ProgramReloadContext::ReloadPhase4CommitFinish() {
|
|
CommitAfterInstanceMorphing();
|
|
PostCommit();
|
|
}
|
|
|
|
void ProgramReloadContext::ReloadPhase4Rollback() {
|
|
IG->RestoreOriginalClassTable();
|
|
RollbackLibraries();
|
|
}
|
|
|
|
void ProgramReloadContext::RegisterClass(const Class& new_cls) {
|
|
const Class& old_cls = Class::Handle(OldClassOrNull(new_cls));
|
|
if (old_cls.IsNull()) {
|
|
if (new_cls.IsTopLevel()) {
|
|
IG->class_table()->RegisterTopLevel(new_cls);
|
|
} else {
|
|
IG->class_table()->Register(new_cls);
|
|
}
|
|
|
|
if (FLAG_identity_reload) {
|
|
TIR_Print("Could not find replacement class for %s\n",
|
|
new_cls.ToCString());
|
|
UNREACHABLE();
|
|
}
|
|
|
|
// New class maps to itself.
|
|
AddClassMapping(new_cls, new_cls);
|
|
return;
|
|
}
|
|
VTIR_Print("Registering class: %s\n", new_cls.ToCString());
|
|
new_cls.set_id(old_cls.id());
|
|
IG->class_table()->SetAt(old_cls.id(), new_cls.ptr());
|
|
if (!old_cls.is_enum_class()) {
|
|
new_cls.CopyCanonicalConstants(old_cls);
|
|
}
|
|
new_cls.CopyDeclarationType(old_cls);
|
|
AddBecomeMapping(old_cls, new_cls);
|
|
AddClassMapping(new_cls, old_cls);
|
|
}
|
|
|
|
void IsolateGroupReloadContext::CommonFinalizeTail(
|
|
intptr_t final_library_count) {
|
|
RELEASE_ASSERT(!reload_finalized_);
|
|
ReportOnJSON(js_, final_library_count);
|
|
reload_finalized_ = true;
|
|
}
|
|
|
|
void IsolateGroupReloadContext::ReportOnJSON(JSONStream* stream,
|
|
intptr_t final_library_count) {
|
|
JSONObject jsobj(stream);
|
|
jsobj.AddProperty("type", "ReloadReport");
|
|
jsobj.AddProperty("success", reload_skipped_ || !HasReasonsForCancelling());
|
|
{
|
|
if (HasReasonsForCancelling()) {
|
|
// Reload was rejected.
|
|
JSONArray array(&jsobj, "notices");
|
|
for (intptr_t i = 0; i < reasons_to_cancel_reload_.length(); i++) {
|
|
ReasonForCancelling* reason = reasons_to_cancel_reload_.At(i);
|
|
reason->AppendTo(&array);
|
|
}
|
|
return;
|
|
}
|
|
|
|
JSONObject details(&jsobj, "details");
|
|
details.AddProperty("finalLibraryCount", final_library_count);
|
|
details.AddProperty("receivedLibraryCount", num_received_libs_);
|
|
details.AddProperty("receivedLibrariesBytes", bytes_received_libs_);
|
|
details.AddProperty("receivedClassesCount", num_received_classes_);
|
|
details.AddProperty("receivedProceduresCount", num_received_procedures_);
|
|
if (reload_skipped_) {
|
|
// Reload was skipped.
|
|
details.AddProperty("savedLibraryCount", final_library_count);
|
|
details.AddProperty("loadedLibraryCount", static_cast<intptr_t>(0));
|
|
} else {
|
|
// Reload was successful.
|
|
const intptr_t loaded_library_count =
|
|
final_library_count - num_saved_libs_;
|
|
details.AddProperty("savedLibraryCount", num_saved_libs_);
|
|
details.AddProperty("loadedLibraryCount", loaded_library_count);
|
|
JSONArray array(&jsobj, "shapeChangeMappings");
|
|
for (intptr_t i = 0; i < instance_morphers_.length(); i++) {
|
|
instance_morphers_.At(i)->AppendTo(&array);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
void ProgramReloadContext::EnsuredUnoptimizedCodeForStack() {
|
|
TIMELINE_SCOPE(EnsuredUnoptimizedCodeForStack);
|
|
|
|
IG->ForEachIsolate([](Isolate* isolate) {
|
|
auto thread = isolate->mutator_thread();
|
|
StackFrameIterator it(ValidationPolicy::kDontValidateFrames, thread,
|
|
StackFrameIterator::kAllowCrossThreadIteration);
|
|
|
|
Function& func = Function::Handle();
|
|
while (it.HasNextFrame()) {
|
|
StackFrame* frame = it.NextFrame();
|
|
if (frame->IsDartFrame()) {
|
|
func = frame->LookupDartFunction();
|
|
ASSERT(!func.IsNull());
|
|
// Force-optimized functions don't need unoptimized code because their
|
|
// optimized code cannot deopt.
|
|
if (!func.ForceOptimize()) {
|
|
func.EnsureHasCompiledUnoptimizedCode();
|
|
}
|
|
}
|
|
}
|
|
});
|
|
}
|
|
|
|
void ProgramReloadContext::DeoptimizeDependentCode() {
|
|
TIMELINE_SCOPE(DeoptimizeDependentCode);
|
|
ClassTable* class_table = IG->class_table();
|
|
|
|
const intptr_t bottom = Dart::vm_isolate_group()->class_table()->NumCids();
|
|
const intptr_t top = IG->class_table()->NumCids();
|
|
Class& cls = Class::Handle();
|
|
Array& fields = Array::Handle();
|
|
Field& field = Field::Handle();
|
|
Thread* thread = Thread::Current();
|
|
SafepointWriteRwLocker ml(thread, IG->program_lock());
|
|
for (intptr_t cls_idx = bottom; cls_idx < top; cls_idx++) {
|
|
if (!class_table->HasValidClassAt(cls_idx)) {
|
|
// Skip.
|
|
continue;
|
|
}
|
|
|
|
// Deoptimize CHA code.
|
|
cls = class_table->At(cls_idx);
|
|
ASSERT(!cls.IsNull());
|
|
|
|
cls.DisableAllCHAOptimizedCode();
|
|
|
|
// Deoptimize field guard code.
|
|
fields = cls.fields();
|
|
ASSERT(!fields.IsNull());
|
|
for (intptr_t field_idx = 0; field_idx < fields.Length(); field_idx++) {
|
|
field = Field::RawCast(fields.At(field_idx));
|
|
ASSERT(!field.IsNull());
|
|
field.DeoptimizeDependentCode();
|
|
}
|
|
}
|
|
|
|
DeoptimizeTypeTestingStubs();
|
|
|
|
// TODO(rmacnak): Also call LibraryPrefix::InvalidateDependentCode.
|
|
}
|
|
|
|
void ProgramReloadContext::CheckpointClasses() {
|
|
TIR_Print("---- CHECKPOINTING CLASSES\n");
|
|
// Checkpoint classes before a reload.
|
|
|
|
// Before this operation class table which is used for heap scanning and
|
|
// the class table used for program loading are the same. After this step
|
|
// they will become different until reload is commited (or rolled back).
|
|
//
|
|
// Note that because GC is always reading from heap_walk_class_table and
|
|
// we are not changing that, there is no reason to wait for sweeping
|
|
// threads or marking to complete.
|
|
RELEASE_ASSERT(IG->class_table() == IG->heap_walk_class_table());
|
|
|
|
IG->CloneClassTableForReload();
|
|
|
|
// IG->class_table() is now the clone of heap_walk_class_table.
|
|
RELEASE_ASSERT(IG->class_table() != IG->heap_walk_class_table());
|
|
|
|
ClassTable* class_table = IG->class_table();
|
|
|
|
// For efficiency, we build a set of classes before the reload. This set
|
|
// is used to pair new classes with old classes.
|
|
// Add classes to the set. Set is stored in the Array, so adding an element
|
|
// may allocate Dart object on the heap and trigger GC.
|
|
Class& cls = Class::Handle();
|
|
UnorderedHashSet<ClassMapTraits> old_classes_set(old_classes_set_storage_);
|
|
for (intptr_t i = 0; i < class_table->NumCids(); i++) {
|
|
if (class_table->IsValidIndex(i) && class_table->HasValidClassAt(i)) {
|
|
if (i != kFreeListElement && i != kForwardingCorpse) {
|
|
cls = class_table->At(i);
|
|
bool already_present = old_classes_set.Insert(cls);
|
|
ASSERT(!already_present);
|
|
}
|
|
}
|
|
}
|
|
for (intptr_t i = 0; i < class_table->NumTopLevelCids(); i++) {
|
|
const intptr_t cid = ClassTable::CidFromTopLevelIndex(i);
|
|
if (class_table->IsValidIndex(cid) && class_table->HasValidClassAt(cid)) {
|
|
cls = class_table->At(cid);
|
|
bool already_present = old_classes_set.Insert(cls);
|
|
ASSERT(!already_present);
|
|
}
|
|
}
|
|
old_classes_set_storage_ = old_classes_set.Release().ptr();
|
|
TIR_Print("---- System had %" Pd " classes\n",
|
|
class_table->NumCids() + class_table->NumTopLevelCids());
|
|
}
|
|
|
|
Dart_FileModifiedCallback IsolateGroupReloadContext::file_modified_callback_ =
|
|
nullptr;
|
|
|
|
bool IsolateGroupReloadContext::ScriptModifiedSince(const Script& script,
|
|
int64_t since) {
|
|
if (IsolateGroupReloadContext::file_modified_callback_ == NULL) {
|
|
return true;
|
|
}
|
|
// We use the resolved url to determine if the script has been modified.
|
|
const String& url = String::Handle(script.resolved_url());
|
|
const char* url_chars = url.ToCString();
|
|
return (*IsolateGroupReloadContext::file_modified_callback_)(url_chars,
|
|
since);
|
|
}
|
|
|
|
static bool ContainsScriptUri(const GrowableArray<const char*>& seen_uris,
|
|
const char* uri) {
|
|
for (intptr_t i = 0; i < seen_uris.length(); i++) {
|
|
const char* seen_uri = seen_uris.At(i);
|
|
size_t seen_len = strlen(seen_uri);
|
|
if (seen_len != strlen(uri)) {
|
|
continue;
|
|
} else if (strncmp(seen_uri, uri, seen_len) == 0) {
|
|
return true;
|
|
}
|
|
}
|
|
return false;
|
|
}
|
|
|
|
void IsolateGroupReloadContext::FindModifiedSources(
|
|
bool force_reload,
|
|
Dart_SourceFile** modified_sources,
|
|
intptr_t* count,
|
|
const char* packages_url) {
|
|
const int64_t last_reload = isolate_group_->last_reload_timestamp();
|
|
GrowableArray<const char*> modified_sources_uris;
|
|
const auto& libs =
|
|
GrowableObjectArray::Handle(IG->object_store()->libraries());
|
|
Library& lib = Library::Handle(Z);
|
|
Array& scripts = Array::Handle(Z);
|
|
Script& script = Script::Handle(Z);
|
|
String& uri = String::Handle(Z);
|
|
|
|
for (intptr_t lib_idx = 0; lib_idx < libs.Length(); lib_idx++) {
|
|
lib ^= libs.At(lib_idx);
|
|
if (lib.is_dart_scheme()) {
|
|
// We don't consider dart scheme libraries during reload.
|
|
continue;
|
|
}
|
|
scripts = lib.LoadedScripts();
|
|
for (intptr_t script_idx = 0; script_idx < scripts.Length(); script_idx++) {
|
|
script ^= scripts.At(script_idx);
|
|
uri = script.url();
|
|
const bool dart_scheme = uri.StartsWith(Symbols::DartScheme());
|
|
if (dart_scheme) {
|
|
// If a user-defined class mixes in a mixin from dart:*, it's list of
|
|
// scripts will have a dart:* script as well. We don't consider those
|
|
// during reload.
|
|
continue;
|
|
}
|
|
if (ContainsScriptUri(modified_sources_uris, uri.ToCString())) {
|
|
// We've already accounted for this script in a prior library.
|
|
continue;
|
|
}
|
|
|
|
if (force_reload || ScriptModifiedSince(script, last_reload)) {
|
|
modified_sources_uris.Add(uri.ToCString());
|
|
}
|
|
}
|
|
}
|
|
|
|
// In addition to all sources, we need to check if the .packages file
|
|
// contents have been modified.
|
|
if (packages_url != NULL) {
|
|
if (IsolateGroupReloadContext::file_modified_callback_ == NULL ||
|
|
(*IsolateGroupReloadContext::file_modified_callback_)(packages_url,
|
|
last_reload)) {
|
|
modified_sources_uris.Add(packages_url);
|
|
}
|
|
}
|
|
|
|
*count = modified_sources_uris.length();
|
|
if (*count == 0) {
|
|
return;
|
|
}
|
|
|
|
*modified_sources = Z->Alloc<Dart_SourceFile>(*count);
|
|
for (intptr_t i = 0; i < *count; ++i) {
|
|
(*modified_sources)[i].uri = modified_sources_uris[i];
|
|
(*modified_sources)[i].source = NULL;
|
|
}
|
|
}
|
|
|
|
void ProgramReloadContext::CheckpointLibraries() {
|
|
TIMELINE_SCOPE(CheckpointLibraries);
|
|
TIR_Print("---- CHECKPOINTING LIBRARIES\n");
|
|
// Save the root library in case we abort the reload.
|
|
const Library& root_lib = Library::Handle(object_store()->root_library());
|
|
saved_root_library_ = root_lib.ptr();
|
|
|
|
// Save the old libraries array in case we abort the reload.
|
|
const GrowableObjectArray& libs =
|
|
GrowableObjectArray::Handle(object_store()->libraries());
|
|
saved_libraries_ = libs.ptr();
|
|
|
|
// Make a filtered copy of the old libraries array. Keep "clean" libraries
|
|
// that we will use instead of reloading.
|
|
const GrowableObjectArray& new_libs =
|
|
GrowableObjectArray::Handle(GrowableObjectArray::New(Heap::kOld));
|
|
Library& lib = Library::Handle();
|
|
UnorderedHashSet<LibraryMapTraits> old_libraries_set(
|
|
old_libraries_set_storage_);
|
|
|
|
group_reload_context_->saved_libs_transitive_updated_ = new (Z)
|
|
BitVector(Z, group_reload_context_->modified_libs_transitive_->length());
|
|
for (intptr_t i = 0; i < libs.Length(); i++) {
|
|
lib ^= libs.At(i);
|
|
if (group_reload_context_->modified_libs_->Contains(i)) {
|
|
// We are going to reload this library. Clear the index.
|
|
lib.set_index(-1);
|
|
} else {
|
|
// We are preserving this library across the reload, assign its new index
|
|
lib.set_index(new_libs.Length());
|
|
new_libs.Add(lib, Heap::kOld);
|
|
|
|
if (group_reload_context_->modified_libs_transitive_->Contains(i)) {
|
|
// Remember the new index.
|
|
group_reload_context_->saved_libs_transitive_updated_->Add(lib.index());
|
|
}
|
|
}
|
|
// Add old library to old libraries set.
|
|
bool already_present = old_libraries_set.Insert(lib);
|
|
ASSERT(!already_present);
|
|
}
|
|
old_libraries_set_storage_ = old_libraries_set.Release().ptr();
|
|
|
|
// Reset the registered libraries to the filtered array.
|
|
Library::RegisterLibraries(Thread::Current(), new_libs);
|
|
// Reset the root library to null.
|
|
object_store()->set_root_library(Library::Handle());
|
|
}
|
|
|
|
|
|
void ProgramReloadContext::RollbackLibraries() {
|
|
TIR_Print("---- ROLLING BACK LIBRARY CHANGES\n");
|
|
Thread* thread = Thread::Current();
|
|
Library& lib = Library::Handle();
|
|
const auto& saved_libs = GrowableObjectArray::Handle(Z, saved_libraries_);
|
|
if (!saved_libs.IsNull()) {
|
|
for (intptr_t i = 0; i < saved_libs.Length(); i++) {
|
|
lib = Library::RawCast(saved_libs.At(i));
|
|
// Restore indexes that were modified in CheckpointLibraries.
|
|
lib.set_index(i);
|
|
}
|
|
|
|
// Reset the registered libraries to the filtered array.
|
|
Library::RegisterLibraries(thread, saved_libs);
|
|
}
|
|
|
|
Library& saved_root_lib = Library::Handle(Z, saved_root_library_);
|
|
if (!saved_root_lib.IsNull()) {
|
|
object_store()->set_root_library(saved_root_lib);
|
|
}
|
|
|
|
saved_root_library_ = Library::null();
|
|
saved_libraries_ = GrowableObjectArray::null();
|
|
}
|
|
|
|
#ifdef DEBUG
|
|
void ProgramReloadContext::VerifyMaps() {
|
|
TIMELINE_SCOPE(VerifyMaps);
|
|
Class& cls = Class::Handle();
|
|
Class& new_cls = Class::Handle();
|
|
Class& cls2 = Class::Handle();
|
|
|
|
// Verify that two old classes aren't both mapped to the same new
|
|
// class. This could happen is the IsSameClass function is broken.
|
|
UnorderedHashMap<ClassMapTraits> class_map(class_map_storage_);
|
|
UnorderedHashMap<ClassMapTraits> reverse_class_map(
|
|
HashTables::New<UnorderedHashMap<ClassMapTraits> >(
|
|
class_map.NumOccupied()));
|
|
{
|
|
UnorderedHashMap<ClassMapTraits>::Iterator it(&class_map);
|
|
while (it.MoveNext()) {
|
|
const intptr_t entry = it.Current();
|
|
new_cls = Class::RawCast(class_map.GetKey(entry));
|
|
cls = Class::RawCast(class_map.GetPayload(entry, 0));
|
|
cls2 ^= reverse_class_map.GetOrNull(new_cls);
|
|
if (!cls2.IsNull()) {
|
|
OS::PrintErr(
|
|
"Classes '%s' and '%s' are distinct classes but both map "
|
|
" to class '%s'\n",
|
|
cls.ToCString(), cls2.ToCString(), new_cls.ToCString());
|
|
UNREACHABLE();
|
|
}
|
|
bool update = reverse_class_map.UpdateOrInsert(cls, new_cls);
|
|
ASSERT(!update);
|
|
}
|
|
}
|
|
class_map.Release();
|
|
reverse_class_map.Release();
|
|
}
|
|
#endif
|
|
|
|
void ProgramReloadContext::CommitBeforeInstanceMorphing() {
|
|
TIMELINE_SCOPE(Commit);
|
|
|
|
#ifdef DEBUG
|
|
VerifyMaps();
|
|
#endif
|
|
|
|
// Copy over certain properties of libraries, e.g. is the library
|
|
// debuggable?
|
|
{
|
|
TIMELINE_SCOPE(CopyLibraryBits);
|
|
Library& lib = Library::Handle();
|
|
Library& new_lib = Library::Handle();
|
|
|
|
UnorderedHashMap<LibraryMapTraits> lib_map(library_map_storage_);
|
|
|
|
{
|
|
// Reload existing libraries.
|
|
UnorderedHashMap<LibraryMapTraits>::Iterator it(&lib_map);
|
|
|
|
while (it.MoveNext()) {
|
|
const intptr_t entry = it.Current();
|
|
ASSERT(entry != -1);
|
|
new_lib = Library::RawCast(lib_map.GetKey(entry));
|
|
lib = Library::RawCast(lib_map.GetPayload(entry, 0));
|
|
new_lib.set_debuggable(lib.IsDebuggable());
|
|
// Native extension support.
|
|
new_lib.set_native_entry_resolver(lib.native_entry_resolver());
|
|
new_lib.set_native_entry_symbol_resolver(
|
|
lib.native_entry_symbol_resolver());
|
|
}
|
|
}
|
|
|
|
// Release the library map.
|
|
lib_map.Release();
|
|
}
|
|
|
|
{
|
|
TIMELINE_SCOPE(CopyStaticFieldsAndPatchFieldsAndFunctions);
|
|
// Copy static field values from the old classes to the new classes.
|
|
// Patch fields and functions in the old classes so that they retain
|
|
// the old script.
|
|
Class& old_cls = Class::Handle();
|
|
Class& new_cls = Class::Handle();
|
|
UnorderedHashMap<ClassMapTraits> class_map(class_map_storage_);
|
|
|
|
{
|
|
UnorderedHashMap<ClassMapTraits>::Iterator it(&class_map);
|
|
while (it.MoveNext()) {
|
|
const intptr_t entry = it.Current();
|
|
new_cls = Class::RawCast(class_map.GetKey(entry));
|
|
old_cls = Class::RawCast(class_map.GetPayload(entry, 0));
|
|
if (new_cls.ptr() != old_cls.ptr()) {
|
|
ASSERT(new_cls.is_enum_class() == old_cls.is_enum_class());
|
|
if (new_cls.is_enum_class() && new_cls.is_finalized()) {
|
|
new_cls.ReplaceEnum(this, old_cls);
|
|
}
|
|
new_cls.CopyStaticFieldValues(this, old_cls);
|
|
old_cls.PatchFieldsAndFunctions();
|
|
old_cls.MigrateImplicitStaticClosures(this, new_cls);
|
|
}
|
|
}
|
|
}
|
|
|
|
class_map.Release();
|
|
|
|
{
|
|
UnorderedHashSet<ClassMapTraits> removed_class_set(
|
|
removed_class_set_storage_);
|
|
UnorderedHashSet<ClassMapTraits>::Iterator it(&removed_class_set);
|
|
while (it.MoveNext()) {
|
|
const intptr_t entry = it.Current();
|
|
old_cls ^= removed_class_set.GetKey(entry);
|
|
old_cls.PatchFieldsAndFunctions();
|
|
}
|
|
removed_class_set.Release();
|
|
}
|
|
}
|
|
|
|
{
|
|
TIMELINE_SCOPE(UpdateLibrariesArray);
|
|
// Update the libraries array.
|
|
Library& lib = Library::Handle();
|
|
const GrowableObjectArray& libs =
|
|
GrowableObjectArray::Handle(IG->object_store()->libraries());
|
|
for (intptr_t i = 0; i < libs.Length(); i++) {
|
|
lib = Library::RawCast(libs.At(i));
|
|
VTIR_Print("Lib '%s' at index %" Pd "\n", lib.ToCString(), i);
|
|
lib.set_index(i);
|
|
}
|
|
|
|
// Initialize library side table.
|
|
library_infos_.SetLength(libs.Length());
|
|
for (intptr_t i = 0; i < libs.Length(); i++) {
|
|
lib = Library::RawCast(libs.At(i));
|
|
// Mark the library dirty if it comes after the libraries we saved.
|
|
library_infos_[i].dirty =
|
|
i >= group_reload_context_->num_saved_libs_ ||
|
|
group_reload_context_->saved_libs_transitive_updated_->Contains(
|
|
lib.index());
|
|
}
|
|
}
|
|
}
|
|
|
|
void ProgramReloadContext::CommitAfterInstanceMorphing() {
|
|
become_.Forward();
|
|
|
|
// Rehash constants map for all classes. Constants are hashed by content, and
|
|
// content may have changed from fields being added or removed.
|
|
{
|
|
TIMELINE_SCOPE(RehashConstants);
|
|
IG->RehashConstants();
|
|
}
|
|
|
|
#ifdef DEBUG
|
|
IG->ValidateConstants();
|
|
#endif
|
|
|
|
if (FLAG_identity_reload) {
|
|
const auto& saved_libs = GrowableObjectArray::Handle(saved_libraries_);
|
|
const GrowableObjectArray& libs =
|
|
GrowableObjectArray::Handle(IG->object_store()->libraries());
|
|
if (saved_libs.Length() != libs.Length()) {
|
|
TIR_Print("Identity reload failed! B#L=%" Pd " A#L=%" Pd "\n",
|
|
saved_libs.Length(), libs.Length());
|
|
}
|
|
}
|
|
}
|
|
|
|
bool ProgramReloadContext::IsDirty(const Library& lib) {
|
|
const intptr_t index = lib.index();
|
|
if (index == static_cast<classid_t>(-1)) {
|
|
// Treat deleted libraries as dirty.
|
|
return true;
|
|
}
|
|
ASSERT((index >= 0) && (index < library_infos_.length()));
|
|
return library_infos_[index].dirty;
|
|
}
|
|
|
|
void ProgramReloadContext::PostCommit() {
|
|
TIMELINE_SCOPE(PostCommit);
|
|
saved_root_library_ = Library::null();
|
|
saved_libraries_ = GrowableObjectArray::null();
|
|
InvalidateWorld();
|
|
}
|
|
|
|
void IsolateGroupReloadContext::AddReasonForCancelling(
|
|
ReasonForCancelling* reason) {
|
|
reasons_to_cancel_reload_.Add(reason);
|
|
}
|
|
|
|
void IsolateGroupReloadContext::EnsureHasInstanceMorpherFor(
|
|
classid_t cid,
|
|
InstanceMorpher* instance_morpher) {
|
|
for (intptr_t i = 0; i < instance_morphers_.length(); ++i) {
|
|
if (instance_morphers_[i]->cid() == cid) {
|
|
return;
|
|
}
|
|
}
|
|
instance_morphers_.Add(instance_morpher);
|
|
instance_morpher_by_cid_.Insert(instance_morpher);
|
|
ASSERT(instance_morphers_[instance_morphers_.length() - 1]->cid() == cid);
|
|
}
|
|
|
|
void IsolateGroupReloadContext::ReportReasonsForCancelling() {
|
|
ASSERT(FLAG_reload_force_rollback || HasReasonsForCancelling());
|
|
for (int i = 0; i < reasons_to_cancel_reload_.length(); i++) {
|
|
reasons_to_cancel_reload_.At(i)->Report(this);
|
|
}
|
|
}
|
|
|
|
void IsolateGroupReloadContext::MorphInstancesPhase1Allocate(
|
|
ObjectLocator* locator,
|
|
Become* become) {
|
|
ASSERT(HasInstanceMorphers());
|
|
|
|
if (FLAG_trace_reload) {
|
|
LogBlock blocker;
|
|
TIR_Print("MorphInstance: \n");
|
|
for (intptr_t i = 0; i < instance_morphers_.length(); i++) {
|
|
instance_morphers_.At(i)->Dump();
|
|
}
|
|
}
|
|
|
|
const intptr_t count = locator->count();
|
|
TIR_Print("Found %" Pd " object%s subject to morphing.\n", count,
|
|
(count > 1) ? "s" : "");
|
|
|
|
for (intptr_t i = 0; i < instance_morphers_.length(); i++) {
|
|
instance_morphers_.At(i)->CreateMorphedCopies(become);
|
|
}
|
|
}
|
|
|
|
void IsolateGroupReloadContext::MorphInstancesPhase2Become(Become* become) {
|
|
ASSERT(HasInstanceMorphers());
|
|
|
|
become->Forward();
|
|
// The heap now contains only instances with the new layout.
|
|
// Ordinary GC is safe again.
|
|
}
|
|
|
|
void IsolateGroupReloadContext::ForEachIsolate(
|
|
std::function<void(Isolate*)> callback) {
|
|
isolate_group_->ForEachIsolate(callback);
|
|
}
|
|
|
|
void ProgramReloadContext::ValidateReload() {
|
|
TIMELINE_SCOPE(ValidateReload);
|
|
|
|
TIR_Print("---- VALIDATING RELOAD\n");
|
|
|
|
// Validate libraries.
|
|
{
|
|
ASSERT(library_map_storage_ != Array::null());
|
|
UnorderedHashMap<LibraryMapTraits> map(library_map_storage_);
|
|
UnorderedHashMap<LibraryMapTraits>::Iterator it(&map);
|
|
Library& lib = Library::Handle();
|
|
Library& new_lib = Library::Handle();
|
|
while (it.MoveNext()) {
|
|
const intptr_t entry = it.Current();
|
|
new_lib = Library::RawCast(map.GetKey(entry));
|
|
lib = Library::RawCast(map.GetPayload(entry, 0));
|
|
if (new_lib.ptr() != lib.ptr()) {
|
|
lib.CheckReload(new_lib, this);
|
|
}
|
|
}
|
|
map.Release();
|
|
}
|
|
|
|
// Validate classes.
|
|
{
|
|
ASSERT(class_map_storage_ != Array::null());
|
|
UnorderedHashMap<ClassMapTraits> map(class_map_storage_);
|
|
UnorderedHashMap<ClassMapTraits>::Iterator it(&map);
|
|
Class& cls = Class::Handle();
|
|
Class& new_cls = Class::Handle();
|
|
while (it.MoveNext()) {
|
|
const intptr_t entry = it.Current();
|
|
new_cls = Class::RawCast(map.GetKey(entry));
|
|
cls = Class::RawCast(map.GetPayload(entry, 0));
|
|
if (new_cls.ptr() != cls.ptr()) {
|
|
cls.CheckReload(new_cls, this);
|
|
}
|
|
}
|
|
map.Release();
|
|
}
|
|
}
|
|
|
|
void IsolateGroupReloadContext::VisitObjectPointers(
|
|
ObjectPointerVisitor* visitor) {
|
|
visitor->VisitPointers(from(), to());
|
|
}
|
|
|
|
void ProgramReloadContext::VisitObjectPointers(ObjectPointerVisitor* visitor) {
|
|
visitor->VisitPointers(from(), to());
|
|
}
|
|
|
|
ObjectStore* ProgramReloadContext::object_store() {
|
|
return IG->object_store();
|
|
}
|
|
|
|
void ProgramReloadContext::ResetUnoptimizedICsOnStack() {
|
|
Thread* thread = Thread::Current();
|
|
StackZone stack_zone(thread);
|
|
Zone* zone = stack_zone.GetZone();
|
|
Code& code = Code::Handle(zone);
|
|
Function& function = Function::Handle(zone);
|
|
CallSiteResetter resetter(zone);
|
|
|
|
IG->ForEachIsolate([&](Isolate* isolate) {
|
|
DartFrameIterator iterator(isolate->mutator_thread(),
|
|
StackFrameIterator::kAllowCrossThreadIteration);
|
|
StackFrame* frame = iterator.NextFrame();
|
|
while (frame != nullptr) {
|
|
code = frame->LookupDartCode();
|
|
if (code.is_optimized() && !code.is_force_optimized()) {
|
|
// If this code is optimized, we need to reset the ICs in the
|
|
// corresponding unoptimized code, which will be executed when the stack
|
|
// unwinds to the optimized code.
|
|
function = code.function();
|
|
code = function.unoptimized_code();
|
|
ASSERT(!code.IsNull());
|
|
resetter.ResetSwitchableCalls(code);
|
|
resetter.ResetCaches(code);
|
|
} else {
|
|
resetter.ResetSwitchableCalls(code);
|
|
resetter.ResetCaches(code);
|
|
}
|
|
frame = iterator.NextFrame();
|
|
}
|
|
});
|
|
}
|
|
|
|
void ProgramReloadContext::ResetMegamorphicCaches() {
|
|
object_store()->set_megamorphic_cache_table(GrowableObjectArray::Handle());
|
|
// Since any current optimized code will not make any more calls, it may be
|
|
// better to clear the table instead of clearing each of the caches, allow
|
|
// the current megamorphic caches get GC'd and any new optimized code allocate
|
|
// new ones.
|
|
}
|
|
|
|
class InvalidationCollector : public ObjectVisitor {
|
|
public:
|
|
InvalidationCollector(Zone* zone,
|
|
GrowableArray<const Function*>* functions,
|
|
GrowableArray<const KernelProgramInfo*>* kernel_infos,
|
|
GrowableArray<const Field*>* fields,
|
|
GrowableArray<const SuspendState*>* suspend_states,
|
|
GrowableArray<const Instance*>* instances)
|
|
: zone_(zone),
|
|
functions_(functions),
|
|
kernel_infos_(kernel_infos),
|
|
fields_(fields),
|
|
suspend_states_(suspend_states),
|
|
instances_(instances) {}
|
|
virtual ~InvalidationCollector() {}
|
|
|
|
void VisitObject(ObjectPtr obj) {
|
|
intptr_t cid = obj->GetClassId();
|
|
if (cid == kFunctionCid) {
|
|
const Function& func =
|
|
Function::Handle(zone_, static_cast<FunctionPtr>(obj));
|
|
if (!func.ForceOptimize()) {
|
|
// Force-optimized functions cannot deoptimize.
|
|
functions_->Add(&func);
|
|
}
|
|
} else if (cid == kKernelProgramInfoCid) {
|
|
kernel_infos_->Add(&KernelProgramInfo::Handle(
|
|
zone_, static_cast<KernelProgramInfoPtr>(obj)));
|
|
} else if (cid == kFieldCid) {
|
|
fields_->Add(&Field::Handle(zone_, static_cast<FieldPtr>(obj)));
|
|
} else if (cid == kSuspendStateCid) {
|
|
const auto& suspend_state =
|
|
SuspendState::Handle(zone_, static_cast<SuspendStatePtr>(obj));
|
|
if (suspend_state.pc() != 0) {
|
|
suspend_states_->Add(&suspend_state);
|
|
}
|
|
} else if (cid > kNumPredefinedCids) {
|
|
instances_->Add(&Instance::Handle(zone_, static_cast<InstancePtr>(obj)));
|
|
}
|
|
}
|
|
|
|
private:
|
|
Zone* const zone_;
|
|
GrowableArray<const Function*>* const functions_;
|
|
GrowableArray<const KernelProgramInfo*>* const kernel_infos_;
|
|
GrowableArray<const Field*>* const fields_;
|
|
GrowableArray<const SuspendState*>* const suspend_states_;
|
|
GrowableArray<const Instance*>* const instances_;
|
|
};
|
|
|
|
typedef UnorderedHashMap<SmiTraits> IntHashMap;
|
|
|
|
void ProgramReloadContext::RunInvalidationVisitors() {
|
|
TIR_Print("---- RUNNING INVALIDATION HEAP VISITORS\n");
|
|
Thread* thread = Thread::Current();
|
|
StackZone stack_zone(thread);
|
|
Zone* zone = stack_zone.GetZone();
|
|
|
|
GrowableArray<const Function*> functions(4 * KB);
|
|
GrowableArray<const KernelProgramInfo*> kernel_infos(KB);
|
|
GrowableArray<const Field*> fields(4 * KB);
|
|
GrowableArray<const SuspendState*> suspend_states(4 * KB);
|
|
GrowableArray<const Instance*> instances(4 * KB);
|
|
|
|
{
|
|
HeapIterationScope iteration(thread);
|
|
InvalidationCollector visitor(zone, &functions, &kernel_infos, &fields,
|
|
&suspend_states, &instances);
|
|
iteration.IterateObjects(&visitor);
|
|
}
|
|
|
|
InvalidateKernelInfos(zone, kernel_infos);
|
|
InvalidateSuspendStates(zone, suspend_states);
|
|
InvalidateFunctions(zone, functions);
|
|
InvalidateFields(zone, fields, instances);
|
|
}
|
|
|
|
void ProgramReloadContext::InvalidateKernelInfos(
|
|
Zone* zone,
|
|
const GrowableArray<const KernelProgramInfo*>& kernel_infos) {
|
|
TIMELINE_SCOPE(InvalidateKernelInfos);
|
|
HANDLESCOPE(Thread::Current());
|
|
|
|
Array& data = Array::Handle(zone);
|
|
Object& key = Object::Handle(zone);
|
|
Smi& value = Smi::Handle(zone);
|
|
for (intptr_t i = 0; i < kernel_infos.length(); i++) {
|
|
const KernelProgramInfo& info = *kernel_infos[i];
|
|
// Clear the libraries cache.
|
|
{
|
|
data = info.libraries_cache();
|
|
ASSERT(!data.IsNull());
|
|
IntHashMap table(&key, &value, &data);
|
|
table.Clear();
|
|
info.set_libraries_cache(table.Release());
|
|
}
|
|
// Clear the classes cache.
|
|
{
|
|
data = info.classes_cache();
|
|
ASSERT(!data.IsNull());
|
|
IntHashMap table(&key, &value, &data);
|
|
table.Clear();
|
|
info.set_classes_cache(table.Release());
|
|
}
|
|
}
|
|
}
|
|
|
|
void ProgramReloadContext::InvalidateFunctions(
|
|
Zone* zone,
|
|
const GrowableArray<const Function*>& functions) {
|
|
TIMELINE_SCOPE(InvalidateFunctions);
|
|
auto thread = Thread::Current();
|
|
HANDLESCOPE(thread);
|
|
|
|
CallSiteResetter resetter(zone);
|
|
|
|
Class& owning_class = Class::Handle(zone);
|
|
Library& owning_lib = Library::Handle(zone);
|
|
Code& code = Code::Handle(zone);
|
|
SafepointWriteRwLocker ml(thread, thread->isolate_group()->program_lock());
|
|
for (intptr_t i = 0; i < functions.length(); i++) {
|
|
const Function& func = *functions[i];
|
|
|
|
// Switch to unoptimized code or the lazy compilation stub.
|
|
func.SwitchToLazyCompiledUnoptimizedCode();
|
|
|
|
// Grab the current code.
|
|
code = func.CurrentCode();
|
|
ASSERT(!code.IsNull());
|
|
|
|
owning_class = func.Owner();
|
|
owning_lib = owning_class.library();
|
|
const bool clear_code = IsDirty(owning_lib);
|
|
const bool stub_code = code.IsStubCode();
|
|
|
|
// Zero edge counters, before clearing the ICDataArray, since that's where
|
|
// they're held.
|
|
resetter.ZeroEdgeCounters(func);
|
|
|
|
if (stub_code) {
|
|
// Nothing to reset.
|
|
} else if (clear_code) {
|
|
VTIR_Print("Marking %s for recompilation, clearing code\n",
|
|
func.ToCString());
|
|
// Null out the ICData array and code.
|
|
func.ClearICDataArray();
|
|
func.ClearCode();
|
|
func.SetWasCompiled(false);
|
|
} else {
|
|
// We are preserving the unoptimized code, reset instance calls and type
|
|
// test caches.
|
|
resetter.ResetSwitchableCalls(code);
|
|
resetter.ResetCaches(code);
|
|
}
|
|
|
|
// Clear counters.
|
|
func.set_usage_counter(0);
|
|
func.set_deoptimization_counter(0);
|
|
func.set_optimized_instruction_count(0);
|
|
func.set_optimized_call_site_count(0);
|
|
}
|
|
}
|
|
|
|
void ProgramReloadContext::InvalidateSuspendStates(
|
|
Zone* zone,
|
|
const GrowableArray<const SuspendState*>& suspend_states) {
|
|
TIMELINE_SCOPE(InvalidateSuspendStates);
|
|
auto thread = Thread::Current();
|
|
HANDLESCOPE(thread);
|
|
|
|
CallSiteResetter resetter(zone);
|
|
Code& code = Code::Handle(zone);
|
|
Function& function = Function::Handle(zone);
|
|
|
|
SafepointWriteRwLocker ml(thread, thread->isolate_group()->program_lock());
|
|
for (intptr_t i = 0, n = suspend_states.length(); i < n; ++i) {
|
|
const SuspendState& suspend_state = *suspend_states[i];
|
|
ASSERT(suspend_state.pc() != 0);
|
|
code = suspend_state.GetCodeObject();
|
|
ASSERT(!code.IsNull());
|
|
if (code.is_optimized() && !code.is_force_optimized()) {
|
|
function = code.function();
|
|
// Before disabling [code], function needs to
|
|
// switch to unoptimized code first.
|
|
function.SwitchToLazyCompiledUnoptimizedCode();
|
|
// Disable [code] in order to trigger lazy deoptimization.
|
|
// Unless [code] is compiled for OSR, it may be already
|
|
// disabled in SwitchToLazyCompiledUnoptimizedCode.
|
|
if (!code.IsDisabled()) {
|
|
code.DisableDartCode();
|
|
}
|
|
// Reset switchable calls and caches for unoptimized
|
|
// code (if any), as it is going to be used to continue
|
|
// execution of the suspended function.
|
|
code = function.unoptimized_code();
|
|
if (!code.IsNull()) {
|
|
resetter.ResetSwitchableCalls(code);
|
|
resetter.ResetCaches(code);
|
|
}
|
|
} else {
|
|
function = code.function();
|
|
// ResetSwitchableCalls uses ICData array, which
|
|
// can be cleared along with the code in InvalidateFunctions
|
|
// during previous hot reloads.
|
|
// Rebuild an unoptimized code in order to recreate ICData array.
|
|
function.EnsureHasCompiledUnoptimizedCode();
|
|
resetter.ResetSwitchableCalls(code);
|
|
resetter.ResetCaches(code);
|
|
}
|
|
}
|
|
}
|
|
|
|
// Finds fields that are initialized or have a value that does not conform to
|
|
// the field's static type, setting Field::needs_load_guard(). Accessors for
|
|
// such fields are compiled with additional checks to handle lazy initialization
|
|
// and to preserve type soundness.
|
|
class FieldInvalidator {
|
|
public:
|
|
explicit FieldInvalidator(Zone* zone)
|
|
: zone_(zone),
|
|
cls_(Class::Handle(zone)),
|
|
cls_fields_(Array::Handle(zone)),
|
|
entry_(Object::Handle(zone)),
|
|
value_(Object::Handle(zone)),
|
|
instance_(Instance::Handle(zone)),
|
|
type_(AbstractType::Handle(zone)),
|
|
cache_(SubtypeTestCache::Handle(zone)),
|
|
entries_(Array::Handle(zone)),
|
|
closure_function_(Function::Handle(zone)),
|
|
instantiator_type_arguments_(TypeArguments::Handle(zone)),
|
|
function_type_arguments_(TypeArguments::Handle(zone)),
|
|
instance_cid_or_signature_(Object::Handle(zone)),
|
|
instance_type_arguments_(TypeArguments::Handle(zone)),
|
|
parent_function_type_arguments_(TypeArguments::Handle(zone)),
|
|
delayed_function_type_arguments_(TypeArguments::Handle(zone)) {}
|
|
|
|
void CheckStatics(const GrowableArray<const Field*>& fields) {
|
|
Thread* thread = Thread::Current();
|
|
const bool null_safety = thread->isolate_group()->null_safety();
|
|
HANDLESCOPE(thread);
|
|
instantiator_type_arguments_ = TypeArguments::null();
|
|
for (intptr_t i = 0; i < fields.length(); i++) {
|
|
const Field& field = *fields[i];
|
|
if (!field.is_static()) {
|
|
continue;
|
|
}
|
|
if (field.needs_load_guard()) {
|
|
continue; // Already guarding.
|
|
}
|
|
const intptr_t field_id = field.field_id();
|
|
thread->isolate_group()->ForEachIsolate([&](Isolate* isolate) {
|
|
auto field_table = isolate->field_table();
|
|
// The isolate might've just been created and is now participating in
|
|
// the reload request inside `IsolateGroup::RegisterIsolate()`.
|
|
// At that point it doesn't have the field table setup yet.
|
|
if (field_table->IsReadyToUse()) {
|
|
value_ = field_table->At(field_id);
|
|
if ((value_.ptr() != Object::sentinel().ptr()) &&
|
|
(value_.ptr() != Object::transition_sentinel().ptr())) {
|
|
CheckValueType(null_safety, value_, field);
|
|
}
|
|
}
|
|
});
|
|
}
|
|
}
|
|
|
|
void CheckInstances(const GrowableArray<const Instance*>& instances) {
|
|
Thread* thread = Thread::Current();
|
|
const bool null_safety = thread->isolate_group()->null_safety();
|
|
HANDLESCOPE(thread);
|
|
for (intptr_t i = 0; i < instances.length(); i++) {
|
|
CheckInstance(null_safety, *instances[i]);
|
|
}
|
|
}
|
|
|
|
private:
|
|
DART_FORCE_INLINE
|
|
void CheckInstance(bool null_safety, const Instance& instance) {
|
|
cls_ = instance.clazz();
|
|
if (cls_.NumTypeArguments() > 0) {
|
|
instantiator_type_arguments_ = instance.GetTypeArguments();
|
|
} else {
|
|
instantiator_type_arguments_ = TypeArguments::null();
|
|
}
|
|
cls_fields_ = cls_.OffsetToFieldMap();
|
|
for (intptr_t i = 0; i < cls_fields_.Length(); i++) {
|
|
entry_ = cls_fields_.At(i);
|
|
if (!entry_.IsField()) {
|
|
continue;
|
|
}
|
|
const Field& field = Field::Cast(entry_);
|
|
CheckInstanceField(null_safety, instance, field);
|
|
}
|
|
}
|
|
|
|
DART_FORCE_INLINE
|
|
void CheckInstanceField(bool null_safety,
|
|
const Instance& instance,
|
|
const Field& field) {
|
|
if (field.needs_load_guard()) {
|
|
return; // Already guarding.
|
|
}
|
|
if (field.is_unboxed()) {
|
|
// Unboxed fields are guaranteed to match.
|
|
return;
|
|
}
|
|
value_ = instance.GetField(field);
|
|
if (value_.ptr() == Object::sentinel().ptr()) {
|
|
if (field.is_late()) {
|
|
// Late fields already have lazy initialization logic.
|
|
return;
|
|
}
|
|
// Needs guard for initialization.
|
|
ASSERT(!FLAG_identity_reload);
|
|
field.set_needs_load_guard(true);
|
|
return;
|
|
}
|
|
CheckValueType(null_safety, value_, field);
|
|
}
|
|
|
|
DART_FORCE_INLINE
|
|
bool CheckAssignabilityUsingCache(bool null_safety,
|
|
const Object& value,
|
|
const AbstractType& type,
|
|
const Field& field) {
|
|
ASSERT(!value.IsSentinel());
|
|
if (!null_safety && value.IsNull()) {
|
|
return true;
|
|
}
|
|
|
|
if (type.IsDynamicType()) {
|
|
return true;
|
|
}
|
|
|
|
if (type.IsRecordType()) {
|
|
return CheckAssignabilityForRecordType(null_safety, value,
|
|
RecordType::Cast(type), field);
|
|
}
|
|
|
|
cls_ = value.clazz();
|
|
const intptr_t cid = cls_.id();
|
|
if (cid == kClosureCid) {
|
|
const auto& closure = Closure::Cast(value);
|
|
closure_function_ = closure.function();
|
|
instance_cid_or_signature_ = closure_function_.signature();
|
|
instance_type_arguments_ = closure.instantiator_type_arguments();
|
|
parent_function_type_arguments_ = closure.function_type_arguments();
|
|
delayed_function_type_arguments_ = closure.delayed_type_arguments();
|
|
} else {
|
|
instance_cid_or_signature_ = Smi::New(cid);
|
|
if (cls_.NumTypeArguments() > 0) {
|
|
instance_type_arguments_ = Instance::Cast(value).GetTypeArguments();
|
|
} else {
|
|
instance_type_arguments_ = TypeArguments::null();
|
|
}
|
|
parent_function_type_arguments_ = TypeArguments::null();
|
|
delayed_function_type_arguments_ = TypeArguments::null();
|
|
}
|
|
|
|
cache_ = field.type_test_cache();
|
|
if (cache_.IsNull()) {
|
|
cache_ = SubtypeTestCache::New();
|
|
field.set_type_test_cache(cache_);
|
|
}
|
|
entries_ = cache_.cache();
|
|
|
|
for (intptr_t i = 0; entries_.At(i) != Object::null();
|
|
i += SubtypeTestCache::kTestEntryLength) {
|
|
if ((entries_.At(i + SubtypeTestCache::kInstanceCidOrSignature) ==
|
|
instance_cid_or_signature_.ptr()) &&
|
|
(entries_.At(i + SubtypeTestCache::kDestinationType) == type.ptr()) &&
|
|
(entries_.At(i + SubtypeTestCache::kInstanceTypeArguments) ==
|
|
instance_type_arguments_.ptr()) &&
|
|
(entries_.At(i + SubtypeTestCache::kInstantiatorTypeArguments) ==
|
|
instantiator_type_arguments_.ptr()) &&
|
|
(entries_.At(i + SubtypeTestCache::kFunctionTypeArguments) ==
|
|
function_type_arguments_.ptr()) &&
|
|
(entries_.At(
|
|
i + SubtypeTestCache::kInstanceParentFunctionTypeArguments) ==
|
|
parent_function_type_arguments_.ptr()) &&
|
|
(entries_.At(
|
|
i + SubtypeTestCache::kInstanceDelayedFunctionTypeArguments) ==
|
|
delayed_function_type_arguments_.ptr())) {
|
|
return entries_.At(i + SubtypeTestCache::kTestResult) ==
|
|
Bool::True().ptr();
|
|
}
|
|
}
|
|
|
|
instance_ ^= value.ptr();
|
|
if (instance_.IsAssignableTo(type, instantiator_type_arguments_,
|
|
function_type_arguments_)) {
|
|
// Do not add record instances to cache as they don't have a valid
|
|
// key (type of a record depends on types of all its fields).
|
|
if (cid != kRecordCid) {
|
|
cache_.AddCheck(instance_cid_or_signature_, type,
|
|
instance_type_arguments_, instantiator_type_arguments_,
|
|
function_type_arguments_,
|
|
parent_function_type_arguments_,
|
|
delayed_function_type_arguments_, Bool::True());
|
|
}
|
|
return true;
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
bool CheckAssignabilityForRecordType(bool null_safety,
|
|
const Object& value,
|
|
const RecordType& type,
|
|
const Field& field) {
|
|
if (!value.IsRecord()) {
|
|
return false;
|
|
}
|
|
|
|
const Record& record = Record::Cast(value);
|
|
const intptr_t num_fields = record.num_fields();
|
|
if (num_fields != type.NumFields() ||
|
|
record.field_names() != type.field_names()) {
|
|
return false;
|
|
}
|
|
|
|
// This method can be called recursively, so cannot reuse handles.
|
|
auto& field_value = Object::Handle(zone_);
|
|
auto& field_type = AbstractType::Handle(zone_);
|
|
for (intptr_t i = 0; i < num_fields; ++i) {
|
|
field_value = record.FieldAt(i);
|
|
field_type = type.FieldTypeAt(i);
|
|
if (!CheckAssignabilityUsingCache(null_safety, field_value, field_type,
|
|
field)) {
|
|
return false;
|
|
}
|
|
}
|
|
return true;
|
|
}
|
|
|
|
DART_FORCE_INLINE
|
|
void CheckValueType(bool null_safety,
|
|
const Object& value,
|
|
const Field& field) {
|
|
ASSERT(!value.IsSentinel());
|
|
if (!null_safety && value.IsNull()) {
|
|
return;
|
|
}
|
|
type_ = field.type();
|
|
if (!CheckAssignabilityUsingCache(null_safety, value, type_, field)) {
|
|
// Even if doing an identity reload, type check can fail if hot reload
|
|
// happens while constructor is still running and field is not
|
|
// initialized yet, so it has a null value.
|
|
ASSERT(!FLAG_identity_reload || value.IsNull());
|
|
field.set_needs_load_guard(true);
|
|
}
|
|
}
|
|
|
|
Zone* zone_;
|
|
Class& cls_;
|
|
Array& cls_fields_;
|
|
Object& entry_;
|
|
Object& value_;
|
|
Instance& instance_;
|
|
AbstractType& type_;
|
|
SubtypeTestCache& cache_;
|
|
Array& entries_;
|
|
Function& closure_function_;
|
|
TypeArguments& instantiator_type_arguments_;
|
|
TypeArguments& function_type_arguments_;
|
|
Object& instance_cid_or_signature_;
|
|
TypeArguments& instance_type_arguments_;
|
|
TypeArguments& parent_function_type_arguments_;
|
|
TypeArguments& delayed_function_type_arguments_;
|
|
};
|
|
|
|
void ProgramReloadContext::InvalidateFields(
|
|
Zone* zone,
|
|
const GrowableArray<const Field*>& fields,
|
|
const GrowableArray<const Instance*>& instances) {
|
|
TIMELINE_SCOPE(InvalidateFields);
|
|
SafepointMutexLocker ml(IG->subtype_test_cache_mutex());
|
|
FieldInvalidator invalidator(zone);
|
|
invalidator.CheckStatics(fields);
|
|
invalidator.CheckInstances(instances);
|
|
}
|
|
|
|
void ProgramReloadContext::InvalidateWorld() {
|
|
TIMELINE_SCOPE(InvalidateWorld);
|
|
TIR_Print("---- INVALIDATING WORLD\n");
|
|
ResetMegamorphicCaches();
|
|
if (FLAG_trace_deoptimization) {
|
|
THR_Print("Deopt for reload\n");
|
|
}
|
|
DeoptimizeFunctionsOnStack();
|
|
ResetUnoptimizedICsOnStack();
|
|
RunInvalidationVisitors();
|
|
}
|
|
|
|
ClassPtr ProgramReloadContext::OldClassOrNull(const Class& replacement_or_new) {
|
|
UnorderedHashSet<ClassMapTraits> old_classes_set(old_classes_set_storage_);
|
|
Class& cls = Class::Handle();
|
|
cls ^= old_classes_set.GetOrNull(replacement_or_new);
|
|
old_classes_set_storage_ = old_classes_set.Release().ptr();
|
|
return cls.ptr();
|
|
}
|
|
|
|
StringPtr ProgramReloadContext::FindLibraryPrivateKey(
|
|
const Library& replacement_or_new) {
|
|
const Library& old = Library::Handle(OldLibraryOrNull(replacement_or_new));
|
|
if (old.IsNull()) {
|
|
return String::null();
|
|
}
|
|
#if defined(DEBUG)
|
|
VTIR_Print("`%s` is getting `%s`'s private key.\n",
|
|
String::Handle(replacement_or_new.url()).ToCString(),
|
|
String::Handle(old.url()).ToCString());
|
|
#endif
|
|
return old.private_key();
|
|
}
|
|
|
|
LibraryPtr ProgramReloadContext::OldLibraryOrNull(
|
|
const Library& replacement_or_new) {
|
|
UnorderedHashSet<LibraryMapTraits> old_libraries_set(
|
|
old_libraries_set_storage_);
|
|
Library& lib = Library::Handle();
|
|
lib ^= old_libraries_set.GetOrNull(replacement_or_new);
|
|
old_libraries_set.Release();
|
|
|
|
if (lib.IsNull() &&
|
|
(group_reload_context_->root_url_prefix_ != String::null()) &&
|
|
(group_reload_context_->old_root_url_prefix_ != String::null())) {
|
|
return OldLibraryOrNullBaseMoved(replacement_or_new);
|
|
}
|
|
return lib.ptr();
|
|
}
|
|
|
|
// Attempt to find the pair to |replacement_or_new| with the knowledge that
|
|
// the base url prefix has moved.
|
|
LibraryPtr ProgramReloadContext::OldLibraryOrNullBaseMoved(
|
|
const Library& replacement_or_new) {
|
|
const String& url_prefix =
|
|
String::Handle(group_reload_context_->root_url_prefix_);
|
|
const String& old_url_prefix =
|
|
String::Handle(group_reload_context_->old_root_url_prefix_);
|
|
const intptr_t prefix_length = url_prefix.Length();
|
|
const intptr_t old_prefix_length = old_url_prefix.Length();
|
|
const String& new_url = String::Handle(replacement_or_new.url());
|
|
const String& suffix =
|
|
String::Handle(String::SubString(new_url, prefix_length));
|
|
if (!new_url.StartsWith(url_prefix)) {
|
|
return Library::null();
|
|
}
|
|
Library& old = Library::Handle();
|
|
String& old_url = String::Handle();
|
|
String& old_suffix = String::Handle();
|
|
const auto& saved_libs = GrowableObjectArray::Handle(saved_libraries_);
|
|
ASSERT(!saved_libs.IsNull());
|
|
for (intptr_t i = 0; i < saved_libs.Length(); i++) {
|
|
old = Library::RawCast(saved_libs.At(i));
|
|
old_url = old.url();
|
|
if (!old_url.StartsWith(old_url_prefix)) {
|
|
continue;
|
|
}
|
|
old_suffix = String::SubString(old_url, old_prefix_length);
|
|
if (old_suffix.IsNull()) {
|
|
continue;
|
|
}
|
|
if (old_suffix.Equals(suffix)) {
|
|
TIR_Print("`%s` is moving to `%s`\n", old_url.ToCString(),
|
|
new_url.ToCString());
|
|
return old.ptr();
|
|
}
|
|
}
|
|
return Library::null();
|
|
}
|
|
|
|
void ProgramReloadContext::BuildLibraryMapping() {
|
|
const GrowableObjectArray& libs =
|
|
GrowableObjectArray::Handle(object_store()->libraries());
|
|
|
|
Library& replacement_or_new = Library::Handle();
|
|
Library& old = Library::Handle();
|
|
for (intptr_t i = group_reload_context_->num_saved_libs_; i < libs.Length();
|
|
i++) {
|
|
replacement_or_new = Library::RawCast(libs.At(i));
|
|
old = OldLibraryOrNull(replacement_or_new);
|
|
if (old.IsNull()) {
|
|
if (FLAG_identity_reload) {
|
|
TIR_Print("Could not find original library for %s\n",
|
|
replacement_or_new.ToCString());
|
|
UNREACHABLE();
|
|
}
|
|
// New library.
|
|
AddLibraryMapping(replacement_or_new, replacement_or_new);
|
|
} else {
|
|
ASSERT(!replacement_or_new.is_dart_scheme());
|
|
// Replaced class.
|
|
AddLibraryMapping(replacement_or_new, old);
|
|
|
|
AddBecomeMapping(old, replacement_or_new);
|
|
}
|
|
}
|
|
}
|
|
|
|
// Find classes that have been removed from the program.
|
|
// Instances of these classes may still be referenced from variables, so the
|
|
// functions of these class may still execute in the future, and they need to
|
|
// be given patch class owners still they correctly reference their (old) kernel
|
|
// data even after the library's kernel data is updated.
|
|
//
|
|
// Note that all such classes must belong to a library that has either been
|
|
// changed or removed.
|
|
void ProgramReloadContext::BuildRemovedClassesSet() {
|
|
// Find all old classes [mapped_old_classes_set].
|
|
UnorderedHashMap<ClassMapTraits> class_map(class_map_storage_);
|
|
UnorderedHashSet<ClassMapTraits> mapped_old_classes_set(
|
|
HashTables::New<UnorderedHashSet<ClassMapTraits> >(
|
|
class_map.NumOccupied()));
|
|
{
|
|
UnorderedHashMap<ClassMapTraits>::Iterator it(&class_map);
|
|
Class& cls = Class::Handle();
|
|
Class& new_cls = Class::Handle();
|
|
while (it.MoveNext()) {
|
|
const intptr_t entry = it.Current();
|
|
new_cls = Class::RawCast(class_map.GetKey(entry));
|
|
cls = Class::RawCast(class_map.GetPayload(entry, 0));
|
|
mapped_old_classes_set.InsertOrGet(cls);
|
|
}
|
|
}
|
|
class_map.Release();
|
|
|
|
// Find all reloaded libraries [mapped_old_library_set].
|
|
UnorderedHashMap<LibraryMapTraits> library_map(library_map_storage_);
|
|
UnorderedHashMap<LibraryMapTraits>::Iterator it_library(&library_map);
|
|
UnorderedHashSet<LibraryMapTraits> mapped_old_library_set(
|
|
HashTables::New<UnorderedHashSet<LibraryMapTraits> >(
|
|
library_map.NumOccupied()));
|
|
{
|
|
Library& old_library = Library::Handle();
|
|
Library& new_library = Library::Handle();
|
|
while (it_library.MoveNext()) {
|
|
const intptr_t entry = it_library.Current();
|
|
new_library ^= library_map.GetKey(entry);
|
|
old_library ^= library_map.GetPayload(entry, 0);
|
|
if (new_library.ptr() != old_library.ptr()) {
|
|
mapped_old_library_set.InsertOrGet(old_library);
|
|
}
|
|
}
|
|
}
|
|
|
|
// For every old class, check if it's library was reloaded and if
|
|
// the class was mapped. If the class wasn't mapped - add it to
|
|
// [removed_class_set].
|
|
UnorderedHashSet<ClassMapTraits> old_classes_set(old_classes_set_storage_);
|
|
UnorderedHashSet<ClassMapTraits>::Iterator it(&old_classes_set);
|
|
UnorderedHashSet<ClassMapTraits> removed_class_set(
|
|
removed_class_set_storage_);
|
|
Class& old_cls = Class::Handle();
|
|
Class& new_cls = Class::Handle();
|
|
Library& old_library = Library::Handle();
|
|
Library& mapped_old_library = Library::Handle();
|
|
while (it.MoveNext()) {
|
|
const intptr_t entry = it.Current();
|
|
old_cls ^= Class::RawCast(old_classes_set.GetKey(entry));
|
|
old_library = old_cls.library();
|
|
if (old_library.IsNull()) {
|
|
continue;
|
|
}
|
|
mapped_old_library ^= mapped_old_library_set.GetOrNull(old_library);
|
|
if (!mapped_old_library.IsNull()) {
|
|
new_cls ^= mapped_old_classes_set.GetOrNull(old_cls);
|
|
if (new_cls.IsNull()) {
|
|
removed_class_set.InsertOrGet(old_cls);
|
|
}
|
|
}
|
|
}
|
|
removed_class_set_storage_ = removed_class_set.Release().ptr();
|
|
|
|
old_classes_set.Release();
|
|
mapped_old_classes_set.Release();
|
|
mapped_old_library_set.Release();
|
|
library_map.Release();
|
|
}
|
|
|
|
void ProgramReloadContext::AddClassMapping(const Class& replacement_or_new,
|
|
const Class& original) {
|
|
UnorderedHashMap<ClassMapTraits> map(class_map_storage_);
|
|
bool update = map.UpdateOrInsert(replacement_or_new, original);
|
|
ASSERT(!update);
|
|
// The storage given to the map may have been reallocated, remember the new
|
|
// address.
|
|
class_map_storage_ = map.Release().ptr();
|
|
}
|
|
|
|
void ProgramReloadContext::AddLibraryMapping(const Library& replacement_or_new,
|
|
const Library& original) {
|
|
UnorderedHashMap<LibraryMapTraits> map(library_map_storage_);
|
|
bool update = map.UpdateOrInsert(replacement_or_new, original);
|
|
ASSERT(!update);
|
|
// The storage given to the map may have been reallocated, remember the new
|
|
// address.
|
|
library_map_storage_ = map.Release().ptr();
|
|
}
|
|
|
|
void ProgramReloadContext::AddStaticFieldMapping(const Field& old_field,
|
|
const Field& new_field) {
|
|
ASSERT(old_field.is_static());
|
|
ASSERT(new_field.is_static());
|
|
AddBecomeMapping(old_field, new_field);
|
|
}
|
|
|
|
void ProgramReloadContext::AddBecomeMapping(const Object& old,
|
|
const Object& neu) {
|
|
become_.Add(old, neu);
|
|
}
|
|
|
|
void ProgramReloadContext::RebuildDirectSubclasses() {
|
|
ClassTable* class_table = IG->class_table();
|
|
intptr_t num_cids = class_table->NumCids();
|
|
|
|
// Clear the direct subclasses for all classes.
|
|
Class& cls = Class::Handle();
|
|
const GrowableObjectArray& null_list = GrowableObjectArray::Handle();
|
|
for (intptr_t i = 1; i < num_cids; i++) {
|
|
if (class_table->HasValidClassAt(i)) {
|
|
cls = class_table->At(i);
|
|
if (!cls.is_declaration_loaded()) {
|
|
continue; // Can't have any subclasses or implementors yet.
|
|
}
|
|
// Testing for null to prevent attempting to write to read-only classes
|
|
// in the VM isolate.
|
|
if (cls.direct_subclasses() != GrowableObjectArray::null()) {
|
|
cls.set_direct_subclasses(null_list);
|
|
}
|
|
if (cls.direct_implementors() != GrowableObjectArray::null()) {
|
|
cls.set_direct_implementors(null_list);
|
|
}
|
|
}
|
|
}
|
|
|
|
// Recompute the direct subclasses / implementors.
|
|
|
|
AbstractType& super_type = AbstractType::Handle();
|
|
Class& super_cls = Class::Handle();
|
|
|
|
Array& interface_types = Array::Handle();
|
|
AbstractType& interface_type = AbstractType::Handle();
|
|
Class& interface_class = Class::Handle();
|
|
|
|
for (intptr_t i = 1; i < num_cids; i++) {
|
|
if (class_table->HasValidClassAt(i)) {
|
|
cls = class_table->At(i);
|
|
if (!cls.is_declaration_loaded()) {
|
|
continue; // Will register itself later when loaded.
|
|
}
|
|
super_type = cls.super_type();
|
|
if (!super_type.IsNull() && !super_type.IsObjectType()) {
|
|
super_cls = cls.SuperClass();
|
|
ASSERT(!super_cls.IsNull());
|
|
super_cls.AddDirectSubclass(cls);
|
|
}
|
|
|
|
interface_types = cls.interfaces();
|
|
if (!interface_types.IsNull()) {
|
|
const intptr_t mixin_index = cls.is_transformed_mixin_application()
|
|
? interface_types.Length() - 1
|
|
: -1;
|
|
for (intptr_t j = 0; j < interface_types.Length(); ++j) {
|
|
interface_type ^= interface_types.At(j);
|
|
interface_class = interface_type.type_class();
|
|
interface_class.AddDirectImplementor(
|
|
cls, /* is_mixin = */ i == mixin_index);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
void ReloadHandler::RegisterIsolate() {
|
|
SafepointMonitorLocker ml(&monitor_);
|
|
ParticipateIfReloadRequested(&ml, /*is_registered=*/false,
|
|
/*allow_later_retry=*/false);
|
|
ASSERT(reloading_thread_ == nullptr);
|
|
++registered_isolate_count_;
|
|
}
|
|
|
|
void ReloadHandler::UnregisterIsolate() {
|
|
SafepointMonitorLocker ml(&monitor_);
|
|
ParticipateIfReloadRequested(&ml, /*is_registered=*/true,
|
|
/*allow_later_retry=*/false);
|
|
ASSERT(reloading_thread_ == nullptr);
|
|
--registered_isolate_count_;
|
|
}
|
|
|
|
void ReloadHandler::CheckForReload() {
|
|
SafepointMonitorLocker ml(&monitor_);
|
|
ParticipateIfReloadRequested(&ml, /*is_registered=*/true,
|
|
/*allow_later_retry=*/true);
|
|
}
|
|
|
|
void ReloadHandler::ParticipateIfReloadRequested(SafepointMonitorLocker* ml,
|
|
bool is_registered,
|
|
bool allow_later_retry) {
|
|
if (reloading_thread_ != nullptr) {
|
|
auto thread = Thread::Current();
|
|
auto isolate = thread->isolate();
|
|
|
|
// If the current thread is in a no reload scope, we'll not participate here
|
|
// and instead delay to a point (further up the stack, namely in the main
|
|
// message handling loop) where this isolate can participate.
|
|
if (thread->IsInNoReloadScope()) {
|
|
RELEASE_ASSERT(allow_later_retry);
|
|
isolate->SendInternalLibMessage(Isolate::kCheckForReload, /*ignored=*/-1);
|
|
return;
|
|
}
|
|
|
|
if (is_registered) {
|
|
SafepointMonitorLocker ml(&checkin_monitor_);
|
|
++isolates_checked_in_;
|
|
ml.NotifyAll();
|
|
}
|
|
// While we're waiting for the reload to be performed, we'll exit the
|
|
// isolate. That will transition into a safepoint - which a blocking `Wait`
|
|
// would also do - but it does something in addition: It will release it's
|
|
// current TLAB and decrease the mutator count. We want this in order to let
|
|
// all isolates in the group participate in the reload, despite our parallel
|
|
// mutator limit.
|
|
while (reloading_thread_ != nullptr) {
|
|
SafepointMonitorUnlockScope ml_unlocker(ml);
|
|
Thread::ExitIsolate(/*nested=*/true);
|
|
{
|
|
MonitorLocker ml(&monitor_);
|
|
while (reloading_thread_ != nullptr) {
|
|
ml.Wait();
|
|
}
|
|
}
|
|
Thread::EnterIsolate(isolate, /*nested=*/true);
|
|
}
|
|
if (is_registered) {
|
|
SafepointMonitorLocker ml(&checkin_monitor_);
|
|
--isolates_checked_in_;
|
|
}
|
|
}
|
|
}
|
|
|
|
void ReloadHandler::PauseIsolatesForReloadLocked() {
|
|
intptr_t registered = -1;
|
|
{
|
|
SafepointMonitorLocker ml(&monitor_);
|
|
|
|
// Maybe participate in existing reload requested by another isolate.
|
|
ParticipateIfReloadRequested(&ml, /*registered=*/true,
|
|
/*allow_later_retry=*/false);
|
|
|
|
// Now it's our turn to request reload.
|
|
ASSERT(reloading_thread_ == nullptr);
|
|
reloading_thread_ = Thread::Current();
|
|
|
|
// At this point no isolate register/unregister, so we save the current
|
|
// number of registered isolates.
|
|
registered = registered_isolate_count_;
|
|
}
|
|
|
|
// Send OOB to a superset of all registered isolates and make them participate
|
|
// in this reload.
|
|
reloading_thread_->isolate_group()->ForEachIsolate([](Isolate* isolate) {
|
|
isolate->SendInternalLibMessage(Isolate::kCheckForReload, /*ignored=*/-1);
|
|
});
|
|
|
|
{
|
|
SafepointMonitorLocker ml(&checkin_monitor_);
|
|
while (isolates_checked_in_ < (registered - /*reload_requester=*/1)) {
|
|
ml.Wait();
|
|
}
|
|
}
|
|
}
|
|
|
|
void ReloadHandler::ResumeIsolatesLocked() {
|
|
{
|
|
SafepointMonitorLocker ml(&monitor_);
|
|
ASSERT(reloading_thread_ == Thread::Current());
|
|
reloading_thread_ = nullptr;
|
|
ml.NotifyAll();
|
|
}
|
|
}
|
|
|
|
#endif // !defined(PRODUCT) && !defined(DART_PRECOMPILED_RUNTIME)
|
|
|
|
} // namespace dart
|