Skip to content
Snippets Groups Projects

Compare revisions

Changes are shown as if the source revision was being merged into the target revision. Learn more about comparing revisions.

Source

Select target project
No results found
Select Git revision

Target

Select target project
  • hive/haf
  • dan/haf
2 results
Select Git revision
Show changes
Commits on Source (3)
Subproject commit 3a297f9ad2f44ad7dc23c8f229849c1d7ef2c10d
Subproject commit 5c7a9f728cb5454f16849900b50598a404e04cde
......@@ -669,7 +669,8 @@ BEGIN
ARRAY[__context_id]
)
ON CONFLICT (table_name, index_constraint_name) DO UPDATE
SET contexts = array_append(hafd.indexes_constraints.contexts, __context_id);
SET contexts = array_append(hafd.indexes_constraints.contexts, __context_id)
WHERE NOT (__context_id = ANY(hafd.indexes_constraints.contexts));
END;
$BODY$
;
......
......@@ -55,7 +55,7 @@ data_processor::data_processor( std::string description, std::string short_descr
{
auto body = [this, dataProcessor]() -> void
{
ilog("Entering data processor thread: ${d}", ("d", _description));
dlog("Entering data processor thread: ${d}", ("d", _description));
const std::string thread_name = "sql[" + _short_description + "]";
fc::set_thread_name(thread_name.c_str());
fc::thread::current().set_name(thread_name);
......@@ -111,7 +111,7 @@ data_processor::data_processor( std::string description, std::string short_descr
auto current_exception = std::current_exception();
handle_exception( current_exception );
}
ilog("Leaving data processor thread: ${d}", ("d", _description));
dlog("Leaving data processor thread: ${d}", ("d", _description));
};
_future = std::async(std::launch::async, body);
......@@ -185,9 +185,9 @@ void data_processor::join()
_continue.store(false);
{
ilog("Trying to resume data processor: ${d}...", ("d", _description));
dlog("Trying to resume data processor: ${d}...", ("d", _description));
std::lock_guard<std::mutex> lk(_mtx);
ilog("Data processor: ${d} resumed...", ("d", _description));
dlog("Data processor: ${d} resumed...", ("d", _description));
}
_cv.notify_one();
......
......@@ -28,7 +28,7 @@ namespace hive::plugins::sql_serializer {
~cached_data_t()
{
ilog(
dlog(
"blocks: ${b} trx: ${t} operations: ${o} total size: ${ts}...",
("b", blocks.size() )
("t", transactions.size() )
......
......@@ -380,7 +380,14 @@ indexation_state::update_state(
break;
}
case INDEXATION::WAIT:
ilog("PROFILE: Entering WAIT sync, creating indexes/constraints as needed: ${t} s",("t",(fc::time_point::now() - _start_state_time).to_seconds()));
ilog("PROFILE: Entering WAIT sync: ${t} s",("t",(fc::time_point::now() - _start_state_time).to_seconds()));
_trigger.reset();
_dumper.reset();
_trigger = std::make_shared< fake_flush_trigger >();
_dumper = std::make_shared< fake_data_dumper >();
ilog("PROFILE: Entered WAIT sync from start state: ${t} s",("t",(fc::time_point::now() - _start_state_time).to_seconds()));
break;
ilog("PROFILE: Entering WAIT sync: ${t} s",("t",(fc::time_point::now() - _start_state_time).to_seconds()));
_trigger.reset();
_dumper.reset();
_trigger = std::make_shared< fake_flush_trigger >();
......@@ -388,12 +395,12 @@ indexation_state::update_state(
ilog("PROFILE: Entered WAIT sync from start state: ${t} s",("t",(fc::time_point::now() - _start_state_time).to_seconds()));
break;
case INDEXATION::REINDEX_WAIT:
ilog("PROFILE: Entering REINDEX_WAIT sync, creating indexes/constraints as needed: ${t} s",("t",(fc::time_point::now() - _start_state_time).to_seconds()));
ilog("PROFILE: Entering REINDEX_WAIT sync: ${t} s",("t",(fc::time_point::now() - _start_state_time).to_seconds()));
_trigger.reset();
_dumper.reset();
_trigger = std::make_shared< fake_flush_trigger >();
_dumper = std::make_shared< fake_data_dumper >();
ilog("PROFILE: Entered REINDEX_WAIT sync, creating indexes/constraints as needed: ${t} s",("t",(fc::time_point::now() - _start_state_time).to_seconds()));
ilog("PROFILE: Entered REINDEX_WAIT sync: ${t} s",("t",(fc::time_point::now() - _start_state_time).to_seconds()));
break;
default:
FC_ASSERT( false, "Unknown INDEXATION state" );
......