Remove the use of atomics and use thread-local accumulation instead

This commit is contained in:
Patrick Niklaus 2016-05-28 17:36:31 +02:00
parent 4ae8abe6a3
commit cd30f370b8
No known key found for this signature in database
GPG Key ID: E426891B5F978B1B

View File

@ -25,6 +25,7 @@
#include <tbb/blocked_range.h> #include <tbb/blocked_range.h>
#include <tbb/concurrent_unordered_map.h> #include <tbb/concurrent_unordered_map.h>
#include <tbb/enumerable_thread_specific.h>
#include <tbb/parallel_for.h> #include <tbb/parallel_for.h>
#include <tbb/parallel_for_each.h> #include <tbb/parallel_for_each.h>
#include <tbb/parallel_invoke.h> #include <tbb/parallel_invoke.h>
@ -32,7 +33,6 @@
#include <tbb/spin_mutex.h> #include <tbb/spin_mutex.h>
#include <algorithm> #include <algorithm>
#include <atomic>
#include <bitset> #include <bitset>
#include <cstdint> #include <cstdint>
#include <fstream> #include <fstream>
@ -516,15 +516,14 @@ std::size_t Contractor::LoadEdgeExpandedGraph(
// vector to count used speeds for logging // vector to count used speeds for logging
// size offset by one since index 0 is used for speeds not from external file // size offset by one since index 0 is used for speeds not from external file
std::vector<std::atomic<std::uint64_t>> segment_speeds_counters; using counters_type = std::vector<std::size_t>;
for (std::size_t i = 0; i < segment_speed_filenames.size() + 1; ++i) std::size_t num_counters = segment_speed_filenames.size() + 1;
{ tbb::enumerable_thread_specific<counters_type> segment_speeds_counters(
segment_speeds_counters.emplace_back(); counters_type(num_counters, 0));
segment_speeds_counters[i].store(0);
}
const constexpr auto LUA_SOURCE = 0; const constexpr auto LUA_SOURCE = 0;
tbb::parallel_for_each(first, last, [&](const LeafNode &current_node) { tbb::parallel_for_each(first, last, [&](const LeafNode &current_node) {
auto &counters = segment_speeds_counters.local();
for (size_t i = 0; i < current_node.object_count; i++) for (size_t i = 0; i < current_node.object_count; i++)
{ {
const auto &leaf_object = current_node.objects[i]; const auto &leaf_object = current_node.objects[i];
@ -571,12 +570,12 @@ std::size_t Contractor::LoadEdgeExpandedGraph(
forward_speed_iter->speed_source.source; forward_speed_iter->speed_source.source;
// count statistics for logging // count statistics for logging
segment_speeds_counters[forward_speed_iter->speed_source.source] += 1; counters[forward_speed_iter->speed_source.source] += 1;
} }
else else
{ {
// count statistics for logging // count statistics for logging
segment_speeds_counters[LUA_SOURCE] += 1; counters[LUA_SOURCE] += 1;
} }
} }
if (leaf_object.reverse_packed_geometry_id != SPECIAL_EDGEID) if (leaf_object.reverse_packed_geometry_id != SPECIAL_EDGEID)
@ -622,30 +621,39 @@ std::size_t Contractor::LoadEdgeExpandedGraph(
reverse_speed_iter->speed_source.source; reverse_speed_iter->speed_source.source;
// count statistics for logging // count statistics for logging
segment_speeds_counters[reverse_speed_iter->speed_source.source] += 1; counters[reverse_speed_iter->speed_source.source] += 1;
} }
else else
{ {
// count statistics for logging // count statistics for logging
segment_speeds_counters[LUA_SOURCE] += 1; counters[LUA_SOURCE] += 1;
} }
} }
} }
}); // parallel_for_each }); // parallel_for_each
for (std::size_t i = 0; i < segment_speeds_counters.size(); i++) counters_type merged_counters(num_counters, 0);
for (const auto &counters : segment_speeds_counters)
{
for (std::size_t i = 0; i < counters.size(); i++)
{
merged_counters[i] += counters[i];
}
}
for (std::size_t i = 0; i < merged_counters.size(); i++)
{ {
if (i == LUA_SOURCE) if (i == LUA_SOURCE)
{ {
util::SimpleLogger().Write() << "Used " << segment_speeds_counters[LUA_SOURCE] util::SimpleLogger().Write() << "Used " << merged_counters[LUA_SOURCE]
<< " speeds from LUA profile or input map"; << " speeds from LUA profile or input map";
} }
else else
{ {
// segments_speeds_counters has 0 as LUA, segment_speed_filenames not, thus we need // segments_speeds_counters has 0 as LUA, segment_speed_filenames not, thus we need
// to susbstract 1 to avoid off-by-one error // to susbstract 1 to avoid off-by-one error
util::SimpleLogger().Write() << "Used " << segment_speeds_counters[i] util::SimpleLogger().Write() << "Used " << merged_counters[i] << " speeds from "
<< " speeds from " << segment_speed_filenames[i - 1]; << segment_speed_filenames[i - 1];
} }
} }
} }