1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
|
Patch status: fixed in >=chromium-127.0.6524.0
https://bugs.gentoo.org/937541
https://github.com/google/perfetto/commit/e2f661907a717551235563389977b7468da6d45e
https://android-review.googlesource.com/c/platform/external/perfetto/+/3114454
https://chromium-review.googlesource.com/c/chromium/src/+/5604664
--- a/src/3rdparty/chromium/third_party/blink/renderer/platform/wtf/hash_table.h
+++ b/src/3rdparty/chromium/third_party/blink/renderer/platform/wtf/hash_table.h
@@ -1647,5 +1647,5 @@
}
table_ = temporary_table;
- Allocator::template BackingWriteBarrier(&table_);
+ Allocator::BackingWriteBarrier(&table_);
HashTableBucketInitializer<Traits, Allocator, Value>::InitializeTable(
@@ -1701,5 +1701,5 @@
// the current table has to be atomic to prevent races with concurrent marker.
AsAtomicPtr(&table_)->store(new_hash_table.table_, std::memory_order_relaxed);
- Allocator::template BackingWriteBarrier(&table_);
+ Allocator::BackingWriteBarrier(&table_);
table_size_ = new_table_size;
@@ -1853,6 +1853,6 @@
// there is *no* risk of data races when reading.
AtomicWriteSwap(table_, other.table_);
- Allocator::template BackingWriteBarrier(&table_);
- Allocator::template BackingWriteBarrier(&other.table_);
+ Allocator::BackingWriteBarrier(&table_);
+ Allocator::BackingWriteBarrier(&other.table_);
if (IsWeak<ValueType>::value) {
// Weak processing is omitted when no backing store is present. In case such
--- a/src/3rdparty/chromium/third_party/perfetto/include/perfetto/tracing/internal/track_event_data_source.h
+++ b/src/3rdparty/chromium/third_party/perfetto/include/perfetto/tracing/internal/track_event_data_source.h
@@ -329,5 +329,5 @@
static void Flush() {
- Base::template Trace([](typename Base::TraceContext ctx) { ctx.Flush(); });
+ Base::Trace([](typename Base::TraceContext ctx) { ctx.Flush(); });
}
@@ -335,6 +335,5 @@
static bool IsEnabled() {
bool enabled = false;
- Base::template CallIfEnabled(
- [&](uint32_t /*instances*/) { enabled = true; });
+ Base::CallIfEnabled([&](uint32_t /*instances*/) { enabled = true; });
return enabled;
}
@@ -350,5 +349,5 @@
const DynamicCategory& dynamic_category) {
bool enabled = false;
- Base::template Trace([&](typename Base::TraceContext ctx) {
+ Base::Trace([&](typename Base::TraceContext ctx) {
enabled = enabled || IsDynamicCategoryEnabled(&ctx, dynamic_category);
});
@@ -497,5 +496,5 @@
PERFETTO_DCHECK(track.uuid == desc.uuid());
TrackRegistry::Get()->UpdateTrack(track, desc.SerializeAsString());
- Base::template Trace([&](typename Base::TraceContext ctx) {
+ Base::Trace([&](typename Base::TraceContext ctx) {
TrackEventInternal::WriteTrackDescriptor(
track, ctx.tls_inst_->trace_writer.get(), ctx.GetIncrementalState(),
@@ -1048,5 +1047,5 @@
using CatTraits = CategoryTraits<CategoryType>;
if (CatTraits::kIsDynamic) {
- Base::template TraceWithInstances(instances, std::move(lambda));
+ Base::TraceWithInstances(instances, std::move(lambda));
} else {
Base::template TraceWithInstances<CategoryTracePointTraits>(
@@ -1062,5 +1061,5 @@
std::function<void(protos::pbzero::TrackDescriptor*)> callback) {
TrackRegistry::Get()->UpdateTrack(track, std::move(callback));
- Base::template Trace([&](typename Base::TraceContext ctx) {
+ Base::Trace([&](typename Base::TraceContext ctx) {
TrackEventInternal::WriteTrackDescriptor(
track, ctx.tls_inst_->trace_writer.get(), ctx.GetIncrementalState(),
|