1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
|
// This Source Code Form is subject to the terms of the Mozilla Public
// License, v. 2.0. If a copy of the MPL was not distributed with this
// file, You can obtain one at https://mozilla.org/MPL/2.0/.
//! IPC Implementation, Rust part
use crate::private::MetricId;
use once_cell::sync::Lazy;
use serde::{Deserialize, Serialize};
use std::collections::HashMap;
#[cfg(not(feature = "with_gecko"))]
use std::sync::atomic::AtomicBool;
use std::sync::atomic::{AtomicUsize, Ordering};
use std::sync::Mutex;
#[cfg(feature = "with_gecko")]
use {std::convert::TryInto, std::sync::atomic::AtomicU32, xpcom::interfaces::nsIXULRuntime};
use super::metrics::__glean_metric_maps;
type EventRecord = (u64, HashMap<String, String>);
/// Contains all the information necessary to update the metrics on the main
/// process.
#[derive(Debug, Default, Deserialize, Serialize)]
pub struct IPCPayload {
pub counters: HashMap<MetricId, i32>,
pub custom_samples: HashMap<MetricId, Vec<i64>>,
pub denominators: HashMap<MetricId, i32>,
pub events: HashMap<MetricId, Vec<EventRecord>>,
pub labeled_counters: HashMap<MetricId, HashMap<String, i32>>,
pub memory_samples: HashMap<MetricId, Vec<u64>>,
pub numerators: HashMap<MetricId, i32>,
pub rates: HashMap<MetricId, (i32, i32)>,
pub string_lists: HashMap<MetricId, Vec<String>>,
pub timing_samples: HashMap<MetricId, Vec<u64>>,
}
/// Global singleton: pending IPC payload.
static PAYLOAD: Lazy<Mutex<IPCPayload>> = Lazy::new(|| Mutex::new(IPCPayload::default()));
/// Global singleton: number of times the IPC payload was accessed.
static PAYLOAD_ACCESS_COUNT: AtomicUsize = AtomicUsize::new(0);
// The maximum size of an IPC message in Firefox Desktop is 256MB.
// (See IPC::Channel::kMaximumMessageSize)
// In `IPCPayload` the largest size can be attained in the fewest accesses via events.
// Each event could be its own u64 id, u64 timestamp, and HashMap of ten i32 to ten 100-byte strings.
// That's 1056B = 8 + 8 + 10(4 + 100)
// In 256MB we can fit 254200 or so of these, not counting overhead.
// Let's take a conservative estimate of 100000 to
// 0) Account for overhead
// 1) Not be greedy
// 2) Allow time for the dispatch to main thread which will actually perform the flush
// "Why the -1?" Because fetch_add returns the value before the addition.
const PAYLOAD_ACCESS_WATERMARK: usize = 100000 - 1;
pub fn with_ipc_payload<F, R>(f: F) -> R
where
F: FnOnce(&mut IPCPayload) -> R,
{
if PAYLOAD_ACCESS_COUNT.fetch_add(1, Ordering::SeqCst) > PAYLOAD_ACCESS_WATERMARK {
// We reset this before the actual flush to keep all the logic together.
// Otherwise the count reset would need to happen down in take_buf().
// This may overcount (resulting in undersized payloads) which is okay.
PAYLOAD_ACCESS_COUNT.store(0, Ordering::SeqCst);
handle_payload_filling();
}
let mut payload = PAYLOAD.lock().unwrap();
f(&mut payload)
}
/// Do we need IPC?
///
/// Thread-safe.
#[cfg(feature = "with_gecko")]
static PROCESS_TYPE: Lazy<AtomicU32> = Lazy::new(|| {
extern "C" {
fn FOG_GetProcessType() -> i32;
}
// SAFETY NOTE: Safe because it returns a primitive by value.
let process_type = unsafe { FOG_GetProcessType() };
// It's impossible for i32 to overflow u32, but maybe someone got clever
// and introduced a negative process type constant. Default to parent.
let process_type = process_type
.try_into()
.unwrap_or(nsIXULRuntime::PROCESS_TYPE_DEFAULT);
// We don't have process-specific init locations outside of the main
// process, so we introduce this side-effect to a global static init.
// This is the absolute first time we decide which process type we're
// treating this process as, so this is the earliest we can do this.
register_process_shutdown(process_type);
AtomicU32::new(process_type)
});
#[cfg(feature = "with_gecko")]
pub fn need_ipc() -> bool {
PROCESS_TYPE.load(Ordering::Relaxed) != nsIXULRuntime::PROCESS_TYPE_DEFAULT
}
/// The first time we're used in a process,
/// we'll need to start thinking about cleanup.
///
/// Please only call once per process.
/// Multiple calls may register multiple handlers.
#[cfg(feature = "with_gecko")]
fn register_process_shutdown(process_type: u32) {
match process_type {
nsIXULRuntime::PROCESS_TYPE_DEFAULT => {
// Parent process shutdown is handled by the FOG XPCOM Singleton.
}
nsIXULRuntime::PROCESS_TYPE_CONTENT => {
// Content child shutdown is in C++ for access to RunOnShutdown().
extern "C" {
fn FOG_RegisterContentChildShutdown();
}
unsafe {
FOG_RegisterContentChildShutdown();
};
}
nsIXULRuntime::PROCESS_TYPE_GMPLUGIN => {
// GMP process shutdown is handled in GMPChild::ActorDestroy.
}
nsIXULRuntime::PROCESS_TYPE_GPU => {
// GPU process shutdown is handled in GPUParent::ActorDestroy.
}
nsIXULRuntime::PROCESS_TYPE_RDD => {
// RDD process shutdown is handled in RDDParent::ActorDestroy.
}
nsIXULRuntime::PROCESS_TYPE_SOCKET => {
// Socket process shutdown is handled in SocketProcessChild::ActorDestroy.
}
nsIXULRuntime::PROCESS_TYPE_UTILITY => {
// Utility process shutdown is handled in UtilityProcessChild::ActorDestroy.
}
_ => {
// We don't yet support other process types.
log::error!("Process type {} tried to use FOG, but isn't supported! (Process type constants are in nsIXULRuntime.rs)", process_type);
}
}
}
/// An RAII that, on drop, restores the value used to determine whether FOG
/// needs IPC. Used in tests.
/// ```rust,ignore
/// #[test]
/// fn test_need_ipc_raii() {
/// assert!(false == ipc::need_ipc());
/// {
/// let _raii = ipc::test_set_need_ipc(true);
/// assert!(ipc::need_ipc());
/// }
/// assert!(false == ipc::need_ipc());
/// }
/// ```
#[cfg(not(feature = "with_gecko"))]
pub struct TestNeedIpcRAII {
prev_value: bool,
}
#[cfg(not(feature = "with_gecko"))]
impl Drop for TestNeedIpcRAII {
fn drop(&mut self) {
TEST_NEED_IPC.store(self.prev_value, Ordering::Relaxed);
}
}
#[cfg(not(feature = "with_gecko"))]
static TEST_NEED_IPC: AtomicBool = AtomicBool::new(false);
/// Test-only API for telling FOG to use IPC mechanisms even if the test has
/// only the one process. See TestNeedIpcRAII for an example.
#[cfg(not(feature = "with_gecko"))]
pub fn test_set_need_ipc(need_ipc: bool) -> TestNeedIpcRAII {
TestNeedIpcRAII {
prev_value: TEST_NEED_IPC.swap(need_ipc, Ordering::Relaxed),
}
}
#[cfg(not(feature = "with_gecko"))]
pub fn need_ipc() -> bool {
TEST_NEED_IPC.load(Ordering::Relaxed)
}
pub fn take_buf() -> Option<Vec<u8>> {
with_ipc_payload(move |payload| {
let buf = bincode::serialize(&payload).ok();
*payload = IPCPayload {
..Default::default()
};
buf
})
}
#[cfg(not(feature = "with_gecko"))]
fn handle_payload_filling() {
// Space intentionally left blank.
// Without Gecko IPC to drain the buffer, there's nothing we can do.
}
#[cfg(feature = "with_gecko")]
fn handle_payload_filling() {
extern "C" {
fn FOG_IPCPayloadFull();
}
// SAFETY NOTE: Safe because it doesn't take or return values.
unsafe { FOG_IPCPayloadFull() };
}
#[cfg(not(feature = "with_gecko"))]
pub fn is_in_automation() -> bool {
// Without Gecko IPC to drain the buffer, there's nothing we can do.
false
}
#[cfg(feature = "with_gecko")]
pub fn is_in_automation() -> bool {
extern "C" {
fn FOG_IPCIsInAutomation() -> bool;
}
// SAFETY NOTE: Safe because it returns a primitive by value.
unsafe { FOG_IPCIsInAutomation() }
}
// Reason: We instrument the error counts,
// but don't need more detailed error information at the moment.
#[allow(clippy::result_unit_err)]
pub fn replay_from_buf(buf: &[u8]) -> Result<(), ()> {
// TODO: Instrument failures to find metrics by id.
let ipc_payload: IPCPayload = bincode::deserialize(buf).map_err(|_| ())?;
for (id, value) in ipc_payload.counters.into_iter() {
if id.0 & (1 << crate::factory::DYNAMIC_METRIC_BIT) > 0 {
let map = crate::factory::__jog_metric_maps::COUNTER_MAP
.read()
.expect("Read lock for dynamic counter map was poisoned");
if let Some(metric) = map.get(&id) {
metric.add(value);
}
} else if let Some(metric) = __glean_metric_maps::COUNTER_MAP.get(&id) {
metric.add(value);
}
}
for (id, samples) in ipc_payload.custom_samples.into_iter() {
if id.0 & (1 << crate::factory::DYNAMIC_METRIC_BIT) > 0 {
let map = crate::factory::__jog_metric_maps::CUSTOM_DISTRIBUTION_MAP
.read()
.expect("Read lock for dynamic custom distribution map was poisoned");
if let Some(metric) = map.get(&id) {
metric.accumulate_samples_signed(samples);
}
} else if let Some(metric) = __glean_metric_maps::CUSTOM_DISTRIBUTION_MAP.get(&id) {
metric.accumulate_samples_signed(samples);
}
}
for (id, value) in ipc_payload.denominators.into_iter() {
if id.0 & (1 << crate::factory::DYNAMIC_METRIC_BIT) > 0 {
let map = crate::factory::__jog_metric_maps::DENOMINATOR_MAP
.read()
.expect("Read lock for dynamic denominator map was poisoned");
if let Some(metric) = map.get(&id) {
metric.add(value);
}
} else if let Some(metric) = __glean_metric_maps::DENOMINATOR_MAP.get(&id) {
metric.add(value);
}
}
for (id, records) in ipc_payload.events.into_iter() {
if id.0 & (1 << crate::factory::DYNAMIC_METRIC_BIT) > 0 {
let map = crate::factory::__jog_metric_maps::EVENT_MAP
.read()
.expect("Read lock for dynamic event map was poisoned");
if let Some(metric) = map.get(&id) {
for (timestamp, extra) in records.into_iter() {
metric.record_with_time(timestamp, extra);
}
}
} else {
for (timestamp, extra) in records.into_iter() {
let _ = __glean_metric_maps::record_event_by_id_with_time(id, timestamp, extra);
}
}
}
for (id, labeled_counts) in ipc_payload.labeled_counters.into_iter() {
if id.0 & (1 << crate::factory::DYNAMIC_METRIC_BIT) > 0 {
let map = crate::factory::__jog_metric_maps::LABELED_COUNTER_MAP
.read()
.expect("Read lock for dynamic labeled counter map was poisoned");
if let Some(metric) = map.get(&id) {
for (label, count) in labeled_counts.into_iter() {
metric.get(&label).add(count);
}
}
} else {
for (label, count) in labeled_counts.into_iter() {
__glean_metric_maps::labeled_counter_get(id.0, &label).add(count);
}
}
}
for (id, samples) in ipc_payload.memory_samples.into_iter() {
if id.0 & (1 << crate::factory::DYNAMIC_METRIC_BIT) > 0 {
let map = crate::factory::__jog_metric_maps::MEMORY_DISTRIBUTION_MAP
.read()
.expect("Read lock for dynamic memory dist map was poisoned");
if let Some(metric) = map.get(&id) {
samples
.into_iter()
.for_each(|sample| metric.accumulate(sample));
}
} else if let Some(metric) = __glean_metric_maps::MEMORY_DISTRIBUTION_MAP.get(&id) {
samples
.into_iter()
.for_each(|sample| metric.accumulate(sample));
}
}
for (id, value) in ipc_payload.numerators.into_iter() {
if id.0 & (1 << crate::factory::DYNAMIC_METRIC_BIT) > 0 {
let map = crate::factory::__jog_metric_maps::NUMERATOR_MAP
.read()
.expect("Read lock for dynamic numerator map was poisoned");
if let Some(metric) = map.get(&id) {
metric.add_to_numerator(value);
}
} else if let Some(metric) = __glean_metric_maps::NUMERATOR_MAP.get(&id) {
metric.add_to_numerator(value);
}
}
for (id, (n, d)) in ipc_payload.rates.into_iter() {
if id.0 & (1 << crate::factory::DYNAMIC_METRIC_BIT) > 0 {
let map = crate::factory::__jog_metric_maps::RATE_MAP
.read()
.expect("Read lock for dynamic rate map was poisoned");
if let Some(metric) = map.get(&id) {
metric.add_to_numerator(n);
metric.add_to_denominator(d);
}
} else if let Some(metric) = __glean_metric_maps::RATE_MAP.get(&id) {
metric.add_to_numerator(n);
metric.add_to_denominator(d);
}
}
for (id, strings) in ipc_payload.string_lists.into_iter() {
if id.0 & (1 << crate::factory::DYNAMIC_METRIC_BIT) > 0 {
let map = crate::factory::__jog_metric_maps::STRING_LIST_MAP
.read()
.expect("Read lock for dynamic string list map was poisoned");
if let Some(metric) = map.get(&id) {
strings.iter().for_each(|s| metric.add(s));
}
} else if let Some(metric) = __glean_metric_maps::STRING_LIST_MAP.get(&id) {
strings.iter().for_each(|s| metric.add(s));
}
}
for (id, samples) in ipc_payload.timing_samples.into_iter() {
if id.0 & (1 << crate::factory::DYNAMIC_METRIC_BIT) > 0 {
let map = crate::factory::__jog_metric_maps::TIMING_DISTRIBUTION_MAP
.read()
.expect("Read lock for dynamic timing distribution map was poisoned");
if let Some(metric) = map.get(&id) {
metric.accumulate_raw_samples_nanos(samples);
}
} else if let Some(metric) = __glean_metric_maps::TIMING_DISTRIBUTION_MAP.get(&id) {
metric.accumulate_raw_samples_nanos(samples);
}
}
Ok(())
}
|