1 /*
2  * Copyright (C) 2017 The Android Open Source Project
3  *
4  * Licensed under the Apache License, Version 2.0 (the "License");
5  * you may not use this file except in compliance with the License.
6  * You may obtain a copy of the License at
7  *
8  *      http://www.apache.org/licenses/LICENSE-2.0
9  *
10  * Unless required by applicable law or agreed to in writing, software
11  * distributed under the License is distributed on an "AS IS" BASIS,
12  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13  * See the License for the specific language governing permissions and
14  * limitations under the License.
15  */
16 
17 #define DEBUG false  // STOPSHIP if true
18 #include "Log.h"
19 
20 #include "ValueMetricProducer.h"
21 #include "../guardrail/StatsdStats.h"
22 #include "../stats_log_util.h"
23 
24 #include <cutils/log.h>
25 #include <limits.h>
26 #include <stdlib.h>
27 
28 using android::util::FIELD_COUNT_REPEATED;
29 using android::util::FIELD_TYPE_BOOL;
30 using android::util::FIELD_TYPE_FLOAT;
31 using android::util::FIELD_TYPE_INT32;
32 using android::util::FIELD_TYPE_INT64;
33 using android::util::FIELD_TYPE_MESSAGE;
34 using android::util::FIELD_TYPE_STRING;
35 using android::util::ProtoOutputStream;
36 using std::list;
37 using std::make_pair;
38 using std::make_shared;
39 using std::map;
40 using std::shared_ptr;
41 using std::unique_ptr;
42 using std::unordered_map;
43 
44 namespace android {
45 namespace os {
46 namespace statsd {
47 
48 // for StatsLogReport
49 const int FIELD_ID_ID = 1;
50 const int FIELD_ID_VALUE_METRICS = 7;
51 const int FIELD_ID_TIME_BASE = 9;
52 const int FIELD_ID_BUCKET_SIZE = 10;
53 const int FIELD_ID_DIMENSION_PATH_IN_WHAT = 11;
54 const int FIELD_ID_DIMENSION_PATH_IN_CONDITION = 12;
55 // for ValueMetricDataWrapper
56 const int FIELD_ID_DATA = 1;
57 const int FIELD_ID_SKIPPED = 2;
58 const int FIELD_ID_SKIPPED_START_MILLIS = 3;
59 const int FIELD_ID_SKIPPED_END_MILLIS = 4;
60 // for ValueMetricData
61 const int FIELD_ID_DIMENSION_IN_WHAT = 1;
62 const int FIELD_ID_DIMENSION_IN_CONDITION = 2;
63 const int FIELD_ID_BUCKET_INFO = 3;
64 const int FIELD_ID_DIMENSION_LEAF_IN_WHAT = 4;
65 const int FIELD_ID_DIMENSION_LEAF_IN_CONDITION = 5;
66 // for ValueBucketInfo
67 const int FIELD_ID_VALUE = 3;
68 const int FIELD_ID_BUCKET_NUM = 4;
69 const int FIELD_ID_START_BUCKET_ELAPSED_MILLIS = 5;
70 const int FIELD_ID_END_BUCKET_ELAPSED_MILLIS = 6;
71 
72 // ValueMetric has a minimum bucket size of 10min so that we don't pull too frequently
ValueMetricProducer(const ConfigKey & key,const ValueMetric & metric,const int conditionIndex,const sp<ConditionWizard> & wizard,const int pullTagId,const int64_t timeBaseNs,const int64_t startTimestampNs,shared_ptr<StatsPullerManager> statsPullerManager)73 ValueMetricProducer::ValueMetricProducer(const ConfigKey& key, const ValueMetric& metric,
74                                          const int conditionIndex,
75                                          const sp<ConditionWizard>& wizard, const int pullTagId,
76                                          const int64_t timeBaseNs, const int64_t startTimestampNs,
77                                          shared_ptr<StatsPullerManager> statsPullerManager)
78     : MetricProducer(metric.id(), key, timeBaseNs, conditionIndex, wizard),
79       mValueField(metric.value_field()),
80       mStatsPullerManager(statsPullerManager),
81       mPullTagId(pullTagId),
82       mMinBucketSizeNs(metric.min_bucket_size_nanos()),
83       mDimensionSoftLimit(StatsdStats::kAtomDimensionKeySizeLimitMap.find(pullTagId) !=
84                                           StatsdStats::kAtomDimensionKeySizeLimitMap.end()
85                                   ? StatsdStats::kAtomDimensionKeySizeLimitMap.at(pullTagId).first
86                                   : StatsdStats::kDimensionKeySizeSoftLimit),
87       mDimensionHardLimit(StatsdStats::kAtomDimensionKeySizeLimitMap.find(pullTagId) !=
88                                           StatsdStats::kAtomDimensionKeySizeLimitMap.end()
89                                   ? StatsdStats::kAtomDimensionKeySizeLimitMap.at(pullTagId).second
90                                   : StatsdStats::kDimensionKeySizeHardLimit),
91       mUseAbsoluteValueOnReset(metric.use_absolute_value_on_reset()) {
92     // TODO: valuemetric for pushed events may need unlimited bucket length
93     int64_t bucketSizeMills = 0;
94     if (metric.has_bucket()) {
95         bucketSizeMills = TimeUnitToBucketSizeInMillisGuardrailed(key.GetUid(), metric.bucket());
96     } else {
97         bucketSizeMills = TimeUnitToBucketSizeInMillis(ONE_HOUR);
98     }
99 
100     mBucketSizeNs = bucketSizeMills * 1000000;
101     if (metric.has_dimensions_in_what()) {
102         translateFieldMatcher(metric.dimensions_in_what(), &mDimensionsInWhat);
103         mContainANYPositionInDimensionsInWhat = HasPositionANY(metric.dimensions_in_what());
104     }
105 
106     if (metric.has_dimensions_in_condition()) {
107         translateFieldMatcher(metric.dimensions_in_condition(), &mDimensionsInCondition);
108     }
109 
110     if (metric.links().size() > 0) {
111         for (const auto& link : metric.links()) {
112             Metric2Condition mc;
113             mc.conditionId = link.condition();
114             translateFieldMatcher(link.fields_in_what(), &mc.metricFields);
115             translateFieldMatcher(link.fields_in_condition(), &mc.conditionFields);
116             mMetric2ConditionLinks.push_back(mc);
117         }
118     }
119 
120     if (mValueField.child_size() > 0) {
121         mField = mValueField.child(0).field();
122     }
123     mConditionSliced = (metric.links().size() > 0) || (mDimensionsInCondition.size() > 0);
124     mSliceByPositionALL = HasPositionALL(metric.dimensions_in_what()) ||
125             HasPositionALL(metric.dimensions_in_condition());
126 
127     // Kicks off the puller immediately.
128     flushIfNeededLocked(startTimestampNs);
129     if (mPullTagId != -1) {
130         mStatsPullerManager->RegisterReceiver(
131             mPullTagId, this, mCurrentBucketStartTimeNs + mBucketSizeNs, mBucketSizeNs);
132     }
133 
134     VLOG("value metric %lld created. bucket size %lld start_time: %lld",
135         (long long)metric.id(), (long long)mBucketSizeNs, (long long)mTimeBaseNs);
136 }
137 
138 // for testing
ValueMetricProducer(const ConfigKey & key,const ValueMetric & metric,const int conditionIndex,const sp<ConditionWizard> & wizard,const int pullTagId,const int64_t timeBaseNs,const int64_t startTimeNs)139 ValueMetricProducer::ValueMetricProducer(const ConfigKey& key, const ValueMetric& metric,
140                                          const int conditionIndex,
141                                          const sp<ConditionWizard>& wizard, const int pullTagId,
142                                          const int64_t timeBaseNs, const int64_t startTimeNs)
143     : ValueMetricProducer(key, metric, conditionIndex, wizard, pullTagId, timeBaseNs, startTimeNs,
144                           make_shared<StatsPullerManager>()) {
145 }
146 
~ValueMetricProducer()147 ValueMetricProducer::~ValueMetricProducer() {
148     VLOG("~ValueMetricProducer() called");
149     if (mPullTagId != -1) {
150         mStatsPullerManager->UnRegisterReceiver(mPullTagId, this);
151     }
152 }
153 
onSlicedConditionMayChangeLocked(bool overallCondition,const int64_t eventTime)154 void ValueMetricProducer::onSlicedConditionMayChangeLocked(bool overallCondition,
155                                                            const int64_t eventTime) {
156     VLOG("Metric %lld onSlicedConditionMayChange", (long long)mMetricId);
157 }
158 
dropDataLocked(const int64_t dropTimeNs)159 void ValueMetricProducer::dropDataLocked(const int64_t dropTimeNs) {
160     flushIfNeededLocked(dropTimeNs);
161     mPastBuckets.clear();
162 }
163 
clearPastBucketsLocked(const int64_t dumpTimeNs)164 void ValueMetricProducer::clearPastBucketsLocked(const int64_t dumpTimeNs) {
165     flushIfNeededLocked(dumpTimeNs);
166     mPastBuckets.clear();
167     mSkippedBuckets.clear();
168 }
169 
onDumpReportLocked(const int64_t dumpTimeNs,const bool include_current_partial_bucket,std::set<string> * str_set,ProtoOutputStream * protoOutput)170 void ValueMetricProducer::onDumpReportLocked(const int64_t dumpTimeNs,
171                                              const bool include_current_partial_bucket,
172                                              std::set<string> *str_set,
173                                              ProtoOutputStream* protoOutput) {
174     VLOG("metric %lld dump report now...", (long long)mMetricId);
175     if (include_current_partial_bucket) {
176         flushLocked(dumpTimeNs);
177     } else {
178         flushIfNeededLocked(dumpTimeNs);
179     }
180     if (mPastBuckets.empty() && mSkippedBuckets.empty()) {
181         return;
182     }
183     protoOutput->write(FIELD_TYPE_INT64 | FIELD_ID_ID, (long long)mMetricId);
184     protoOutput->write(FIELD_TYPE_INT64 | FIELD_ID_TIME_BASE, (long long)mTimeBaseNs);
185     protoOutput->write(FIELD_TYPE_INT64 | FIELD_ID_BUCKET_SIZE, (long long)mBucketSizeNs);
186     // Fills the dimension path if not slicing by ALL.
187     if (!mSliceByPositionALL) {
188         if (!mDimensionsInWhat.empty()) {
189             uint64_t dimenPathToken = protoOutput->start(
190                     FIELD_TYPE_MESSAGE | FIELD_ID_DIMENSION_PATH_IN_WHAT);
191             writeDimensionPathToProto(mDimensionsInWhat, protoOutput);
192             protoOutput->end(dimenPathToken);
193         }
194         if (!mDimensionsInCondition.empty()) {
195             uint64_t dimenPathToken = protoOutput->start(
196                     FIELD_TYPE_MESSAGE | FIELD_ID_DIMENSION_PATH_IN_CONDITION);
197             writeDimensionPathToProto(mDimensionsInCondition, protoOutput);
198             protoOutput->end(dimenPathToken);
199         }
200     }
201 
202     uint64_t protoToken = protoOutput->start(FIELD_TYPE_MESSAGE | FIELD_ID_VALUE_METRICS);
203 
204     for (const auto& pair : mSkippedBuckets) {
205         uint64_t wrapperToken =
206                 protoOutput->start(FIELD_TYPE_MESSAGE | FIELD_COUNT_REPEATED | FIELD_ID_SKIPPED);
207         protoOutput->write(FIELD_TYPE_INT64 | FIELD_ID_SKIPPED_START_MILLIS,
208                            (long long)(NanoToMillis(pair.first)));
209         protoOutput->write(FIELD_TYPE_INT64 | FIELD_ID_SKIPPED_END_MILLIS,
210                            (long long)(NanoToMillis(pair.second)));
211         protoOutput->end(wrapperToken);
212     }
213     mSkippedBuckets.clear();
214 
215     for (const auto& pair : mPastBuckets) {
216         const MetricDimensionKey& dimensionKey = pair.first;
217         VLOG("  dimension key %s", dimensionKey.toString().c_str());
218         uint64_t wrapperToken =
219                 protoOutput->start(FIELD_TYPE_MESSAGE | FIELD_COUNT_REPEATED | FIELD_ID_DATA);
220 
221         // First fill dimension.
222         if (mSliceByPositionALL) {
223             uint64_t dimensionToken = protoOutput->start(
224                     FIELD_TYPE_MESSAGE | FIELD_ID_DIMENSION_IN_WHAT);
225             writeDimensionToProto(dimensionKey.getDimensionKeyInWhat(), str_set, protoOutput);
226             protoOutput->end(dimensionToken);
227             if (dimensionKey.hasDimensionKeyInCondition()) {
228                 uint64_t dimensionInConditionToken = protoOutput->start(
229                         FIELD_TYPE_MESSAGE | FIELD_ID_DIMENSION_IN_CONDITION);
230                 writeDimensionToProto(dimensionKey.getDimensionKeyInCondition(),
231                                       str_set, protoOutput);
232                 protoOutput->end(dimensionInConditionToken);
233             }
234         } else {
235             writeDimensionLeafNodesToProto(dimensionKey.getDimensionKeyInWhat(),
236                                            FIELD_ID_DIMENSION_LEAF_IN_WHAT, str_set, protoOutput);
237             if (dimensionKey.hasDimensionKeyInCondition()) {
238                 writeDimensionLeafNodesToProto(dimensionKey.getDimensionKeyInCondition(),
239                                                FIELD_ID_DIMENSION_LEAF_IN_CONDITION,
240                                                str_set, protoOutput);
241             }
242         }
243 
244         // Then fill bucket_info (ValueBucketInfo).
245         for (const auto& bucket : pair.second) {
246             uint64_t bucketInfoToken = protoOutput->start(
247                     FIELD_TYPE_MESSAGE | FIELD_COUNT_REPEATED | FIELD_ID_BUCKET_INFO);
248 
249             if (bucket.mBucketEndNs - bucket.mBucketStartNs != mBucketSizeNs) {
250                 protoOutput->write(FIELD_TYPE_INT64 | FIELD_ID_START_BUCKET_ELAPSED_MILLIS,
251                                    (long long)NanoToMillis(bucket.mBucketStartNs));
252                 protoOutput->write(FIELD_TYPE_INT64 | FIELD_ID_END_BUCKET_ELAPSED_MILLIS,
253                                    (long long)NanoToMillis(bucket.mBucketEndNs));
254             } else {
255                 protoOutput->write(FIELD_TYPE_INT64 | FIELD_ID_BUCKET_NUM,
256                                    (long long)(getBucketNumFromEndTimeNs(bucket.mBucketEndNs)));
257             }
258 
259             protoOutput->write(FIELD_TYPE_INT64 | FIELD_ID_VALUE, (long long)bucket.mValue);
260             protoOutput->end(bucketInfoToken);
261             VLOG("\t bucket [%lld - %lld] count: %lld", (long long)bucket.mBucketStartNs,
262                  (long long)bucket.mBucketEndNs, (long long)bucket.mValue);
263         }
264         protoOutput->end(wrapperToken);
265     }
266     protoOutput->end(protoToken);
267 
268     VLOG("metric %lld dump report now...", (long long)mMetricId);
269     mPastBuckets.clear();
270 }
271 
onConditionChangedLocked(const bool condition,const int64_t eventTimeNs)272 void ValueMetricProducer::onConditionChangedLocked(const bool condition,
273                                                    const int64_t eventTimeNs) {
274     mCondition = condition;
275 
276     if (eventTimeNs < mCurrentBucketStartTimeNs) {
277         VLOG("Skip event due to late arrival: %lld vs %lld", (long long)eventTimeNs,
278              (long long)mCurrentBucketStartTimeNs);
279         return;
280     }
281 
282     flushIfNeededLocked(eventTimeNs);
283 
284     if (mPullTagId != -1) {
285         vector<shared_ptr<LogEvent>> allData;
286         if (mStatsPullerManager->Pull(mPullTagId, eventTimeNs, &allData)) {
287             if (allData.size() == 0) {
288                 return;
289             }
290             for (const auto& data : allData) {
291                 onMatchedLogEventLocked(0, *data);
292             }
293         }
294         return;
295     }
296 }
297 
onDataPulled(const std::vector<std::shared_ptr<LogEvent>> & allData)298 void ValueMetricProducer::onDataPulled(const std::vector<std::shared_ptr<LogEvent>>& allData) {
299     std::lock_guard<std::mutex> lock(mMutex);
300 
301     if (mCondition == true || mConditionTrackerIndex < 0) {
302         if (allData.size() == 0) {
303             return;
304         }
305         // For scheduled pulled data, the effective event time is snap to the nearest
306         // bucket boundary to make bucket finalize.
307         int64_t realEventTime = allData.at(0)->GetElapsedTimestampNs();
308         int64_t eventTime = mTimeBaseNs +
309             ((realEventTime - mTimeBaseNs) / mBucketSizeNs) * mBucketSizeNs;
310 
311         mCondition = false;
312         for (const auto& data : allData) {
313             data->setElapsedTimestampNs(eventTime - 1);
314             onMatchedLogEventLocked(0, *data);
315         }
316 
317         mCondition = true;
318         for (const auto& data : allData) {
319             data->setElapsedTimestampNs(eventTime);
320             onMatchedLogEventLocked(0, *data);
321         }
322     }
323 }
324 
dumpStatesLocked(FILE * out,bool verbose) const325 void ValueMetricProducer::dumpStatesLocked(FILE* out, bool verbose) const {
326     if (mCurrentSlicedBucket.size() == 0) {
327         return;
328     }
329 
330     fprintf(out, "ValueMetric %lld dimension size %lu\n", (long long)mMetricId,
331             (unsigned long)mCurrentSlicedBucket.size());
332     if (verbose) {
333         for (const auto& it : mCurrentSlicedBucket) {
334             fprintf(out, "\t(what)%s\t(condition)%s  (value)%lld\n",
335                 it.first.getDimensionKeyInWhat().toString().c_str(),
336                 it.first.getDimensionKeyInCondition().toString().c_str(),
337                 (unsigned long long)it.second.sum);
338         }
339     }
340 }
341 
hitGuardRailLocked(const MetricDimensionKey & newKey)342 bool ValueMetricProducer::hitGuardRailLocked(const MetricDimensionKey& newKey) {
343     // ===========GuardRail==============
344     // 1. Report the tuple count if the tuple count > soft limit
345     if (mCurrentSlicedBucket.find(newKey) != mCurrentSlicedBucket.end()) {
346         return false;
347     }
348     if (mCurrentSlicedBucket.size() > mDimensionSoftLimit - 1) {
349         size_t newTupleCount = mCurrentSlicedBucket.size() + 1;
350         StatsdStats::getInstance().noteMetricDimensionSize(mConfigKey, mMetricId, newTupleCount);
351         // 2. Don't add more tuples, we are above the allowed threshold. Drop the data.
352         if (newTupleCount > mDimensionHardLimit) {
353             ALOGE("ValueMetric %lld dropping data for dimension key %s",
354                 (long long)mMetricId, newKey.toString().c_str());
355             return true;
356         }
357     }
358 
359     return false;
360 }
361 
onMatchedLogEventInternalLocked(const size_t matcherIndex,const MetricDimensionKey & eventKey,const ConditionKey & conditionKey,bool condition,const LogEvent & event)362 void ValueMetricProducer::onMatchedLogEventInternalLocked(
363         const size_t matcherIndex, const MetricDimensionKey& eventKey,
364         const ConditionKey& conditionKey, bool condition,
365         const LogEvent& event) {
366     int64_t eventTimeNs = event.GetElapsedTimestampNs();
367     if (eventTimeNs < mCurrentBucketStartTimeNs) {
368         VLOG("Skip event due to late arrival: %lld vs %lld", (long long)eventTimeNs,
369              (long long)mCurrentBucketStartTimeNs);
370         return;
371     }
372 
373     flushIfNeededLocked(eventTimeNs);
374 
375     if (hitGuardRailLocked(eventKey)) {
376         return;
377     }
378     Interval& interval = mCurrentSlicedBucket[eventKey];
379 
380     int error = 0;
381     const int64_t value = event.GetLong(mField, &error);
382     if (error < 0) {
383         return;
384     }
385 
386     if (mPullTagId != -1) { // for pulled events
387         if (mCondition == true) {
388             if (!interval.startUpdated) {
389                 interval.start = value;
390                 interval.startUpdated = true;
391             } else {
392                 // skip it if there is already value recorded for the start
393                 VLOG("Already recorded value for this dimension %s", eventKey.toString().c_str());
394             }
395         } else {
396             // Generally we expect value to be monotonically increasing.
397             // If not, take absolute value or drop it, based on config.
398             if (interval.startUpdated) {
399                 if (value >= interval.start) {
400                     interval.sum += (value - interval.start);
401                     interval.hasValue = true;
402                 } else {
403                     if (mUseAbsoluteValueOnReset) {
404                         interval.sum += value;
405                         interval.hasValue = true;
406                     } else {
407                         VLOG("Dropping data for atom %d, prev: %lld, now: %lld", mPullTagId,
408                              (long long)interval.start, (long long)value);
409                     }
410                 }
411                 interval.startUpdated = false;
412             } else {
413                 VLOG("No start for matching end %lld", (long long)value);
414                 interval.tainted += 1;
415             }
416         }
417     } else {    // for pushed events, only accumulate when condition is true
418         if (mCondition == true || mConditionTrackerIndex < 0) {
419             interval.sum += value;
420             interval.hasValue = true;
421         }
422     }
423 
424     long wholeBucketVal = interval.sum;
425     auto prev = mCurrentFullBucket.find(eventKey);
426     if (prev != mCurrentFullBucket.end()) {
427         wholeBucketVal += prev->second;
428     }
429     for (auto& tracker : mAnomalyTrackers) {
430         tracker->detectAndDeclareAnomaly(eventTimeNs, mCurrentBucketNum, eventKey, wholeBucketVal);
431     }
432 }
433 
flushIfNeededLocked(const int64_t & eventTimeNs)434 void ValueMetricProducer::flushIfNeededLocked(const int64_t& eventTimeNs) {
435     int64_t currentBucketEndTimeNs = getCurrentBucketEndTimeNs();
436 
437     if (eventTimeNs < currentBucketEndTimeNs) {
438         VLOG("eventTime is %lld, less than next bucket start time %lld", (long long)eventTimeNs,
439              (long long)(currentBucketEndTimeNs));
440         return;
441     }
442 
443     flushCurrentBucketLocked(eventTimeNs);
444 
445     int64_t numBucketsForward = 1 + (eventTimeNs - currentBucketEndTimeNs) / mBucketSizeNs;
446     mCurrentBucketStartTimeNs = currentBucketEndTimeNs + (numBucketsForward - 1) * mBucketSizeNs;
447     mCurrentBucketNum += numBucketsForward;
448 
449     if (numBucketsForward > 1) {
450         VLOG("Skipping forward %lld buckets", (long long)numBucketsForward);
451     }
452     VLOG("metric %lld: new bucket start time: %lld", (long long)mMetricId,
453          (long long)mCurrentBucketStartTimeNs);
454 }
455 
flushCurrentBucketLocked(const int64_t & eventTimeNs)456 void ValueMetricProducer::flushCurrentBucketLocked(const int64_t& eventTimeNs) {
457     VLOG("finalizing bucket for %ld, dumping %d slices", (long)mCurrentBucketStartTimeNs,
458          (int)mCurrentSlicedBucket.size());
459     int64_t fullBucketEndTimeNs = getCurrentBucketEndTimeNs();
460 
461     ValueBucket info;
462     info.mBucketStartNs = mCurrentBucketStartTimeNs;
463     if (eventTimeNs < fullBucketEndTimeNs) {
464         info.mBucketEndNs = eventTimeNs;
465     } else {
466         info.mBucketEndNs = fullBucketEndTimeNs;
467     }
468 
469     if (info.mBucketEndNs - mCurrentBucketStartTimeNs >= mMinBucketSizeNs) {
470         // The current bucket is large enough to keep.
471         int tainted = 0;
472         for (const auto& slice : mCurrentSlicedBucket) {
473             tainted += slice.second.tainted;
474             tainted += slice.second.startUpdated;
475             if (slice.second.hasValue) {
476                 info.mValue = slice.second.sum;
477                 // it will auto create new vector of ValuebucketInfo if the key is not found.
478                 auto& bucketList = mPastBuckets[slice.first];
479                 bucketList.push_back(info);
480             }
481         }
482         VLOG("%d tainted pairs in the bucket", tainted);
483     } else {
484         mSkippedBuckets.emplace_back(info.mBucketStartNs, info.mBucketEndNs);
485     }
486 
487     if (eventTimeNs > fullBucketEndTimeNs) {  // If full bucket, send to anomaly tracker.
488         // Accumulate partial buckets with current value and then send to anomaly tracker.
489         if (mCurrentFullBucket.size() > 0) {
490             for (const auto& slice : mCurrentSlicedBucket) {
491                 mCurrentFullBucket[slice.first] += slice.second.sum;
492             }
493             for (const auto& slice : mCurrentFullBucket) {
494                 for (auto& tracker : mAnomalyTrackers) {
495                     if (tracker != nullptr) {
496                         tracker->addPastBucket(slice.first, slice.second, mCurrentBucketNum);
497                     }
498                 }
499             }
500             mCurrentFullBucket.clear();
501         } else {
502             // Skip aggregating the partial buckets since there's no previous partial bucket.
503             for (const auto& slice : mCurrentSlicedBucket) {
504                 for (auto& tracker : mAnomalyTrackers) {
505                     if (tracker != nullptr) {
506                         tracker->addPastBucket(slice.first, slice.second.sum, mCurrentBucketNum);
507                     }
508                 }
509             }
510         }
511     } else {
512         // Accumulate partial bucket.
513         for (const auto& slice : mCurrentSlicedBucket) {
514             mCurrentFullBucket[slice.first] += slice.second.sum;
515         }
516     }
517 
518     // Reset counters
519     mCurrentSlicedBucket.clear();
520 }
521 
byteSizeLocked() const522 size_t ValueMetricProducer::byteSizeLocked() const {
523     size_t totalSize = 0;
524     for (const auto& pair : mPastBuckets) {
525         totalSize += pair.second.size() * kBucketSize;
526     }
527     return totalSize;
528 }
529 
530 }  // namespace statsd
531 }  // namespace os
532 }  // namespace android
533