| /* |
| * Copyright (C) 2017 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| #define STATSD_DEBUG false // STOPSHIP if true |
| #include "Log.h" |
| |
| #include "StatsLogProcessor.h" |
| |
| #include <android-base/file.h> |
| #include <cutils/multiuser.h> |
| #include <src/active_config_list.pb.h> |
| #include <src/experiment_ids.pb.h> |
| |
| #include "StatsService.h" |
| #include "android-base/stringprintf.h" |
| #include "external/StatsPullerManager.h" |
| #include "flags/FlagProvider.h" |
| #include "guardrail/StatsdStats.h" |
| #include "logd/LogEvent.h" |
| #include "metrics/CountMetricProducer.h" |
| #include "state/StateManager.h" |
| #include "stats_log_util.h" |
| #include "stats_util.h" |
| #include "statslog_statsd.h" |
| #include "storage/StorageManager.h" |
| #include "utils/api_tracing.h" |
| |
| using namespace android; |
| using android::base::StringPrintf; |
| using android::util::FIELD_COUNT_REPEATED; |
| using android::util::FIELD_TYPE_BOOL; |
| using android::util::FIELD_TYPE_FLOAT; |
| using android::util::FIELD_TYPE_INT32; |
| using android::util::FIELD_TYPE_INT64; |
| using android::util::FIELD_TYPE_MESSAGE; |
| using android::util::FIELD_TYPE_STRING; |
| using android::util::ProtoOutputStream; |
| using std::vector; |
| |
| namespace android { |
| namespace os { |
| namespace statsd { |
| |
| using aidl::android::os::IStatsQueryCallback; |
| |
| // for ConfigMetricsReportList |
| const int FIELD_ID_CONFIG_KEY = 1; |
| const int FIELD_ID_REPORTS = 2; |
| // for ConfigKey |
| const int FIELD_ID_UID = 1; |
| const int FIELD_ID_ID = 2; |
| const int FIELD_ID_REPORT_NUMBER = 3; |
| const int FIELD_ID_STATSD_STATS_ID = 4; |
| // for ConfigMetricsReport |
| // const int FIELD_ID_METRICS = 1; // written in MetricsManager.cpp |
| const int FIELD_ID_UID_MAP = 2; |
| const int FIELD_ID_LAST_REPORT_ELAPSED_NANOS = 3; |
| const int FIELD_ID_CURRENT_REPORT_ELAPSED_NANOS = 4; |
| const int FIELD_ID_LAST_REPORT_WALL_CLOCK_NANOS = 5; |
| const int FIELD_ID_CURRENT_REPORT_WALL_CLOCK_NANOS = 6; |
| const int FIELD_ID_DUMP_REPORT_REASON = 8; |
| const int FIELD_ID_STRINGS = 9; |
| const int FIELD_ID_DATA_CORRUPTED_REASON = 11; |
| const int FIELD_ID_ESTIMATED_DATA_BYTES = 12; |
| |
| // for ActiveConfigList |
| const int FIELD_ID_ACTIVE_CONFIG_LIST_CONFIG = 1; |
| |
| // for permissions checks |
| constexpr const char* kPermissionDump = "android.permission.DUMP"; |
| constexpr const char* kPermissionUsage = "android.permission.PACKAGE_USAGE_STATS"; |
| |
| #define NS_PER_HOUR 3600 * NS_PER_SEC |
| |
| #define STATS_ACTIVE_METRIC_DIR "/data/misc/stats-active-metric" |
| #define STATS_METADATA_DIR "/data/misc/stats-metadata" |
| |
| // Cool down period for writing data to disk to avoid overwriting files. |
| #define WRITE_DATA_COOL_DOWN_SEC 15 |
| |
| namespace { |
| |
| const char* getOnLogEventCallName(int32_t tagId) { |
| static std::string name; |
| // to avoid new string allocation on each call |
| name.reserve(30); |
| name = "OnLogEvent-" + std::to_string(tagId); |
| return name.c_str(); |
| } |
| |
| } // namespace |
| |
| StatsLogProcessor::StatsLogProcessor( |
| const sp<UidMap>& uidMap, const sp<StatsPullerManager>& pullerManager, |
| const sp<AlarmMonitor>& anomalyAlarmMonitor, const sp<AlarmMonitor>& periodicAlarmMonitor, |
| const int64_t timeBaseNs, const std::function<bool(const ConfigKey&)>& sendBroadcast, |
| const std::function<bool(const int&, const vector<int64_t>&)>& activateBroadcast, |
| const std::function<void(const ConfigKey&, const string&, const vector<int64_t>&)>& |
| sendRestrictedMetricsBroadcast, |
| const std::shared_ptr<LogEventFilter>& logEventFilter) |
| : mLastTtlTime(0), |
| mLastFlushRestrictedTime(0), |
| mLastDbGuardrailEnforcementTime(0), |
| mUidMap(uidMap), |
| mPullerManager(pullerManager), |
| mAnomalyAlarmMonitor(anomalyAlarmMonitor), |
| mPeriodicAlarmMonitor(periodicAlarmMonitor), |
| mLogEventFilter(logEventFilter), |
| mSendBroadcast(sendBroadcast), |
| mSendActivationBroadcast(activateBroadcast), |
| mSendRestrictedMetricsBroadcast(sendRestrictedMetricsBroadcast), |
| mTimeBaseNs(timeBaseNs), |
| mLargestTimestampSeen(0), |
| mLastTimestampSeen(0) { |
| mPullerManager->ForceClearPullerCache(); |
| StateManager::getInstance().updateLogSources(uidMap); |
| // It is safe called locked version at constructor - no concurrent access possible |
| updateLogEventFilterLocked(); |
| } |
| |
| StatsLogProcessor::~StatsLogProcessor() { |
| } |
| |
| static void flushProtoToBuffer(ProtoOutputStream& proto, vector<uint8_t>* outData) { |
| outData->clear(); |
| outData->resize(proto.size()); |
| size_t pos = 0; |
| sp<android::util::ProtoReader> reader = proto.data(); |
| while (reader->readBuffer() != NULL) { |
| size_t toRead = reader->currentToRead(); |
| std::memcpy(&((*outData)[pos]), reader->readBuffer(), toRead); |
| pos += toRead; |
| reader->move(toRead); |
| } |
| } |
| |
| void StatsLogProcessor::processFiredAnomalyAlarmsLocked( |
| const int64_t timestampNs, |
| unordered_set<sp<const InternalAlarm>, SpHash<InternalAlarm>>& alarmSet) { |
| for (const auto& itr : mMetricsManagers) { |
| itr.second->onAnomalyAlarmFired(timestampNs, alarmSet); |
| } |
| } |
| void StatsLogProcessor::onPeriodicAlarmFired( |
| const int64_t timestampNs, |
| unordered_set<sp<const InternalAlarm>, SpHash<InternalAlarm>>& alarmSet) { |
| std::lock_guard<std::mutex> lock(mMetricsMutex); |
| for (const auto& itr : mMetricsManagers) { |
| itr.second->onPeriodicAlarmFired(timestampNs, alarmSet); |
| } |
| } |
| |
| void StatsLogProcessor::mapIsolatedUidToHostUidIfNecessaryLocked(LogEvent* event) const { |
| if (std::pair<size_t, size_t> indexRange; event->hasAttributionChain(&indexRange)) { |
| vector<FieldValue>* const fieldValues = event->getMutableValues(); |
| for (size_t i = indexRange.first; i <= indexRange.second; i++) { |
| FieldValue& fieldValue = fieldValues->at(i); |
| if (isAttributionUidField(fieldValue)) { |
| const int hostUid = mUidMap->getHostUidOrSelf(fieldValue.mValue.int_value); |
| fieldValue.mValue.setInt(hostUid); |
| } |
| } |
| } else { |
| mapIsolatedUidsToHostUidInLogEvent(mUidMap, *event); |
| } |
| } |
| |
| void StatsLogProcessor::onIsolatedUidChangedEventLocked(const LogEvent& event) { |
| status_t err = NO_ERROR, err2 = NO_ERROR, err3 = NO_ERROR; |
| bool is_create = event.GetBool(3, &err); |
| auto parent_uid = int(event.GetLong(1, &err2)); |
| auto isolated_uid = int(event.GetLong(2, &err3)); |
| if (err == NO_ERROR && err2 == NO_ERROR && err3 == NO_ERROR) { |
| if (is_create) { |
| mUidMap->assignIsolatedUid(isolated_uid, parent_uid); |
| } else { |
| mUidMap->removeIsolatedUid(isolated_uid); |
| } |
| } else { |
| ALOGE("Failed to parse uid in the isolated uid change event."); |
| } |
| } |
| |
| void StatsLogProcessor::onBinaryPushStateChangedEventLocked(LogEvent* event) { |
| pid_t pid = event->GetPid(); |
| uid_t uid = event->GetUid(); |
| if (!checkPermissionForIds(kPermissionDump, pid, uid) || |
| !checkPermissionForIds(kPermissionUsage, pid, uid)) { |
| return; |
| } |
| // The Get* functions don't modify the status on success, they only write in |
| // failure statuses, so we can use one status variable for all calls then |
| // check if it is no longer NO_ERROR. |
| status_t err = NO_ERROR; |
| InstallTrainInfo trainInfo; |
| trainInfo.trainName = string(event->GetString(1 /*train name field id*/, &err)); |
| trainInfo.trainVersionCode = event->GetLong(2 /*train version field id*/, &err); |
| trainInfo.requiresStaging = event->GetBool(3 /*requires staging field id*/, &err); |
| trainInfo.rollbackEnabled = event->GetBool(4 /*rollback enabled field id*/, &err); |
| trainInfo.requiresLowLatencyMonitor = |
| event->GetBool(5 /*requires low latency monitor field id*/, &err); |
| trainInfo.status = int32_t(event->GetLong(6 /*state field id*/, &err)); |
| std::vector<uint8_t> trainExperimentIdBytes = |
| event->GetStorage(7 /*experiment ids field id*/, &err); |
| bool is_rollback = event->GetBool(10 /*is rollback field id*/, &err); |
| |
| if (err != NO_ERROR) { |
| ALOGE("Failed to parse fields in binary push state changed log event"); |
| return; |
| } |
| ExperimentIds trainExperimentIds; |
| if (!trainExperimentIds.ParseFromArray(trainExperimentIdBytes.data(), |
| trainExperimentIdBytes.size())) { |
| ALOGE("Failed to parse experimentids in binary push state changed."); |
| return; |
| } |
| trainInfo.experimentIds = {trainExperimentIds.experiment_id().begin(), |
| trainExperimentIds.experiment_id().end()}; |
| |
| // Update the train info on disk and get any data the logevent is missing. |
| getAndUpdateTrainInfoOnDisk(is_rollback, &trainInfo); |
| |
| std::vector<uint8_t> trainExperimentIdProto; |
| writeExperimentIdsToProto(trainInfo.experimentIds, &trainExperimentIdProto); |
| int32_t userId = multiuser_get_user_id(uid); |
| |
| event->updateValue(2 /*train version field id*/, trainInfo.trainVersionCode, LONG); |
| event->updateValue(7 /*experiment ids field id*/, trainExperimentIdProto, STORAGE); |
| event->updateValue(8 /*user id field id*/, userId, INT); |
| |
| // If this event is a rollback event, then the following bits in the event |
| // are invalid and we will need to update them with the values we pulled |
| // from disk. |
| if (is_rollback) { |
| int bit = trainInfo.requiresStaging ? 1 : 0; |
| event->updateValue(3 /*requires staging field id*/, bit, INT); |
| bit = trainInfo.rollbackEnabled ? 1 : 0; |
| event->updateValue(4 /*rollback enabled field id*/, bit, INT); |
| bit = trainInfo.requiresLowLatencyMonitor ? 1 : 0; |
| event->updateValue(5 /*requires low latency monitor field id*/, bit, INT); |
| } |
| } |
| |
| void StatsLogProcessor::getAndUpdateTrainInfoOnDisk(bool is_rollback, |
| InstallTrainInfo* trainInfo) { |
| // If the train name is empty, we don't know which train to attribute the |
| // event to, so return early. |
| if (trainInfo->trainName.empty()) { |
| return; |
| } |
| bool readTrainInfoSuccess = false; |
| InstallTrainInfo trainInfoOnDisk; |
| readTrainInfoSuccess = StorageManager::readTrainInfo(trainInfo->trainName, trainInfoOnDisk); |
| |
| bool resetExperimentIds = false; |
| if (readTrainInfoSuccess) { |
| // Keep the old train version if we received an empty version. |
| if (trainInfo->trainVersionCode == -1) { |
| trainInfo->trainVersionCode = trainInfoOnDisk.trainVersionCode; |
| } else if (trainInfo->trainVersionCode != trainInfoOnDisk.trainVersionCode) { |
| // Reset experiment ids if we receive a new non-empty train version. |
| resetExperimentIds = true; |
| } |
| |
| // Reset if we received a different experiment id. |
| if (!trainInfo->experimentIds.empty() && |
| (trainInfoOnDisk.experimentIds.empty() || |
| trainInfo->experimentIds.at(0) != trainInfoOnDisk.experimentIds[0])) { |
| resetExperimentIds = true; |
| } |
| } |
| |
| // Find the right experiment IDs |
| if ((!resetExperimentIds || is_rollback) && readTrainInfoSuccess) { |
| trainInfo->experimentIds = trainInfoOnDisk.experimentIds; |
| } |
| |
| if (!trainInfo->experimentIds.empty()) { |
| int64_t firstId = trainInfo->experimentIds.at(0); |
| auto& ids = trainInfo->experimentIds; |
| switch (trainInfo->status) { |
| case util::BINARY_PUSH_STATE_CHANGED__STATE__INSTALL_SUCCESS: |
| if (find(ids.begin(), ids.end(), firstId + 1) == ids.end()) { |
| ids.push_back(firstId + 1); |
| } |
| break; |
| case util::BINARY_PUSH_STATE_CHANGED__STATE__INSTALLER_ROLLBACK_INITIATED: |
| if (find(ids.begin(), ids.end(), firstId + 2) == ids.end()) { |
| ids.push_back(firstId + 2); |
| } |
| break; |
| case util::BINARY_PUSH_STATE_CHANGED__STATE__INSTALLER_ROLLBACK_SUCCESS: |
| if (find(ids.begin(), ids.end(), firstId + 3) == ids.end()) { |
| ids.push_back(firstId + 3); |
| } |
| break; |
| } |
| } |
| |
| // If this event is a rollback event, the following fields are invalid and |
| // need to be replaced by the fields stored to disk. |
| if (is_rollback) { |
| trainInfo->requiresStaging = trainInfoOnDisk.requiresStaging; |
| trainInfo->rollbackEnabled = trainInfoOnDisk.rollbackEnabled; |
| trainInfo->requiresLowLatencyMonitor = trainInfoOnDisk.requiresLowLatencyMonitor; |
| } |
| |
| StorageManager::writeTrainInfo(*trainInfo); |
| } |
| |
| void StatsLogProcessor::onWatchdogRollbackOccurredLocked(LogEvent* event) { |
| pid_t pid = event->GetPid(); |
| uid_t uid = event->GetUid(); |
| if (!checkPermissionForIds(kPermissionDump, pid, uid) || |
| !checkPermissionForIds(kPermissionUsage, pid, uid)) { |
| return; |
| } |
| // The Get* functions don't modify the status on success, they only write in |
| // failure statuses, so we can use one status variable for all calls then |
| // check if it is no longer NO_ERROR. |
| status_t err = NO_ERROR; |
| int32_t rollbackType = int32_t(event->GetInt(1 /*rollback type field id*/, &err)); |
| string packageName = string(event->GetString(2 /*package name field id*/, &err)); |
| |
| if (err != NO_ERROR) { |
| ALOGE("Failed to parse fields in watchdog rollback occurred log event"); |
| return; |
| } |
| |
| vector<int64_t> experimentIds = |
| processWatchdogRollbackOccurred(rollbackType, packageName); |
| vector<uint8_t> experimentIdProto; |
| writeExperimentIdsToProto(experimentIds, &experimentIdProto); |
| |
| event->updateValue(6 /*experiment ids field id*/, experimentIdProto, STORAGE); |
| } |
| |
| vector<int64_t> StatsLogProcessor::processWatchdogRollbackOccurred(const int32_t rollbackTypeIn, |
| const string& packageNameIn) { |
| // If the package name is empty, we can't attribute it to any train, so |
| // return early. |
| if (packageNameIn.empty()) { |
| return vector<int64_t>(); |
| } |
| bool readTrainInfoSuccess = false; |
| InstallTrainInfo trainInfoOnDisk; |
| // We use the package name of the event as the train name. |
| readTrainInfoSuccess = StorageManager::readTrainInfo(packageNameIn, trainInfoOnDisk); |
| |
| if (!readTrainInfoSuccess) { |
| return vector<int64_t>(); |
| } |
| |
| if (trainInfoOnDisk.experimentIds.empty()) { |
| return vector<int64_t>(); |
| } |
| |
| int64_t firstId = trainInfoOnDisk.experimentIds[0]; |
| auto& ids = trainInfoOnDisk.experimentIds; |
| switch (rollbackTypeIn) { |
| case util::WATCHDOG_ROLLBACK_OCCURRED__ROLLBACK_TYPE__ROLLBACK_INITIATE: |
| if (find(ids.begin(), ids.end(), firstId + 4) == ids.end()) { |
| ids.push_back(firstId + 4); |
| } |
| StorageManager::writeTrainInfo(trainInfoOnDisk); |
| break; |
| case util::WATCHDOG_ROLLBACK_OCCURRED__ROLLBACK_TYPE__ROLLBACK_SUCCESS: |
| if (find(ids.begin(), ids.end(), firstId + 5) == ids.end()) { |
| ids.push_back(firstId + 5); |
| } |
| StorageManager::writeTrainInfo(trainInfoOnDisk); |
| break; |
| } |
| |
| return trainInfoOnDisk.experimentIds; |
| } |
| |
| void StatsLogProcessor::resetConfigs() { |
| std::lock_guard<std::mutex> lock(mMetricsMutex); |
| resetConfigsLocked(getElapsedRealtimeNs()); |
| } |
| |
| void StatsLogProcessor::resetConfigsLocked(const int64_t timestampNs) { |
| std::vector<ConfigKey> configKeys; |
| for (auto it = mMetricsManagers.begin(); it != mMetricsManagers.end(); it++) { |
| configKeys.push_back(it->first); |
| } |
| resetConfigsLocked(timestampNs, configKeys); |
| } |
| |
| void StatsLogProcessor::OnLogEvent(LogEvent* event) { |
| ATRACE_NAME(getOnLogEventCallName(event->GetTagId())); |
| OnLogEvent(event, getElapsedRealtimeNs()); |
| } |
| |
| void StatsLogProcessor::OnLogEvent(LogEvent* event, int64_t elapsedRealtimeNs) { |
| const int64_t eventElapsedTimeNs = event->GetElapsedTimestampNs(); |
| const int atomId = event->GetTagId(); |
| |
| if (!event->isValid()) { |
| StatsdStats::getInstance().noteAtomError(atomId); |
| return; |
| } |
| |
| std::lock_guard<std::mutex> lock(mMetricsMutex); |
| |
| // Hard-coded logic to update train info on disk and fill in any information |
| // this log event may be missing. |
| if (atomId == util::BINARY_PUSH_STATE_CHANGED) { |
| onBinaryPushStateChangedEventLocked(event); |
| } |
| |
| // Hard-coded logic to update experiment ids on disk for certain rollback |
| // types and fill the rollback atom with experiment ids |
| if (atomId == util::WATCHDOG_ROLLBACK_OCCURRED) { |
| onWatchdogRollbackOccurredLocked(event); |
| } |
| |
| if (mPrintAllLogs) { |
| ALOGI("%s", event->ToString().c_str()); |
| } |
| resetIfConfigTtlExpiredLocked(eventElapsedTimeNs); |
| |
| // Hard-coded logic to update the isolated uid's in the uid-map. |
| // The field numbers need to be currently updated by hand with atoms.proto |
| if (atomId == util::ISOLATED_UID_CHANGED) { |
| onIsolatedUidChangedEventLocked(*event); |
| } else { |
| // Map the isolated uid to host uid if necessary. |
| mapIsolatedUidToHostUidIfNecessaryLocked(event); |
| } |
| |
| StateManager::getInstance().onLogEvent(*event); |
| |
| if (mMetricsManagers.empty()) { |
| return; |
| } |
| |
| bool fireAlarm = false; |
| { |
| std::lock_guard<std::mutex> anomalyLock(mAnomalyAlarmMutex); |
| if (mNextAnomalyAlarmTime != 0 && |
| MillisToNano(mNextAnomalyAlarmTime) <= elapsedRealtimeNs) { |
| mNextAnomalyAlarmTime = 0; |
| VLOG("informing anomaly alarm at time %lld", (long long)elapsedRealtimeNs); |
| fireAlarm = true; |
| } |
| } |
| if (fireAlarm) { |
| informAnomalyAlarmFiredLocked(NanoToMillis(elapsedRealtimeNs)); |
| } |
| |
| const int64_t curTimeSec = NanoToSeconds(elapsedRealtimeNs); |
| if (curTimeSec - mLastPullerCacheClearTimeSec > StatsdStats::kPullerCacheClearIntervalSec) { |
| mPullerManager->ClearPullerCacheIfNecessary(curTimeSec * NS_PER_SEC); |
| mLastPullerCacheClearTimeSec = curTimeSec; |
| } |
| |
| flushRestrictedDataIfNecessaryLocked(elapsedRealtimeNs); |
| enforceDataTtlsIfNecessaryLocked(getWallClockNs(), elapsedRealtimeNs); |
| enforceDbGuardrailsIfNecessaryLocked(getWallClockNs(), elapsedRealtimeNs); |
| |
| if (!validateAppBreadcrumbEvent(*event)) { |
| return; |
| } |
| |
| std::unordered_set<int> uidsWithActiveConfigsChanged; |
| std::unordered_map<int, std::vector<int64_t>> activeConfigsPerUid; |
| |
| // pass the event to metrics managers. |
| for (auto& pair : mMetricsManagers) { |
| if (event->isRestricted() && !pair.second->hasRestrictedMetricsDelegate()) { |
| continue; |
| } |
| int uid = pair.first.GetUid(); |
| int64_t configId = pair.first.GetId(); |
| bool isPrevActive = pair.second->isActive(); |
| pair.second->onLogEvent(*event); |
| bool isCurActive = pair.second->isActive(); |
| // Map all active configs by uid. |
| if (isCurActive) { |
| auto activeConfigs = activeConfigsPerUid.find(uid); |
| if (activeConfigs != activeConfigsPerUid.end()) { |
| activeConfigs->second.push_back(configId); |
| } else { |
| vector<int64_t> newActiveConfigs; |
| newActiveConfigs.push_back(configId); |
| activeConfigsPerUid[uid] = newActiveConfigs; |
| } |
| } |
| // The activation state of this config changed. |
| if (isPrevActive != isCurActive) { |
| VLOG("Active status changed for uid %d", uid); |
| uidsWithActiveConfigsChanged.insert(uid); |
| StatsdStats::getInstance().noteActiveStatusChanged(pair.first, isCurActive); |
| } |
| flushIfNecessaryLocked(pair.first, *(pair.second)); |
| } |
| |
| // Don't use the event timestamp for the guardrail. |
| for (int uid : uidsWithActiveConfigsChanged) { |
| // Send broadcast so that receivers can pull data. |
| auto lastBroadcastTime = mLastActivationBroadcastTimes.find(uid); |
| if (lastBroadcastTime != mLastActivationBroadcastTimes.end()) { |
| if (elapsedRealtimeNs - lastBroadcastTime->second < |
| StatsdStats::kMinActivationBroadcastPeriodNs) { |
| StatsdStats::getInstance().noteActivationBroadcastGuardrailHit(uid); |
| VLOG("StatsD would've sent an activation broadcast but the rate limit stopped us."); |
| return; |
| } |
| } |
| auto activeConfigs = activeConfigsPerUid.find(uid); |
| if (activeConfigs != activeConfigsPerUid.end()) { |
| if (mSendActivationBroadcast(uid, activeConfigs->second)) { |
| VLOG("StatsD sent activation notice for uid %d", uid); |
| mLastActivationBroadcastTimes[uid] = elapsedRealtimeNs; |
| } |
| } else { |
| std::vector<int64_t> emptyActiveConfigs; |
| if (mSendActivationBroadcast(uid, emptyActiveConfigs)) { |
| VLOG("StatsD sent EMPTY activation notice for uid %d", uid); |
| mLastActivationBroadcastTimes[uid] = elapsedRealtimeNs; |
| } |
| } |
| } |
| } |
| |
| void StatsLogProcessor::GetActiveConfigs(const int uid, vector<int64_t>& outActiveConfigs) { |
| std::lock_guard<std::mutex> lock(mMetricsMutex); |
| GetActiveConfigsLocked(uid, outActiveConfigs); |
| } |
| |
| void StatsLogProcessor::GetActiveConfigsLocked(const int uid, vector<int64_t>& outActiveConfigs) { |
| outActiveConfigs.clear(); |
| for (auto& pair : mMetricsManagers) { |
| if (pair.first.GetUid() == uid && pair.second->isActive()) { |
| outActiveConfigs.push_back(pair.first.GetId()); |
| } |
| } |
| } |
| |
| void StatsLogProcessor::OnConfigUpdated(const int64_t timestampNs, const int64_t wallClockNs, |
| const ConfigKey& key, const StatsdConfig& config, |
| bool modularUpdate) { |
| std::lock_guard<std::mutex> lock(mMetricsMutex); |
| WriteDataToDiskLocked(key, timestampNs, wallClockNs, CONFIG_UPDATED, NO_TIME_CONSTRAINTS); |
| OnConfigUpdatedLocked(timestampNs, key, config, modularUpdate); |
| } |
| |
| void StatsLogProcessor::OnConfigUpdated(const int64_t timestampNs, const ConfigKey& key, |
| const StatsdConfig& config, bool modularUpdate) { |
| OnConfigUpdated(timestampNs, getWallClockNs(), key, config, modularUpdate); |
| } |
| |
| void StatsLogProcessor::OnConfigUpdatedLocked(const int64_t timestampNs, const ConfigKey& key, |
| const StatsdConfig& config, bool modularUpdate) { |
| VLOG("Updated configuration for key %s", key.ToString().c_str()); |
| const auto& it = mMetricsManagers.find(key); |
| bool configValid = false; |
| if (isAtLeastU() && it != mMetricsManagers.end()) { |
| if (it->second->hasRestrictedMetricsDelegate() != |
| config.has_restricted_metrics_delegate_package_name()) { |
| // Not a modular update if has_restricted_metrics_delegate changes |
| modularUpdate = false; |
| } |
| if (!modularUpdate && it->second->hasRestrictedMetricsDelegate()) { |
| StatsdStats::getInstance().noteDbDeletionConfigUpdated(key); |
| // Always delete the old db if restricted metrics config is not a |
| // modular update. |
| dbutils::deleteDb(key); |
| } |
| } |
| // Create new config if this is not a modular update or if this is a new config. |
| if (!modularUpdate || it == mMetricsManagers.end()) { |
| sp<MetricsManager> newMetricsManager = |
| new MetricsManager(key, config, mTimeBaseNs, timestampNs, mUidMap, mPullerManager, |
| mAnomalyAlarmMonitor, mPeriodicAlarmMonitor); |
| configValid = newMetricsManager->isConfigValid(); |
| if (configValid) { |
| newMetricsManager->init(); |
| newMetricsManager->refreshTtl(timestampNs); |
| // Sdk check for U+ is unnecessary because config with restricted metrics delegate |
| // will be invalid on non U+ devices. |
| if (newMetricsManager->hasRestrictedMetricsDelegate()) { |
| mSendRestrictedMetricsBroadcast(key, |
| newMetricsManager->getRestrictedMetricsDelegate(), |
| newMetricsManager->getAllMetricIds()); |
| string err; |
| if (!dbutils::updateDeviceInfoTable(key, err)) { |
| ALOGE("Failed to create device_info table for configKey %s, err: %s", |
| key.ToString().c_str(), err.c_str()); |
| StatsdStats::getInstance().noteDeviceInfoTableCreationFailed(key); |
| } |
| } else if (it != mMetricsManagers.end() && it->second->hasRestrictedMetricsDelegate()) { |
| mSendRestrictedMetricsBroadcast(key, it->second->getRestrictedMetricsDelegate(), |
| {}); |
| } |
| mMetricsManagers[key] = newMetricsManager; |
| VLOG("StatsdConfig valid"); |
| } |
| } else { |
| // Preserve the existing MetricsManager, update necessary components and metadata in place. |
| configValid = it->second->updateConfig(config, mTimeBaseNs, timestampNs, |
| mAnomalyAlarmMonitor, mPeriodicAlarmMonitor); |
| if (configValid && it->second->hasRestrictedMetricsDelegate()) { |
| mSendRestrictedMetricsBroadcast(key, it->second->getRestrictedMetricsDelegate(), |
| it->second->getAllMetricIds()); |
| } |
| } |
| |
| if (configValid && !config.has_restricted_metrics_delegate_package_name()) { |
| // We do not need to track uid map changes for restricted metrics since the uidmap is not |
| // stored in the sqlite db. |
| mUidMap->OnConfigUpdated(key); |
| } else if (configValid && config.has_restricted_metrics_delegate_package_name()) { |
| mUidMap->OnConfigRemoved(key); |
| } |
| if (!configValid) { |
| // If there is any error in the config, don't use it. |
| // Remove any existing config with the same key. |
| ALOGE("StatsdConfig NOT valid"); |
| // Send an empty restricted metrics broadcast if the previous config was restricted. |
| if (isAtLeastU() && it != mMetricsManagers.end() && |
| it->second->hasRestrictedMetricsDelegate()) { |
| mSendRestrictedMetricsBroadcast(key, it->second->getRestrictedMetricsDelegate(), {}); |
| StatsdStats::getInstance().noteDbConfigInvalid(key); |
| dbutils::deleteDb(key); |
| } |
| mMetricsManagers.erase(key); |
| mUidMap->OnConfigRemoved(key); |
| } |
| |
| updateLogEventFilterLocked(); |
| } |
| |
| size_t StatsLogProcessor::GetMetricsSize(const ConfigKey& key) const { |
| std::lock_guard<std::mutex> lock(mMetricsMutex); |
| auto it = mMetricsManagers.find(key); |
| if (it == mMetricsManagers.end()) { |
| ALOGW("Config source %s does not exist", key.ToString().c_str()); |
| return 0; |
| } |
| return it->second->byteSize(); |
| } |
| |
| void StatsLogProcessor::dumpStates(int out, bool verbose) const { |
| std::lock_guard<std::mutex> lock(mMetricsMutex); |
| dprintf(out, "MetricsManager count: %lu\n", (unsigned long)mMetricsManagers.size()); |
| for (const auto& metricsManager : mMetricsManagers) { |
| metricsManager.second->dumpStates(out, verbose); |
| } |
| } |
| |
| /* |
| * onDumpReport dumps serialized ConfigMetricsReportList into proto. |
| */ |
| void StatsLogProcessor::onDumpReport(const ConfigKey& key, const int64_t dumpTimeStampNs, |
| const int64_t wallClockNs, |
| const bool include_current_partial_bucket, |
| const bool erase_data, const DumpReportReason dumpReportReason, |
| const DumpLatency dumpLatency, ProtoOutputStream* proto) { |
| std::lock_guard<std::mutex> lock(mMetricsMutex); |
| |
| auto it = mMetricsManagers.find(key); |
| if (it != mMetricsManagers.end() && it->second->hasRestrictedMetricsDelegate()) { |
| VLOG("Unexpected call to StatsLogProcessor::onDumpReport for restricted metrics."); |
| return; |
| } |
| |
| // Start of ConfigKey. |
| uint64_t configKeyToken = proto->start(FIELD_TYPE_MESSAGE | FIELD_ID_CONFIG_KEY); |
| proto->write(FIELD_TYPE_INT32 | FIELD_ID_UID, key.GetUid()); |
| proto->write(FIELD_TYPE_INT64 | FIELD_ID_ID, (long long)key.GetId()); |
| proto->end(configKeyToken); |
| // End of ConfigKey. |
| |
| bool keepFile = false; |
| if (it != mMetricsManagers.end() && it->second->shouldPersistLocalHistory()) { |
| keepFile = true; |
| } |
| |
| // Then, check stats-data directory to see there's any file containing |
| // ConfigMetricsReport from previous shutdowns to concatenate to reports. |
| StorageManager::appendConfigMetricsReport( |
| key, proto, erase_data && !keepFile /* should remove file after appending it */, |
| dumpReportReason == ADB_DUMP /*if caller is adb*/); |
| |
| if (it != mMetricsManagers.end()) { |
| // This allows another broadcast to be sent within the rate-limit period if we get close to |
| // filling the buffer again soon. |
| mLastBroadcastTimes.erase(key); |
| |
| vector<uint8_t> buffer; |
| onConfigMetricsReportLocked(key, dumpTimeStampNs, wallClockNs, |
| include_current_partial_bucket, erase_data, dumpReportReason, |
| dumpLatency, false /* is this data going to be saved on disk */, |
| &buffer); |
| proto->write(FIELD_TYPE_MESSAGE | FIELD_COUNT_REPEATED | FIELD_ID_REPORTS, |
| reinterpret_cast<char*>(buffer.data()), buffer.size()); |
| } else { |
| ALOGW("Config source %s does not exist", key.ToString().c_str()); |
| } |
| |
| if (erase_data) { |
| ++mDumpReportNumbers[key]; |
| } |
| proto->write(FIELD_TYPE_INT32 | FIELD_ID_REPORT_NUMBER, mDumpReportNumbers[key]); |
| |
| proto->write(FIELD_TYPE_INT32 | FIELD_ID_STATSD_STATS_ID, |
| StatsdStats::getInstance().getStatsdStatsId()); |
| if (erase_data) { |
| StatsdStats::getInstance().noteMetricsReportSent(key, proto->size(), |
| mDumpReportNumbers[key]); |
| } |
| } |
| |
| /* |
| * onDumpReport dumps serialized ConfigMetricsReportList into outData. |
| */ |
| void StatsLogProcessor::onDumpReport(const ConfigKey& key, const int64_t dumpTimeStampNs, |
| const int64_t wallClockNs, |
| const bool include_current_partial_bucket, |
| const bool erase_data, const DumpReportReason dumpReportReason, |
| const DumpLatency dumpLatency, vector<uint8_t>* outData) { |
| ProtoOutputStream proto; |
| onDumpReport(key, dumpTimeStampNs, wallClockNs, include_current_partial_bucket, erase_data, |
| dumpReportReason, dumpLatency, &proto); |
| |
| if (outData != nullptr) { |
| flushProtoToBuffer(proto, outData); |
| VLOG("output data size %zu", outData->size()); |
| } |
| } |
| |
| /* |
| * For test use only. Excludes wallclockNs. |
| * onDumpReport dumps serialized ConfigMetricsReportList into outData. |
| */ |
| void StatsLogProcessor::onDumpReport(const ConfigKey& key, const int64_t dumpTimeStampNs, |
| const bool include_current_partial_bucket, |
| const bool erase_data, const DumpReportReason dumpReportReason, |
| const DumpLatency dumpLatency, vector<uint8_t>* outData) { |
| onDumpReport(key, dumpTimeStampNs, getWallClockNs(), include_current_partial_bucket, erase_data, |
| dumpReportReason, dumpLatency, outData); |
| } |
| |
| /* |
| * onConfigMetricsReportLocked dumps serialized ConfigMetricsReport into outData. |
| */ |
| void StatsLogProcessor::onConfigMetricsReportLocked( |
| const ConfigKey& key, const int64_t dumpTimeStampNs, const int64_t wallClockNs, |
| const bool include_current_partial_bucket, const bool erase_data, |
| const DumpReportReason dumpReportReason, const DumpLatency dumpLatency, |
| const bool dataSavedOnDisk, vector<uint8_t>* buffer) { |
| // We already checked whether key exists in mMetricsManagers in |
| // WriteDataToDisk. |
| auto it = mMetricsManagers.find(key); |
| if (it == mMetricsManagers.end()) { |
| return; |
| } |
| if (it->second->hasRestrictedMetricsDelegate()) { |
| VLOG("Unexpected call to StatsLogProcessor::onConfigMetricsReportLocked for restricted " |
| "metrics."); |
| // Do not call onDumpReport for restricted metrics. |
| return; |
| } |
| |
| // get & forward queue overflow stats to StateManager only when |
| // there is a metric report to be collected, the data loss flags |
| // are not used otherwise |
| processQueueOverflowStatsLocked(); |
| |
| int64_t lastReportTimeNs = it->second->getLastReportTimeNs(); |
| int64_t lastReportWallClockNs = it->second->getLastReportWallClockNs(); |
| |
| std::set<string> strSet; |
| std::set<int32_t> usedUids; |
| |
| int64_t totalSize = it->second->byteSize(); |
| |
| ProtoOutputStream tempProto; |
| // First, fill in ConfigMetricsReport using current data on memory, which |
| // starts from filling in StatsLogReport's. |
| it->second->onDumpReport(dumpTimeStampNs, wallClockNs, include_current_partial_bucket, |
| erase_data, dumpLatency, &strSet, usedUids, &tempProto); |
| |
| // Fill in UidMap if there is at least one metric to report. |
| // This skips the uid map if it's an empty config. |
| if (it->second->getNumMetrics() > 0) { |
| uint64_t uidMapToken = tempProto.start(FIELD_TYPE_MESSAGE | FIELD_ID_UID_MAP); |
| UidMapOptions uidMapOptions = it->second->getUidMapOptions(); |
| uidMapOptions.usedUids = std::move(usedUids); |
| mUidMap->appendUidMap(dumpTimeStampNs, key, uidMapOptions, |
| it->second->hashStringInReport() ? &strSet : nullptr, &tempProto); |
| tempProto.end(uidMapToken); |
| } |
| |
| // Fill in the timestamps. |
| tempProto.write(FIELD_TYPE_INT64 | FIELD_ID_LAST_REPORT_ELAPSED_NANOS, |
| (long long)lastReportTimeNs); |
| tempProto.write(FIELD_TYPE_INT64 | FIELD_ID_CURRENT_REPORT_ELAPSED_NANOS, |
| (long long)dumpTimeStampNs); |
| tempProto.write(FIELD_TYPE_INT64 | FIELD_ID_LAST_REPORT_WALL_CLOCK_NANOS, |
| (long long)lastReportWallClockNs); |
| tempProto.write(FIELD_TYPE_INT64 | FIELD_ID_CURRENT_REPORT_WALL_CLOCK_NANOS, |
| (long long)wallClockNs); |
| // Dump report reason |
| tempProto.write(FIELD_TYPE_INT32 | FIELD_ID_DUMP_REPORT_REASON, dumpReportReason); |
| |
| for (const auto& str : strSet) { |
| tempProto.write(FIELD_TYPE_STRING | FIELD_COUNT_REPEATED | FIELD_ID_STRINGS, str); |
| } |
| |
| // Data corrupted reason |
| writeDataCorruptedReasons(tempProto, FIELD_ID_DATA_CORRUPTED_REASON, |
| StatsdStats::getInstance().hasEventQueueOverflow(), |
| StatsdStats::getInstance().hasSocketLoss()); |
| |
| // Estimated memory bytes |
| tempProto.write(FIELD_TYPE_INT64 | FIELD_ID_ESTIMATED_DATA_BYTES, totalSize); |
| |
| flushProtoToBuffer(tempProto, buffer); |
| |
| // save buffer to disk if needed |
| if (erase_data && !dataSavedOnDisk && it->second->shouldPersistLocalHistory()) { |
| VLOG("save history to disk"); |
| string file_name = StorageManager::getDataHistoryFileName((long)getWallClockSec(), |
| key.GetUid(), key.GetId()); |
| StorageManager::writeFile(file_name.c_str(), buffer->data(), buffer->size()); |
| } |
| } |
| |
| void StatsLogProcessor::resetConfigsLocked(const int64_t timestampNs, |
| const std::vector<ConfigKey>& configs) { |
| for (const auto& key : configs) { |
| StatsdConfig config; |
| if (StorageManager::readConfigFromDisk(key, &config)) { |
| // Force a full update when resetting a config. |
| OnConfigUpdatedLocked(timestampNs, key, config, /*modularUpdate=*/false); |
| StatsdStats::getInstance().noteConfigReset(key); |
| } else { |
| ALOGE("Failed to read backup config from disk for : %s", key.ToString().c_str()); |
| auto it = mMetricsManagers.find(key); |
| if (it != mMetricsManagers.end()) { |
| it->second->refreshTtl(timestampNs); |
| } |
| } |
| } |
| } |
| |
| void StatsLogProcessor::resetIfConfigTtlExpiredLocked(const int64_t eventTimeNs) { |
| std::vector<ConfigKey> configKeysTtlExpired; |
| for (auto it = mMetricsManagers.begin(); it != mMetricsManagers.end(); it++) { |
| if (it->second != nullptr && !it->second->isInTtl(eventTimeNs)) { |
| configKeysTtlExpired.push_back(it->first); |
| } |
| } |
| if (configKeysTtlExpired.size() > 0) { |
| WriteDataToDiskLocked(CONFIG_RESET, NO_TIME_CONSTRAINTS, getElapsedRealtimeNs(), |
| getWallClockNs()); |
| resetConfigsLocked(eventTimeNs, configKeysTtlExpired); |
| } |
| } |
| |
| void StatsLogProcessor::OnConfigRemoved(const ConfigKey& key) { |
| std::lock_guard<std::mutex> lock(mMetricsMutex); |
| auto it = mMetricsManagers.find(key); |
| if (it != mMetricsManagers.end()) { |
| WriteDataToDiskLocked(key, getElapsedRealtimeNs(), getWallClockNs(), CONFIG_REMOVED, |
| NO_TIME_CONSTRAINTS); |
| if (isAtLeastU() && it->second->hasRestrictedMetricsDelegate()) { |
| StatsdStats::getInstance().noteDbDeletionConfigRemoved(key); |
| dbutils::deleteDb(key); |
| mSendRestrictedMetricsBroadcast(key, it->second->getRestrictedMetricsDelegate(), {}); |
| } |
| mMetricsManagers.erase(it); |
| mUidMap->OnConfigRemoved(key); |
| } |
| StatsdStats::getInstance().noteConfigRemoved(key); |
| |
| mLastBroadcastTimes.erase(key); |
| mLastByteSizeTimes.erase(key); |
| mDumpReportNumbers.erase(key); |
| |
| int uid = key.GetUid(); |
| bool lastConfigForUid = true; |
| for (const auto& it : mMetricsManagers) { |
| if (it.first.GetUid() == uid) { |
| lastConfigForUid = false; |
| break; |
| } |
| } |
| if (lastConfigForUid) { |
| mLastActivationBroadcastTimes.erase(uid); |
| } |
| |
| if (mMetricsManagers.empty()) { |
| mPullerManager->ForceClearPullerCache(); |
| } |
| |
| updateLogEventFilterLocked(); |
| } |
| |
| // TODO(b/267501143): Add unit tests when metric producer is ready |
| void StatsLogProcessor::enforceDataTtlsIfNecessaryLocked(const int64_t wallClockNs, |
| const int64_t elapsedRealtimeNs) { |
| if (!isAtLeastU()) { |
| return; |
| } |
| if (elapsedRealtimeNs - mLastTtlTime < StatsdStats::kMinTtlCheckPeriodNs) { |
| return; |
| } |
| enforceDataTtlsLocked(wallClockNs, elapsedRealtimeNs); |
| } |
| |
| void StatsLogProcessor::flushRestrictedDataIfNecessaryLocked(const int64_t elapsedRealtimeNs) { |
| if (!isAtLeastU()) { |
| return; |
| } |
| if (elapsedRealtimeNs - mLastFlushRestrictedTime < StatsdStats::kMinFlushRestrictedPeriodNs) { |
| return; |
| } |
| flushRestrictedDataLocked(elapsedRealtimeNs); |
| } |
| |
| void StatsLogProcessor::querySql(const string& sqlQuery, const int32_t minSqlClientVersion, |
| const optional<vector<uint8_t>>& policyConfig, |
| const shared_ptr<IStatsQueryCallback>& callback, |
| const int64_t configId, const string& configPackage, |
| const int32_t callingUid) { |
| std::lock_guard<std::mutex> lock(mMetricsMutex); |
| string err = ""; |
| |
| if (!isAtLeastU()) { |
| ALOGW("Restricted metrics query invoked on U- device"); |
| StatsdStats::getInstance().noteQueryRestrictedMetricFailed( |
| configId, configPackage, std::nullopt, callingUid, |
| InvalidQueryReason(FLAG_DISABLED)); |
| return; |
| } |
| |
| const int64_t elapsedRealtimeNs = getElapsedRealtimeNs(); |
| |
| // TODO(b/268416460): validate policyConfig here |
| |
| if (minSqlClientVersion > dbutils::getDbVersion()) { |
| callback->sendFailure(StringPrintf( |
| "Unsupported sqlite version. Installed Version: %d, Requested Version: %d.", |
| dbutils::getDbVersion(), minSqlClientVersion)); |
| StatsdStats::getInstance().noteQueryRestrictedMetricFailed( |
| configId, configPackage, std::nullopt, callingUid, |
| InvalidQueryReason(UNSUPPORTED_SQLITE_VERSION)); |
| return; |
| } |
| |
| set<int32_t> configPackageUids; |
| const auto& uidMapItr = UidMap::sAidToUidMapping.find(configPackage); |
| if (uidMapItr != UidMap::sAidToUidMapping.end()) { |
| configPackageUids.insert(uidMapItr->second); |
| } else { |
| configPackageUids = mUidMap->getAppUid(configPackage); |
| } |
| |
| InvalidQueryReason invalidQueryReason; |
| set<ConfigKey> keysToQuery = getRestrictedConfigKeysToQueryLocked( |
| callingUid, configId, configPackageUids, err, invalidQueryReason); |
| |
| if (keysToQuery.empty()) { |
| callback->sendFailure(err); |
| StatsdStats::getInstance().noteQueryRestrictedMetricFailed( |
| configId, configPackage, std::nullopt, callingUid, |
| InvalidQueryReason(invalidQueryReason)); |
| return; |
| } |
| |
| if (keysToQuery.size() > 1) { |
| err = "Ambiguous ConfigKey"; |
| callback->sendFailure(err); |
| StatsdStats::getInstance().noteQueryRestrictedMetricFailed( |
| configId, configPackage, std::nullopt, callingUid, |
| InvalidQueryReason(AMBIGUOUS_CONFIG_KEY)); |
| return; |
| } |
| |
| flushRestrictedDataLocked(elapsedRealtimeNs); |
| enforceDataTtlsLocked(getWallClockNs(), elapsedRealtimeNs); |
| |
| std::vector<std::vector<std::string>> rows; |
| std::vector<int32_t> columnTypes; |
| std::vector<string> columnNames; |
| if (!dbutils::query(*(keysToQuery.begin()), sqlQuery, rows, columnTypes, columnNames, err)) { |
| callback->sendFailure(StringPrintf("failed to query db %s:", err.c_str())); |
| StatsdStats::getInstance().noteQueryRestrictedMetricFailed( |
| configId, configPackage, keysToQuery.begin()->GetUid(), callingUid, |
| InvalidQueryReason(QUERY_FAILURE), err.c_str()); |
| return; |
| } |
| |
| vector<string> queryData; |
| queryData.reserve(rows.size() * columnNames.size()); |
| // TODO(b/268415904): avoid this vector transformation. |
| if (columnNames.size() != columnTypes.size()) { |
| callback->sendFailure("Inconsistent row sizes"); |
| StatsdStats::getInstance().noteQueryRestrictedMetricFailed( |
| configId, configPackage, keysToQuery.begin()->GetUid(), callingUid, |
| InvalidQueryReason(INCONSISTENT_ROW_SIZE)); |
| } |
| for (size_t i = 0; i < rows.size(); ++i) { |
| if (rows[i].size() != columnNames.size()) { |
| callback->sendFailure("Inconsistent row sizes"); |
| StatsdStats::getInstance().noteQueryRestrictedMetricFailed( |
| configId, configPackage, keysToQuery.begin()->GetUid(), callingUid, |
| InvalidQueryReason(INCONSISTENT_ROW_SIZE)); |
| return; |
| } |
| queryData.insert(std::end(queryData), std::make_move_iterator(std::begin(rows[i])), |
| std::make_move_iterator(std::end(rows[i]))); |
| } |
| callback->sendResults(queryData, columnNames, columnTypes, rows.size()); |
| StatsdStats::getInstance().noteQueryRestrictedMetricSucceed( |
| configId, configPackage, keysToQuery.begin()->GetUid(), callingUid, |
| /*queryLatencyNs=*/getElapsedRealtimeNs() - elapsedRealtimeNs); |
| } |
| |
| set<ConfigKey> StatsLogProcessor::getRestrictedConfigKeysToQueryLocked( |
| const int32_t callingUid, const int64_t configId, const set<int32_t>& configPackageUids, |
| string& err, InvalidQueryReason& invalidQueryReason) { |
| set<ConfigKey> matchedConfigKeys; |
| for (auto uid : configPackageUids) { |
| ConfigKey configKey(uid, configId); |
| if (mMetricsManagers.find(configKey) != mMetricsManagers.end()) { |
| matchedConfigKeys.insert(configKey); |
| } |
| } |
| |
| set<ConfigKey> excludedKeys; |
| for (auto& configKey : matchedConfigKeys) { |
| auto it = mMetricsManagers.find(configKey); |
| if (!it->second->validateRestrictedMetricsDelegate(callingUid)) { |
| excludedKeys.insert(configKey); |
| }; |
| } |
| |
| set<ConfigKey> result; |
| std::set_difference(matchedConfigKeys.begin(), matchedConfigKeys.end(), excludedKeys.begin(), |
| excludedKeys.end(), std::inserter(result, result.end())); |
| if (matchedConfigKeys.empty()) { |
| err = "No configs found matching the config key"; |
| invalidQueryReason = InvalidQueryReason(CONFIG_KEY_NOT_FOUND); |
| } else if (result.empty()) { |
| err = "No matching configs for restricted metrics delegate"; |
| invalidQueryReason = InvalidQueryReason(CONFIG_KEY_WITH_UNMATCHED_DELEGATE); |
| } |
| |
| return result; |
| } |
| |
| void StatsLogProcessor::EnforceDataTtls(const int64_t wallClockNs, |
| const int64_t elapsedRealtimeNs) { |
| if (!isAtLeastU()) { |
| return; |
| } |
| std::lock_guard<std::mutex> lock(mMetricsMutex); |
| enforceDataTtlsLocked(wallClockNs, elapsedRealtimeNs); |
| } |
| |
| void StatsLogProcessor::enforceDataTtlsLocked(const int64_t wallClockNs, |
| const int64_t elapsedRealtimeNs) { |
| for (const auto& itr : mMetricsManagers) { |
| itr.second->enforceRestrictedDataTtls(wallClockNs); |
| } |
| mLastTtlTime = elapsedRealtimeNs; |
| } |
| |
| void StatsLogProcessor::enforceDbGuardrailsIfNecessaryLocked(const int64_t wallClockNs, |
| const int64_t elapsedRealtimeNs) { |
| if (elapsedRealtimeNs - mLastDbGuardrailEnforcementTime < |
| StatsdStats::kMinDbGuardrailEnforcementPeriodNs) { |
| return; |
| } |
| StorageManager::enforceDbGuardrails(STATS_RESTRICTED_DATA_DIR, wallClockNs / NS_PER_SEC, |
| StatsdStats::kMaxFileSize); |
| mLastDbGuardrailEnforcementTime = elapsedRealtimeNs; |
| } |
| |
| void StatsLogProcessor::fillRestrictedMetrics(const int64_t configId, const string& configPackage, |
| const int32_t delegateUid, vector<int64_t>* output) { |
| std::lock_guard<std::mutex> lock(mMetricsMutex); |
| |
| set<int32_t> configPackageUids; |
| const auto& uidMapItr = UidMap::sAidToUidMapping.find(configPackage); |
| if (uidMapItr != UidMap::sAidToUidMapping.end()) { |
| configPackageUids.insert(uidMapItr->second); |
| } else { |
| configPackageUids = mUidMap->getAppUid(configPackage); |
| } |
| string err; |
| InvalidQueryReason invalidQueryReason; |
| set<ConfigKey> keysToGetMetrics = getRestrictedConfigKeysToQueryLocked( |
| delegateUid, configId, configPackageUids, err, invalidQueryReason); |
| |
| for (const ConfigKey& key : keysToGetMetrics) { |
| vector<int64_t> metricIds = mMetricsManagers[key]->getAllMetricIds(); |
| output->insert(output->end(), metricIds.begin(), metricIds.end()); |
| } |
| } |
| |
| void StatsLogProcessor::flushRestrictedDataLocked(const int64_t elapsedRealtimeNs) { |
| for (const auto& it : mMetricsManagers) { |
| // no-op if metricsManager is not restricted |
| it.second->flushRestrictedData(); |
| } |
| |
| mLastFlushRestrictedTime = elapsedRealtimeNs; |
| } |
| |
| void StatsLogProcessor::flushIfNecessaryLocked(const ConfigKey& key, |
| MetricsManager& metricsManager) { |
| int64_t elapsedRealtimeNs = getElapsedRealtimeNs(); |
| auto lastCheckTime = mLastByteSizeTimes.find(key); |
| int64_t minCheckPeriodNs = metricsManager.useV2SoftMemoryCalculation() |
| ? StatsdStats::kMinByteSizeV2CheckPeriodNs |
| : StatsdStats::kMinByteSizeCheckPeriodNs; |
| if (lastCheckTime != mLastByteSizeTimes.end()) { |
| if (elapsedRealtimeNs - lastCheckTime->second < minCheckPeriodNs) { |
| return; |
| } |
| } |
| |
| // We suspect that the byteSize() computation is expensive, so we set a rate limit. |
| size_t totalBytes = metricsManager.byteSize(); |
| |
| mLastByteSizeTimes[key] = elapsedRealtimeNs; |
| const size_t kBytesPerConfig = metricsManager.hasRestrictedMetricsDelegate() |
| ? StatsdStats::kBytesPerRestrictedConfigTriggerFlush |
| : metricsManager.getTriggerGetDataBytes(); |
| bool requestDump = false; |
| if (totalBytes > metricsManager.getMaxMetricsBytes()) { |
| // Too late. We need to start clearing data. |
| metricsManager.dropData(elapsedRealtimeNs); |
| StatsdStats::getInstance().noteDataDropped(key, totalBytes); |
| VLOG("StatsD had to toss out metrics for %s", key.ToString().c_str()); |
| } else if ((totalBytes > kBytesPerConfig) || |
| (mOnDiskDataConfigs.find(key) != mOnDiskDataConfigs.end())) { |
| // Request to dump if: |
| // 1. in memory data > threshold OR |
| // 2. config has old data report on disk. |
| requestDump = true; |
| } |
| |
| if (requestDump) { |
| if (metricsManager.hasRestrictedMetricsDelegate()) { |
| metricsManager.flushRestrictedData(); |
| // No need to send broadcast for restricted metrics. |
| return; |
| } |
| // Send broadcast so that receivers can pull data. |
| auto lastBroadcastTime = mLastBroadcastTimes.find(key); |
| if (lastBroadcastTime != mLastBroadcastTimes.end()) { |
| if (elapsedRealtimeNs - lastBroadcastTime->second < |
| StatsdStats::kMinBroadcastPeriodNs) { |
| VLOG("StatsD would've sent a broadcast but the rate limit stopped us."); |
| return; |
| } |
| } |
| if (mSendBroadcast(key)) { |
| mOnDiskDataConfigs.erase(key); |
| VLOG("StatsD triggered data fetch for %s", key.ToString().c_str()); |
| mLastBroadcastTimes[key] = elapsedRealtimeNs; |
| StatsdStats::getInstance().noteBroadcastSent(key); |
| } |
| } |
| } |
| |
| void StatsLogProcessor::WriteDataToDiskLocked(const ConfigKey& key, const int64_t timestampNs, |
| const int64_t wallClockNs, |
| const DumpReportReason dumpReportReason, |
| const DumpLatency dumpLatency) { |
| if (mMetricsManagers.find(key) == mMetricsManagers.end() || |
| !mMetricsManagers.find(key)->second->shouldWriteToDisk()) { |
| return; |
| } |
| if (mMetricsManagers.find(key)->second->hasRestrictedMetricsDelegate()) { |
| mMetricsManagers.find(key)->second->flushRestrictedData(); |
| return; |
| } |
| vector<uint8_t> buffer; |
| onConfigMetricsReportLocked(key, timestampNs, wallClockNs, |
| true /* include_current_partial_bucket*/, true /* erase_data */, |
| dumpReportReason, dumpLatency, true, &buffer); |
| string file_name = |
| StorageManager::getDataFileName((long)getWallClockSec(), key.GetUid(), key.GetId()); |
| StorageManager::writeFile(file_name.c_str(), buffer.data(), buffer.size()); |
| |
| // We were able to write the ConfigMetricsReport to disk, so we should trigger collection ASAP. |
| mOnDiskDataConfigs.insert(key); |
| } |
| |
| void StatsLogProcessor::SaveActiveConfigsToDisk(int64_t currentTimeNs) { |
| std::lock_guard<std::mutex> lock(mMetricsMutex); |
| const int64_t timeNs = getElapsedRealtimeNs(); |
| // Do not write to disk if we already have in the last few seconds. |
| if (static_cast<unsigned long long> (timeNs) < |
| mLastActiveMetricsWriteNs + WRITE_DATA_COOL_DOWN_SEC * NS_PER_SEC) { |
| ALOGI("Statsd skipping writing active metrics to disk. Already wrote data in last %d seconds", |
| WRITE_DATA_COOL_DOWN_SEC); |
| return; |
| } |
| mLastActiveMetricsWriteNs = timeNs; |
| |
| ProtoOutputStream proto; |
| WriteActiveConfigsToProtoOutputStreamLocked(currentTimeNs, DEVICE_SHUTDOWN, &proto); |
| |
| string file_name = StringPrintf("%s/active_metrics", STATS_ACTIVE_METRIC_DIR); |
| StorageManager::deleteFile(file_name.c_str()); |
| android::base::unique_fd fd( |
| open(file_name.c_str(), O_WRONLY | O_CREAT | O_CLOEXEC, S_IRUSR | S_IWUSR)); |
| if (fd == -1) { |
| ALOGE("Attempt to write %s but failed", file_name.c_str()); |
| return; |
| } |
| proto.flush(fd.get()); |
| } |
| |
| void StatsLogProcessor::SaveMetadataToDisk(int64_t currentWallClockTimeNs, |
| int64_t systemElapsedTimeNs) { |
| std::lock_guard<std::mutex> lock(mMetricsMutex); |
| // Do not write to disk if we already have in the last few seconds. |
| if (static_cast<unsigned long long> (systemElapsedTimeNs) < |
| mLastMetadataWriteNs + WRITE_DATA_COOL_DOWN_SEC * NS_PER_SEC) { |
| ALOGI("Statsd skipping writing metadata to disk. Already wrote data in last %d seconds", |
| WRITE_DATA_COOL_DOWN_SEC); |
| return; |
| } |
| mLastMetadataWriteNs = systemElapsedTimeNs; |
| |
| metadata::StatsMetadataList metadataList; |
| WriteMetadataToProtoLocked( |
| currentWallClockTimeNs, systemElapsedTimeNs, &metadataList); |
| |
| string file_name = StringPrintf("%s/metadata", STATS_METADATA_DIR); |
| StorageManager::deleteFile(file_name.c_str()); |
| |
| if (metadataList.stats_metadata_size() == 0) { |
| // Skip the write if we have nothing to write. |
| return; |
| } |
| |
| std::string data; |
| metadataList.SerializeToString(&data); |
| StorageManager::writeFile(file_name.c_str(), data.c_str(), data.size()); |
| } |
| |
| void StatsLogProcessor::WriteMetadataToProto(int64_t currentWallClockTimeNs, |
| int64_t systemElapsedTimeNs, |
| metadata::StatsMetadataList* metadataList) { |
| std::lock_guard<std::mutex> lock(mMetricsMutex); |
| WriteMetadataToProtoLocked(currentWallClockTimeNs, systemElapsedTimeNs, metadataList); |
| } |
| |
| void StatsLogProcessor::WriteMetadataToProtoLocked(int64_t currentWallClockTimeNs, |
| int64_t systemElapsedTimeNs, |
| metadata::StatsMetadataList* metadataList) { |
| for (const auto& pair : mMetricsManagers) { |
| const sp<MetricsManager>& metricsManager = pair.second; |
| metadata::StatsMetadata* statsMetadata = metadataList->add_stats_metadata(); |
| bool metadataWritten = metricsManager->writeMetadataToProto(currentWallClockTimeNs, |
| systemElapsedTimeNs, statsMetadata); |
| if (!metadataWritten) { |
| metadataList->mutable_stats_metadata()->RemoveLast(); |
| } |
| } |
| } |
| |
| void StatsLogProcessor::LoadMetadataFromDisk(int64_t currentWallClockTimeNs, |
| int64_t systemElapsedTimeNs) { |
| std::lock_guard<std::mutex> lock(mMetricsMutex); |
| string file_name = StringPrintf("%s/metadata", STATS_METADATA_DIR); |
| int fd = open(file_name.c_str(), O_RDONLY | O_CLOEXEC); |
| if (-1 == fd) { |
| VLOG("Attempt to read %s but failed", file_name.c_str()); |
| StorageManager::deleteFile(file_name.c_str()); |
| return; |
| } |
| string content; |
| if (!android::base::ReadFdToString(fd, &content)) { |
| ALOGE("Attempt to read %s but failed", file_name.c_str()); |
| close(fd); |
| StorageManager::deleteFile(file_name.c_str()); |
| return; |
| } |
| |
| close(fd); |
| |
| metadata::StatsMetadataList statsMetadataList; |
| if (!statsMetadataList.ParseFromString(content)) { |
| ALOGE("Attempt to read %s but failed; failed to metadata", file_name.c_str()); |
| StorageManager::deleteFile(file_name.c_str()); |
| return; |
| } |
| SetMetadataStateLocked(statsMetadataList, currentWallClockTimeNs, systemElapsedTimeNs); |
| StorageManager::deleteFile(file_name.c_str()); |
| } |
| |
| void StatsLogProcessor::SetMetadataState(const metadata::StatsMetadataList& statsMetadataList, |
| int64_t currentWallClockTimeNs, |
| int64_t systemElapsedTimeNs) { |
| std::lock_guard<std::mutex> lock(mMetricsMutex); |
| SetMetadataStateLocked(statsMetadataList, currentWallClockTimeNs, systemElapsedTimeNs); |
| } |
| |
| void StatsLogProcessor::SetMetadataStateLocked( |
| const metadata::StatsMetadataList& statsMetadataList, |
| int64_t currentWallClockTimeNs, |
| int64_t systemElapsedTimeNs) { |
| for (const metadata::StatsMetadata& metadata : statsMetadataList.stats_metadata()) { |
| ConfigKey key(metadata.config_key().uid(), metadata.config_key().config_id()); |
| auto it = mMetricsManagers.find(key); |
| if (it == mMetricsManagers.end()) { |
| ALOGE("No config found for configKey %s", key.ToString().c_str()); |
| continue; |
| } |
| VLOG("Setting metadata %s", key.ToString().c_str()); |
| it->second->loadMetadata(metadata, currentWallClockTimeNs, systemElapsedTimeNs); |
| } |
| VLOG("Successfully loaded %d metadata.", statsMetadataList.stats_metadata_size()); |
| } |
| |
| void StatsLogProcessor::WriteActiveConfigsToProtoOutputStream( |
| int64_t currentTimeNs, const DumpReportReason reason, ProtoOutputStream* proto) { |
| std::lock_guard<std::mutex> lock(mMetricsMutex); |
| WriteActiveConfigsToProtoOutputStreamLocked(currentTimeNs, reason, proto); |
| } |
| |
| void StatsLogProcessor::WriteActiveConfigsToProtoOutputStreamLocked( |
| int64_t currentTimeNs, const DumpReportReason reason, ProtoOutputStream* proto) { |
| for (const auto& pair : mMetricsManagers) { |
| const sp<MetricsManager>& metricsManager = pair.second; |
| uint64_t configToken = proto->start(FIELD_TYPE_MESSAGE | FIELD_COUNT_REPEATED | |
| FIELD_ID_ACTIVE_CONFIG_LIST_CONFIG); |
| metricsManager->writeActiveConfigToProtoOutputStream(currentTimeNs, reason, proto); |
| proto->end(configToken); |
| } |
| } |
| void StatsLogProcessor::LoadActiveConfigsFromDisk() { |
| std::lock_guard<std::mutex> lock(mMetricsMutex); |
| string file_name = StringPrintf("%s/active_metrics", STATS_ACTIVE_METRIC_DIR); |
| int fd = open(file_name.c_str(), O_RDONLY | O_CLOEXEC); |
| if (-1 == fd) { |
| VLOG("Attempt to read %s but failed", file_name.c_str()); |
| StorageManager::deleteFile(file_name.c_str()); |
| return; |
| } |
| string content; |
| if (!android::base::ReadFdToString(fd, &content)) { |
| ALOGE("Attempt to read %s but failed", file_name.c_str()); |
| close(fd); |
| StorageManager::deleteFile(file_name.c_str()); |
| return; |
| } |
| |
| close(fd); |
| |
| ActiveConfigList activeConfigList; |
| if (!activeConfigList.ParseFromString(content)) { |
| ALOGE("Attempt to read %s but failed; failed to load active configs", file_name.c_str()); |
| StorageManager::deleteFile(file_name.c_str()); |
| return; |
| } |
| // Passing in mTimeBaseNs only works as long as we only load from disk is when statsd starts. |
| SetConfigsActiveStateLocked(activeConfigList, mTimeBaseNs); |
| StorageManager::deleteFile(file_name.c_str()); |
| } |
| |
| void StatsLogProcessor::SetConfigsActiveState(const ActiveConfigList& activeConfigList, |
| int64_t currentTimeNs) { |
| std::lock_guard<std::mutex> lock(mMetricsMutex); |
| SetConfigsActiveStateLocked(activeConfigList, currentTimeNs); |
| } |
| |
| void StatsLogProcessor::SetConfigsActiveStateLocked(const ActiveConfigList& activeConfigList, |
| int64_t currentTimeNs) { |
| for (int i = 0; i < activeConfigList.config_size(); i++) { |
| const auto& config = activeConfigList.config(i); |
| ConfigKey key(config.uid(), config.id()); |
| auto it = mMetricsManagers.find(key); |
| if (it == mMetricsManagers.end()) { |
| ALOGE("No config found for config %s", key.ToString().c_str()); |
| continue; |
| } |
| VLOG("Setting active config %s", key.ToString().c_str()); |
| it->second->loadActiveConfig(config, currentTimeNs); |
| } |
| VLOG("Successfully loaded %d active configs.", activeConfigList.config_size()); |
| } |
| |
| void StatsLogProcessor::WriteDataToDiskLocked(const DumpReportReason dumpReportReason, |
| const DumpLatency dumpLatency, |
| const int64_t elapsedRealtimeNs, |
| const int64_t wallClockNs) { |
| // Do not write to disk if we already have in the last few seconds. |
| // This is to avoid overwriting files that would have the same name if we |
| // write twice in the same second. |
| if (static_cast<unsigned long long>(elapsedRealtimeNs) < |
| mLastWriteTimeNs + WRITE_DATA_COOL_DOWN_SEC * NS_PER_SEC) { |
| ALOGI("Statsd skipping writing data to disk. Already wrote data in last %d seconds", |
| WRITE_DATA_COOL_DOWN_SEC); |
| return; |
| } |
| mLastWriteTimeNs = elapsedRealtimeNs; |
| for (auto& pair : mMetricsManagers) { |
| WriteDataToDiskLocked(pair.first, elapsedRealtimeNs, wallClockNs, dumpReportReason, |
| dumpLatency); |
| } |
| } |
| |
| void StatsLogProcessor::WriteDataToDisk(const DumpReportReason dumpReportReason, |
| const DumpLatency dumpLatency, |
| const int64_t elapsedRealtimeNs, |
| const int64_t wallClockNs) { |
| std::lock_guard<std::mutex> lock(mMetricsMutex); |
| WriteDataToDiskLocked(dumpReportReason, dumpLatency, elapsedRealtimeNs, wallClockNs); |
| } |
| |
| void StatsLogProcessor::informPullAlarmFired(const int64_t timestampNs) { |
| std::lock_guard<std::mutex> lock(mMetricsMutex); |
| mPullerManager->OnAlarmFired(timestampNs); |
| } |
| |
| int64_t StatsLogProcessor::getLastReportTimeNs(const ConfigKey& key) { |
| auto it = mMetricsManagers.find(key); |
| if (it == mMetricsManagers.end()) { |
| return 0; |
| } else { |
| return it->second->getLastReportTimeNs(); |
| } |
| } |
| |
| void StatsLogProcessor::notifyAppUpgrade(const int64_t eventTimeNs, const string& apk, |
| const int uid, const int64_t version) { |
| std::lock_guard<std::mutex> lock(mMetricsMutex); |
| VLOG("Received app upgrade"); |
| StateManager::getInstance().notifyAppChanged(apk, mUidMap); |
| for (const auto& it : mMetricsManagers) { |
| it.second->notifyAppUpgrade(eventTimeNs, apk, uid, version); |
| } |
| } |
| |
| void StatsLogProcessor::notifyAppRemoved(const int64_t eventTimeNs, const string& apk, |
| const int uid) { |
| std::lock_guard<std::mutex> lock(mMetricsMutex); |
| VLOG("Received app removed"); |
| StateManager::getInstance().notifyAppChanged(apk, mUidMap); |
| for (const auto& it : mMetricsManagers) { |
| it.second->notifyAppRemoved(eventTimeNs, apk, uid); |
| } |
| } |
| |
| void StatsLogProcessor::onUidMapReceived(const int64_t eventTimeNs) { |
| std::lock_guard<std::mutex> lock(mMetricsMutex); |
| VLOG("Received uid map"); |
| StateManager::getInstance().updateLogSources(mUidMap); |
| for (const auto& it : mMetricsManagers) { |
| it.second->onUidMapReceived(eventTimeNs); |
| } |
| } |
| |
| void StatsLogProcessor::onStatsdInitCompleted(const int64_t elapsedTimeNs) { |
| ATRACE_CALL(); |
| std::lock_guard<std::mutex> lock(mMetricsMutex); |
| VLOG("Received boot completed signal"); |
| for (const auto& it : mMetricsManagers) { |
| it.second->onStatsdInitCompleted(elapsedTimeNs); |
| } |
| } |
| |
| void StatsLogProcessor::noteOnDiskData(const ConfigKey& key) { |
| std::lock_guard<std::mutex> lock(mMetricsMutex); |
| mOnDiskDataConfigs.insert(key); |
| } |
| |
| void StatsLogProcessor::setAnomalyAlarm(const int64_t elapsedTimeMillis) { |
| std::lock_guard<std::mutex> lock(mAnomalyAlarmMutex); |
| mNextAnomalyAlarmTime = elapsedTimeMillis; |
| } |
| |
| void StatsLogProcessor::cancelAnomalyAlarm() { |
| std::lock_guard<std::mutex> lock(mAnomalyAlarmMutex); |
| mNextAnomalyAlarmTime = 0; |
| } |
| |
| void StatsLogProcessor::informAnomalyAlarmFiredLocked(const int64_t elapsedTimeMillis) { |
| VLOG("StatsService::informAlarmForSubscriberTriggeringFired was called"); |
| unordered_set<sp<const InternalAlarm>, SpHash<InternalAlarm>> alarmSet = |
| mAnomalyAlarmMonitor->popSoonerThan(static_cast<uint32_t>(elapsedTimeMillis / 1000)); |
| if (alarmSet.size() > 0) { |
| VLOG("Found periodic alarm fired."); |
| processFiredAnomalyAlarmsLocked(MillisToNano(elapsedTimeMillis), alarmSet); |
| } else { |
| ALOGW("Cannot find an periodic alarm that fired. Perhaps it was recently cancelled."); |
| } |
| } |
| |
| LogEventFilter::AtomIdSet StatsLogProcessor::getDefaultAtomIdSet() { |
| // populate hard-coded list of useful atoms |
| // we add also atoms which could be pushed by statsd itself to simplify the logic |
| // to handle metric configs update: APP_BREADCRUMB_REPORTED & ANOMALY_DETECTED |
| LogEventFilter::AtomIdSet allAtomIds{ |
| util::BINARY_PUSH_STATE_CHANGED, util::ISOLATED_UID_CHANGED, |
| util::APP_BREADCRUMB_REPORTED, util::WATCHDOG_ROLLBACK_OCCURRED, |
| util::ANOMALY_DETECTED, util::STATS_SOCKET_LOSS_REPORTED}; |
| return allAtomIds; |
| } |
| |
| void StatsLogProcessor::updateLogEventFilterLocked() const { |
| VLOG("StatsLogProcessor: Updating allAtomIds at %lld", (long long)getElapsedRealtimeNs()); |
| LogEventFilter::AtomIdSet allAtomIds = getDefaultAtomIdSet(); |
| for (const auto& metricsManager : mMetricsManagers) { |
| metricsManager.second->addAllAtomIds(allAtomIds); |
| } |
| StateManager::getInstance().addAllAtomIds(allAtomIds); |
| VLOG("StatsLogProcessor: Updating allAtomIds done. Total atoms %d", (int)allAtomIds.size()); |
| #ifdef STATSD_DEBUG |
| for (auto atomId : allAtomIds) { |
| VLOG("Atom in use %d", atomId); |
| } |
| #endif // STATSD_DEBUG |
| mLogEventFilter->setAtomIds(std::move(allAtomIds), this); |
| } |
| |
| bool StatsLogProcessor::validateAppBreadcrumbEvent(const LogEvent& event) const { |
| if (event.GetTagId() == util::APP_BREADCRUMB_REPORTED) { |
| // Check that app breadcrumb reported fields are valid. |
| status_t err = NO_ERROR; |
| |
| // Uid is 3rd from last field and must match the caller's uid, |
| // unless that caller is statsd itself (statsd is allowed to spoof uids). |
| const long appHookUid = event.GetLong(event.size() - 2, &err); |
| if (err != NO_ERROR) { |
| VLOG("APP_BREADCRUMB_REPORTED had error when parsing the uid"); |
| return false; |
| } |
| |
| // Because the uid within the LogEvent may have been mapped from |
| // isolated to host, map the loggerUid similarly before comparing. |
| const int32_t loggerUid = mUidMap->getHostUidOrSelf(event.GetUid()); |
| if (loggerUid != appHookUid && loggerUid != AID_STATSD) { |
| VLOG("APP_BREADCRUMB_REPORTED has invalid uid: claimed %ld but caller is %d", |
| appHookUid, loggerUid); |
| return false; |
| } |
| |
| // The state must be from 0,3. This part of code must be manually updated. |
| const long appHookState = event.GetLong(event.size(), &err); |
| if (err != NO_ERROR) { |
| VLOG("APP_BREADCRUMB_REPORTED had error when parsing the state field"); |
| return false; |
| } else if (appHookState < 0 || appHookState > 3) { |
| VLOG("APP_BREADCRUMB_REPORTED does not have valid state %ld", appHookState); |
| return false; |
| } |
| } |
| |
| return true; |
| } |
| |
| void StatsLogProcessor::processQueueOverflowStatsLocked() { |
| auto queueOverflowStats = StatsdStats::getInstance().getQueueOverflowAtomsStats(); |
| |
| for (const auto [atomId, count] : queueOverflowStats) { |
| // are there new atoms dropped due to queue overflow since previous request |
| auto droppedAtomStatsIt = mQueueOverflowAtomsStats.find(atomId); |
| if (droppedAtomStatsIt != mQueueOverflowAtomsStats.end() && |
| droppedAtomStatsIt->second == count) { |
| // no new dropped atoms detected for the atomId |
| continue; |
| } |
| |
| StateManager::getInstance().onLogEventLost(atomId, DATA_CORRUPTED_EVENT_QUEUE_OVERFLOW); |
| } |
| mQueueOverflowAtomsStats = std::move(queueOverflowStats); |
| } |
| |
| } // namespace statsd |
| } // namespace os |
| } // namespace android |