| /* |
| * Copyright (C) 2018 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| package com.android.server.appop; |
| |
| import static android.app.AppOpsManager.FILTER_BY_ATTRIBUTION_TAG; |
| import static android.app.AppOpsManager.FILTER_BY_OP_NAMES; |
| import static android.app.AppOpsManager.FILTER_BY_PACKAGE_NAME; |
| import static android.app.AppOpsManager.FILTER_BY_UID; |
| |
| import android.annotation.NonNull; |
| import android.annotation.Nullable; |
| import android.app.AppOpsManager; |
| import android.app.AppOpsManager.HistoricalMode; |
| import android.app.AppOpsManager.HistoricalOp; |
| import android.app.AppOpsManager.HistoricalOps; |
| import android.app.AppOpsManager.HistoricalOpsRequestFilter; |
| import android.app.AppOpsManager.HistoricalPackageOps; |
| import android.app.AppOpsManager.HistoricalUidOps; |
| import android.app.AppOpsManager.OpFlags; |
| import android.app.AppOpsManager.UidState; |
| import android.content.ContentResolver; |
| import android.database.ContentObserver; |
| import android.net.Uri; |
| import android.os.Binder; |
| import android.os.Build; |
| import android.os.Bundle; |
| import android.os.Debug; |
| import android.os.Environment; |
| import android.os.Message; |
| import android.os.Process; |
| import android.os.RemoteCallback; |
| import android.os.UserHandle; |
| import android.provider.DeviceConfig; |
| import android.provider.Settings; |
| import android.util.ArraySet; |
| import android.util.LongSparseArray; |
| import android.util.Slog; |
| import android.util.TimeUtils; |
| import android.util.Xml; |
| |
| import com.android.internal.annotations.GuardedBy; |
| import com.android.internal.os.AtomicDirectory; |
| import com.android.internal.os.BackgroundThread; |
| import com.android.internal.util.ArrayUtils; |
| import com.android.internal.util.XmlUtils; |
| import com.android.internal.util.function.pooled.PooledLambda; |
| import com.android.server.FgThread; |
| |
| import org.xmlpull.v1.XmlPullParser; |
| import org.xmlpull.v1.XmlPullParserException; |
| import org.xmlpull.v1.XmlSerializer; |
| |
| import java.io.File; |
| import java.io.FileInputStream; |
| import java.io.FileNotFoundException; |
| import java.io.FileOutputStream; |
| import java.io.IOException; |
| import java.io.PrintWriter; |
| import java.nio.charset.StandardCharsets; |
| import java.nio.file.Files; |
| import java.text.SimpleDateFormat; |
| import java.util.ArrayList; |
| import java.util.Arrays; |
| import java.util.Collections; |
| import java.util.Date; |
| import java.util.LinkedList; |
| import java.util.List; |
| import java.util.Objects; |
| import java.util.Set; |
| import java.util.concurrent.TimeUnit; |
| |
| /** |
| * This class managers historical app op state. This includes reading, persistence, |
| * accounting, querying. |
| * <p> |
| * The history is kept forever in multiple files. Each file time contains the |
| * relative offset from the current time which time is encoded in the file name. |
| * The files contain historical app op state snapshots which have times that |
| * are relative to the time of the container file. |
| * |
| * The data in the files are stored in a logarithmic fashion where where every |
| * subsequent file would contain data for ten times longer interval with ten |
| * times more time distance between snapshots. Hence, the more time passes |
| * the lesser the fidelity. |
| * <p> |
| * For example, the first file would contain data for 1 days with snapshots |
| * every 0.1 days, the next file would contain data for the period 1 to 10 |
| * days with snapshots every 1 days, and so on. |
| * <p> |
| * THREADING AND LOCKING: Reported ops must be processed as quickly as possible. |
| * We keep ops pending to be persisted in memory and write to disk on a background |
| * thread. Hence, methods that report op changes are locking only the in memory |
| * state guarded by the mInMemoryLock which happens to be the app ops service lock |
| * avoiding a lock addition on the critical path. When a query comes we need to |
| * evaluate it based off both in memory and on disk state. This means they need to |
| * be frozen with respect to each other and not change from the querying caller's |
| * perspective. To achieve this we add a dedicated mOnDiskLock to guard the on |
| * disk state. To have fast critical path we need to limit the locking of the |
| * mInMemoryLock, thus for operations that touch in memory and on disk state one |
| * must grab first the mOnDiskLock and then the mInMemoryLock and limit the |
| * in memory lock to extraction of relevant data. Locking order is critical to |
| * avoid deadlocks. The convention is that xxxDLocked suffix means the method |
| * must be called with the mOnDiskLock lock, xxxMLocked suffix means the method |
| * must be called with the mInMemoryLock, xxxDMLocked suffix means the method |
| * must be called with the mOnDiskLock and mInMemoryLock locks acquired in that |
| * exact order. |
| * <p> |
| * INITIALIZATION: We can initialize persistence only after the system is ready |
| * as we need to check the optional configuration override from the settings |
| * database which is not initialized at the time the app ops service is created. |
| * This means that all entry points that touch persistence should be short |
| * circuited via isPersistenceInitialized() check. |
| */ |
| // TODO (bug:122218838): Make sure we handle start of epoch time |
| // TODO (bug:122218838): Validate changed time is handled correctly |
| final class HistoricalRegistry { |
| private static final boolean DEBUG = false; |
| private static final boolean KEEP_WTF_LOG = Build.IS_DEBUGGABLE; |
| |
| private static final String LOG_TAG = HistoricalRegistry.class.getSimpleName(); |
| |
| private static final String PARAMETER_DELIMITER = ","; |
| private static final String PARAMETER_ASSIGNMENT = "="; |
| private static final String PROPERTY_PERMISSIONS_HUB_ENABLED = "permissions_hub_enabled"; |
| |
| @GuardedBy("mLock") |
| private @NonNull LinkedList<HistoricalOps> mPendingWrites = new LinkedList<>(); |
| |
| // Lock for read/write access to on disk state |
| private final Object mOnDiskLock = new Object(); |
| |
| //Lock for read/write access to in memory state |
| private final @NonNull Object mInMemoryLock; |
| |
| private static final int MSG_WRITE_PENDING_HISTORY = 1; |
| |
| // See mMode |
| private static final int DEFAULT_MODE = AppOpsManager.HISTORICAL_MODE_ENABLED_ACTIVE; |
| |
| // See mBaseSnapshotInterval |
| private static final long DEFAULT_SNAPSHOT_INTERVAL_MILLIS = TimeUnit.MINUTES.toMillis(15); |
| |
| // See mIntervalCompressionMultiplier |
| private static final long DEFAULT_COMPRESSION_STEP = 10; |
| |
| private static final String HISTORY_FILE_SUFFIX = ".xml"; |
| |
| /** |
| * Whether history is enabled. |
| */ |
| @GuardedBy("mInMemoryLock") |
| private int mMode = AppOpsManager.HISTORICAL_MODE_ENABLED_ACTIVE; |
| |
| /** |
| * This granularity has been chosen to allow clean delineation for intervals |
| * humans understand, 15 min, 60, min, a day, a week, a month (30 days). |
| */ |
| @GuardedBy("mInMemoryLock") |
| private long mBaseSnapshotInterval = DEFAULT_SNAPSHOT_INTERVAL_MILLIS; |
| |
| /** |
| * The compression between steps. Each subsequent step is this much longer |
| * in terms of duration and each snapshot is this much more apart from the |
| * previous step. |
| */ |
| @GuardedBy("mInMemoryLock") |
| private long mIntervalCompressionMultiplier = DEFAULT_COMPRESSION_STEP; |
| |
| // The current ops to which to add statistics. |
| @GuardedBy("mInMemoryLock") |
| private @Nullable HistoricalOps mCurrentHistoricalOps; |
| |
| // The time we should write the next snapshot. |
| @GuardedBy("mInMemoryLock") |
| private long mNextPersistDueTimeMillis; |
| |
| // How much to offset the history on the next write. |
| @GuardedBy("mInMemoryLock") |
| private long mPendingHistoryOffsetMillis; |
| |
| // Object managing persistence (read/write) |
| @GuardedBy("mOnDiskLock") |
| private Persistence mPersistence; |
| |
| HistoricalRegistry(@NonNull Object lock) { |
| mInMemoryLock = lock; |
| } |
| |
| HistoricalRegistry(@NonNull HistoricalRegistry other) { |
| this(other.mInMemoryLock); |
| mMode = other.mMode; |
| mBaseSnapshotInterval = other.mBaseSnapshotInterval; |
| mIntervalCompressionMultiplier = other.mIntervalCompressionMultiplier; |
| } |
| |
| void systemReady(@NonNull ContentResolver resolver) { |
| final Uri uri = Settings.Global.getUriFor(Settings.Global.APPOP_HISTORY_PARAMETERS); |
| resolver.registerContentObserver(uri, false, new ContentObserver( |
| FgThread.getHandler()) { |
| @Override |
| public void onChange(boolean selfChange) { |
| updateParametersFromSetting(resolver); |
| } |
| }); |
| |
| updateParametersFromSetting(resolver); |
| |
| synchronized (mOnDiskLock) { |
| synchronized (mInMemoryLock) { |
| if (mMode != AppOpsManager.HISTORICAL_MODE_DISABLED) { |
| // Can be uninitialized if there is no config in the settings table. |
| if (!isPersistenceInitializedMLocked()) { |
| mPersistence = new Persistence(mBaseSnapshotInterval, |
| mIntervalCompressionMultiplier); |
| } |
| |
| // When starting always adjust history to now. |
| final long lastPersistTimeMills = |
| mPersistence.getLastPersistTimeMillisDLocked(); |
| |
| if (lastPersistTimeMills > 0) { |
| mPendingHistoryOffsetMillis = System.currentTimeMillis() |
| - lastPersistTimeMills; |
| |
| if (DEBUG) { |
| Slog.i(LOG_TAG, "Time since last write: " |
| + TimeUtils.formatDuration(mPendingHistoryOffsetMillis) |
| + " by which to push history on next write"); |
| } |
| } |
| } |
| } |
| } |
| } |
| |
| private boolean isPersistenceInitializedMLocked() { |
| return mPersistence != null; |
| } |
| |
| private void updateParametersFromSetting(@NonNull ContentResolver resolver) { |
| final String setting = Settings.Global.getString(resolver, |
| Settings.Global.APPOP_HISTORY_PARAMETERS); |
| if (setting == null) { |
| return; |
| } |
| String modeValue = null; |
| String baseSnapshotIntervalValue = null; |
| String intervalMultiplierValue = null; |
| final String[] parameters = setting.split(PARAMETER_DELIMITER); |
| for (String parameter : parameters) { |
| final String[] parts = parameter.split(PARAMETER_ASSIGNMENT); |
| if (parts.length == 2) { |
| final String key = parts[0].trim(); |
| switch (key) { |
| case Settings.Global.APPOP_HISTORY_MODE: { |
| modeValue = parts[1].trim(); |
| } break; |
| case Settings.Global.APPOP_HISTORY_BASE_INTERVAL_MILLIS: { |
| baseSnapshotIntervalValue = parts[1].trim(); |
| } break; |
| case Settings.Global.APPOP_HISTORY_INTERVAL_MULTIPLIER: { |
| intervalMultiplierValue = parts[1].trim(); |
| } break; |
| default: { |
| Slog.w(LOG_TAG, "Unknown parameter: " + parameter); |
| } |
| } |
| } |
| } |
| if (modeValue != null && baseSnapshotIntervalValue != null |
| && intervalMultiplierValue != null) { |
| try { |
| final int mode = AppOpsManager.parseHistoricalMode(modeValue); |
| final long baseSnapshotInterval = Long.parseLong(baseSnapshotIntervalValue); |
| final int intervalCompressionMultiplier = Integer.parseInt(intervalMultiplierValue); |
| setHistoryParameters(mode, baseSnapshotInterval,intervalCompressionMultiplier); |
| return; |
| } catch (NumberFormatException ignored) {} |
| } |
| Slog.w(LOG_TAG, "Bad value for" + Settings.Global.APPOP_HISTORY_PARAMETERS |
| + "=" + setting + " resetting!"); |
| } |
| |
| void dump(String prefix, PrintWriter pw, int filterUid, @Nullable String filterPackage, |
| @Nullable String filterAttributionTag, int filterOp, |
| @HistoricalOpsRequestFilter int filter) { |
| if (!isApiEnabled()) { |
| return; |
| } |
| |
| synchronized (mOnDiskLock) { |
| synchronized (mInMemoryLock) { |
| pw.println(); |
| pw.print(prefix); |
| pw.print("History:"); |
| |
| pw.print(" mode="); |
| pw.println(AppOpsManager.historicalModeToString(mMode)); |
| |
| final StringDumpVisitor visitor = new StringDumpVisitor(prefix + " ", |
| pw, filterUid, filterPackage, filterAttributionTag, filterOp, filter); |
| final long nowMillis = System.currentTimeMillis(); |
| |
| // Dump in memory state first |
| final HistoricalOps currentOps = getUpdatedPendingHistoricalOpsMLocked( |
| nowMillis); |
| makeRelativeToEpochStart(currentOps, nowMillis); |
| currentOps.accept(visitor); |
| |
| if (!isPersistenceInitializedMLocked()) { |
| Slog.e(LOG_TAG, "Interaction before persistence initialized"); |
| return; |
| } |
| |
| final List<HistoricalOps> ops = mPersistence.readHistoryDLocked(); |
| if (ops != null) { |
| // TODO (bug:122218838): Make sure this is properly dumped |
| final long remainingToFillBatchMillis = mNextPersistDueTimeMillis |
| - nowMillis - mBaseSnapshotInterval; |
| final int opCount = ops.size(); |
| for (int i = 0; i < opCount; i++) { |
| final HistoricalOps op = ops.get(i); |
| op.offsetBeginAndEndTime(remainingToFillBatchMillis); |
| makeRelativeToEpochStart(op, nowMillis); |
| op.accept(visitor); |
| } |
| } else { |
| pw.println(" Empty"); |
| } |
| } |
| } |
| } |
| |
| @HistoricalMode int getMode() { |
| synchronized (mInMemoryLock) { |
| return mMode; |
| } |
| } |
| |
| void getHistoricalOpsFromDiskRaw(int uid, @NonNull String packageName, |
| @Nullable String attributionTag, @Nullable String[] opNames, |
| @HistoricalOpsRequestFilter int filter, long beginTimeMillis, long endTimeMillis, |
| @OpFlags int flags, @NonNull RemoteCallback callback) { |
| if (!isApiEnabled()) { |
| callback.sendResult(new Bundle()); |
| return; |
| } |
| |
| synchronized (mOnDiskLock) { |
| synchronized (mInMemoryLock) { |
| if (!isPersistenceInitializedMLocked()) { |
| Slog.e(LOG_TAG, "Interaction before persistence initialized"); |
| callback.sendResult(new Bundle()); |
| return; |
| } |
| final HistoricalOps result = new HistoricalOps(beginTimeMillis, endTimeMillis); |
| mPersistence.collectHistoricalOpsDLocked(result, uid, packageName, attributionTag, |
| opNames, filter, beginTimeMillis, endTimeMillis, flags); |
| final Bundle payload = new Bundle(); |
| payload.putParcelable(AppOpsManager.KEY_HISTORICAL_OPS, result); |
| callback.sendResult(payload); |
| } |
| } |
| } |
| |
| void getHistoricalOps(int uid, @NonNull String packageName, @Nullable String attributionTag, |
| @Nullable String[] opNames, @HistoricalOpsRequestFilter int filter, |
| long beginTimeMillis, long endTimeMillis, @OpFlags int flags, |
| @NonNull RemoteCallback callback) { |
| if (!isApiEnabled()) { |
| callback.sendResult(new Bundle()); |
| return; |
| } |
| |
| final long currentTimeMillis = System.currentTimeMillis(); |
| if (endTimeMillis == Long.MAX_VALUE) { |
| endTimeMillis = currentTimeMillis; |
| } |
| |
| // Argument times are based off epoch start while our internal store is |
| // based off now, so take this into account. |
| final long inMemoryAdjBeginTimeMillis = Math.max(currentTimeMillis - endTimeMillis, 0); |
| final long inMemoryAdjEndTimeMillis = Math.max(currentTimeMillis - beginTimeMillis, 0); |
| final HistoricalOps result = new HistoricalOps(inMemoryAdjBeginTimeMillis, |
| inMemoryAdjEndTimeMillis); |
| |
| synchronized (mOnDiskLock) { |
| final List<HistoricalOps> pendingWrites; |
| final HistoricalOps currentOps; |
| boolean collectOpsFromDisk; |
| |
| synchronized (mInMemoryLock) { |
| if (!isPersistenceInitializedMLocked()) { |
| Slog.e(LOG_TAG, "Interaction before persistence initialized"); |
| callback.sendResult(new Bundle()); |
| return; |
| } |
| |
| currentOps = getUpdatedPendingHistoricalOpsMLocked(currentTimeMillis); |
| if (!(inMemoryAdjBeginTimeMillis >= currentOps.getEndTimeMillis() |
| || inMemoryAdjEndTimeMillis <= currentOps.getBeginTimeMillis())) { |
| // Some of the current batch falls into the query, so extract that. |
| final HistoricalOps currentOpsCopy = new HistoricalOps(currentOps); |
| currentOpsCopy.filter(uid, packageName, attributionTag, opNames, filter, |
| inMemoryAdjBeginTimeMillis, inMemoryAdjEndTimeMillis); |
| result.merge(currentOpsCopy); |
| } |
| pendingWrites = new ArrayList<>(mPendingWrites); |
| mPendingWrites.clear(); |
| collectOpsFromDisk = inMemoryAdjEndTimeMillis > currentOps.getEndTimeMillis(); |
| } |
| |
| // If the query was only for in-memory state - done. |
| if (collectOpsFromDisk) { |
| // If there is a write in flight we need to force it now |
| persistPendingHistory(pendingWrites); |
| // Collect persisted state. |
| final long onDiskAndInMemoryOffsetMillis = currentTimeMillis |
| - mNextPersistDueTimeMillis + mBaseSnapshotInterval; |
| final long onDiskAdjBeginTimeMillis = Math.max(inMemoryAdjBeginTimeMillis |
| - onDiskAndInMemoryOffsetMillis, 0); |
| final long onDiskAdjEndTimeMillis = Math.max(inMemoryAdjEndTimeMillis |
| - onDiskAndInMemoryOffsetMillis, 0); |
| mPersistence.collectHistoricalOpsDLocked(result, uid, packageName, attributionTag, |
| opNames, filter, onDiskAdjBeginTimeMillis, onDiskAdjEndTimeMillis, flags); |
| } |
| |
| // Rebase the result time to be since epoch. |
| result.setBeginAndEndTime(beginTimeMillis, endTimeMillis); |
| |
| // Send back the result. |
| final Bundle payload = new Bundle(); |
| payload.putParcelable(AppOpsManager.KEY_HISTORICAL_OPS, result); |
| callback.sendResult(payload); |
| } |
| } |
| |
| void incrementOpAccessedCount(int op, int uid, @NonNull String packageName, |
| @Nullable String attributionTag, @UidState int uidState, @OpFlags int flags) { |
| synchronized (mInMemoryLock) { |
| if (mMode == AppOpsManager.HISTORICAL_MODE_ENABLED_ACTIVE) { |
| if (!isPersistenceInitializedMLocked()) { |
| Slog.e(LOG_TAG, "Interaction before persistence initialized"); |
| return; |
| } |
| getUpdatedPendingHistoricalOpsMLocked( |
| System.currentTimeMillis()).increaseAccessCount(op, uid, packageName, |
| attributionTag, uidState, flags, 1); |
| } |
| } |
| } |
| |
| void incrementOpRejected(int op, int uid, @NonNull String packageName, |
| @Nullable String attributionTag, @UidState int uidState, @OpFlags int flags) { |
| synchronized (mInMemoryLock) { |
| if (mMode == AppOpsManager.HISTORICAL_MODE_ENABLED_ACTIVE) { |
| if (!isPersistenceInitializedMLocked()) { |
| Slog.e(LOG_TAG, "Interaction before persistence initialized"); |
| return; |
| } |
| getUpdatedPendingHistoricalOpsMLocked( |
| System.currentTimeMillis()).increaseRejectCount(op, uid, packageName, |
| attributionTag, uidState, flags, 1); |
| } |
| } |
| } |
| |
| void increaseOpAccessDuration(int op, int uid, @NonNull String packageName, |
| @Nullable String attributionTag, @UidState int uidState, @OpFlags int flags, |
| long increment) { |
| synchronized (mInMemoryLock) { |
| if (mMode == AppOpsManager.HISTORICAL_MODE_ENABLED_ACTIVE) { |
| if (!isPersistenceInitializedMLocked()) { |
| Slog.e(LOG_TAG, "Interaction before persistence initialized"); |
| return; |
| } |
| getUpdatedPendingHistoricalOpsMLocked( |
| System.currentTimeMillis()).increaseAccessDuration(op, uid, packageName, |
| attributionTag, uidState, flags, increment); |
| } |
| } |
| } |
| |
| void setHistoryParameters(@HistoricalMode int mode, |
| long baseSnapshotInterval, long intervalCompressionMultiplier) { |
| synchronized (mOnDiskLock) { |
| synchronized (mInMemoryLock) { |
| // NOTE: We allow this call if persistence is not initialized as |
| // it is a part of the persistence initialization process. |
| boolean resampleHistory = false; |
| Slog.i(LOG_TAG, "New history parameters: mode:" |
| + AppOpsManager.historicalModeToString(mode) + " baseSnapshotInterval:" |
| + baseSnapshotInterval + " intervalCompressionMultiplier:" |
| + intervalCompressionMultiplier); |
| if (mMode != mode) { |
| mMode = mode; |
| if (mMode == AppOpsManager.HISTORICAL_MODE_DISABLED) { |
| clearHistoryOnDiskDLocked(); |
| } |
| } |
| if (mBaseSnapshotInterval != baseSnapshotInterval) { |
| mBaseSnapshotInterval = baseSnapshotInterval; |
| resampleHistory = true; |
| } |
| if (mIntervalCompressionMultiplier != intervalCompressionMultiplier) { |
| mIntervalCompressionMultiplier = intervalCompressionMultiplier; |
| resampleHistory = true; |
| } |
| if (resampleHistory) { |
| resampleHistoryOnDiskInMemoryDMLocked(0); |
| } |
| } |
| } |
| } |
| |
| void offsetHistory(long offsetMillis) { |
| synchronized (mOnDiskLock) { |
| synchronized (mInMemoryLock) { |
| if (!isPersistenceInitializedMLocked()) { |
| Slog.e(LOG_TAG, "Interaction before persistence initialized"); |
| return; |
| } |
| final List<HistoricalOps> history = mPersistence.readHistoryDLocked(); |
| clearHistory(); |
| if (history != null) { |
| final int historySize = history.size(); |
| for (int i = 0; i < historySize; i++) { |
| final HistoricalOps ops = history.get(i); |
| ops.offsetBeginAndEndTime(offsetMillis); |
| } |
| if (offsetMillis < 0) { |
| pruneFutureOps(history); |
| } |
| mPersistence.persistHistoricalOpsDLocked(history); |
| } |
| } |
| } |
| } |
| |
| void addHistoricalOps(HistoricalOps ops) { |
| final List<HistoricalOps> pendingWrites; |
| synchronized (mInMemoryLock) { |
| if (!isPersistenceInitializedMLocked()) { |
| Slog.e(LOG_TAG, "Interaction before persistence initialized"); |
| return; |
| } |
| // The history files start from mBaseSnapshotInterval - take this into account. |
| ops.offsetBeginAndEndTime(mBaseSnapshotInterval); |
| mPendingWrites.offerFirst(ops); |
| pendingWrites = new ArrayList<>(mPendingWrites); |
| mPendingWrites.clear(); |
| } |
| persistPendingHistory(pendingWrites); |
| } |
| |
| private void resampleHistoryOnDiskInMemoryDMLocked(long offsetMillis) { |
| mPersistence = new Persistence(mBaseSnapshotInterval, mIntervalCompressionMultiplier); |
| offsetHistory(offsetMillis); |
| } |
| |
| void resetHistoryParameters() { |
| if (!isPersistenceInitializedMLocked()) { |
| Slog.e(LOG_TAG, "Interaction before persistence initialized"); |
| return; |
| } |
| setHistoryParameters(DEFAULT_MODE, DEFAULT_SNAPSHOT_INTERVAL_MILLIS, |
| DEFAULT_COMPRESSION_STEP); |
| } |
| |
| void clearHistory(int uid, String packageName) { |
| synchronized (mOnDiskLock) { |
| synchronized (mInMemoryLock) { |
| if (!isPersistenceInitializedMLocked()) { |
| Slog.e(LOG_TAG, "Interaction before persistence initialized"); |
| return; |
| } |
| if (mMode != AppOpsManager.HISTORICAL_MODE_ENABLED_ACTIVE) { |
| return; |
| } |
| |
| for (int index = 0; index < mPendingWrites.size(); index++) { |
| mPendingWrites.get(index).clearHistory(uid, packageName); |
| } |
| |
| getUpdatedPendingHistoricalOpsMLocked(System.currentTimeMillis()) |
| .clearHistory(uid, packageName); |
| |
| mPersistence.clearHistoryDLocked(uid, packageName); |
| } |
| } |
| } |
| |
| void clearHistory() { |
| synchronized (mOnDiskLock) { |
| synchronized (mInMemoryLock) { |
| if (!isPersistenceInitializedMLocked()) { |
| Slog.e(LOG_TAG, "Interaction before persistence initialized"); |
| return; |
| } |
| clearHistoryOnDiskDLocked(); |
| mNextPersistDueTimeMillis = 0; |
| mPendingHistoryOffsetMillis = 0; |
| mCurrentHistoricalOps = null; |
| } |
| } |
| } |
| |
| private void clearHistoryOnDiskDLocked() { |
| BackgroundThread.getHandler().removeMessages(MSG_WRITE_PENDING_HISTORY); |
| synchronized (mInMemoryLock) { |
| mCurrentHistoricalOps = null; |
| mNextPersistDueTimeMillis = System.currentTimeMillis(); |
| mPendingWrites.clear(); |
| } |
| Persistence.clearHistoryDLocked(); |
| } |
| |
| private @NonNull HistoricalOps getUpdatedPendingHistoricalOpsMLocked(long now) { |
| if (mCurrentHistoricalOps != null) { |
| final long remainingTimeMillis = mNextPersistDueTimeMillis - now; |
| if (remainingTimeMillis > mBaseSnapshotInterval) { |
| // If time went backwards we need to push history to the future with the |
| // overflow over our snapshot interval. If time went forward do nothing |
| // as we would naturally push history into the past on the next write. |
| mPendingHistoryOffsetMillis = remainingTimeMillis - mBaseSnapshotInterval; |
| } |
| final long elapsedTimeMillis = mBaseSnapshotInterval - remainingTimeMillis; |
| mCurrentHistoricalOps.setEndTime(elapsedTimeMillis); |
| if (remainingTimeMillis > 0) { |
| if (DEBUG) { |
| Slog.i(LOG_TAG, "Returning current in-memory state"); |
| } |
| return mCurrentHistoricalOps; |
| } |
| if (mCurrentHistoricalOps.isEmpty()) { |
| mCurrentHistoricalOps.setBeginAndEndTime(0, 0); |
| mNextPersistDueTimeMillis = now + mBaseSnapshotInterval; |
| return mCurrentHistoricalOps; |
| } |
| // The current batch is full, so persist taking into account overdue persist time. |
| mCurrentHistoricalOps.offsetBeginAndEndTime(mBaseSnapshotInterval); |
| mCurrentHistoricalOps.setBeginTime(mCurrentHistoricalOps.getEndTimeMillis() |
| - mBaseSnapshotInterval); |
| final long overdueTimeMillis = Math.abs(remainingTimeMillis); |
| mCurrentHistoricalOps.offsetBeginAndEndTime(overdueTimeMillis); |
| schedulePersistHistoricalOpsMLocked(mCurrentHistoricalOps); |
| } |
| // The current batch is in the future, i.e. not complete yet. |
| mCurrentHistoricalOps = new HistoricalOps(0, 0); |
| mNextPersistDueTimeMillis = now + mBaseSnapshotInterval; |
| if (DEBUG) { |
| Slog.i(LOG_TAG, "Returning new in-memory state"); |
| } |
| return mCurrentHistoricalOps; |
| } |
| |
| void shutdown() { |
| synchronized (mInMemoryLock) { |
| if (mMode == AppOpsManager.HISTORICAL_MODE_DISABLED) { |
| return; |
| } |
| } |
| // Do not call persistPendingHistory inside the memory lock, due to possible deadlock |
| persistPendingHistory(); |
| } |
| |
| void persistPendingHistory() { |
| final List<HistoricalOps> pendingWrites; |
| synchronized (mOnDiskLock) { |
| synchronized (mInMemoryLock) { |
| pendingWrites = new ArrayList<>(mPendingWrites); |
| mPendingWrites.clear(); |
| if (mPendingHistoryOffsetMillis != 0) { |
| resampleHistoryOnDiskInMemoryDMLocked(mPendingHistoryOffsetMillis); |
| mPendingHistoryOffsetMillis = 0; |
| } |
| } |
| persistPendingHistory(pendingWrites); |
| } |
| } |
| |
| private void persistPendingHistory(@NonNull List<HistoricalOps> pendingWrites) { |
| synchronized (mOnDiskLock) { |
| BackgroundThread.getHandler().removeMessages(MSG_WRITE_PENDING_HISTORY); |
| if (pendingWrites.isEmpty()) { |
| return; |
| } |
| final int opCount = pendingWrites.size(); |
| // Pending writes are offset relative to each other, so take this |
| // into account to persist everything in one shot - single write. |
| for (int i = 0; i < opCount; i++) { |
| final HistoricalOps current = pendingWrites.get(i); |
| if (i > 0) { |
| final HistoricalOps previous = pendingWrites.get(i - 1); |
| current.offsetBeginAndEndTime(previous.getBeginTimeMillis()); |
| } |
| } |
| mPersistence.persistHistoricalOpsDLocked(pendingWrites); |
| } |
| } |
| |
| private void schedulePersistHistoricalOpsMLocked(@NonNull HistoricalOps ops) { |
| final Message message = PooledLambda.obtainMessage( |
| HistoricalRegistry::persistPendingHistory, HistoricalRegistry.this); |
| message.what = MSG_WRITE_PENDING_HISTORY; |
| BackgroundThread.getHandler().sendMessage(message); |
| mPendingWrites.offerFirst(ops); |
| } |
| |
| private static void makeRelativeToEpochStart(@NonNull HistoricalOps ops, long nowMillis) { |
| ops.setBeginAndEndTime(nowMillis - ops.getEndTimeMillis(), |
| nowMillis- ops.getBeginTimeMillis()); |
| } |
| |
| private void pruneFutureOps(@NonNull List<HistoricalOps> ops) { |
| final int opCount = ops.size(); |
| for (int i = opCount - 1; i >= 0; i--) { |
| final HistoricalOps op = ops.get(i); |
| if (op.getEndTimeMillis() <= mBaseSnapshotInterval) { |
| ops.remove(i); |
| } else if (op.getBeginTimeMillis() < mBaseSnapshotInterval) { |
| final double filterScale = (double) (op.getEndTimeMillis() - mBaseSnapshotInterval) |
| / (double) op.getDurationMillis(); |
| Persistence.spliceFromBeginning(op, filterScale); |
| } |
| } |
| } |
| |
| private static boolean isApiEnabled() { |
| return Binder.getCallingUid() == Process.myUid() |
| || DeviceConfig.getBoolean(DeviceConfig.NAMESPACE_PRIVACY, |
| PROPERTY_PERMISSIONS_HUB_ENABLED, false); |
| } |
| |
| private static final class Persistence { |
| private static final boolean DEBUG = false; |
| |
| private static final String LOG_TAG = Persistence.class.getSimpleName(); |
| |
| private static final String TAG_HISTORY = "history"; |
| private static final String TAG_OPS = "ops"; |
| private static final String TAG_UID = "uid"; |
| private static final String TAG_PACKAGE = "pkg"; |
| private static final String TAG_ATTRIBUTION = "ftr"; |
| private static final String TAG_OP = "op"; |
| private static final String TAG_STATE = "st"; |
| |
| private static final String ATTR_VERSION = "ver"; |
| private static final String ATTR_NAME = "na"; |
| private static final String ATTR_ACCESS_COUNT = "ac"; |
| private static final String ATTR_REJECT_COUNT = "rc"; |
| private static final String ATTR_ACCESS_DURATION = "du"; |
| private static final String ATTR_BEGIN_TIME = "beg"; |
| private static final String ATTR_END_TIME = "end"; |
| private static final String ATTR_OVERFLOW = "ov"; |
| |
| private static final int CURRENT_VERSION = 2; |
| |
| private final long mBaseSnapshotInterval; |
| private final long mIntervalCompressionMultiplier; |
| |
| Persistence(long baseSnapshotInterval, long intervalCompressionMultiplier) { |
| mBaseSnapshotInterval = baseSnapshotInterval; |
| mIntervalCompressionMultiplier = intervalCompressionMultiplier; |
| } |
| |
| private static final AtomicDirectory sHistoricalAppOpsDir = new AtomicDirectory( |
| new File(new File(Environment.getDataSystemDirectory(), "appops"), "history")); |
| |
| private File generateFile(@NonNull File baseDir, int depth) { |
| final long globalBeginMillis = computeGlobalIntervalBeginMillis(depth); |
| return new File(baseDir, Long.toString(globalBeginMillis) + HISTORY_FILE_SUFFIX); |
| } |
| |
| void clearHistoryDLocked(int uid, String packageName) { |
| List<HistoricalOps> historicalOps = readHistoryDLocked(); |
| |
| if (historicalOps == null) { |
| return; |
| } |
| |
| for (int index = 0; index < historicalOps.size(); index++) { |
| historicalOps.get(index).clearHistory(uid, packageName); |
| } |
| |
| clearHistoryDLocked(); |
| |
| persistHistoricalOpsDLocked(historicalOps); |
| } |
| |
| static void clearHistoryDLocked() { |
| sHistoricalAppOpsDir.delete(); |
| } |
| |
| void persistHistoricalOpsDLocked(@NonNull List<HistoricalOps> ops) { |
| if (DEBUG) { |
| Slog.i(LOG_TAG, "Persisting ops:\n" + opsToDebugString(ops)); |
| enforceOpsWellFormed(ops); |
| } |
| try { |
| final File newBaseDir = sHistoricalAppOpsDir.startWrite(); |
| final File oldBaseDir = sHistoricalAppOpsDir.getBackupDirectory(); |
| final HistoricalFilesInvariant filesInvariant; |
| if (DEBUG) { |
| filesInvariant = new HistoricalFilesInvariant(); |
| filesInvariant.startTracking(oldBaseDir); |
| } |
| final Set<String> oldFileNames = getHistoricalFileNames(oldBaseDir); |
| handlePersistHistoricalOpsRecursiveDLocked(newBaseDir, oldBaseDir, ops, |
| oldFileNames, 0); |
| if (DEBUG) { |
| filesInvariant.stopTracking(newBaseDir); |
| } |
| sHistoricalAppOpsDir.finishWrite(); |
| } catch (Throwable t) { |
| wtf("Failed to write historical app ops, restoring backup", t, null); |
| sHistoricalAppOpsDir.failWrite(); |
| } |
| } |
| |
| @Nullable List<HistoricalOps> readHistoryRawDLocked() { |
| return collectHistoricalOpsBaseDLocked(Process.INVALID_UID /*filterUid*/, |
| null /*filterPackageName*/, null /*filterAttributionTag*/, |
| null /*filterOpNames*/, 0 /*filter*/, 0 /*filterBeginTimeMills*/, |
| Long.MAX_VALUE /*filterEndTimeMills*/, AppOpsManager.OP_FLAGS_ALL); |
| } |
| |
| @Nullable List<HistoricalOps> readHistoryDLocked() { |
| final List<HistoricalOps> result = readHistoryRawDLocked(); |
| // Take into account in memory state duration. |
| if (result != null) { |
| final int opCount = result.size(); |
| for (int i = 0; i < opCount; i++) { |
| result.get(i).offsetBeginAndEndTime(mBaseSnapshotInterval); |
| } |
| } |
| return result; |
| } |
| |
| long getLastPersistTimeMillisDLocked() { |
| File baseDir = null; |
| try { |
| baseDir = sHistoricalAppOpsDir.startRead(); |
| final File[] files = baseDir.listFiles(); |
| if (files != null && files.length > 0) { |
| File shortestFile = null; |
| for (File candidate : files) { |
| final String candidateName = candidate.getName(); |
| if (!candidateName.endsWith(HISTORY_FILE_SUFFIX)) { |
| continue; |
| } |
| if (shortestFile == null) { |
| shortestFile = candidate; |
| } else if (candidateName.length() < shortestFile.getName().length()) { |
| shortestFile = candidate; |
| } |
| } |
| if (shortestFile == null) { |
| return 0; |
| } |
| return shortestFile.lastModified(); |
| } |
| sHistoricalAppOpsDir.finishRead(); |
| } catch (Throwable e) { |
| wtf("Error reading historical app ops. Deleting history.", e, baseDir); |
| sHistoricalAppOpsDir.delete(); |
| } |
| return 0; |
| } |
| |
| private void collectHistoricalOpsDLocked(@NonNull HistoricalOps currentOps, int filterUid, |
| @Nullable String filterPackageName, @Nullable String filterAttributionTag, |
| @Nullable String[] filterOpNames, @HistoricalOpsRequestFilter int filter, |
| long filterBeingMillis, long filterEndMillis, @OpFlags int filterFlags) { |
| final List<HistoricalOps> readOps = collectHistoricalOpsBaseDLocked(filterUid, |
| filterPackageName, filterAttributionTag, filterOpNames, filter, |
| filterBeingMillis, filterEndMillis, filterFlags); |
| if (readOps != null) { |
| final int readCount = readOps.size(); |
| for (int i = 0; i < readCount; i++) { |
| final HistoricalOps readOp = readOps.get(i); |
| currentOps.merge(readOp); |
| } |
| } |
| } |
| |
| private @Nullable LinkedList<HistoricalOps> collectHistoricalOpsBaseDLocked(int filterUid, |
| @Nullable String filterPackageName, @Nullable String filterAttributionTag, |
| @Nullable String[] filterOpNames, @HistoricalOpsRequestFilter int filter, |
| long filterBeginTimeMillis, long filterEndTimeMillis, @OpFlags int filterFlags) { |
| File baseDir = null; |
| try { |
| baseDir = sHistoricalAppOpsDir.startRead(); |
| final HistoricalFilesInvariant filesInvariant; |
| if (DEBUG) { |
| filesInvariant = new HistoricalFilesInvariant(); |
| filesInvariant.startTracking(baseDir); |
| } |
| final Set<String> historyFiles = getHistoricalFileNames(baseDir); |
| final long[] globalContentOffsetMillis = {0}; |
| final LinkedList<HistoricalOps> ops = collectHistoricalOpsRecursiveDLocked( |
| baseDir, filterUid, filterPackageName, filterAttributionTag, filterOpNames, |
| filter, filterBeginTimeMillis, filterEndTimeMillis, filterFlags, |
| globalContentOffsetMillis, null /*outOps*/, 0 /*depth*/, historyFiles); |
| if (DEBUG) { |
| filesInvariant.stopTracking(baseDir); |
| } |
| sHistoricalAppOpsDir.finishRead(); |
| return ops; |
| } catch (Throwable t) { |
| wtf("Error reading historical app ops. Deleting history.", t, baseDir); |
| sHistoricalAppOpsDir.delete(); |
| } |
| return null; |
| } |
| |
| private @Nullable LinkedList<HistoricalOps> collectHistoricalOpsRecursiveDLocked( |
| @NonNull File baseDir, int filterUid, @Nullable String filterPackageName, |
| @Nullable String filterAttributionTag, @Nullable String[] filterOpNames, |
| @HistoricalOpsRequestFilter int filter, long filterBeginTimeMillis, |
| long filterEndTimeMillis, @OpFlags int filterFlags, |
| @NonNull long[] globalContentOffsetMillis, |
| @Nullable LinkedList<HistoricalOps> outOps, int depth, |
| @NonNull Set<String> historyFiles) |
| throws IOException, XmlPullParserException { |
| final long previousIntervalEndMillis = (long) Math.pow(mIntervalCompressionMultiplier, |
| depth) * mBaseSnapshotInterval; |
| final long currentIntervalEndMillis = (long) Math.pow(mIntervalCompressionMultiplier, |
| depth + 1) * mBaseSnapshotInterval; |
| |
| filterBeginTimeMillis = Math.max(filterBeginTimeMillis - previousIntervalEndMillis, 0); |
| filterEndTimeMillis = filterEndTimeMillis - previousIntervalEndMillis; |
| |
| // Read historical data at this level |
| final List<HistoricalOps> readOps = readHistoricalOpsLocked(baseDir, |
| previousIntervalEndMillis, currentIntervalEndMillis, filterUid, |
| filterPackageName, filterAttributionTag, filterOpNames, filter, |
| filterBeginTimeMillis, filterEndTimeMillis, filterFlags, |
| globalContentOffsetMillis, depth, historyFiles); |
| // Empty is a special signal to stop diving |
| if (readOps != null && readOps.isEmpty()) { |
| return outOps; |
| } |
| |
| // Collect older historical data from subsequent levels |
| outOps = collectHistoricalOpsRecursiveDLocked(baseDir, filterUid, filterPackageName, |
| filterAttributionTag, filterOpNames, filter, filterBeginTimeMillis, |
| filterEndTimeMillis, filterFlags, globalContentOffsetMillis, outOps, depth + 1, |
| historyFiles); |
| |
| // Make older historical data relative to the current historical level |
| if (outOps != null) { |
| final int opCount = outOps.size(); |
| for (int i = 0; i < opCount; i++) { |
| final HistoricalOps collectedOp = outOps.get(i); |
| collectedOp.offsetBeginAndEndTime(currentIntervalEndMillis); |
| } |
| } |
| |
| if (readOps != null) { |
| if (outOps == null) { |
| outOps = new LinkedList<>(); |
| } |
| // Add the read ops to output |
| final int opCount = readOps.size(); |
| for (int i = opCount - 1; i >= 0; i--) { |
| outOps.offerFirst(readOps.get(i)); |
| } |
| } |
| |
| return outOps; |
| } |
| |
| private void handlePersistHistoricalOpsRecursiveDLocked(@NonNull File newBaseDir, |
| @NonNull File oldBaseDir, @Nullable List<HistoricalOps> passedOps, |
| @NonNull Set<String> oldFileNames, int depth) |
| throws IOException, XmlPullParserException { |
| final long previousIntervalEndMillis = (long) Math.pow(mIntervalCompressionMultiplier, |
| depth) * mBaseSnapshotInterval; |
| final long currentIntervalEndMillis = (long) Math.pow(mIntervalCompressionMultiplier, |
| depth + 1) * mBaseSnapshotInterval; |
| |
| if (passedOps == null || passedOps.isEmpty()) { |
| if (!oldFileNames.isEmpty()) { |
| // If there is an old file we need to copy it over to the new state. |
| final File oldFile = generateFile(oldBaseDir, depth); |
| if (oldFileNames.remove(oldFile.getName())) { |
| final File newFile = generateFile(newBaseDir, depth); |
| Files.createLink(newFile.toPath(), oldFile.toPath()); |
| } |
| handlePersistHistoricalOpsRecursiveDLocked(newBaseDir, oldBaseDir, |
| passedOps, oldFileNames, depth + 1); |
| } |
| return; |
| } |
| |
| if (DEBUG) { |
| enforceOpsWellFormed(passedOps); |
| } |
| |
| // Normalize passed ops time to be based off this interval start |
| final int passedOpCount = passedOps.size(); |
| for (int i = 0; i < passedOpCount; i++) { |
| final HistoricalOps passedOp = passedOps.get(i); |
| passedOp.offsetBeginAndEndTime(-previousIntervalEndMillis); |
| } |
| |
| if (DEBUG) { |
| enforceOpsWellFormed(passedOps); |
| } |
| |
| // Read persisted ops for this interval |
| final List<HistoricalOps> existingOps = readHistoricalOpsLocked(oldBaseDir, |
| previousIntervalEndMillis, currentIntervalEndMillis, |
| Process.INVALID_UID /*filterUid*/, null /*filterPackageName*/, |
| null /*filterAttributionTag*/, null /*filterOpNames*/, 0 /*filter*/, |
| Long.MIN_VALUE /*filterBeginTimeMillis*/, |
| Long.MAX_VALUE /*filterEndTimeMillis*/, AppOpsManager.OP_FLAGS_ALL, null, depth, |
| null /*historyFiles*/); |
| if (DEBUG) { |
| enforceOpsWellFormed(existingOps); |
| } |
| |
| // Offset existing ops to account for elapsed time |
| if (existingOps != null) { |
| final int existingOpCount = existingOps.size(); |
| if (existingOpCount > 0) { |
| // Compute elapsed time |
| final long elapsedTimeMillis = passedOps.get(passedOps.size() - 1) |
| .getEndTimeMillis(); |
| for (int i = 0; i < existingOpCount; i++) { |
| final HistoricalOps existingOp = existingOps.get(i); |
| existingOp.offsetBeginAndEndTime(elapsedTimeMillis); |
| } |
| } |
| } |
| |
| if (DEBUG) { |
| enforceOpsWellFormed(existingOps); |
| } |
| |
| final long slotDurationMillis = previousIntervalEndMillis; |
| |
| // Consolidate passed ops at the current slot duration ensuring each snapshot is |
| // full. To achieve this we put all passed and existing ops in a list and will |
| // merge them to ensure each represents a snapshot at the current granularity. |
| final List<HistoricalOps> allOps = new LinkedList<>(passedOps); |
| if (existingOps != null) { |
| allOps.addAll(existingOps); |
| } |
| |
| if (DEBUG) { |
| enforceOpsWellFormed(allOps); |
| } |
| |
| // Compute ops to persist and overflow ops |
| List<HistoricalOps> persistedOps = null; |
| List<HistoricalOps> overflowedOps = null; |
| |
| // We move a snapshot into the next level only if the start time is |
| // after the end of the current interval. This avoids rewriting all |
| // files to propagate the information added to the history on every |
| // iteration. Instead, we would rewrite the next level file only if |
| // an entire snapshot from the previous level is being propagated. |
| // The trade off is that we need to store how much the last snapshot |
| // of the current interval overflows past the interval end. We write |
| // the overflow data to avoid parsing all snapshots on query. |
| long intervalOverflowMillis = 0; |
| final int opCount = allOps.size(); |
| for (int i = 0; i < opCount; i++) { |
| final HistoricalOps op = allOps.get(i); |
| final HistoricalOps persistedOp; |
| final HistoricalOps overflowedOp; |
| if (op.getEndTimeMillis() <= currentIntervalEndMillis) { |
| persistedOp = op; |
| overflowedOp = null; |
| } else if (op.getBeginTimeMillis() < currentIntervalEndMillis) { |
| persistedOp = op; |
| intervalOverflowMillis = op.getEndTimeMillis() - currentIntervalEndMillis; |
| if (intervalOverflowMillis > previousIntervalEndMillis) { |
| final double splitScale = (double) intervalOverflowMillis |
| / op.getDurationMillis(); |
| overflowedOp = spliceFromEnd(op, splitScale); |
| intervalOverflowMillis = op.getEndTimeMillis() - currentIntervalEndMillis; |
| } else { |
| overflowedOp = null; |
| } |
| } else { |
| persistedOp = null; |
| overflowedOp = op; |
| } |
| if (persistedOp != null) { |
| if (persistedOps == null) { |
| persistedOps = new ArrayList<>(); |
| } |
| persistedOps.add(persistedOp); |
| } |
| if (overflowedOp != null) { |
| if (overflowedOps == null) { |
| overflowedOps = new ArrayList<>(); |
| } |
| overflowedOps.add(overflowedOp); |
| } |
| } |
| |
| if (DEBUG) { |
| enforceOpsWellFormed(persistedOps); |
| enforceOpsWellFormed(overflowedOps); |
| } |
| |
| final File newFile = generateFile(newBaseDir, depth); |
| oldFileNames.remove(newFile.getName()); |
| |
| if (persistedOps != null) { |
| normalizeSnapshotForSlotDuration(persistedOps, slotDurationMillis); |
| writeHistoricalOpsDLocked(persistedOps, intervalOverflowMillis, newFile); |
| if (DEBUG) { |
| Slog.i(LOG_TAG, "Persisted at depth: " + depth + " file: " + newFile |
| + " ops:\n" + opsToDebugString(persistedOps)); |
| enforceOpsWellFormed(persistedOps); |
| } |
| } |
| |
| handlePersistHistoricalOpsRecursiveDLocked(newBaseDir, oldBaseDir, |
| overflowedOps, oldFileNames, depth + 1); |
| } |
| |
| private @Nullable List<HistoricalOps> readHistoricalOpsLocked(File baseDir, |
| long intervalBeginMillis, long intervalEndMillis, int filterUid, |
| @Nullable String filterPackageName, @Nullable String filterAttributionTag, |
| @Nullable String[] filterOpNames, @HistoricalOpsRequestFilter int filter, |
| long filterBeginTimeMillis, long filterEndTimeMillis, @OpFlags int filterFlags, |
| @Nullable long[] cumulativeOverflowMillis, int depth, |
| @NonNull Set<String> historyFiles) |
| throws IOException, XmlPullParserException { |
| final File file = generateFile(baseDir, depth); |
| if (historyFiles != null) { |
| historyFiles.remove(file.getName()); |
| } |
| if (filterBeginTimeMillis >= filterEndTimeMillis |
| || filterEndTimeMillis < intervalBeginMillis) { |
| // Don't go deeper |
| return Collections.emptyList(); |
| } |
| if (filterBeginTimeMillis >= (intervalEndMillis |
| + ((intervalEndMillis - intervalBeginMillis) / mIntervalCompressionMultiplier) |
| + (cumulativeOverflowMillis != null ? cumulativeOverflowMillis[0] : 0)) |
| || !file.exists()) { |
| if (historyFiles == null || historyFiles.isEmpty()) { |
| // Don't go deeper |
| return Collections.emptyList(); |
| } else { |
| // Keep diving |
| return null; |
| } |
| } |
| return readHistoricalOpsLocked(file, filterUid, filterPackageName, filterAttributionTag, |
| filterOpNames, filter, filterBeginTimeMillis, filterEndTimeMillis, filterFlags, |
| cumulativeOverflowMillis); |
| } |
| |
| private @Nullable List<HistoricalOps> readHistoricalOpsLocked(@NonNull File file, |
| int filterUid, @Nullable String filterPackageName, |
| @Nullable String filterAttributionTag, @Nullable String[] filterOpNames, |
| @HistoricalOpsRequestFilter int filter, long filterBeginTimeMillis, |
| long filterEndTimeMillis, @OpFlags int filterFlags, |
| @Nullable long[] cumulativeOverflowMillis) |
| throws IOException, XmlPullParserException { |
| if (DEBUG) { |
| Slog.i(LOG_TAG, "Reading ops from:" + file); |
| } |
| List<HistoricalOps> allOps = null; |
| try (FileInputStream stream = new FileInputStream(file)) { |
| final XmlPullParser parser = Xml.newPullParser(); |
| parser.setInput(stream, StandardCharsets.UTF_8.name()); |
| XmlUtils.beginDocument(parser, TAG_HISTORY); |
| |
| // We haven't released version 1 and have more detailed |
| // accounting - just nuke the current state |
| final int version = XmlUtils.readIntAttribute(parser, ATTR_VERSION); |
| if (CURRENT_VERSION == 2 && version < CURRENT_VERSION) { |
| throw new IllegalStateException("Dropping unsupported history " |
| + "version 1 for file:" + file); |
| } |
| |
| final long overflowMillis = XmlUtils.readLongAttribute(parser, ATTR_OVERFLOW, 0); |
| final int depth = parser.getDepth(); |
| while (XmlUtils.nextElementWithin(parser, depth)) { |
| if (TAG_OPS.equals(parser.getName())) { |
| final HistoricalOps ops = readeHistoricalOpsDLocked(parser, filterUid, |
| filterPackageName, filterAttributionTag, filterOpNames, filter, |
| filterBeginTimeMillis, filterEndTimeMillis, filterFlags, |
| cumulativeOverflowMillis); |
| if (ops == null) { |
| continue; |
| } |
| if (ops.isEmpty()) { |
| XmlUtils.skipCurrentTag(parser); |
| continue; |
| } |
| if (allOps == null) { |
| allOps = new ArrayList<>(); |
| } |
| allOps.add(ops); |
| } |
| } |
| if (cumulativeOverflowMillis != null) { |
| cumulativeOverflowMillis[0] += overflowMillis; |
| } |
| } catch (FileNotFoundException e) { |
| Slog.i(LOG_TAG, "No history file: " + file.getName()); |
| return Collections.emptyList(); |
| } |
| if (DEBUG) { |
| if (allOps != null) { |
| Slog.i(LOG_TAG, "Read from file: " + file + " ops:\n" |
| + opsToDebugString(allOps)); |
| enforceOpsWellFormed(allOps); |
| } |
| } |
| return allOps; |
| } |
| |
| private @Nullable HistoricalOps readeHistoricalOpsDLocked( |
| @NonNull XmlPullParser parser, int filterUid, @Nullable String filterPackageName, |
| @Nullable String filterAttributionTag, @Nullable String[] filterOpNames, |
| @HistoricalOpsRequestFilter int filter, long filterBeginTimeMillis, |
| long filterEndTimeMillis, @OpFlags int filterFlags, |
| @Nullable long[] cumulativeOverflowMillis) |
| throws IOException, XmlPullParserException { |
| final long beginTimeMillis = XmlUtils.readLongAttribute(parser, ATTR_BEGIN_TIME, 0) |
| + (cumulativeOverflowMillis != null ? cumulativeOverflowMillis[0] : 0); |
| final long endTimeMillis = XmlUtils.readLongAttribute(parser, ATTR_END_TIME, 0) |
| + (cumulativeOverflowMillis != null ? cumulativeOverflowMillis[0] : 0); |
| // Keep reading as subsequent records may start matching |
| if (filterEndTimeMillis < beginTimeMillis) { |
| return null; |
| } |
| // Stop reading as subsequent records will not match |
| if (filterBeginTimeMillis > endTimeMillis) { |
| return new HistoricalOps(0, 0); |
| } |
| final long filteredBeginTimeMillis = Math.max(beginTimeMillis, filterBeginTimeMillis); |
| final long filteredEndTimeMillis = Math.min(endTimeMillis, filterEndTimeMillis); |
| // // Keep reading as subsequent records may start matching |
| // if (filteredEndTimeMillis - filterBeginTimeMillis <= 0) { |
| // return null; |
| // } |
| final double filterScale = (double) (filteredEndTimeMillis - filteredBeginTimeMillis) |
| / (double) (endTimeMillis - beginTimeMillis); |
| HistoricalOps ops = null; |
| final int depth = parser.getDepth(); |
| while (XmlUtils.nextElementWithin(parser, depth)) { |
| if (TAG_UID.equals(parser.getName())) { |
| final HistoricalOps returnedOps = readHistoricalUidOpsDLocked(ops, parser, |
| filterUid, filterPackageName, filterAttributionTag, filterOpNames, |
| filter, filterFlags, filterScale); |
| if (ops == null) { |
| ops = returnedOps; |
| } |
| } |
| } |
| if (ops != null) { |
| ops.setBeginAndEndTime(filteredBeginTimeMillis, filteredEndTimeMillis); |
| } |
| return ops; |
| } |
| |
| private @Nullable HistoricalOps readHistoricalUidOpsDLocked( |
| @Nullable HistoricalOps ops, @NonNull XmlPullParser parser, int filterUid, |
| @Nullable String filterPackageName, @Nullable String filterAttributionTag, |
| @Nullable String[] filterOpNames, @HistoricalOpsRequestFilter int filter, |
| @OpFlags int filterFlags, double filterScale) |
| throws IOException, XmlPullParserException { |
| final int uid = XmlUtils.readIntAttribute(parser, ATTR_NAME); |
| if ((filter & FILTER_BY_UID) != 0 && filterUid != uid) { |
| XmlUtils.skipCurrentTag(parser); |
| return null; |
| } |
| final int depth = parser.getDepth(); |
| while (XmlUtils.nextElementWithin(parser, depth)) { |
| if (TAG_PACKAGE.equals(parser.getName())) { |
| final HistoricalOps returnedOps = readHistoricalPackageOpsDLocked(ops, uid, |
| parser, filterPackageName, filterAttributionTag, filterOpNames, filter, |
| filterFlags, filterScale); |
| if (ops == null) { |
| ops = returnedOps; |
| } |
| } |
| } |
| return ops; |
| } |
| |
| private @Nullable HistoricalOps readHistoricalPackageOpsDLocked( |
| @Nullable HistoricalOps ops, int uid, @NonNull XmlPullParser parser, |
| @Nullable String filterPackageName, @Nullable String filterAttributionTag, |
| @Nullable String[] filterOpNames, @HistoricalOpsRequestFilter int filter, |
| @OpFlags int filterFlags, double filterScale) |
| throws IOException, XmlPullParserException { |
| final String packageName = XmlUtils.readStringAttribute(parser, ATTR_NAME); |
| if ((filter & FILTER_BY_PACKAGE_NAME) != 0 && !filterPackageName.equals(packageName)) { |
| XmlUtils.skipCurrentTag(parser); |
| return null; |
| } |
| final int depth = parser.getDepth(); |
| while (XmlUtils.nextElementWithin(parser, depth)) { |
| if (TAG_ATTRIBUTION.equals(parser.getName())) { |
| final HistoricalOps returnedOps = readHistoricalAttributionOpsDLocked(ops, uid, |
| packageName, parser, filterAttributionTag, filterOpNames, filter, |
| filterFlags, filterScale); |
| if (ops == null) { |
| ops = returnedOps; |
| } |
| } |
| } |
| return ops; |
| } |
| |
| private @Nullable HistoricalOps readHistoricalAttributionOpsDLocked( |
| @Nullable HistoricalOps ops, int uid, String packageName, |
| @NonNull XmlPullParser parser, @Nullable String filterAttributionTag, |
| @Nullable String[] filterOpNames, @HistoricalOpsRequestFilter int filter, |
| @OpFlags int filterFlags, double filterScale) |
| throws IOException, XmlPullParserException { |
| final String attributionTag = XmlUtils.readStringAttribute(parser, ATTR_NAME); |
| if ((filter & FILTER_BY_ATTRIBUTION_TAG) != 0 && !Objects.equals(filterAttributionTag, |
| attributionTag)) { |
| XmlUtils.skipCurrentTag(parser); |
| return null; |
| } |
| final int depth = parser.getDepth(); |
| while (XmlUtils.nextElementWithin(parser, depth)) { |
| if (TAG_OP.equals(parser.getName())) { |
| final HistoricalOps returnedOps = readHistoricalOpDLocked(ops, uid, packageName, |
| attributionTag, parser, filterOpNames, filter, filterFlags, |
| filterScale); |
| if (ops == null) { |
| ops = returnedOps; |
| } |
| } |
| } |
| return ops; |
| } |
| |
| private @Nullable HistoricalOps readHistoricalOpDLocked(@Nullable HistoricalOps ops, |
| int uid, @NonNull String packageName, @Nullable String attributionTag, |
| @NonNull XmlPullParser parser, @Nullable String[] filterOpNames, |
| @HistoricalOpsRequestFilter int filter, @OpFlags int filterFlags, |
| double filterScale) |
| throws IOException, XmlPullParserException { |
| final int op = XmlUtils.readIntAttribute(parser, ATTR_NAME); |
| if ((filter & FILTER_BY_OP_NAMES) != 0 && !ArrayUtils.contains(filterOpNames, |
| AppOpsManager.opToPublicName(op))) { |
| XmlUtils.skipCurrentTag(parser); |
| return null; |
| } |
| final int depth = parser.getDepth(); |
| while (XmlUtils.nextElementWithin(parser, depth)) { |
| if (TAG_STATE.equals(parser.getName())) { |
| final HistoricalOps returnedOps = readStateDLocked(ops, uid, |
| packageName, attributionTag, op, parser, filterFlags, filterScale); |
| if (ops == null) { |
| ops = returnedOps; |
| } |
| } |
| } |
| return ops; |
| } |
| |
| private @Nullable HistoricalOps readStateDLocked(@Nullable HistoricalOps ops, |
| int uid, @NonNull String packageName, @Nullable String attributionTag, int op, |
| @NonNull XmlPullParser parser, @OpFlags int filterFlags, double filterScale) |
| throws IOException { |
| final long key = XmlUtils.readLongAttribute(parser, ATTR_NAME); |
| final int flags = AppOpsManager.extractFlagsFromKey(key) & filterFlags; |
| if (flags == 0) { |
| return null; |
| } |
| final int uidState = AppOpsManager.extractUidStateFromKey(key); |
| long accessCount = XmlUtils.readLongAttribute(parser, ATTR_ACCESS_COUNT, 0); |
| if (accessCount > 0) { |
| if (!Double.isNaN(filterScale)) { |
| accessCount = (long) HistoricalOps.round( |
| (double) accessCount * filterScale); |
| } |
| if (ops == null) { |
| ops = new HistoricalOps(0, 0); |
| } |
| ops.increaseAccessCount(op, uid, packageName, attributionTag, uidState, flags, |
| accessCount); |
| } |
| long rejectCount = XmlUtils.readLongAttribute(parser, ATTR_REJECT_COUNT, 0); |
| if (rejectCount > 0) { |
| if (!Double.isNaN(filterScale)) { |
| rejectCount = (long) HistoricalOps.round( |
| (double) rejectCount * filterScale); |
| } |
| if (ops == null) { |
| ops = new HistoricalOps(0, 0); |
| } |
| ops.increaseRejectCount(op, uid, packageName, attributionTag, uidState, flags, |
| rejectCount); |
| } |
| long accessDuration = XmlUtils.readLongAttribute(parser, ATTR_ACCESS_DURATION, 0); |
| if (accessDuration > 0) { |
| if (!Double.isNaN(filterScale)) { |
| accessDuration = (long) HistoricalOps.round( |
| (double) accessDuration * filterScale); |
| } |
| if (ops == null) { |
| ops = new HistoricalOps(0, 0); |
| } |
| ops.increaseAccessDuration(op, uid, packageName, attributionTag, uidState, flags, |
| accessDuration); |
| } |
| return ops; |
| } |
| |
| private void writeHistoricalOpsDLocked(@Nullable List<HistoricalOps> allOps, |
| long intervalOverflowMillis, @NonNull File file) throws IOException { |
| final FileOutputStream output = sHistoricalAppOpsDir.openWrite(file); |
| try { |
| final XmlSerializer serializer = Xml.newSerializer(); |
| serializer.setOutput(output, StandardCharsets.UTF_8.name()); |
| serializer.setFeature("http://xmlpull.org/v1/doc/features.html#indent-output", |
| true); |
| serializer.startDocument(null, true); |
| serializer.startTag(null, TAG_HISTORY); |
| serializer.attribute(null, ATTR_VERSION, String.valueOf(CURRENT_VERSION)); |
| if (intervalOverflowMillis != 0) { |
| serializer.attribute(null, ATTR_OVERFLOW, |
| Long.toString(intervalOverflowMillis)); |
| } |
| if (allOps != null) { |
| final int opsCount = allOps.size(); |
| for (int i = 0; i < opsCount; i++) { |
| final HistoricalOps ops = allOps.get(i); |
| writeHistoricalOpDLocked(ops, serializer); |
| } |
| } |
| serializer.endTag(null, TAG_HISTORY); |
| serializer.endDocument(); |
| sHistoricalAppOpsDir.closeWrite(output); |
| } catch (IOException e) { |
| sHistoricalAppOpsDir.failWrite(output); |
| throw e; |
| } |
| } |
| |
| private void writeHistoricalOpDLocked(@NonNull HistoricalOps ops, |
| @NonNull XmlSerializer serializer) throws IOException { |
| serializer.startTag(null, TAG_OPS); |
| serializer.attribute(null, ATTR_BEGIN_TIME, Long.toString(ops.getBeginTimeMillis())); |
| serializer.attribute(null, ATTR_END_TIME, Long.toString(ops.getEndTimeMillis())); |
| final int uidCount = ops.getUidCount(); |
| for (int i = 0; i < uidCount; i++) { |
| final HistoricalUidOps uidOp = ops.getUidOpsAt(i); |
| writeHistoricalUidOpsDLocked(uidOp, serializer); |
| } |
| serializer.endTag(null, TAG_OPS); |
| } |
| |
| private void writeHistoricalUidOpsDLocked(@NonNull HistoricalUidOps uidOps, |
| @NonNull XmlSerializer serializer) throws IOException { |
| serializer.startTag(null, TAG_UID); |
| serializer.attribute(null, ATTR_NAME, Integer.toString(uidOps.getUid())); |
| final int packageCount = uidOps.getPackageCount(); |
| for (int i = 0; i < packageCount; i++) { |
| final HistoricalPackageOps packageOps = uidOps.getPackageOpsAt(i); |
| writeHistoricalPackageOpsDLocked(packageOps, serializer); |
| } |
| serializer.endTag(null, TAG_UID); |
| } |
| |
| private void writeHistoricalPackageOpsDLocked(@NonNull HistoricalPackageOps packageOps, |
| @NonNull XmlSerializer serializer) throws IOException { |
| serializer.startTag(null, TAG_PACKAGE); |
| serializer.attribute(null, ATTR_NAME, packageOps.getPackageName()); |
| final int numAttributions = packageOps.getAttributedOpsCount(); |
| for (int i = 0; i < numAttributions; i++) { |
| final AppOpsManager.AttributedHistoricalOps op = packageOps.getAttributedOpsAt(i); |
| writeHistoricalAttributionOpsDLocked(op, serializer); |
| } |
| serializer.endTag(null, TAG_PACKAGE); |
| } |
| |
| private void writeHistoricalAttributionOpsDLocked( |
| @NonNull AppOpsManager.AttributedHistoricalOps attributionOps, |
| @NonNull XmlSerializer serializer) throws IOException { |
| serializer.startTag(null, TAG_ATTRIBUTION); |
| XmlUtils.writeStringAttribute(serializer, ATTR_NAME, attributionOps.getTag()); |
| final int opCount = attributionOps.getOpCount(); |
| for (int i = 0; i < opCount; i++) { |
| final HistoricalOp op = attributionOps.getOpAt(i); |
| writeHistoricalOpDLocked(op, serializer); |
| } |
| serializer.endTag(null, TAG_ATTRIBUTION); |
| } |
| |
| private void writeHistoricalOpDLocked(@NonNull HistoricalOp op, |
| @NonNull XmlSerializer serializer) throws IOException { |
| final LongSparseArray keys = op.collectKeys(); |
| if (keys == null || keys.size() <= 0) { |
| return; |
| } |
| serializer.startTag(null, TAG_OP); |
| serializer.attribute(null, ATTR_NAME, Integer.toString(op.getOpCode())); |
| final int keyCount = keys.size(); |
| for (int i = 0; i < keyCount; i++) { |
| writeStateOnLocked(op, keys.keyAt(i), serializer); |
| } |
| serializer.endTag(null, TAG_OP); |
| } |
| |
| private void writeStateOnLocked(@NonNull HistoricalOp op, long key, |
| @NonNull XmlSerializer serializer) throws IOException { |
| final int uidState = AppOpsManager.extractUidStateFromKey(key); |
| final int flags = AppOpsManager.extractFlagsFromKey(key); |
| |
| final long accessCount = op.getAccessCount(uidState, uidState, flags); |
| final long rejectCount = op.getRejectCount(uidState, uidState, flags); |
| final long accessDuration = op.getAccessDuration(uidState, uidState, flags); |
| |
| if (accessCount <= 0 && rejectCount <= 0 && accessDuration <= 0) { |
| return; |
| } |
| |
| serializer.startTag(null, TAG_STATE); |
| serializer.attribute(null, ATTR_NAME, Long.toString(key)); |
| if (accessCount > 0) { |
| serializer.attribute(null, ATTR_ACCESS_COUNT, Long.toString(accessCount)); |
| } |
| if (rejectCount > 0) { |
| serializer.attribute(null, ATTR_REJECT_COUNT, Long.toString(rejectCount)); |
| } |
| if (accessDuration > 0) { |
| serializer.attribute(null, ATTR_ACCESS_DURATION, Long.toString(accessDuration)); |
| } |
| serializer.endTag(null, TAG_STATE); |
| } |
| |
| private static void enforceOpsWellFormed(@NonNull List<HistoricalOps> ops) { |
| if (ops == null) { |
| return; |
| } |
| HistoricalOps previous; |
| HistoricalOps current = null; |
| final int opsCount = ops.size(); |
| for (int i = 0; i < opsCount; i++) { |
| previous = current; |
| current = ops.get(i); |
| if (current.isEmpty()) { |
| throw new IllegalStateException("Empty ops:\n" |
| + opsToDebugString(ops)); |
| } |
| if (current.getEndTimeMillis() < current.getBeginTimeMillis()) { |
| throw new IllegalStateException("Begin after end:\n" |
| + opsToDebugString(ops)); |
| } |
| if (previous != null) { |
| if (previous.getEndTimeMillis() > current.getBeginTimeMillis()) { |
| throw new IllegalStateException("Intersecting ops:\n" |
| + opsToDebugString(ops)); |
| } |
| if (previous.getBeginTimeMillis() > current.getBeginTimeMillis()) { |
| throw new IllegalStateException("Non increasing ops:\n" |
| + opsToDebugString(ops)); |
| } |
| } |
| } |
| } |
| |
| private long computeGlobalIntervalBeginMillis(int depth) { |
| long beginTimeMillis = 0; |
| for (int i = 0; i < depth + 1; i++) { |
| beginTimeMillis += Math.pow(mIntervalCompressionMultiplier, i); |
| } |
| return beginTimeMillis * mBaseSnapshotInterval; |
| } |
| |
| private static @NonNull HistoricalOps spliceFromEnd(@NonNull HistoricalOps ops, |
| double spliceRatio) { |
| if (DEBUG) { |
| Slog.w(LOG_TAG, "Splicing from end:" + ops + " ratio:" + spliceRatio); |
| } |
| final HistoricalOps splice = ops.spliceFromEnd(spliceRatio); |
| if (DEBUG) { |
| Slog.w(LOG_TAG, "Spliced into:" + ops + " and:" + splice); |
| } |
| return splice; |
| } |
| |
| |
| private static @NonNull HistoricalOps spliceFromBeginning(@NonNull HistoricalOps ops, |
| double spliceRatio) { |
| if (DEBUG) { |
| Slog.w(LOG_TAG, "Splicing from beginning:" + ops + " ratio:" + spliceRatio); |
| } |
| final HistoricalOps splice = ops.spliceFromBeginning(spliceRatio); |
| if (DEBUG) { |
| Slog.w(LOG_TAG, "Spliced into:" + ops + " and:" + splice); |
| } |
| return splice; |
| } |
| |
| private static void normalizeSnapshotForSlotDuration(@NonNull List<HistoricalOps> ops, |
| long slotDurationMillis) { |
| if (DEBUG) { |
| Slog.i(LOG_TAG, "Normalizing for slot duration: " + slotDurationMillis |
| + " ops:\n" + opsToDebugString(ops)); |
| enforceOpsWellFormed(ops); |
| } |
| long slotBeginTimeMillis; |
| final int opCount = ops.size(); |
| for (int processedIdx = opCount - 1; processedIdx >= 0; processedIdx--) { |
| final HistoricalOps processedOp = ops.get(processedIdx); |
| slotBeginTimeMillis = Math.max(processedOp.getEndTimeMillis() |
| - slotDurationMillis, 0); |
| for (int candidateIdx = processedIdx - 1; candidateIdx >= 0; candidateIdx--) { |
| final HistoricalOps candidateOp = ops.get(candidateIdx); |
| final long candidateSlotIntersectionMillis = candidateOp.getEndTimeMillis() |
| - Math.min(slotBeginTimeMillis, processedOp.getBeginTimeMillis()); |
| if (candidateSlotIntersectionMillis <= 0) { |
| break; |
| } |
| final float candidateSplitRatio = candidateSlotIntersectionMillis |
| / (float) candidateOp.getDurationMillis(); |
| if (Float.compare(candidateSplitRatio, 1.0f) >= 0) { |
| ops.remove(candidateIdx); |
| processedIdx--; |
| processedOp.merge(candidateOp); |
| } else { |
| final HistoricalOps endSplice = spliceFromEnd(candidateOp, |
| candidateSplitRatio); |
| if (endSplice != null) { |
| processedOp.merge(endSplice); |
| } |
| if (candidateOp.isEmpty()) { |
| ops.remove(candidateIdx); |
| processedIdx--; |
| } |
| } |
| } |
| } |
| if (DEBUG) { |
| Slog.i(LOG_TAG, "Normalized for slot duration: " + slotDurationMillis |
| + " ops:\n" + opsToDebugString(ops)); |
| enforceOpsWellFormed(ops); |
| } |
| } |
| |
| private static @NonNull String opsToDebugString(@NonNull List<HistoricalOps> ops) { |
| StringBuilder builder = new StringBuilder(); |
| final int opCount = ops.size(); |
| for (int i = 0; i < opCount; i++) { |
| builder.append(" "); |
| builder.append(ops.get(i)); |
| if (i < opCount - 1) { |
| builder.append('\n'); |
| } |
| } |
| return builder.toString(); |
| } |
| |
| private static Set<String> getHistoricalFileNames(@NonNull File historyDir) { |
| final File[] files = historyDir.listFiles(); |
| if (files == null) { |
| return Collections.emptySet(); |
| } |
| final ArraySet<String> fileNames = new ArraySet<>(files.length); |
| for (File file : files) { |
| fileNames.add(file.getName()); |
| |
| } |
| return fileNames; |
| } |
| } |
| |
| private static class HistoricalFilesInvariant { |
| private final @NonNull List<File> mBeginFiles = new ArrayList<>(); |
| |
| public void startTracking(@NonNull File folder) { |
| final File[] files = folder.listFiles(); |
| if (files != null) { |
| Collections.addAll(mBeginFiles, files); |
| } |
| } |
| |
| public void stopTracking(@NonNull File folder) { |
| final List<File> endFiles = new ArrayList<>(); |
| final File[] files = folder.listFiles(); |
| if (files != null) { |
| Collections.addAll(endFiles, files); |
| } |
| final long beginOldestFileOffsetMillis = getOldestFileOffsetMillis(mBeginFiles); |
| final long endOldestFileOffsetMillis = getOldestFileOffsetMillis(endFiles); |
| if (endOldestFileOffsetMillis < beginOldestFileOffsetMillis) { |
| final String message = "History loss detected!" |
| + "\nold files: " + mBeginFiles; |
| wtf(message, null, folder); |
| throw new IllegalStateException(message); |
| } |
| } |
| |
| private static long getOldestFileOffsetMillis(@NonNull List<File> files) { |
| if (files.isEmpty()) { |
| return 0; |
| } |
| String longestName = files.get(0).getName(); |
| final int fileCount = files.size(); |
| for (int i = 1; i < fileCount; i++) { |
| final File file = files.get(i); |
| if (file.getName().length() > longestName.length()) { |
| longestName = file.getName(); |
| } |
| } |
| return Long.parseLong(longestName.replace(HISTORY_FILE_SUFFIX, "")); |
| } |
| } |
| |
| private final class StringDumpVisitor implements AppOpsManager.HistoricalOpsVisitor { |
| private final long mNow = System.currentTimeMillis(); |
| |
| private final SimpleDateFormat mDateFormatter = new SimpleDateFormat( |
| "yyyy-MM-dd HH:mm:ss.SSS"); |
| private final Date mDate = new Date(); |
| |
| private final @NonNull String mOpsPrefix; |
| private final @NonNull String mUidPrefix; |
| private final @NonNull String mPackagePrefix; |
| private final @NonNull String mAttributionPrefix; |
| private final @NonNull String mEntryPrefix; |
| private final @NonNull String mUidStatePrefix; |
| private final @NonNull PrintWriter mWriter; |
| private final int mFilterUid; |
| private final String mFilterPackage; |
| private final String mFilterAttributionTag; |
| private final int mFilterOp; |
| private final @HistoricalOpsRequestFilter int mFilter; |
| |
| StringDumpVisitor(@NonNull String prefix, @NonNull PrintWriter writer, int filterUid, |
| @Nullable String filterPackage, @Nullable String filterAttributionTag, int filterOp, |
| @HistoricalOpsRequestFilter int filter) { |
| mOpsPrefix = prefix + " "; |
| mUidPrefix = mOpsPrefix + " "; |
| mPackagePrefix = mUidPrefix + " "; |
| mAttributionPrefix = mPackagePrefix + " "; |
| mEntryPrefix = mAttributionPrefix + " "; |
| mUidStatePrefix = mEntryPrefix + " "; |
| mWriter = writer; |
| mFilterUid = filterUid; |
| mFilterPackage = filterPackage; |
| mFilterAttributionTag = filterAttributionTag; |
| mFilterOp = filterOp; |
| mFilter = filter; |
| } |
| |
| @Override |
| public void visitHistoricalOps(HistoricalOps ops) { |
| mWriter.println(); |
| mWriter.print(mOpsPrefix); |
| mWriter.println("snapshot:"); |
| mWriter.print(mUidPrefix); |
| mWriter.print("begin = "); |
| mDate.setTime(ops.getBeginTimeMillis()); |
| mWriter.print(mDateFormatter.format(mDate)); |
| mWriter.print(" ("); |
| TimeUtils.formatDuration(ops.getBeginTimeMillis() - mNow, mWriter); |
| mWriter.println(")"); |
| mWriter.print(mUidPrefix); |
| mWriter.print("end = "); |
| mDate.setTime(ops.getEndTimeMillis()); |
| mWriter.print(mDateFormatter.format(mDate)); |
| mWriter.print(" ("); |
| TimeUtils.formatDuration(ops.getEndTimeMillis() - mNow, mWriter); |
| mWriter.println(")"); |
| } |
| |
| @Override |
| public void visitHistoricalUidOps(HistoricalUidOps ops) { |
| if ((mFilter & FILTER_BY_UID) != 0 && mFilterUid != ops.getUid()) { |
| return; |
| } |
| mWriter.println(); |
| mWriter.print(mUidPrefix); |
| mWriter.print("Uid "); |
| UserHandle.formatUid(mWriter, ops.getUid()); |
| mWriter.println(":"); |
| } |
| |
| @Override |
| public void visitHistoricalPackageOps(HistoricalPackageOps ops) { |
| if ((mFilter & FILTER_BY_PACKAGE_NAME) != 0 && !mFilterPackage.equals( |
| ops.getPackageName())) { |
| return; |
| } |
| mWriter.print(mPackagePrefix); |
| mWriter.print("Package "); |
| mWriter.print(ops.getPackageName()); |
| mWriter.println(":"); |
| } |
| |
| @Override |
| public void visitHistoricalAttributionOps(AppOpsManager.AttributedHistoricalOps ops) { |
| if ((mFilter & FILTER_BY_ATTRIBUTION_TAG) != 0 && !Objects.equals(mFilterPackage, |
| ops.getTag())) { |
| return; |
| } |
| mWriter.print(mAttributionPrefix); |
| mWriter.print("Attribution "); |
| mWriter.print(ops.getTag()); |
| mWriter.println(":"); |
| } |
| |
| @Override |
| public void visitHistoricalOp(HistoricalOp ops) { |
| if ((mFilter & FILTER_BY_OP_NAMES) != 0 && mFilterOp != ops.getOpCode()) { |
| return; |
| } |
| mWriter.print(mEntryPrefix); |
| mWriter.print(AppOpsManager.opToName(ops.getOpCode())); |
| mWriter.println(":"); |
| final LongSparseArray keys = ops.collectKeys(); |
| final int keyCount = keys.size(); |
| for (int i = 0; i < keyCount; i++) { |
| final long key = keys.keyAt(i); |
| final int uidState = AppOpsManager.extractUidStateFromKey(key); |
| final int flags = AppOpsManager.extractFlagsFromKey(key); |
| boolean printedUidState = false; |
| final long accessCount = ops.getAccessCount(uidState, uidState, flags); |
| if (accessCount > 0) { |
| if (!printedUidState) { |
| mWriter.print(mUidStatePrefix); |
| mWriter.print(AppOpsManager.keyToString(key)); |
| mWriter.print(" = "); |
| printedUidState = true; |
| } |
| mWriter.print("access="); |
| mWriter.print(accessCount); |
| } |
| final long rejectCount = ops.getRejectCount(uidState, uidState, flags); |
| if (rejectCount > 0) { |
| if (!printedUidState) { |
| mWriter.print(mUidStatePrefix); |
| mWriter.print(AppOpsManager.keyToString(key)); |
| mWriter.print(" = "); |
| printedUidState = true; |
| } else { |
| mWriter.print(", "); |
| } |
| mWriter.print("reject="); |
| mWriter.print(rejectCount); |
| } |
| final long accessDuration = ops.getAccessDuration(uidState, uidState, flags); |
| if (accessDuration > 0) { |
| if (!printedUidState) { |
| mWriter.print(mUidStatePrefix); |
| mWriter.print(AppOpsManager.keyToString(key)); |
| mWriter.print(" = "); |
| printedUidState = true; |
| } else { |
| mWriter.print(", "); |
| } |
| mWriter.print("duration="); |
| TimeUtils.formatDuration(accessDuration, mWriter); |
| } |
| if (printedUidState) { |
| mWriter.println(""); |
| } |
| } |
| } |
| } |
| |
| private static void wtf(@Nullable String message, @Nullable Throwable t, |
| @Nullable File storage) { |
| Slog.wtf(LOG_TAG, message, t); |
| if (KEEP_WTF_LOG) { |
| try { |
| final File file = new File(new File(Environment.getDataSystemDirectory(), "appops"), |
| "wtf" + TimeUtils.formatForLogging(System.currentTimeMillis())); |
| if (file.createNewFile()) { |
| try (PrintWriter writer = new PrintWriter(file)) { |
| if (t != null) { |
| writer.append('\n').append(t.toString()); |
| } |
| writer.append('\n').append(Debug.getCallers(10)); |
| if (storage != null) { |
| writer.append("\nfiles: " + Arrays.toString(storage.listFiles())); |
| } else { |
| writer.append("\nfiles: none"); |
| } |
| } |
| } |
| } catch (IOException e) { |
| /* ignore */ |
| } |
| } |
| } |
| } |