// Copyright 2014 The Bazel Authors. All rights reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package com.google.devtools.build.lib.skyframe; import com.google.common.annotations.VisibleForTesting; import com.google.common.base.Joiner; import com.google.common.base.Objects; import com.google.common.base.Predicate; import com.google.common.collect.ImmutableList; import com.google.common.collect.ImmutableMap; import com.google.common.collect.ImmutableSet; import com.google.common.collect.Iterables; import com.google.common.collect.Lists; import com.google.common.collect.Maps; import com.google.common.collect.Sets; import com.google.devtools.build.lib.analysis.BlazeDirectories; import com.google.devtools.build.lib.analysis.BuildView.Options; import com.google.devtools.build.lib.analysis.ConfiguredTarget; import com.google.devtools.build.lib.analysis.WorkspaceStatusAction.Factory; import com.google.devtools.build.lib.analysis.buildinfo.BuildInfoFactory; import com.google.devtools.build.lib.analysis.config.BinTools; import com.google.devtools.build.lib.cmdline.PackageIdentifier; import com.google.devtools.build.lib.concurrent.Uninterruptibles; import com.google.devtools.build.lib.events.ExtendedEventHandler; import com.google.devtools.build.lib.packages.Package; import com.google.devtools.build.lib.packages.PackageFactory; import com.google.devtools.build.lib.pkgcache.PackageCacheOptions; import com.google.devtools.build.lib.pkgcache.PathPackageLocator; import com.google.devtools.build.lib.profiler.AutoProfiler; import com.google.devtools.build.lib.skyframe.DirtinessCheckerUtils.BasicFilesystemDirtinessChecker; import com.google.devtools.build.lib.skyframe.DirtinessCheckerUtils.ExternalDirtinessChecker; import com.google.devtools.build.lib.skyframe.DirtinessCheckerUtils.MissingDiffDirtinessChecker; import com.google.devtools.build.lib.skyframe.DirtinessCheckerUtils.UnionDirtinessChecker; import com.google.devtools.build.lib.skyframe.ExternalFilesHelper.ExternalFileAction; import com.google.devtools.build.lib.skyframe.ExternalFilesHelper.ExternalFilesKnowledge; import com.google.devtools.build.lib.skyframe.ExternalFilesHelper.FileType; import com.google.devtools.build.lib.skyframe.PackageFunction.ActionOnIOExceptionReadingBuildFile; import com.google.devtools.build.lib.skyframe.PackageLookupFunction.CrossRepositoryLabelViolationStrategy; import com.google.devtools.build.lib.skyframe.PackageLookupValue.BuildFileName; import com.google.devtools.build.lib.syntax.SkylarkSemanticsOptions; import com.google.devtools.build.lib.util.AbruptExitException; import com.google.devtools.build.lib.util.Pair; import com.google.devtools.build.lib.util.Preconditions; import com.google.devtools.build.lib.util.ResourceUsage; import com.google.devtools.build.lib.util.io.TimestampGranularityMonitor; import com.google.devtools.build.lib.vfs.ModifiedFileSet; import com.google.devtools.build.lib.vfs.Path; import com.google.devtools.build.lib.vfs.PathFragment; import com.google.devtools.build.skyframe.BuildDriver; import com.google.devtools.build.skyframe.Differencer; import com.google.devtools.build.skyframe.InMemoryMemoizingEvaluator; import com.google.devtools.build.skyframe.Injectable; import com.google.devtools.build.skyframe.LegacySkyKey; import com.google.devtools.build.skyframe.MemoizingEvaluator.EvaluatorSupplier; import com.google.devtools.build.skyframe.NodeEntry; import com.google.devtools.build.skyframe.RecordingDifferencer; import com.google.devtools.build.skyframe.SequentialBuildDriver; import com.google.devtools.build.skyframe.SkyFunction; import com.google.devtools.build.skyframe.SkyFunctionName; import com.google.devtools.build.skyframe.SkyKey; import com.google.devtools.build.skyframe.SkyValue; import com.google.devtools.common.options.OptionsClassProvider; import java.io.PrintStream; import java.util.ArrayList; import java.util.Collection; import java.util.EnumSet; import java.util.HashSet; import java.util.Iterator; import java.util.List; import java.util.Map; import java.util.Set; import java.util.UUID; import java.util.concurrent.Callable; import java.util.logging.Logger; /** * A SkyframeExecutor that implicitly assumes that builds can be done incrementally from the most * recent build. In other words, builds are "sequenced". */ public final class SequencedSkyframeExecutor extends SkyframeExecutor { private static final Logger LOG = Logger.getLogger(SequencedSkyframeExecutor.class.getName()); private boolean lastAnalysisDiscarded = false; private enum IncrementalState { NORMAL, CLEAR_EDGES_AND_ACTIONS } // Can only be set once over the lifetime of this object. If CLEAR_EDGES or // CLEAR_EDGES_AND_ACTIONS, the graph will not store edges, saving memory but making incremental // builds impossible. If CLEAR_EDGES_AND_ACTIONS, each action will be dereferenced once it is // executed, saving memory. private IncrementalState incrementalState = IncrementalState.NORMAL; private RecordingDifferencer recordingDiffer; private final DiffAwarenessManager diffAwarenessManager; private final Iterable customDirtinessCheckers; private Set previousClientEnvironment = null; private SequencedSkyframeExecutor( EvaluatorSupplier evaluatorSupplier, PackageFactory pkgFactory, BlazeDirectories directories, BinTools binTools, Factory workspaceStatusActionFactory, ImmutableList buildInfoFactories, Iterable diffAwarenessFactories, Predicate allowedMissingInputs, ImmutableMap extraSkyFunctions, ImmutableList extraPrecomputedValues, Iterable customDirtinessCheckers, PathFragment blacklistedPackagePrefixesFile, String productName, CrossRepositoryLabelViolationStrategy crossRepositoryLabelViolationStrategy, List buildFilesByPriority, ActionOnIOExceptionReadingBuildFile actionOnIOExceptionReadingBuildFile) { super( evaluatorSupplier, pkgFactory, directories, binTools, workspaceStatusActionFactory, buildInfoFactories, allowedMissingInputs, extraSkyFunctions, extraPrecomputedValues, ExternalFileAction.DEPEND_ON_EXTERNAL_PKG_FOR_EXTERNAL_REPO_PATHS, blacklistedPackagePrefixesFile, productName, crossRepositoryLabelViolationStrategy, buildFilesByPriority, actionOnIOExceptionReadingBuildFile); this.diffAwarenessManager = new DiffAwarenessManager(diffAwarenessFactories); this.customDirtinessCheckers = customDirtinessCheckers; } public static SequencedSkyframeExecutor create( PackageFactory pkgFactory, BlazeDirectories directories, BinTools binTools, Factory workspaceStatusActionFactory, ImmutableList buildInfoFactories, Iterable diffAwarenessFactories, Predicate allowedMissingInputs, ImmutableMap extraSkyFunctions, ImmutableList extraPrecomputedValues, Iterable customDirtinessCheckers, PathFragment blacklistedPackagePrefixesFile, String productName, CrossRepositoryLabelViolationStrategy crossRepositoryLabelViolationStrategy, List buildFilesByPriority, ActionOnIOExceptionReadingBuildFile actionOnIOExceptionReadingBuildFile) { SequencedSkyframeExecutor skyframeExecutor = new SequencedSkyframeExecutor( InMemoryMemoizingEvaluator.SUPPLIER, pkgFactory, directories, binTools, workspaceStatusActionFactory, buildInfoFactories, diffAwarenessFactories, allowedMissingInputs, extraSkyFunctions, extraPrecomputedValues, customDirtinessCheckers, blacklistedPackagePrefixesFile, productName, crossRepositoryLabelViolationStrategy, buildFilesByPriority, actionOnIOExceptionReadingBuildFile); skyframeExecutor.init(); return skyframeExecutor; } @Override protected BuildDriver getBuildDriver() { return new SequentialBuildDriver(memoizingEvaluator); } @Override protected void init() { // Note that we need to set recordingDiffer first since SkyframeExecutor#init calls // SkyframeExecutor#evaluatorDiffer. recordingDiffer = new RecordingDifferencer(); super.init(); } @Override public void resetEvaluator() { super.resetEvaluator(); diffAwarenessManager.reset(); } @Override protected Differencer evaluatorDiffer() { return recordingDiffer; } @Override protected Injectable injectable() { return recordingDiffer; } @VisibleForTesting public RecordingDifferencer getDifferencerForTesting() { return recordingDiffer; } @Override public void sync( ExtendedEventHandler eventHandler, PackageCacheOptions packageCacheOptions, SkylarkSemanticsOptions skylarkSemanticsOptions, Path outputBase, Path workingDirectory, String defaultsPackageContents, UUID commandId, Map clientEnv, TimestampGranularityMonitor tsgm, OptionsClassProvider options) throws InterruptedException, AbruptExitException { super.sync(eventHandler, packageCacheOptions, skylarkSemanticsOptions, outputBase, workingDirectory, defaultsPackageContents, commandId, clientEnv, tsgm, options); handleDiffs(eventHandler, packageCacheOptions.checkOutputFiles, options); } /** * The value types whose builders have direct access to the package locator, rather than accessing * it via an explicit Skyframe dependency. They need to be invalidated if the package locator * changes. */ private static final ImmutableSet PACKAGE_LOCATOR_DEPENDENT_VALUES = ImmutableSet.of( SkyFunctions.AST_FILE_LOOKUP, SkyFunctions.FILE_STATE, SkyFunctions.FILE, SkyFunctions.DIRECTORY_LISTING_STATE, SkyFunctions.TARGET_PATTERN, SkyFunctions.PREPARE_DEPS_OF_PATTERN, SkyFunctions.WORKSPACE_FILE, SkyFunctions.EXTERNAL_PACKAGE); @Override protected void onNewPackageLocator(PathPackageLocator oldLocator, PathPackageLocator pkgLocator) { invalidate(SkyFunctionName.functionIsIn(PACKAGE_LOCATOR_DEPENDENT_VALUES)); } @Override protected void invalidate(Predicate pred) { recordingDiffer.invalidate(Iterables.filter(memoizingEvaluator.getValues().keySet(), pred)); } private void invalidateDeletedPackages(Iterable deletedPackages) { ArrayList packagesToInvalidate = Lists.newArrayList(); for (PackageIdentifier deletedPackage : deletedPackages) { packagesToInvalidate.add(PackageLookupValue.key(deletedPackage)); } recordingDiffer.invalidate(packagesToInvalidate); } /** * Sets the packages that should be treated as deleted and ignored. */ @Override @VisibleForTesting // productionVisibility = Visibility.PRIVATE public void setDeletedPackages(Iterable pkgs) { // Invalidate the old deletedPackages as they may exist now. invalidateDeletedPackages(deletedPackages.get()); deletedPackages.set(ImmutableSet.copyOf(pkgs)); // Invalidate the new deletedPackages as we need to pretend that they don't exist now. invalidateDeletedPackages(deletedPackages.get()); } /** Uses diff awareness on all the package paths to invalidate changed files. */ @VisibleForTesting public void handleDiffs(ExtendedEventHandler eventHandler) throws InterruptedException { handleDiffs(eventHandler, /*checkOutputFiles=*/false, OptionsClassProvider.EMPTY); } private void handleDiffs( ExtendedEventHandler eventHandler, boolean checkOutputFiles, OptionsClassProvider options) throws InterruptedException { if (lastAnalysisDiscarded) { // Values were cleared last build, but they couldn't be deleted because they were needed for // the execution phase. We can delete them now. dropConfiguredTargetsNow(eventHandler); lastAnalysisDiscarded = false; } TimestampGranularityMonitor tsgm = this.tsgm.get(); modifiedFiles = 0; Map modifiedFilesByPathEntry = Maps.newHashMap(); Set> pathEntriesWithoutDiffInformation = Sets.newHashSet(); for (Path pathEntry : pkgLocator.get().getPathEntries()) { DiffAwarenessManager.ProcessableModifiedFileSet modifiedFileSet = diffAwarenessManager.getDiff(eventHandler, pathEntry, options); if (modifiedFileSet.getModifiedFileSet().treatEverythingAsModified()) { pathEntriesWithoutDiffInformation.add(Pair.of(pathEntry, modifiedFileSet)); } else { modifiedFilesByPathEntry.put(pathEntry, modifiedFileSet); } } handleDiffsWithCompleteDiffInformation(tsgm, modifiedFilesByPathEntry); handleDiffsWithMissingDiffInformation(eventHandler, tsgm, pathEntriesWithoutDiffInformation, checkOutputFiles); handleClientEnvironmentChanges(); } /** Invalidates entries in the client environment that have changed since last sync. */ private void handleClientEnvironmentChanges() { Map values = memoizingEvaluator.getValues(); ImmutableMap.Builder newValuesBuilder = ImmutableMap.builder(); HashSet envToCheck = new HashSet<>(); if (previousClientEnvironment != null) { envToCheck.addAll(previousClientEnvironment); } envToCheck.addAll(clientEnv.get().keySet()); previousClientEnvironment = clientEnv.get().keySet(); for (String env : envToCheck) { SkyKey key = LegacySkyKey.create(SkyFunctions.CLIENT_ENVIRONMENT_VARIABLE, env); if (values.containsKey(key)) { String value = ((ClientEnvironmentValue) values.get(key)).getValue(); String newValue = clientEnv.get().get(env); if (!Objects.equal(newValue, value)) { newValuesBuilder.put(key, new ClientEnvironmentValue(newValue)); } } } recordingDiffer.inject(newValuesBuilder.build()); } /** * Invalidates files under path entries whose corresponding {@link DiffAwareness} gave an exact * diff. Removes entries from the given map as they are processed. All of the files need to be * invalidated, so the map should be empty upon completion of this function. */ private void handleDiffsWithCompleteDiffInformation(TimestampGranularityMonitor tsgm, Map modifiedFilesByPathEntry) throws InterruptedException { for (Path pathEntry : ImmutableSet.copyOf(modifiedFilesByPathEntry.keySet())) { DiffAwarenessManager.ProcessableModifiedFileSet processableModifiedFileSet = modifiedFilesByPathEntry.get(pathEntry); ModifiedFileSet modifiedFileSet = processableModifiedFileSet.getModifiedFileSet(); Preconditions.checkState(!modifiedFileSet.treatEverythingAsModified(), pathEntry); handleChangedFiles(ImmutableList.of(pathEntry), getDiff(tsgm, modifiedFileSet.modifiedSourceFiles(), pathEntry)); processableModifiedFileSet.markProcessed(); } } /** * Finds and invalidates changed files under path entries whose corresponding {@link * DiffAwareness} said all files may have been modified. */ private void handleDiffsWithMissingDiffInformation( ExtendedEventHandler eventHandler, TimestampGranularityMonitor tsgm, Set> pathEntriesWithoutDiffInformation, boolean checkOutputFiles) throws InterruptedException { ExternalFilesKnowledge externalFilesKnowledge = externalFilesHelper.getExternalFilesKnowledge(); if (pathEntriesWithoutDiffInformation.isEmpty() && Iterables.isEmpty(customDirtinessCheckers) && ((!externalFilesKnowledge.anyOutputFilesSeen || !checkOutputFiles) && !externalFilesKnowledge.anyNonOutputExternalFilesSeen)) { // Avoid a full graph scan if we have good diff information for all path entries, there are // no custom checkers that need to look at the whole graph, and no external (not under any // path) files need to be checked. return; } // Before running the FilesystemValueChecker, ensure that all values marked for invalidation // have actually been invalidated (recall that invalidation happens at the beginning of the // next evaluate() call), because checking those is a waste of time. buildDriver.evaluate(ImmutableList.of(), false, DEFAULT_THREAD_COUNT, eventHandler); FilesystemValueChecker fsvc = new FilesystemValueChecker(tsgm, null); // We need to manually check for changes to known files. This entails finding all dirty file // system values under package roots for which we don't have diff information. If at least // one path entry doesn't have diff information, then we're going to have to iterate over // the skyframe values at least once no matter what. Set diffPackageRootsUnderWhichToCheck = new HashSet<>(); for (Pair pair : pathEntriesWithoutDiffInformation) { diffPackageRootsUnderWhichToCheck.add(pair.getFirst()); } // We freshly compute knowledge of the presence of external files in the skyframe graph. We use // a fresh ExternalFilesHelper instance and only set the real instance's knowledge *after* we // are done with the graph scan, lest an interrupt during the graph scan causes us to // incorrectly think there are no longer any external files. ExternalFilesHelper tmpExternalFilesHelper = externalFilesHelper.cloneWithFreshExternalFilesKnowledge(); // See the comment for FileType.OUTPUT for why we need to consider output files here. EnumSet fileTypesToCheck = checkOutputFiles ? EnumSet.of(FileType.EXTERNAL, FileType.EXTERNAL_REPO, FileType.OUTPUT) : EnumSet.of(FileType.EXTERNAL, FileType.EXTERNAL_REPO); LOG.info("About to scan skyframe graph checking for filesystem nodes of types " + Iterables.toString(fileTypesToCheck)); Differencer.Diff diff = fsvc.getDirtyKeys( memoizingEvaluator.getValues(), new UnionDirtinessChecker( Iterables.concat( customDirtinessCheckers, ImmutableList.of( new ExternalDirtinessChecker( tmpExternalFilesHelper, fileTypesToCheck), new MissingDiffDirtinessChecker(diffPackageRootsUnderWhichToCheck))))); handleChangedFiles(diffPackageRootsUnderWhichToCheck, diff); for (Pair pair : pathEntriesWithoutDiffInformation) { pair.getSecond().markProcessed(); } // We use the knowledge gained during the graph scan that just completed. Otherwise, naively, // once an external file gets into the Skyframe graph, we'll overly-conservatively always think // the graph needs to be scanned. externalFilesHelper.setExternalFilesKnowledge( tmpExternalFilesHelper.getExternalFilesKnowledge()); } private void handleChangedFiles( Collection diffPackageRootsUnderWhichToCheck, Differencer.Diff diff) { Collection changedKeysWithoutNewValues = diff.changedKeysWithoutNewValues(); Map changedKeysWithNewValues = diff.changedKeysWithNewValues(); logDiffInfo(diffPackageRootsUnderWhichToCheck, changedKeysWithoutNewValues, changedKeysWithNewValues); recordingDiffer.invalidate(changedKeysWithoutNewValues); recordingDiffer.inject(changedKeysWithNewValues); modifiedFiles += getNumberOfModifiedFiles(changedKeysWithoutNewValues); modifiedFiles += getNumberOfModifiedFiles(changedKeysWithNewValues.keySet()); incrementalBuildMonitor.accrue(changedKeysWithoutNewValues); incrementalBuildMonitor.accrue(changedKeysWithNewValues.keySet()); } private static void logDiffInfo(Iterable pathEntries, Collection changedWithoutNewValue, Map changedWithNewValue) { int numModified = changedWithNewValue.size() + changedWithoutNewValue.size(); StringBuilder result = new StringBuilder("DiffAwareness found ") .append(numModified) .append(" modified source files and directory listings"); if (!Iterables.isEmpty(pathEntries)) { result.append(" for "); result.append(Joiner.on(", ").join(pathEntries)); } if (numModified > 0) { Iterable allModifiedKeys = Iterables.concat(changedWithoutNewValue, changedWithNewValue.keySet()); Iterable trimmed = Iterables.limit(allModifiedKeys, 5); result.append(": ") .append(Joiner.on(", ").join(trimmed)); if (numModified > 5) { result.append(", ..."); } } LOG.info(result.toString()); } private static int getNumberOfModifiedFiles(Iterable modifiedValues) { // We are searching only for changed files, DirectoryListingValues don't depend on // child values, that's why they are invalidated separately return Iterables.size(Iterables.filter(modifiedValues, SkyFunctionName.functionIs(SkyFunctions.FILE_STATE))); } @Override public void decideKeepIncrementalState(boolean batch, Options viewOptions) { Preconditions.checkState(!active); if (viewOptions == null) { // Some blaze commands don't include the view options. Don't bother with them. return; } if (batch && viewOptions.keepGoing && viewOptions.discardAnalysisCache) { Preconditions.checkState( incrementalState == IncrementalState.NORMAL, "May only be called once if successful: %s", incrementalState); incrementalState = IncrementalState.CLEAR_EDGES_AND_ACTIONS; // Graph will be recreated on next sync. LOG.info("Set incremental state to " + incrementalState); } removeActionsAfterEvaluation.set(incrementalState == IncrementalState.CLEAR_EDGES_AND_ACTIONS); } @Override public boolean hasIncrementalState() { // TODO(bazel-team): Combine this method with clearSkyframeRelevantCaches() once legacy // execution is removed [skyframe-execution]. return incrementalState == IncrementalState.NORMAL; } @Override public void invalidateFilesUnderPathForTesting( ExtendedEventHandler eventHandler, ModifiedFileSet modifiedFileSet, Path pathEntry) throws InterruptedException { if (lastAnalysisDiscarded) { // Values were cleared last build, but they couldn't be deleted because they were needed for // the execution phase. We can delete them now. dropConfiguredTargetsNow(eventHandler); lastAnalysisDiscarded = false; } TimestampGranularityMonitor tsgm = this.tsgm.get(); Differencer.Diff diff; if (modifiedFileSet.treatEverythingAsModified()) { diff = new FilesystemValueChecker(tsgm, null).getDirtyKeys(memoizingEvaluator.getValues(), new BasicFilesystemDirtinessChecker()); } else { diff = getDiff(tsgm, modifiedFileSet.modifiedSourceFiles(), pathEntry); } syscalls.set(getPerBuildSyscallCache(/*concurrencyLevel=*/ 42)); recordingDiffer.invalidate(diff.changedKeysWithoutNewValues()); recordingDiffer.inject(diff.changedKeysWithNewValues()); // Blaze invalidates transient errors on every build. invalidateTransientErrors(); } @Override public void invalidateTransientErrors() { checkActive(); recordingDiffer.invalidateTransientErrors(); } @Override protected void invalidateDirtyActions(Iterable dirtyActionValues) { recordingDiffer.invalidate(dirtyActionValues); } private static ImmutableSet LOADING_TYPES = ImmutableSet.of( SkyFunctions.PACKAGE, SkyFunctions.SKYLARK_IMPORTS_LOOKUP, SkyFunctions.AST_FILE_LOOKUP, SkyFunctions.GLOB); /** * Save memory by removing references to configured targets and aspects in Skyframe. * *

These nodes must be recreated on subsequent builds. We do not clear the top-level target * nodes, since their configured targets are needed for the target completion middleman values. * *

The nodes are not deleted during this method call, because they are needed for the execution * phase. Instead, their analysis-time data is cleared while preserving the generating action info * needed for execution. The next build will delete the nodes (and recreate them if necessary). * *

If {@link #hasIncrementalState} is false, then also delete loading-phase nodes (as * determined by {@link #LOADING_TYPES}) from the graph, since there will be no future builds to * use them for. */ private void discardAnalysisCache( Collection topLevelTargets, Collection topLevelAspects) { topLevelTargets = ImmutableSet.copyOf(topLevelTargets); topLevelAspects = ImmutableSet.copyOf(topLevelAspects); try (AutoProfiler p = AutoProfiler.logged("discarding analysis cache", LOG)) { lastAnalysisDiscarded = true; Iterator> it = memoizingEvaluator.getGraphMap().entrySet().iterator(); while (it.hasNext()) { Map.Entry keyAndEntry = it.next(); NodeEntry entry = keyAndEntry.getValue(); if (entry == null || !entry.isDone()) { continue; } SkyKey key = keyAndEntry.getKey(); SkyFunctionName functionName = key.functionName(); if (!hasIncrementalState() && LOADING_TYPES.contains(functionName)) { it.remove(); continue; } if (functionName.equals(SkyFunctions.CONFIGURED_TARGET)) { ConfiguredTargetValue ctValue; try { ctValue = (ConfiguredTargetValue) entry.getValue(); } catch (InterruptedException e) { throw new IllegalStateException("No interruption in sequenced evaluation", e); } // ctValue may be null if target was not successfully analyzed. if (ctValue != null) { ctValue.clear(!topLevelTargets.contains(ctValue.getConfiguredTarget())); } } else if (functionName.equals(SkyFunctions.ASPECT)) { AspectValue aspectValue; try { aspectValue = (AspectValue) entry.getValue(); } catch (InterruptedException e) { throw new IllegalStateException("No interruption in sequenced evaluation", e); } // value may be null if target was not successfully analyzed. if (aspectValue != null) { aspectValue.clear(!topLevelAspects.contains(aspectValue)); } } } } } @Override public void clearAnalysisCache( Collection topLevelTargets, Collection topLevelAspects) { discardAnalysisCache(topLevelTargets, topLevelAspects); } @Override public void dropConfiguredTargets() { skyframeBuildView.clearInvalidatedConfiguredTargets(); skyframeBuildView.clearLegacyData(); memoizingEvaluator.delete( // We delete any value that can hold an action -- all subclasses of ActionLookupValue -- as // well as ActionExecutionValues, since they do not depend on ActionLookupValues. SkyFunctionName.functionIsIn(ImmutableSet.of( SkyFunctions.CONFIGURED_TARGET, SkyFunctions.ACTION_LOOKUP, SkyFunctions.BUILD_INFO, SkyFunctions.TARGET_COMPLETION, SkyFunctions.BUILD_INFO_COLLECTION, SkyFunctions.ACTION_EXECUTION)) ); } /** * Deletes all ConfiguredTarget values from the Skyframe cache. * *

After the execution of this method all invalidated and marked for deletion values (and the * values depending on them) will be deleted from the cache. * *

WARNING: Note that a call to this method leaves legacy data inconsistent with Skyframe. The * next build should clear the legacy caches. */ private void dropConfiguredTargetsNow(final ExtendedEventHandler eventHandler) { dropConfiguredTargets(); // Run the invalidator to actually delete the values. try { progressReceiver.ignoreInvalidations = true; Uninterruptibles.callUninterruptibly(new Callable() { @Override public Void call() throws InterruptedException { buildDriver.evaluate(ImmutableList.of(), false, ResourceUsage.getAvailableProcessors(), eventHandler); return null; } }); } catch (Exception e) { throw new IllegalStateException(e); } finally { progressReceiver.ignoreInvalidations = false; } } @Override public void deleteOldNodes(long versionWindowForDirtyGc) { // TODO(bazel-team): perhaps we should come up with a separate GC class dedicated to maintaining // value garbage. If we ever do so, this logic should be moved there. memoizingEvaluator.deleteDirty(versionWindowForDirtyGc); } @Override public void dumpPackages(PrintStream out) { Iterable packageSkyKeys = Iterables.filter(memoizingEvaluator.getValues().keySet(), SkyFunctions.isSkyFunction(SkyFunctions.PACKAGE)); out.println(Iterables.size(packageSkyKeys) + " packages"); for (SkyKey packageSkyKey : packageSkyKeys) { Package pkg = ((PackageValue) memoizingEvaluator.getValues().get(packageSkyKey)).getPackage(); pkg.dump(out); } } }