Unnamed: 0 int64 0 305k | body stringlengths 7 52.9k | name stringlengths 1 185 |
|---|---|---|
2,200 | boolean (K key) { try { return myMap.containsMapping(key); } catch (IOException e) { throw new BuildDataCorruptedException(e); } } | containsKey |
2,201 | C (K key) { try { C col = myMap.get(key); return col != null? col : myEmptyCollection; } catch (IOException e) { throw new BuildDataCorruptedException(e); } } | get |
2,202 | void (K key, @NotNull Iterable<? extends V> values) { try { //noinspection unchecked C data = ensureCollection(values); if (data.isEmpty()) { myMap.remove(key); } else { myMap.put(key, data); } } catch (IOException e) { throw new BuildDataCorruptedException(e); } } | put |
2,203 | C (Iterable<? extends V> seq) { if (myEmptyCollection instanceof Set && seq instanceof Set) { return (C)seq; } if (myEmptyCollection instanceof List && seq instanceof List) { return (C)seq; } return Iterators.collect(seq, myCollectionFactory.get()); } | ensureCollection |
2,204 | void (K key) { try { myMap.remove(key); } catch (IOException e) { throw new BuildDataCorruptedException(e); } } | remove |
2,205 | void (K key, V value) { appendValues(key, Collections.singleton(value)); } | appendValue |
2,206 | void (K key, @NotNull Iterable<? extends V> values) { try { myMap.appendData(key, new AppendablePersistentMap.ValueDataAppender() { @Override public void append(@NotNull DataOutput out) throws IOException { for (V v : values) { myValuesExternalizer.save(out, v); } } }); } catch (IOException e) { throw new BuildDataCorruptedException(e); } } | appendValues |
2,207 | void (K key, V value) { removeValues(key, Collections.singleton(value)); } | removeValue |
2,208 | void (K key, @NotNull Iterable<? extends V> values) { try { C collection = get(key); if (!collection.isEmpty()) { boolean changes = false; for (V value : values) { changes |= collection.remove(value); } if (changes) { if (collection.isEmpty()) { myMap.remove(key); } else { myMap.put(key, collection); } } } } catch (IOException e) { throw new BuildDataCorruptedException(e); } } | removeValues |
2,209 | Iterable<K> () { try { return myMap.getAllKeysWithExistingMapping(); } catch (IOException e) { throw new RuntimeException(e); } } | getKeys |
2,210 | void () { try { myMap.close(); } catch (IOException e) { throw new BuildDataCorruptedException(e); } } | close |
2,211 | Set<NodeSource> () { return myBaseSources; } | getBaseSources |
2,212 | Set<NodeSource> () { return myDeletedSources; } | getDeletedSources |
2,213 | void (@NotNull Node<?, ?> node, @NotNull Iterable<NodeSource> sources) { ReferenceID nodeID = node.getReferenceID(); for (NodeSource src : sources) { myNodeToSourcesMap.appendValue(nodeID, src); mySourceToNodesMap.appendValue(src, node); } // deduce dependencies for (BackDependencyIndex index : getIndices()) { index.indexNode(node); } } | associate |
2,214 | boolean (K key) { return myMap.containsKey(key); } | containsKey |
2,215 | void (K key, V value) { myMap.put(key, value); } | put |
2,216 | void (K key) { myMap.remove(key); } | remove |
2,217 | Iterable<K> () { return myMap.keySet(); } | getKeys |
2,218 | void () { myMap.clear(); } | close |
2,219 | Path () { return myPath; } | getPath |
2,220 | boolean (Object o) { if (this == o) { return true; } if (o == null || getClass() != o.getClass()) { return false; } final FileSource that = (FileSource)o; if (!myPath.equals(that.myPath)) { return false; } return true; } | equals |
2,221 | int () { return myPath.hashCode(); } | hashCode |
2,222 | String () { return "NodeSource {" + myPath + "}"; } | toString |
2,223 | Iterable<ReferenceID> (Node<?, ?> node) { ReferenceID nodeID = node.getReferenceID(); return Iterators.unique(Iterators.map(Iterators.filter(node.getUsages(), u -> !nodeID.equals(u.getElementOwner())), u -> u.getElementOwner())); } | getIndexedDependencies |
2,224 | DifferentiateResult (Delta delta, DifferentiateParameters params) { Iterable<NodeSource> deltaSources = delta.getSources(); Set<NodeSource> allProcessedSources = Iterators.collect(Iterators.flat(Arrays.asList(delta.getBaseSources(), deltaSources, delta.getDeletedSources())), new HashSet<>()); Set<Node<?, ?>> nodesBefore = Iterators.collect(Iterators.flat(Iterators.map(allProcessedSources, s -> getNodes(s))), Containers.createCustomPolicySet(DiffCapable::isSame, DiffCapable::diffHashCode)); Set<Node<?, ?>> nodesAfter = Iterators.collect(Iterators.flat(Iterators.map(deltaSources, s -> delta.getNodes(s))), Containers.createCustomPolicySet(DiffCapable::isSame, DiffCapable::diffHashCode)); // do not process 'removed' per-source file. This works when a class comes from exactly one source, but might not work, if a class can be associated with several sources // better make a node-diff over all compiled sources => the sets of removed, added, deleted _nodes_ will be more accurate and reflecting reality List<Node<?, ?>> deletedNodes = Iterators.collect(Iterators.filter(nodesBefore, n -> !nodesAfter.contains(n)), new ArrayList<>()); if (!params.isCalculateAffected()) { return new DifferentiateResult() { @Override public Delta getDelta() { return delta; } @Override public Iterable<Node<?, ?>> getDeletedNodes() { return deletedNodes; } @Override public Iterable<NodeSource> getAffectedSources() { return Collections.emptyList(); } }; } var diffContext = new DifferentiateContext() { private final Predicate<Node<?, ?>> ANY_CONSTRAINT = node -> true; final Set<NodeSource> compiledSources = deltaSources instanceof Set? (Set<NodeSource>)deltaSources : Iterators.collect(deltaSources, new HashSet<>()); final Map<Usage, Predicate<Node<?, ?>>> affectedUsages = new HashMap<>(); final Set<BiPredicate<Node<?, ?>, Usage>> usageQueries = new HashSet<>(); final Set<NodeSource> affectedSources = new HashSet<>(); @Override public DifferentiateParameters getParams() { return params; } @Override public @NotNull Graph getGraph() { return DependencyGraphImpl.this; } @Override public @NotNull Delta getDelta() { return delta; } @Override public boolean isCompiled(NodeSource src) { return compiledSources.contains(src); } @Override public void affectUsage(@NotNull Usage usage) { affectedUsages.put(usage, ANY_CONSTRAINT); } @Override public void affectUsage(@NotNull Usage usage, @NotNull Predicate<Node<?, ?>> constraint) { Predicate<Node<?, ?>> prevConstraint = affectedUsages.put(usage, constraint); if (prevConstraint != null) { affectedUsages.put(usage, prevConstraint == ANY_CONSTRAINT? ANY_CONSTRAINT : prevConstraint.or(constraint)); } } @Override public void affectUsage(@NotNull BiPredicate<Node<?, ?>, Usage> usageQuery) { usageQueries.add(usageQuery); } @Override public void affectNodeSource(@NotNull NodeSource source) { affectedSources.add(source); } boolean isNodeAffected(Node<?, ?> node) { for (Usage usage : node.getUsages()) { Predicate<Node<?, ?>> constraint = affectedUsages.get(usage); if (constraint != null && constraint.test(node)) { return true; } for (BiPredicate<Node<?, ?>, Usage> query : usageQueries) { if (query.test(node, usage)) { return true; } } } return false; } }; boolean incremental = true; for (DifferentiateStrategy diffStrategy : ourDifferentiateStrategies) { if (!diffStrategy.differentiate(diffContext, nodesBefore, nodesAfter)) { incremental = false; break; } } if (!incremental) { return DifferentiateResult.createNonIncremental(delta, deletedNodes); } Set<NodeSource> affectedSources = new HashSet<>(); Set<ReferenceID> dependingOnDeleted = Iterators.collect(Iterators.flat(Iterators.map(deletedNodes, n -> getDependingNodes(n.getReferenceID()))), new HashSet<>()); for (ReferenceID dep : dependingOnDeleted) { for (NodeSource src : getSources(dep)) { affectedSources.add(src); } } Iterable<ReferenceID> changedScopeNodes = Iterators.unique(Iterators.flat(Iterators.map(nodesAfter, n -> n.getReferenceID()), Iterators.map(diffContext.affectedUsages.keySet(), u -> u.getElementOwner()))); for (ReferenceID dependent : Iterators.unique(Iterators.filter(Iterators.flat(Iterators.map(changedScopeNodes, id -> getDependingNodes(id))), id -> !dependingOnDeleted.contains(id)))) { for (NodeSource depSrc : getSources(dependent)) { if (!affectedSources.contains(depSrc)) { boolean affectSource = false; for (var depNode : getNodes(depSrc)) { if (diffContext.isNodeAffected(depNode)) { affectSource = true; for (DifferentiateStrategy strategy : ourDifferentiateStrategies) { if (!strategy.isIncremental(diffContext, depNode)) { return DifferentiateResult.createNonIncremental(delta, deletedNodes); } } } } if (affectSource) { affectedSources.add(depSrc); } } } } // do not include sources that were already compiled affectedSources.removeAll(allProcessedSources); // ensure sources explicitly marked by strategies are affected, even if these sources were compiled initially affectedSources.addAll(diffContext.affectedSources); return new DifferentiateResult() { @Override public Delta getDelta() { return delta; } @Override public Iterable<Node<?, ?>> getDeletedNodes() { return deletedNodes; } @Override public Iterable<NodeSource> getAffectedSources() { return affectedSources; } }; } | differentiate |
2,225 | Delta () { return delta; } | getDelta |
2,226 | Iterable<NodeSource> () { return Collections.emptyList(); } | getAffectedSources |
2,227 | DifferentiateParameters () { return params; } | getParams |
2,228 | Graph () { return DependencyGraphImpl.this; } | getGraph |
2,229 | Delta () { return delta; } | getDelta |
2,230 | boolean (NodeSource src) { return compiledSources.contains(src); } | isCompiled |
2,231 | void (@NotNull Usage usage) { affectedUsages.put(usage, ANY_CONSTRAINT); } | affectUsage |
2,232 | void (@NotNull Usage usage, @NotNull Predicate<Node<?, ?>> constraint) { Predicate<Node<?, ?>> prevConstraint = affectedUsages.put(usage, constraint); if (prevConstraint != null) { affectedUsages.put(usage, prevConstraint == ANY_CONSTRAINT? ANY_CONSTRAINT : prevConstraint.or(constraint)); } } | affectUsage |
2,233 | void (@NotNull BiPredicate<Node<?, ?>, Usage> usageQuery) { usageQueries.add(usageQuery); } | affectUsage |
2,234 | void (@NotNull NodeSource source) { affectedSources.add(source); } | affectNodeSource |
2,235 | Delta () { return delta; } | getDelta |
2,236 | Iterable<NodeSource> () { return affectedSources; } | getAffectedSources |
2,237 | void (@NotNull DifferentiateResult diffResult) { final Delta delta = diffResult.getDelta(); // handle deleted nodes and sources if (!Iterators.isEmpty(diffResult.getDeletedNodes())) { Set<NodeSource> differentiatedSources = Iterators.collect(Iterators.flat(List.of(delta.getBaseSources(), delta.getSources(), delta.getDeletedSources())), new HashSet<>()); for (var deletedNode : diffResult.getDeletedNodes()) { // the set of deleted nodes includes ones corresponding to deleted sources Set<NodeSource> nodeSources = Iterators.collect(myNodeToSourcesMap.get(deletedNode.getReferenceID()), new HashSet<>()); nodeSources.removeAll(differentiatedSources); if (nodeSources.isEmpty()) { myNodeToSourcesMap.remove(deletedNode.getReferenceID()); } else { myNodeToSourcesMap.put(deletedNode.getReferenceID(), nodeSources); } } } for (NodeSource deletedSource : delta.getDeletedSources()) { mySourceToNodesMap.remove(deletedSource); } var updatedNodes = Iterators.collect(Iterators.flat(Iterators.map(delta.getSources(), s -> getNodes(s))), new HashSet<>()); for (BackDependencyIndex index : getIndices()) { BackDependencyIndex deltaIndex = delta.getIndex(index.getName()); assert deltaIndex != null; index.integrate(diffResult.getDeletedNodes(), updatedNodes, deltaIndex); } var deltaNodes = Iterators.unique(Iterators.map(Iterators.flat(Iterators.map(delta.getSources(), s -> delta.getNodes(s))), node -> node.getReferenceID())); for (ReferenceID nodeID : deltaNodes) { Set<NodeSource> sources = Iterators.collect(myNodeToSourcesMap.get(nodeID), new HashSet<>()); sources.removeAll(delta.getBaseSources()); Iterators.collect(delta.getSources(nodeID), sources); myNodeToSourcesMap.put(nodeID, sources); } for (NodeSource src : delta.getSources()) { mySourceToNodesMap.put(src, delta.getNodes(src)); } } | integrate |
2,238 | Set<NodeSource> (Iterable<NodeSource> sources, Iterable<NodeSource> deletedSources) { // ensure initial sources are in the result Set<NodeSource> result = Iterators.collect(sources, new HashSet<>()); // todo: check if a special hashing-policy set is required here Set<NodeSource> deleted = Iterators.collect(deletedSources, new HashSet<>()); Set<Node<?, ?>> affectedNodes = Iterators.collect(Iterators.flat(Iterators.map(Iterators.flat(result, deleted), s -> getNodes(s))), new HashSet<>()); for (var node : affectedNodes) { Iterators.collect(Iterators.filter(getSources(node.getReferenceID()), s -> !result.contains(s) && !deleted.contains(s) && Iterators.filter(getNodes(s).iterator(), affectedNodes::contains).hasNext()), result); } return result; } | completeSourceSet |
2,239 | boolean (K key) { return myDelegate.containsKey(key); } | containsKey |
2,240 | Iterable<V> (K key) { return myCache.get(key); } | get |
2,241 | void (K key, @NotNull Iterable<? extends V> values) { myCache.invalidate(key); myDelegate.put(key, values); } | put |
2,242 | void (K key) { myCache.invalidate(key); myDelegate.remove(key); } | remove |
2,243 | void (K key, V value) { appendValues(key, Collections.singleton(value)); } | appendValue |
2,244 | void (K key, @NotNull Iterable<? extends V> values) { if (!Iterators.isEmpty(values)) { myCache.invalidate(key); myDelegate.appendValues(key, values); } } | appendValues |
2,245 | void (K key, V value) { removeValues(key, Collections.singleton(value)); } | removeValue |
2,246 | void (K key, @NotNull Iterable<? extends V> values) { if (Iterators.isEmpty(values)) { return; } Iterable<V> currentData = myCache.get(key); Collection<V> collection = currentData instanceof Collection? (Collection<V>)currentData : Iterators.collect(currentData, new SmartList<>()); if (!collection.isEmpty()) { boolean changes = false; for (V value : values) { changes |= collection.remove(value); } if (changes) { myCache.invalidate(key); if (collection.isEmpty()) { myDelegate.remove(key); } else { myDelegate.put(key, collection); } } } } | removeValues |
2,247 | Iterable<K> () { return myDelegate.getKeys(); } | getKeys |
2,248 | DataInput (DataInput in) { return new GraphDataInput(in); } | wrap |
2,249 | DataInput (DataInput in, StringEnumerator enumerator) { return new GraphDataInput(in) { @Override public @NotNull String readUTF() throws IOException { return enumerator.toString(readInt()); } }; } | wrap |
2,250 | boolean (K key) { try { return myMap.containsMapping(key); } catch (IOException e) { throw new BuildDataCorruptedException(e); } } | containsKey |
2,251 | void (K key, V value) { try { if (value == null) { myMap.remove(key); } else { myMap.put(key, value); } } catch (IOException e) { throw new BuildDataCorruptedException(e); } } | put |
2,252 | void (K key) { try { myMap.remove(key); } catch (IOException e) { throw new BuildDataCorruptedException(e); } } | remove |
2,253 | Iterable<K> () { try { return myMap.getAllKeysWithExistingMapping(); } catch (IOException e) { throw new RuntimeException(e); } } | getKeys |
2,254 | void (BackDependencyIndex index) { myIndices.add(index); } | addIndex |
2,255 | Iterable<ReferenceID> (@NotNull ReferenceID id) { return myDependencyIndex.getDependencies(id); } | getDependingNodes |
2,256 | Iterable<BackDependencyIndex> () { return myIndices; } | getIndices |
2,257 | Iterable<NodeSource> (@NotNull ReferenceID id) { return myNodeToSourcesMap.get(id); } | getSources |
2,258 | Iterable<ReferenceID> () { return myNodeToSourcesMap.getKeys(); } | getRegisteredNodes |
2,259 | Iterable<NodeSource> () { return mySourceToNodesMap.getKeys(); } | getSources |
2,260 | void (@NotNull FileGeneratedEvent event) { final ProjectDescriptor pd = myContext.getProjectDescriptor(); final BuildFSState fsState = pd.fsState; for (Pair<String, String> pair : event.getPaths()) { final String relativePath = pair.getSecond(); final File file = relativePath.equals(".") ? new File(pair.getFirst()) : new File(pair.getFirst(), relativePath); for (BuildRootDescriptor desc : pd.getBuildRootIndex().findAllParentDescriptors(file, myContext)) { if (!event.getSourceTarget().equals(desc.getTarget())) { // do not mark files belonging to the target that originated the event // It is assumed that those files will be explicitly marked dirty by particular builder, if needed. try { fsState.markDirty(myContext, file, desc, pd.getProjectStamps().getStampStorage(), false); } catch (IOException ignored) { } } } } } | filesGenerated |
2,261 | void (@NotNull FileDeletedEvent event) { final BuildFSState state = myContext.getProjectDescriptor().fsState; final BuildRootIndex rootsIndex = myContext.getProjectDescriptor().getBuildRootIndex(); for (String path : event.getFilePaths()) { final File file = new File(FileUtil.toSystemDependentName(path)); for (BuildRootDescriptor desc : rootsIndex.findAllParentDescriptors(file, myContext)) { state.registerDeleted(myContext, desc.getTarget(), file); } } } | filesDeleted |
2,262 | ProfilingMode () { String profilingModeString = System.getProperty(PROFILING_MODE_PROPERTY, "false"); switch (profilingModeString) { case "false": return ProfilingMode.NONE; case "true": return ProfilingMode.YOURKIT_SAMPLING; case "tracing": return ProfilingMode.YOURKIT_TRACING; default: throw new IllegalArgumentException("Invalid value of '" + PROFILING_MODE_PROPERTY + "' system property (accepting only 'false', 'true', 'tracing'): " + profilingModeString); } } | getProfilingMode |
2,263 | File () { return ourSystemRoot; } | getSystemRoot |
2,264 | void (@NotNull File systemRoot) { ourSystemRoot = systemRoot; } | setSystemRoot |
2,265 | File (@NotNull String projectPath) { return getDataStorageRoot(ourSystemRoot, projectPath); } | getDataStorageRoot |
2,266 | File (@NotNull File systemRoot, @NotNull String projectPath) { return getDataStorageRoot(systemRoot, projectPath, s -> s.hashCode()); } | getDataStorageRoot |
2,267 | File (@NotNull File systemRoot, @NotNull String projectPath, @NotNull Function<? super String, Integer> hashFunction) { projectPath = FileUtil.toCanonicalPath(projectPath); String name; final int locationHash; final Path rootFile = Paths.get(projectPath); if (!Files.isDirectory(rootFile) && projectPath.endsWith(".ipr")) { name = StringUtil.trimEnd(rootFile.getFileName().toString(), ".ipr"); locationHash = hashFunction.apply(projectPath); } else { Path directoryBased; if (rootFile.endsWith(PathMacroUtil.DIRECTORY_STORE_NAME)) { directoryBased = rootFile; } else { directoryBased = rootFile.resolve(PathMacroUtil.DIRECTORY_STORE_NAME); } name = PathUtilRt.suggestFileName(JpsProjectLoader.getDirectoryBaseProjectName(directoryBased)); locationHash = hashFunction.apply(directoryBased.toString()); } return new File(systemRoot, StringUtil.toLowerCase(name) + "_" + Integer.toHexString(locationHash)); } | getDataStorageRoot |
2,268 | boolean (CompileContext context) { return ERRORS_DETECTED_KEY.get(context, Boolean.FALSE); } | errorsDetected |
2,269 | String (long duration) { return StringUtil.formatDuration(duration); } | formatDuration |
2,270 | int () { //final JpsProject project = context.getProjectDescriptor().getProject(); //final JpsJavaCompilerConfiguration config = JpsJavaExtensionService.getInstance().getOrCreateCompilerConfiguration(project); //final JpsJavaCompilerOptions options = config.getCurrentCompilerOptions(); //return options.MAXIMUM_HEAP_SIZE; if (FORKED_JAVAC_HEAP_SIZE_MB > 0) { return FORKED_JAVAC_HEAP_SIZE_MB; } final int maxMbytes = (int)(Runtime.getRuntime().maxMemory() / 1048576L); if (maxMbytes < 0 || maxMbytes > 1500) { return -1; // in the size is too big for older VMs or int overflow, return -1 to let VM choose the heap size } return Math.max(maxMbytes, 256); // per-forked process: minimum 256 Mb } | suggestForkedCompilerHeapSize |
2,271 | Runnable (int count, Runnable operation) { return new Runnable() { private final AtomicInteger myCounter = new AtomicInteger(count); @Override public void run() { int currentVal = myCounter.decrementAndGet(); if (currentVal % count == 0) { try { operation.run(); } finally { while (currentVal <= 0 && !myCounter.compareAndSet(currentVal, count + (currentVal % count))) { // restore the counter currentVal = myCounter.get(); } } } } }; } | asCountedRunnable |
2,272 | void () { int currentVal = myCounter.decrementAndGet(); if (currentVal % count == 0) { try { operation.run(); } finally { while (currentVal <= 0 && !myCounter.compareAndSet(currentVal, count + (currentVal % count))) { // restore the counter currentVal = myCounter.get(); } } } } | run |
2,273 | boolean (Iterable<? extends JavaBuilderExtension> builders, File file) { for (JavaBuilderExtension extension : builders) { if (extension.shouldHonorFileEncodingForCompilation(file)) { return true; } } return false; } | shouldHonorEncodingForCompilation |
2,274 | String (JpsModule module) { final Set<String> encodings = getModuleCharsetMap().get(module); return ContainerUtil.getFirstItem(encodings, null); } | getPreferredModuleEncoding |
2,275 | Set<String> (@NotNull ModuleChunk moduleChunk) { final Map<JpsModule, Set<String>> map = getModuleCharsetMap(); Set<String> encodings = new HashSet<>(); for (JpsModule module : moduleChunk.getModules()) { final Set<String> moduleEncodings = map.get(module); if (moduleEncodings != null) { encodings.addAll(moduleEncodings); } } return encodings; } | getAllModuleChunkEncodings |
2,276 | boolean (@NotNull BuildTarget<?> target) { return isWholeTargetAffected(target) || myFiles.containsKey(target) || myIndirectlyAffectedFiles.containsKey(target); } | isAffected |
2,277 | boolean (@NotNull BuildTarget<?> target) { return (myTypes.contains(target.getTargetType()) || myTargets.contains(target) || isAffectedByAssociatedModule(target)) && !myFiles.containsKey(target); } | isWholeTargetAffected |
2,278 | boolean (@NotNull BuildTargetType<?> type) { return myTypes.contains(type) && myFiles.isEmpty(); } | isAllTargetsOfTypeAffected |
2,279 | boolean (@NotNull BuildTarget<?> target) { return myFiles.isEmpty() && myTypesToForceBuild.contains(target.getTargetType()) && isWholeTargetAffected(target); } | isBuildForced |
2,280 | boolean (@NotNull BuildTargetType<?> targetType) { return myTypesToForceBuild.contains(targetType) && isAllTargetsOfTypeAffected(targetType); } | isBuildForcedForAllTargets |
2,281 | boolean (@NotNull BuildTargetType<?> targetType) { return !myTypesToForceBuild.contains(targetType); } | isBuildIncrementally |
2,282 | boolean (BuildTarget<?> target, @NotNull File file) { final Set<File> files = myFiles.isEmpty()? null : myFiles.get(target); if (files == null) { return isWholeTargetAffected(target) || isIndirectlyAffected(target, file); } return files.contains(file) || isIndirectlyAffected(target, file); } | isAffected |
2,283 | boolean (BuildTarget<?> target, @NotNull File file) { synchronized (myIndirectlyAffectedFiles) { final Set<File> indirect = myIndirectlyAffectedFiles.get(target); return indirect != null && indirect.contains(file); } } | isIndirectlyAffected |
2,284 | void (BuildTarget<?> target, @NotNull File file) { synchronized (myIndirectlyAffectedFiles) { Set<File> files = myIndirectlyAffectedFiles.get(target); if (files == null) { files = new HashSet<>(); myIndirectlyAffectedFiles.put(target, files); } files.add(file); } } | markIndirectlyAffected |
2,285 | boolean (BuildTarget<?> target) { if (target instanceof ModuleBasedTarget) { final JpsModule module = ((ModuleBasedTarget<?>)target).getModule(); // this target is associated with module JavaModuleBuildTargetType targetType = JavaModuleBuildTargetType.getInstance(((ModuleBasedTarget<?>)target).isTests()); if (myTypes.contains(targetType) || myTargets.contains(new ModuleBuildTarget(module, targetType))) { return true; } } return false; } | isAffectedByAssociatedModule |
2,286 | File () { return myOutputFile; } | getOutputFile |
2,287 | Collection<File> () { return mySourceFiles; } | getSourceFiles |
2,288 | List<String> () { return ContainerUtil.map(mySourceFiles, file -> file.getPath()); } | getSourceFilesPaths |
2,289 | BinaryContent () { return myContent; } | getContent |
2,290 | void (@NotNull BinaryContent content) { myContent = content; myIsDirty = true; } | setContent |
2,291 | boolean () { return myIsDirty; } | isDirty |
2,292 | boolean (Object o) { if (this == o) return true; if (o == null || getClass() != o.getClass()) return false; CompiledClass aClass = (CompiledClass)o; if (!FileUtil.filesEqual(myOutputFile, aClass.myOutputFile)) return false; return true; } | equals |
2,293 | int () { return FileUtil.fileHashCode(myOutputFile); } | hashCode |
2,294 | String () { return "CompiledClass{" + "myOutputFile=" + myOutputFile + ", mySourceFiles=" + mySourceFiles + ", myIsDirty=" + myIsDirty + '}'; } | toString |
2,295 | String () { return getModule().getName(); } | getId |
2,296 | Set<File> (File root, ModuleExcludeIndex index) { final Collection<File> moduleExcludes = index.getModuleExcludes(getModule()); if (moduleExcludes.isEmpty()) { return Collections.emptySet(); } final Set<File> excludes = FileCollectionFactory.createCanonicalFileSet(); for (File excluded : moduleExcludes) { if (FileUtil.isAncestor(root, excluded, true)) { excludes.add(excluded); } } return excludes; } | computeRootExcludes |
2,297 | R (@NotNull String rootId, @NotNull BuildRootIndex rootIndex) { final List<R> descriptors = rootIndex.getRootDescriptors( new File(rootId), Collections.singletonList((BuildTargetType<? extends JVMModuleBuildTarget<R>>)getTargetType()), null ); return ContainerUtil.getFirstItem(descriptors); } | findRootDescriptor |
2,298 | void (BuildMessage msg) { } | processMessage |
2,299 | boolean (@NotNull String path, @NotNull Collection<? super String> deletedPaths, @Nullable Set<? super File> parentDirs) { File file = new File(path); boolean deleted = deleteRecursively(file, deletedPaths); if (deleted && parentDirs != null) { File parent = file.getParentFile(); if (parent != null) { parentDirs.add(parent); } } return deleted; } | deleteRecursively |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.