Mercurial > jhg
view src/org/tmatesoft/hg/core/HgLogCommand.java @ 515:e6c8b9b654b2
Provide access to HgDataFile being iterated into HgChangesetTreeHandler.TreeElement to give context for renamed files
author | Artem Tikhomirov <tikhomirov.artem@gmail.com> |
---|---|
date | Mon, 17 Dec 2012 20:51:12 +0100 |
parents | 5dcb4581c8ef |
children | 0ae5768081aa |
line wrap: on
line source
/* s * Copyright (c) 2011-2012 TMate Software Ltd * * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation; version 2 of the License. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * For information on how to redistribute this software under * the terms of a license other than GNU General Public License * contact TMate Software at support@hg4j.com */ package org.tmatesoft.hg.core; import static org.tmatesoft.hg.repo.HgRepository.TIP; import static org.tmatesoft.hg.util.LogFacility.Severity.Error; import java.util.ArrayList; import java.util.Arrays; import java.util.Calendar; import java.util.Collection; import java.util.Collections; import java.util.Comparator; import java.util.ConcurrentModificationException; import java.util.Iterator; import java.util.LinkedList; import java.util.List; import java.util.ListIterator; import java.util.Set; import java.util.TreeSet; import org.tmatesoft.hg.internal.IntMap; import org.tmatesoft.hg.internal.IntVector; import org.tmatesoft.hg.repo.HgChangelog; import org.tmatesoft.hg.repo.HgChangelog.RawChangeset; import org.tmatesoft.hg.repo.HgDataFile; import org.tmatesoft.hg.repo.HgInvalidControlFileException; import org.tmatesoft.hg.repo.HgInvalidRevisionException; import org.tmatesoft.hg.repo.HgInvalidStateException; import org.tmatesoft.hg.repo.HgParentChildMap; import org.tmatesoft.hg.repo.HgRepository; import org.tmatesoft.hg.repo.HgRuntimeException; import org.tmatesoft.hg.repo.HgStatusCollector; import org.tmatesoft.hg.util.Adaptable; import org.tmatesoft.hg.util.CancelSupport; import org.tmatesoft.hg.util.CancelledException; import org.tmatesoft.hg.util.Pair; import org.tmatesoft.hg.util.Path; import org.tmatesoft.hg.util.ProgressSupport; /** * Access to changelog, 'hg log' command counterpart. * * <pre> * Usage: * new LogCommand().limit(20).branch("maintenance-2.1").user("me").execute(new MyHandler()); * </pre> * Not thread-safe (each thread has to use own {@link HgLogCommand} instance). * * @author Artem Tikhomirov * @author TMate Software Ltd. */ public class HgLogCommand extends HgAbstractCommand<HgLogCommand> implements HgChangelog.Inspector { private final HgRepository repo; private Set<String> users; private Set<String> branches; private int limit = 0, count = 0; private int startRev = 0, endRev = TIP; private Calendar date; private Path file; /* * Whether to iterate file origins, if any. * Makes sense only when file != null */ private boolean followRenames; /* * Whether to track history of the selected file version (based on file revision * in working dir parent), follow ancestors only. * Note, 'hg log --follow' combines both #followHistory and #followAncestry */ private boolean followAncestry; private ChangesetTransformer csetTransform; private HgParentChildMap<HgChangelog> parentHelper; public HgLogCommand(HgRepository hgRepo) { repo = hgRepo; } /** * Limit search to specified user. Multiple user names may be specified. Once set, user names can't be * cleared, use new command instance in such cases. * @param user - full or partial name of the user, case-insensitive, non-null. * @return <code>this</code> instance for convenience * @throws IllegalArgumentException when argument is null */ public HgLogCommand user(String user) { if (user == null) { throw new IllegalArgumentException(); } if (users == null) { users = new TreeSet<String>(); } users.add(user.toLowerCase()); return this; } /** * Limit search to specified branch. Multiple branch specification possible (changeset from any of these * would be included in result). If unspecified, all branches are considered. There's no way to clean branch selection * once set, create fresh new command instead. * @param branch - branch name, case-sensitive, non-null. * @return <code>this</code> instance for convenience * @throws IllegalArgumentException when branch argument is null */ public HgLogCommand branch(String branch) { if (branch == null) { throw new IllegalArgumentException(); } if (branches == null) { branches = new TreeSet<String>(); } branches.add(branch); return this; } // limit search to specific date // multiple? public HgLogCommand date(Calendar date) { this.date = date; // TODO post-1.0 implement // isSet(field) - false => don't use in detection of 'same date' throw HgRepository.notImplemented(); } /** * * @param num - number of changeset to produce. Pass 0 to clear the limit. * @return <code>this</code> instance for convenience */ public HgLogCommand limit(int num) { limit = num; return this; } /** * Limit to specified subset of Changelog, [min(rev1,rev2), max(rev1,rev2)], inclusive. * Revision may be specified with {@link HgRepository#TIP} * * @param rev1 - local index of start changeset revision * @param rev2 - index of end changeset revision * @return <code>this</code> instance for convenience */ public HgLogCommand range(int rev1, int rev2) { if (rev1 != TIP && rev2 != TIP) { startRev = rev2 < rev1 ? rev2 : rev1; endRev = startRev == rev2 ? rev1 : rev2; } else if (rev1 == TIP && rev2 != TIP) { startRev = rev2; endRev = rev1; } else { startRev = rev1; endRev = rev2; } return this; } /** * Select specific changeset * * @param nid changeset revision * @return <code>this</code> for convenience * @throws HgBadArgumentException if failed to find supplied changeset revision */ public HgLogCommand changeset(Nodeid nid) throws HgBadArgumentException { // XXX perhaps, shall support multiple (...) arguments and extend #execute to handle not only range, but also set of revisions. try { final int csetRevIndex = repo.getChangelog().getRevisionIndex(nid); return range(csetRevIndex, csetRevIndex); } catch (HgInvalidRevisionException ex) { throw new HgBadArgumentException("Can't find revision", ex).setRevision(nid); } } /** * Visit history of a given file only. * @param file path relative to repository root. Pass <code>null</code> to reset. * @param followCopyRename true to report changesets of the original file(-s), if copy/rename ever occured to the file. */ public HgLogCommand file(Path file, boolean followCopyRename) { // multiple? Bad idea, would need to include extra method into Handler to tell start of next file this.file = file; followRenames = followAncestry = followCopyRename; return this; } /** * Handy analog of {@link #file(Path, boolean)} when clients' paths come from filesystem and need conversion to repository's */ public HgLogCommand file(String file, boolean followCopyRename) { return file(Path.create(repo.getToRepoPathHelper().rewrite(file)), followCopyRename); } /** * Similar to {@link #execute(HgChangesetHandler)}, collects and return result as a list. * * @see #execute(HgChangesetHandler) * @throws HgException subclass thereof to indicate specific issue with the command arguments or repository state */ public List<HgChangeset> execute() throws HgException { CollectHandler collector = new CollectHandler(); try { execute(collector); } catch (HgCallbackTargetException ex) { // see below for CanceledException HgInvalidStateException t = new HgInvalidStateException("Internal error"); t.initCause(ex); throw t; } catch (CancelledException ex) { // can't happen as long as our CollectHandler doesn't throw any exception HgInvalidStateException t = new HgInvalidStateException("Internal error"); t.initCause(ex); throw t; } return collector.getChanges(); } /** * Iterate over range of changesets configured in the command. * * @param handler callback to process changesets. * @throws HgCallbackTargetException propagated exception from the handler * @throws HgException subclass thereof to indicate specific issue with the command arguments or repository state * @throws CancelledException if execution of the command was cancelled * @throws IllegalArgumentException when inspector argument is null * @throws ConcurrentModificationException if this log command instance is already running */ public void execute(HgChangesetHandler handler) throws HgCallbackTargetException, HgException, CancelledException { if (handler == null) { throw new IllegalArgumentException(); } if (csetTransform != null) { throw new ConcurrentModificationException(); } final ProgressSupport progressHelper = getProgressSupport(handler); try { count = 0; HgParentChildMap<HgChangelog> pw = getParentHelper(file == null); // leave it uninitialized unless we iterate whole repo // ChangesetTransfrom creates a blank PathPool, and #file(String, boolean) above // may utilize it as well. CommandContext? How about StatusCollector there as well? csetTransform = new ChangesetTransformer(repo, handler, pw, progressHelper, getCancelSupport(handler, true)); if (file == null) { progressHelper.start(endRev - startRev + 1); repo.getChangelog().range(startRev, endRev, this); csetTransform.checkFailure(); } else { progressHelper.start(-1/*XXX enum const, or a dedicated method startUnspecified(). How about startAtLeast(int)?*/); HgDataFile fileNode = repo.getFileNode(file); if (!fileNode.exists()) { throw new HgPathNotFoundException(String.format("File %s not found in the repository", file), file); } // FIXME startRev and endRev ARE CHANGESET REVISIONS, not that of FILE!!! fileNode.history(startRev, endRev, this); csetTransform.checkFailure(); final HgFileRenameHandlerMixin withCopyHandler = Adaptable.Factory.getAdapter(handler, HgFileRenameHandlerMixin.class, null); if (fileNode.isCopy()) { // even if we do not follow history, report file rename do { if (withCopyHandler != null) { HgFileRevision src = new HgFileRevision(repo, fileNode.getCopySourceRevision(), null, fileNode.getCopySourceName()); HgFileRevision dst = new HgFileRevision(repo, fileNode.getRevision(0), null, fileNode.getPath(), src.getPath()); withCopyHandler.copy(src, dst); } if (limit > 0 && count >= limit) { // if limit reach, follow is useless. break; } if (followRenames) { fileNode = repo.getFileNode(fileNode.getCopySourceName()); fileNode.history(this); csetTransform.checkFailure(); } } while (followRenames && fileNode.isCopy()); } } } catch (HgRuntimeException ex) { throw new HgLibraryFailureException(ex); } finally { csetTransform = null; progressHelper.done(); } } /** * Tree-wise iteration of a file history, with handy access to parent-child relations between changesets. * When file history is being followed, handler may additionally implement {@link HgFileRenameHandlerMixin} * to get notified about switching between history chunks that belong to different names. * * @param handler callback to process changesets. * @see HgFileRenameHandlerMixin * @throws HgCallbackTargetException propagated exception from the handler * @throws HgException subclass thereof to indicate specific issue with the command arguments or repository state * @throws CancelledException if execution of the command was cancelled * @throws IllegalArgumentException if command is not satisfied with its arguments * @throws ConcurrentModificationException if this log command instance is already running */ public void execute(final HgChangesetTreeHandler handler) throws HgCallbackTargetException, HgException, CancelledException { if (handler == null) { throw new IllegalArgumentException(); } if (csetTransform != null) { throw new ConcurrentModificationException(); } if (file == null) { throw new IllegalArgumentException("History tree is supported for files only (at least now), please specify file"); } final ProgressSupport progressHelper = getProgressSupport(handler); final CancelSupport cancelHelper = getCancelSupport(handler, true); final HgFileRenameHandlerMixin renameHandler = Adaptable.Factory.getAdapter(handler, HgFileRenameHandlerMixin.class, null); // builds tree of nodes according to parents in file's revlog final TreeBuildInspector treeBuildInspector = new TreeBuildInspector(followRenames); // we iterate separate histories of each filename, need to connect // last node of historyA with first node of historyB (A renamed to B case) // to make overall history smooth. HistoryNode lastFromPrevIteration = null; class HandlerDispatcher { private final int CACHE_CSET_IN_ADVANCE_THRESHOLD = 100; /* XXX is it really worth it? */ private ElementImpl ei = null; private ProgressSupport progress; private HgDataFile currentFileNode; public void prepare(ProgressSupport parentProgress, int historyNodeCount, TreeBuildInspector treeBuildInspector) { if (ei == null) { // when follow is true, changeHistory.size() of the first revision might be quite short // (e.g. bad fname recognized soon), hence ensure at least cache size at once ei = new ElementImpl(Math.max(CACHE_CSET_IN_ADVANCE_THRESHOLD, historyNodeCount)); } if (historyNodeCount < CACHE_CSET_IN_ADVANCE_THRESHOLD ) { int[] commitRevisions = treeBuildInspector.getCommitRevisions(); // read bunch of changesets at once and cache 'em ei.initTransform(); repo.getChangelog().range(ei, commitRevisions); parentProgress.worked(1); progress = new ProgressSupport.Sub(parentProgress, 2); } else { progress = new ProgressSupport.Sub(parentProgress, 3); } progress.start(historyNodeCount); } public void once(HistoryNode n) throws HgCallbackTargetException, CancelledException { handler.treeElement(ei.init(n, currentFileNode)); progress.worked(1); cancelHelper.checkCancelled(); } public void switchTo(HgDataFile df) { // from now on, use df in TreeElement currentFileNode = df; } }; final HandlerDispatcher dispatcher = new HandlerDispatcher(); // renamed files in the queue are placed with respect to #iterateDirection // i.e. if we iterate from new to old, recent filenames come first List<Pair<HgDataFile, Nodeid>> fileRenamesQueue = buildFileRenamesQueue(); progressHelper.start(4 * fileRenamesQueue.size()); for (int namesIndex = 0, renamesQueueSize = fileRenamesQueue.size(); namesIndex < renamesQueueSize; namesIndex++) { final Pair<HgDataFile, Nodeid> renameInfo = fileRenamesQueue.get(namesIndex); cancelHelper.checkCancelled(); final List<HistoryNode> changeHistory = treeBuildInspector.go(renameInfo.first(), renameInfo.second()); assert changeHistory.size() > 0; progressHelper.worked(1); cancelHelper.checkCancelled(); dispatcher.prepare(progressHelper, changeHistory.size(), treeBuildInspector); if (lastFromPrevIteration != null) { if (iterateDirection == IterateDirection.FromOldToNew) { // forward, from old to new: // A(0..n) -> B(0..m). First, report A(0)..A(n-1) // then A(n).bind(B(0)) HistoryNode oldestOfTheNextChunk = changeHistory.get(0); // B(0) lastFromPrevIteration.bindChild(oldestOfTheNextChunk); // lastFromPrevIteration is A(n) dispatcher.once(lastFromPrevIteration); if (renameHandler != null) { // shall report renames assert namesIndex > 0; HgDataFile lastIterationFileNode = fileRenamesQueue.get(namesIndex-1).first(); // A HgFileRevision copiedFrom = new HgFileRevision(lastIterationFileNode, lastFromPrevIteration.fileRevision, null); HgFileRevision copiedTo = new HgFileRevision(renameInfo.first(), oldestOfTheNextChunk.fileRevision, copiedFrom.getPath()); renameHandler.copy(copiedFrom, copiedTo); } } else { assert iterateDirection == IterateDirection.FromNewToOld; // A renamed to B. A(0..n) -> B(0..m). // First, report B(m), B(m-1)...B(1), then A(n).bind(B(0)), report B(0), A(n)... HistoryNode newestOfNextChunk = changeHistory.get(changeHistory.size() - 1); // A(n) newestOfNextChunk.bindChild(lastFromPrevIteration); dispatcher.once(lastFromPrevIteration); if (renameHandler != null) { assert namesIndex > 0; // renameInfo points to chunk of name A now, and lastFromPrevIteration (from namesIndex-1) is B HgFileRevision copiedFrom = new HgFileRevision(renameInfo.first(), newestOfNextChunk.fileRevision, null); HgDataFile lastIterationFileNode = fileRenamesQueue.get(namesIndex-1).first(); // B HgFileRevision copiedTo = new HgFileRevision(lastIterationFileNode, lastFromPrevIteration.fileRevision, copiedFrom.getPath()); renameHandler.copy(copiedFrom, copiedTo); } } } if (namesIndex + 1 < renamesQueueSize) { // there's at least one more name we are going to look at, save // one element for later binding // if (iterateDirection == IterateDirection.FromOldToNew) { // save newest, and exclude it from this iteration (postpone for next) lastFromPrevIteration = changeHistory.remove(changeHistory.size()-1); } else { assert iterateDirection == IterateDirection.FromNewToOld; // save oldest, and exclude it from this iteration (postpone for next) lastFromPrevIteration = changeHistory.remove(0); } } else { lastFromPrevIteration = null; // just for the sake of no references to old items } dispatcher.switchTo(renameInfo.first()); // XXX shall sort changeHistory according to changeset numbers? Iterator<HistoryNode> it; if (iterateDirection == IterateDirection.FromOldToNew) { it = changeHistory.listIterator(); } else { assert iterateDirection == IterateDirection.FromNewToOld; it = new ReverseIterator<HistoryNode>(changeHistory); } while(it.hasNext()) { HistoryNode n = it.next(); dispatcher.once(n); } } // for fileRenamesQueue; progressHelper.done(); } private IterateDirection iterateDirection = IterateDirection.FromOldToNew; private static class ReverseIterator<E> implements Iterator<E> { private final ListIterator<E> listIterator; public ReverseIterator(List<E> list) { listIterator = list.listIterator(list.size()); } public boolean hasNext() { return listIterator.hasPrevious(); } public E next() { return listIterator.previous(); } public void remove() { listIterator.remove(); } } /** * Follows file renames and build a list of all corresponding file nodes and revisions they were * copied/renamed/branched at (IOW, their latest revision to look at). * * If {@link #followRenames} is <code>false</code>, the list contains one element only, * file node with the name of the file as it was specified by the user. * * For the most recent file revision depends on {@link #followAncestry}, and is file revision from working copy parent * in it's true. <code>null</code> indicates file's TIP revision shall be used. * * TODO may use HgFileRevision (after some refactoring to accept HgDataFile and Nodeid) instead of Pair * and possibly reuse this functionality * * @return list of file renames, ordered with respect to {@link #iterateDirection} */ private List<Pair<HgDataFile, Nodeid>> buildFileRenamesQueue() { LinkedList<Pair<HgDataFile, Nodeid>> rv = new LinkedList<Pair<HgDataFile, Nodeid>>(); Nodeid startRev = null; HgDataFile fileNode = repo.getFileNode(file); if (followAncestry) { // TODO subject to dedicated method either in HgRepository (getWorkingCopyParentRevisionIndex) // or in the HgDataFile (getWorkingCopyOriginRevision) Nodeid wdParentChangeset = repo.getWorkingCopyParents().first(); if (!wdParentChangeset.isNull()) { int wdParentRevIndex = repo.getChangelog().getRevisionIndex(wdParentChangeset); startRev = repo.getManifest().getFileRevision(wdParentRevIndex, fileNode.getPath()); } // else fall-through, assume null (eventually, lastRevision()) is ok here } rv.add(new Pair<HgDataFile, Nodeid>(fileNode, startRev)); if (!followRenames) { return rv; } while (fileNode.isCopy()) { Path fp = fileNode.getCopySourceName(); Nodeid copyRev = fileNode.getCopySourceRevision(); fileNode = repo.getFileNode(fp); Pair<HgDataFile, Nodeid> p = new Pair<HgDataFile, Nodeid>(fileNode, copyRev); if (iterateDirection == IterateDirection.FromOldToNew) { rv.addFirst(p); } else { assert iterateDirection == IterateDirection.FromNewToOld; rv.addLast(p); } }; return rv; } private static class TreeBuildInspector implements HgChangelog.ParentInspector, HgChangelog.RevisionInspector { private final boolean followAncestry; private HistoryNode[] completeHistory; private int[] commitRevisions; private List<HistoryNode> resultHistory; TreeBuildInspector(boolean _followAncestry) { followAncestry = _followAncestry; } public void next(int revisionNumber, Nodeid revision, int linkedRevision) { commitRevisions[revisionNumber] = linkedRevision; } public void next(int revisionNumber, Nodeid revision, int parent1, int parent2, Nodeid nidParent1, Nodeid nidParent2) { HistoryNode p1 = null, p2 = null; if (parent1 != -1) { p1 = completeHistory[parent1]; } if (parent2!= -1) { p2 = completeHistory[parent2]; } completeHistory[revisionNumber] = new HistoryNode(commitRevisions[revisionNumber], revision, p1, p2); } /** * Builds history of file changes (in natural order, from oldest to newest) up to (and including) file revision specified. * If {@link TreeBuildInspector} follows ancestry, only elements that are on the line of ancestry of the revision at * lastRevisionIndex would be included. * * @return list of history elements, from oldest to newest. In case {@link #followAncestry} is <code>true</code>, the list * is modifiable (to further augment with last/first elements of renamed file histories) */ List<HistoryNode> go(HgDataFile fileNode, Nodeid fileLastRevisionToVisit) throws HgInvalidControlFileException { resultHistory = null; int fileLastRevIndexToVisit = fileLastRevisionToVisit == null ? fileNode.getLastRevision() : fileNode.getRevisionIndex(fileLastRevisionToVisit); completeHistory = new HistoryNode[fileLastRevIndexToVisit+1]; commitRevisions = new int[completeHistory.length]; fileNode.indexWalk(0, fileLastRevIndexToVisit, this); if (!followAncestry) { // in case when ancestor not followed, it's safe to return unmodifiable list resultHistory = Arrays.asList(completeHistory); completeHistory = null; // keep commitRevisions initialized, no need to recalculate them // as they correspond 1:1 to resultHistory return resultHistory; } /* * Changesets, newest at the top: * o <-- cset from working dir parent (as in dirstate), file not changed (file revision recorded points to that from A) * | x <-- revision with file changed (B') * x / <-- revision with file changed (A) * | x <-- revision with file changed (B) * |/ * o <-- another changeset, where file wasn't changed * | * x <-- revision with file changed (C) * * File history: B', A, B, C * * When "follow", SHALL NOT report B and B', but A and C */ // strippedHistory: only those HistoryNodes from completeHistory that are on the same // line of descendant, in order from older to newer LinkedList<HistoryNode> strippedHistoryList = new LinkedList<HistoryNode>(); LinkedList<HistoryNode> queue = new LinkedList<HistoryNode>(); // look for ancestors of the selected history node queue.add(completeHistory[fileLastRevIndexToVisit]); do { HistoryNode withFileChange = queue.removeFirst(); if (strippedHistoryList.contains(withFileChange)) { // fork point for the change that was later merged (and we traced // both lines of development by now. continue; } if (withFileChange.children != null) { withFileChange.children.retainAll(strippedHistoryList); } strippedHistoryList.addFirst(withFileChange); if (withFileChange.parent1 != null) { queue.addLast(withFileChange.parent1); } if (withFileChange.parent2 != null) { queue.addLast(withFileChange.parent2); } } while (!queue.isEmpty()); Collections.sort(strippedHistoryList, new Comparator<HistoryNode>() { public int compare(HistoryNode o1, HistoryNode o2) { return o1.changeset - o2.changeset; } }); completeHistory = null; commitRevisions = null; // collected values are no longer valid - shall // strip off elements for missing HistoryNodes, but it's easier just to re-create the array // from resultHistory later, once (and if) needed return resultHistory = strippedHistoryList; } /** * handy access to all HistoryNode[i].changeset values */ int[] getCommitRevisions() { if (commitRevisions == null) { commitRevisions = new int[resultHistory.size()]; int i = 0; for (HistoryNode n : resultHistory) { commitRevisions[i++] = n.changeset; } } return commitRevisions; } }; // public void next(int revisionNumber, Nodeid nodeid, RawChangeset cset) { if (limit > 0 && count >= limit) { return; } if (branches != null && !branches.contains(cset.branch())) { return; } if (users != null) { String csetUser = cset.user().toLowerCase(); boolean found = false; for (String u : users) { if (csetUser.indexOf(u) != -1) { found = true; break; } } if (!found) { return; } } if (date != null) { // TODO post-1.0 implement date support for log } count++; csetTransform.next(revisionNumber, nodeid, cset); } private HgParentChildMap<HgChangelog> getParentHelper(boolean create) throws HgInvalidControlFileException { if (parentHelper == null && create) { parentHelper = new HgParentChildMap<HgChangelog>(repo.getChangelog()); parentHelper.init(); } return parentHelper; } public static class CollectHandler implements HgChangesetHandler { private final List<HgChangeset> result = new LinkedList<HgChangeset>(); public List<HgChangeset> getChanges() { return Collections.unmodifiableList(result); } public void cset(HgChangeset changeset) { result.add(changeset.clone()); } } private static class HistoryNode { final int changeset; final Nodeid fileRevision; HistoryNode parent1; // there's special case when we can alter it, see #bindChild() final HistoryNode parent2; List<HistoryNode> children; HistoryNode(int cs, Nodeid revision, HistoryNode p1, HistoryNode p2) { changeset = cs; fileRevision = revision; parent1 = p1; parent2 = p2; if (p1 != null) { p1.addChild(this); } if (p2 != null) { p2.addChild(this); } } private void addChild(HistoryNode child) { if (children == null) { children = new ArrayList<HistoryNode>(2); } children.add(child); } /** * method to merge two history chunks for renamed file so that * this node's history continues with that of child * @param child */ public void bindChild(HistoryNode child) { assert child.parent1 == null && child.parent2 == null; // for the last element in history empty children are by construction: // we don't iterate further than last element of interest in TreeBuildInspector#go assert children == null || children.isEmpty(); child.parent1 = this; addChild(child); } public String toString() { return String.format("<cset:%d, parents: %s, %s>", changeset, parent1 == null ? "-" : String.valueOf(parent1.changeset), parent2 == null ? "-" : String.valueOf(parent2.changeset)); } } private class ElementImpl implements HgChangesetTreeHandler.TreeElement, HgChangelog.Inspector { private HistoryNode historyNode; private HgDataFile fileNode; private Pair<HgChangeset, HgChangeset> parents; private List<HgChangeset> children; private IntMap<HgChangeset> cachedChangesets; private ChangesetTransformer.Transformation transform; private Nodeid changesetRevision; private Pair<Nodeid,Nodeid> parentRevisions; private List<Nodeid> childRevisions; public ElementImpl(int total) { cachedChangesets = new IntMap<HgChangeset>(total); } ElementImpl init(HistoryNode n, HgDataFile df) { historyNode = n; fileNode = df; parents = null; children = null; changesetRevision = null; parentRevisions = null; childRevisions = null; return this; } public Nodeid fileRevision() { return historyNode.fileRevision; } public HgDataFile file() { return fileNode; } public HgChangeset changeset() { return get(historyNode.changeset)[0]; } public Pair<HgChangeset, HgChangeset> parents() { if (parents != null) { return parents; } HistoryNode p; final int p1, p2; if ((p = historyNode.parent1) != null) { p1 = p.changeset; } else { p1 = -1; } if ((p = historyNode.parent2) != null) { p2 = p.changeset; } else { p2 = -1; } HgChangeset[] r = get(p1, p2); return parents = new Pair<HgChangeset, HgChangeset>(r[0], r[1]); } public Collection<HgChangeset> children() { if (children != null) { return children; } if (historyNode.children == null) { children = Collections.emptyList(); } else { int[] childrentChangesetNumbers = new int[historyNode.children.size()]; int j = 0; for (HistoryNode hn : historyNode.children) { childrentChangesetNumbers[j++] = hn.changeset; } children = Arrays.asList(get(childrentChangesetNumbers)); } return children; } void populate(HgChangeset cs) { cachedChangesets.put(cs.getRevisionIndex(), cs); } private HgChangeset[] get(int... changelogRevisionIndex) { HgChangeset[] rv = new HgChangeset[changelogRevisionIndex.length]; IntVector misses = new IntVector(changelogRevisionIndex.length, -1); for (int i = 0; i < changelogRevisionIndex.length; i++) { if (changelogRevisionIndex[i] == -1) { rv[i] = null; continue; } HgChangeset cached = cachedChangesets.get(changelogRevisionIndex[i]); if (cached != null) { rv[i] = cached; } else { misses.add(changelogRevisionIndex[i]); } } if (misses.size() > 0) { final int[] changesets2read = misses.toArray(); initTransform(); repo.getChangelog().range(this, changesets2read); for (int changeset2read : changesets2read) { HgChangeset cs = cachedChangesets.get(changeset2read); if (cs == null) { HgInvalidStateException t = new HgInvalidStateException(String.format("Can't get changeset for revision %d", changeset2read)); throw t.setRevisionIndex(changeset2read); } // HgChangelog.range may reorder changesets according to their order in the changelog // thus need to find original index boolean sanity = false; for (int i = 0; i < changelogRevisionIndex.length; i++) { if (changelogRevisionIndex[i] == cs.getRevisionIndex()) { rv[i] = cs; sanity = true; break; } } if (!sanity) { repo.getSessionContext().getLog().dump(getClass(), Error, "Index of revision %d:%s doesn't match any of requested", cs.getRevisionIndex(), cs.getNodeid().shortNotation()); } assert sanity; } } return rv; } // init only when needed void initTransform() throws HgRuntimeException { if (transform == null) { transform = new ChangesetTransformer.Transformation(new HgStatusCollector(repo)/*XXX try to reuse from context?*/, getParentHelper(false)); } } public void next(int revisionNumber, Nodeid nodeid, RawChangeset cset) { HgChangeset cs = transform.handle(revisionNumber, nodeid, cset); populate(cs.clone()); } public Nodeid changesetRevision() { if (changesetRevision == null) { changesetRevision = getRevision(historyNode.changeset); } return changesetRevision; } public Pair<Nodeid, Nodeid> parentRevisions() { if (parentRevisions == null) { HistoryNode p; final Nodeid p1, p2; if ((p = historyNode.parent1) != null) { p1 = getRevision(p.changeset); } else { p1 = Nodeid.NULL;; } if ((p = historyNode.parent2) != null) { p2 = getRevision(p.changeset); } else { p2 = Nodeid.NULL; } parentRevisions = new Pair<Nodeid, Nodeid>(p1, p2); } return parentRevisions; } public Collection<Nodeid> childRevisions() { if (childRevisions != null) { return childRevisions; } if (historyNode.children == null) { childRevisions = Collections.emptyList(); } else { ArrayList<Nodeid> rv = new ArrayList<Nodeid>(historyNode.children.size()); for (HistoryNode hn : historyNode.children) { rv.add(getRevision(hn.changeset)); } childRevisions = Collections.unmodifiableList(rv); } return childRevisions; } // reading nodeid involves reading index only, guess, can afford not to optimize multiple reads private Nodeid getRevision(int changelogRevisionNumber) { // TODO post-1.0 pipe through pool HgChangeset cs = cachedChangesets.get(changelogRevisionNumber); if (cs != null) { return cs.getNodeid(); } else { return repo.getChangelog().getRevision(changelogRevisionNumber); } } } private enum IterateDirection { FromOldToNew, FromNewToOld } }