tikhomirov@64: /*
tikhomirov@427: s * Copyright (c) 2011-2012 TMate Software Ltd
tikhomirov@64: *
tikhomirov@64: * This program is free software; you can redistribute it and/or modify
tikhomirov@64: * it under the terms of the GNU General Public License as published by
tikhomirov@64: * the Free Software Foundation; version 2 of the License.
tikhomirov@64: *
tikhomirov@64: * This program is distributed in the hope that it will be useful,
tikhomirov@64: * but WITHOUT ANY WARRANTY; without even the implied warranty of
tikhomirov@64: * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
tikhomirov@64: * GNU General Public License for more details.
tikhomirov@64: *
tikhomirov@64: * For information on how to redistribute this software under
tikhomirov@64: * the terms of a license other than GNU General Public License
tikhomirov@102: * contact TMate Software at support@hg4j.com
tikhomirov@64: */
tikhomirov@64: package org.tmatesoft.hg.core;
tikhomirov@64:
tikhomirov@518: import static org.tmatesoft.hg.repo.HgRepository.BAD_REVISION;
tikhomirov@74: import static org.tmatesoft.hg.repo.HgRepository.TIP;
tikhomirov@456: import static org.tmatesoft.hg.util.LogFacility.Severity.Error;
tikhomirov@64:
tikhomirov@328: import java.util.ArrayList;
tikhomirov@328: import java.util.Arrays;
tikhomirov@64: import java.util.Calendar;
tikhomirov@328: import java.util.Collection;
tikhomirov@64: import java.util.Collections;
tikhomirov@511: import java.util.Comparator;
tikhomirov@64: import java.util.ConcurrentModificationException;
tikhomirov@510: import java.util.Iterator;
tikhomirov@64: import java.util.LinkedList;
tikhomirov@64: import java.util.List;
tikhomirov@510: import java.util.ListIterator;
tikhomirov@64: import java.util.Set;
tikhomirov@64: import java.util.TreeSet;
tikhomirov@64:
tikhomirov@520: import org.tmatesoft.hg.internal.AdapterPlug;
tikhomirov@520: import org.tmatesoft.hg.internal.BatchRangeHelper;
tikhomirov@520: import org.tmatesoft.hg.internal.Experimental;
tikhomirov@328: import org.tmatesoft.hg.internal.IntMap;
tikhomirov@328: import org.tmatesoft.hg.internal.IntVector;
tikhomirov@518: import org.tmatesoft.hg.internal.Lifecycle;
tikhomirov@520: import org.tmatesoft.hg.internal.LifecycleProxy;
tikhomirov@215: import org.tmatesoft.hg.repo.HgChangelog;
tikhomirov@154: import org.tmatesoft.hg.repo.HgChangelog.RawChangeset;
tikhomirov@80: import org.tmatesoft.hg.repo.HgDataFile;
tikhomirov@423: import org.tmatesoft.hg.repo.HgInvalidControlFileException;
tikhomirov@457: import org.tmatesoft.hg.repo.HgInvalidRevisionException;
tikhomirov@423: import org.tmatesoft.hg.repo.HgInvalidStateException;
tikhomirov@456: import org.tmatesoft.hg.repo.HgParentChildMap;
tikhomirov@74: import org.tmatesoft.hg.repo.HgRepository;
tikhomirov@423: import org.tmatesoft.hg.repo.HgRuntimeException;
tikhomirov@328: import org.tmatesoft.hg.repo.HgStatusCollector;
tikhomirov@514: import org.tmatesoft.hg.util.Adaptable;
tikhomirov@328: import org.tmatesoft.hg.util.CancelSupport;
tikhomirov@157: import org.tmatesoft.hg.util.CancelledException;
tikhomirov@328: import org.tmatesoft.hg.util.Pair;
tikhomirov@133: import org.tmatesoft.hg.util.Path;
tikhomirov@215: import org.tmatesoft.hg.util.ProgressSupport;
tikhomirov@64:
tikhomirov@64:
tikhomirov@64: /**
tikhomirov@131: * Access to changelog, 'hg log' command counterpart.
tikhomirov@131: *
tikhomirov@64: *
tikhomirov@131: * Usage:
tikhomirov@70: * new LogCommand().limit(20).branch("maintenance-2.1").user("me").execute(new MyHandler());
tikhomirov@64: *
tikhomirov@131: * Not thread-safe (each thread has to use own {@link HgLogCommand} instance).
tikhomirov@64: *
tikhomirov@64: * @author Artem Tikhomirov
tikhomirov@64: * @author TMate Software Ltd.
tikhomirov@64: */
tikhomirov@518: public class HgLogCommand extends HgAbstractCommand {
tikhomirov@64:
tikhomirov@64: private final HgRepository repo;
tikhomirov@64: private Set users;
tikhomirov@64: private Set branches;
tikhomirov@64: private int limit = 0, count = 0;
tikhomirov@64: private int startRev = 0, endRev = TIP;
tikhomirov@64: private Calendar date;
tikhomirov@77: private Path file;
tikhomirov@514: /*
tikhomirov@514: * Whether to iterate file origins, if any.
tikhomirov@514: * Makes sense only when file != null
tikhomirov@514: */
tikhomirov@514: private boolean followRenames;
tikhomirov@514: /*
tikhomirov@514: * Whether to track history of the selected file version (based on file revision
tikhomirov@514: * in working dir parent), follow ancestors only.
tikhomirov@514: * Note, 'hg log --follow' combines both #followHistory and #followAncestry
tikhomirov@514: */
tikhomirov@514: private boolean followAncestry;
tikhomirov@193: private ChangesetTransformer csetTransform;
tikhomirov@432: private HgParentChildMap parentHelper;
tikhomirov@80:
tikhomirov@131: public HgLogCommand(HgRepository hgRepo) {
tikhomirov@107: repo = hgRepo;
tikhomirov@64: }
tikhomirov@64:
tikhomirov@64: /**
tikhomirov@148: * Limit search to specified user. Multiple user names may be specified. Once set, user names can't be
tikhomirov@148: * cleared, use new command instance in such cases.
tikhomirov@64: * @param user - full or partial name of the user, case-insensitive, non-null.
tikhomirov@64: * @return this
instance for convenience
tikhomirov@148: * @throws IllegalArgumentException when argument is null
tikhomirov@64: */
tikhomirov@131: public HgLogCommand user(String user) {
tikhomirov@64: if (user == null) {
tikhomirov@64: throw new IllegalArgumentException();
tikhomirov@64: }
tikhomirov@64: if (users == null) {
tikhomirov@64: users = new TreeSet();
tikhomirov@64: }
tikhomirov@64: users.add(user.toLowerCase());
tikhomirov@64: return this;
tikhomirov@64: }
tikhomirov@64:
tikhomirov@64: /**
tikhomirov@64: * Limit search to specified branch. Multiple branch specification possible (changeset from any of these
tikhomirov@148: * would be included in result). If unspecified, all branches are considered. There's no way to clean branch selection
tikhomirov@148: * once set, create fresh new command instead.
tikhomirov@64: * @param branch - branch name, case-sensitive, non-null.
tikhomirov@64: * @return this
instance for convenience
tikhomirov@148: * @throws IllegalArgumentException when branch argument is null
tikhomirov@64: */
tikhomirov@131: public HgLogCommand branch(String branch) {
tikhomirov@64: if (branch == null) {
tikhomirov@64: throw new IllegalArgumentException();
tikhomirov@64: }
tikhomirov@64: if (branches == null) {
tikhomirov@64: branches = new TreeSet();
tikhomirov@64: }
tikhomirov@64: branches.add(branch);
tikhomirov@64: return this;
tikhomirov@64: }
tikhomirov@64:
tikhomirov@64: // limit search to specific date
tikhomirov@64: // multiple?
tikhomirov@131: public HgLogCommand date(Calendar date) {
tikhomirov@64: this.date = date;
tikhomirov@418: // TODO post-1.0 implement
tikhomirov@64: // isSet(field) - false => don't use in detection of 'same date'
tikhomirov@64: throw HgRepository.notImplemented();
tikhomirov@64: }
tikhomirov@64:
tikhomirov@64: /**
tikhomirov@64: *
tikhomirov@64: * @param num - number of changeset to produce. Pass 0 to clear the limit.
tikhomirov@64: * @return this
instance for convenience
tikhomirov@64: */
tikhomirov@131: public HgLogCommand limit(int num) {
tikhomirov@64: limit = num;
tikhomirov@64: return this;
tikhomirov@64: }
tikhomirov@64:
tikhomirov@64: /**
tikhomirov@64: * Limit to specified subset of Changelog, [min(rev1,rev2), max(rev1,rev2)], inclusive.
tikhomirov@64: * Revision may be specified with {@link HgRepository#TIP}
tikhomirov@427: *
tikhomirov@427: * @param rev1 - local index of start changeset revision
tikhomirov@427: * @param rev2 - index of end changeset revision
tikhomirov@64: * @return this
instance for convenience
tikhomirov@64: */
tikhomirov@131: public HgLogCommand range(int rev1, int rev2) {
tikhomirov@64: if (rev1 != TIP && rev2 != TIP) {
tikhomirov@64: startRev = rev2 < rev1 ? rev2 : rev1;
tikhomirov@64: endRev = startRev == rev2 ? rev1 : rev2;
tikhomirov@64: } else if (rev1 == TIP && rev2 != TIP) {
tikhomirov@64: startRev = rev2;
tikhomirov@64: endRev = rev1;
tikhomirov@64: } else {
tikhomirov@64: startRev = rev1;
tikhomirov@64: endRev = rev2;
tikhomirov@64: }
tikhomirov@64: return this;
tikhomirov@64: }
tikhomirov@64:
tikhomirov@77: /**
tikhomirov@253: * Select specific changeset
tikhomirov@253: *
tikhomirov@253: * @param nid changeset revision
tikhomirov@253: * @return this
for convenience
tikhomirov@427: * @throws HgBadArgumentException if failed to find supplied changeset revision
tikhomirov@253: */
tikhomirov@427: public HgLogCommand changeset(Nodeid nid) throws HgBadArgumentException {
tikhomirov@253: // XXX perhaps, shall support multiple (...) arguments and extend #execute to handle not only range, but also set of revisions.
tikhomirov@427: try {
tikhomirov@427: final int csetRevIndex = repo.getChangelog().getRevisionIndex(nid);
tikhomirov@427: return range(csetRevIndex, csetRevIndex);
tikhomirov@457: } catch (HgInvalidRevisionException ex) {
tikhomirov@427: throw new HgBadArgumentException("Can't find revision", ex).setRevision(nid);
tikhomirov@427: }
tikhomirov@253: }
tikhomirov@253:
tikhomirov@253: /**
tikhomirov@516: * Visit history of a given file only. Note, unlike native hg log
command argument --follow
, this method doesn't
tikhomirov@516: * follow file ancestry, but reports complete file history (with followCopyRenames == true
, for each
tikhomirov@516: * name of the file known in sequence). To achieve output similar to that of hg log --follow filePath
, use
tikhomirov@516: * {@link #file(Path, boolean, boolean) file(filePath, true, true)} alternative.
tikhomirov@516: *
tikhomirov@516: * @param filePath path relative to repository root. Pass null
to reset.
tikhomirov@516: * @param followCopyRename true to report changesets of the original file(-s), if copy/rename ever occured to the file.
tikhomirov@516: * @return this
for convenience
tikhomirov@77: */
tikhomirov@516: public HgLogCommand file(Path filePath, boolean followCopyRename) {
tikhomirov@516: return file(filePath, followCopyRename, false);
tikhomirov@516: }
tikhomirov@516:
tikhomirov@516: /**
tikhomirov@516: * Full control over file history iteration.
tikhomirov@516: *
tikhomirov@516: * @param filePath path relative to repository root. Pass null
to reset.
tikhomirov@516: * @param followCopyRename true to report changesets of the original file(-s), if copy/rename ever occured to the file.
tikhomirov@516: * @param followFileAncestry true to follow file history starting from revision at working copy parent. Note, only revisions
tikhomirov@516: * accessible (i.e. on direct parent line) from the selected one will be reported. This is how hg log --follow filePath
tikhomirov@516: * behaves, with the difference that this method allows separate control whether to follow renames or not.
tikhomirov@516: *
tikhomirov@516: * @return this
for convenience
tikhomirov@516: */
tikhomirov@516: public HgLogCommand file(Path filePath, boolean followCopyRename, boolean followFileAncestry) {
tikhomirov@516: file = filePath;
tikhomirov@516: followRenames = followCopyRename;
tikhomirov@516: followAncestry = followFileAncestry;
tikhomirov@77: return this;
tikhomirov@64: }
tikhomirov@142:
tikhomirov@142: /**
tikhomirov@516: * Handy analog to {@link #file(Path, boolean)} when clients' paths come from filesystem and need conversion to repository's
tikhomirov@516: * @return this
for convenience
tikhomirov@142: */
tikhomirov@142: public HgLogCommand file(String file, boolean followCopyRename) {
tikhomirov@142: return file(Path.create(repo.getToRepoPathHelper().rewrite(file)), followCopyRename);
tikhomirov@142: }
tikhomirov@64:
tikhomirov@64: /**
tikhomirov@516: * Handy analog to {@link #file(Path, boolean, boolean)} when clients' paths come from filesystem and need conversion to repository's
tikhomirov@516: * @return this
for convenience
tikhomirov@516: */
tikhomirov@516: public HgLogCommand file(String file, boolean followCopyRename, boolean followFileAncestry) {
tikhomirov@516: return file(Path.create(repo.getToRepoPathHelper().rewrite(file)), followCopyRename, followFileAncestry);
tikhomirov@516: }
tikhomirov@516:
tikhomirov@516: /**
tikhomirov@419: * Similar to {@link #execute(HgChangesetHandler)}, collects and return result as a list.
tikhomirov@427: *
tikhomirov@427: * @see #execute(HgChangesetHandler)
tikhomirov@427: * @throws HgException subclass thereof to indicate specific issue with the command arguments or repository state
tikhomirov@64: */
tikhomirov@396: public List execute() throws HgException {
tikhomirov@64: CollectHandler collector = new CollectHandler();
tikhomirov@215: try {
tikhomirov@215: execute(collector);
tikhomirov@423: } catch (HgCallbackTargetException ex) {
tikhomirov@423: // see below for CanceledException
tikhomirov@423: HgInvalidStateException t = new HgInvalidStateException("Internal error");
tikhomirov@423: t.initCause(ex);
tikhomirov@423: throw t;
tikhomirov@396: } catch (CancelledException ex) {
tikhomirov@215: // can't happen as long as our CollectHandler doesn't throw any exception
tikhomirov@423: HgInvalidStateException t = new HgInvalidStateException("Internal error");
tikhomirov@423: t.initCause(ex);
tikhomirov@423: throw t;
tikhomirov@215: }
tikhomirov@64: return collector.getChanges();
tikhomirov@64: }
tikhomirov@64:
tikhomirov@64: /**
tikhomirov@402: * Iterate over range of changesets configured in the command.
tikhomirov@64: *
tikhomirov@205: * @param handler callback to process changesets.
tikhomirov@427: * @throws HgCallbackTargetException propagated exception from the handler
tikhomirov@427: * @throws HgException subclass thereof to indicate specific issue with the command arguments or repository state
tikhomirov@380: * @throws CancelledException if execution of the command was cancelled
tikhomirov@64: * @throws IllegalArgumentException when inspector argument is null
tikhomirov@64: * @throws ConcurrentModificationException if this log command instance is already running
tikhomirov@64: */
tikhomirov@370: public void execute(HgChangesetHandler handler) throws HgCallbackTargetException, HgException, CancelledException {
tikhomirov@64: if (handler == null) {
tikhomirov@64: throw new IllegalArgumentException();
tikhomirov@64: }
tikhomirov@193: if (csetTransform != null) {
tikhomirov@64: throw new ConcurrentModificationException();
tikhomirov@64: }
tikhomirov@518: final int lastCset = endRev == TIP ? repo.getChangelog().getLastRevision() : endRev;
tikhomirov@518: // XXX pretty much like HgInternals.checkRevlogRange
tikhomirov@518: if (lastCset < 0 || lastCset > repo.getChangelog().getLastRevision()) {
tikhomirov@518: throw new HgBadArgumentException(String.format("Bad value %d for end revision", endRev), null);
tikhomirov@518: }
tikhomirov@518: if (startRev < 0 || startRev > lastCset) {
tikhomirov@518: throw new HgBadArgumentException(String.format("Bad value %d for start revision for range [%1$d..%d]", startRev, lastCset), null);
tikhomirov@518: }
tikhomirov@215: final ProgressSupport progressHelper = getProgressSupport(handler);
tikhomirov@520: final int BATCH_SIZE = 100;
tikhomirov@64: try {
tikhomirov@64: count = 0;
tikhomirov@432: HgParentChildMap pw = getParentHelper(file == null); // leave it uninitialized unless we iterate whole repo
tikhomirov@193: // ChangesetTransfrom creates a blank PathPool, and #file(String, boolean) above
tikhomirov@193: // may utilize it as well. CommandContext? How about StatusCollector there as well?
tikhomirov@322: csetTransform = new ChangesetTransformer(repo, handler, pw, progressHelper, getCancelSupport(handler, true));
tikhomirov@520: // FilteringInspector is responsible to check command arguments: users, branches, limit, etc.
tikhomirov@520: // prior to passing cset to next Inspector, which is either (a) collector to reverse cset order, then invokes
tikhomirov@520: // transformer from (b), below, with alternative cset order or (b) transformer to hi-level csets.
tikhomirov@518: FilteringInspector filterInsp = new FilteringInspector();
tikhomirov@518: filterInsp.changesets(startRev, lastCset);
tikhomirov@77: if (file == null) {
tikhomirov@520: progressHelper.start(lastCset - startRev + 1);
tikhomirov@520: if (iterateDirection == IterateDirection.FromOldToNew) {
tikhomirov@520: filterInsp.delegateTo(csetTransform);
tikhomirov@520: repo.getChangelog().range(startRev, lastCset, filterInsp);
tikhomirov@520: csetTransform.checkFailure();
tikhomirov@520: } else {
tikhomirov@520: assert iterateDirection == IterateDirection.FromNewToOld;
tikhomirov@520: BatchRangeHelper brh = new BatchRangeHelper(startRev, lastCset, BATCH_SIZE, true);
tikhomirov@520: BatchChangesetInspector batchInspector = new BatchChangesetInspector(Math.min(lastCset-startRev+1, BATCH_SIZE));
tikhomirov@520: filterInsp.delegateTo(batchInspector);
tikhomirov@520: while (brh.hasNext()) {
tikhomirov@520: brh.next();
tikhomirov@520: repo.getChangelog().range(brh.start(), brh.end(), filterInsp);
tikhomirov@520: for (BatchChangesetInspector.BatchRecord br : batchInspector.iterate(true)) {
tikhomirov@520: csetTransform.next(br.csetIndex, br.csetRevision, br.cset);
tikhomirov@520: csetTransform.checkFailure();
tikhomirov@520: }
tikhomirov@520: batchInspector.reset();
tikhomirov@520: }
tikhomirov@520: }
tikhomirov@77: } else {
tikhomirov@520: filterInsp.delegateTo(csetTransform);
tikhomirov@514: final HgFileRenameHandlerMixin withCopyHandler = Adaptable.Factory.getAdapter(handler, HgFileRenameHandlerMixin.class, null);
tikhomirov@518: List> fileRenames = buildFileRenamesQueue();
tikhomirov@518: progressHelper.start(-1/*XXX enum const, or a dedicated method startUnspecified(). How about startAtLeast(int)?*/);
tikhomirov@518:
tikhomirov@518: for (int nameIndex = 0, fileRenamesSize = fileRenames.size(); nameIndex < fileRenamesSize; nameIndex++) {
tikhomirov@518: Pair curRename = fileRenames.get(nameIndex);
tikhomirov@518: HgDataFile fileNode = curRename.first();
tikhomirov@518: if (followAncestry) {
tikhomirov@518: TreeBuildInspector treeBuilder = new TreeBuildInspector(followAncestry);
tikhomirov@520: @SuppressWarnings("unused")
tikhomirov@518: List fileAncestry = treeBuilder.go(fileNode, curRename.second());
tikhomirov@518: int[] commitRevisions = narrowChangesetRange(treeBuilder.getCommitRevisions(), startRev, lastCset);
tikhomirov@518: if (iterateDirection == IterateDirection.FromOldToNew) {
tikhomirov@518: repo.getChangelog().range(filterInsp, commitRevisions);
tikhomirov@520: csetTransform.checkFailure();
tikhomirov@518: } else {
tikhomirov@518: assert iterateDirection == IterateDirection.FromNewToOld;
tikhomirov@518: // visit one by one in the opposite direction
tikhomirov@518: for (int i = commitRevisions.length-1; i >= 0; i--) {
tikhomirov@518: int csetWithFileChange = commitRevisions[i];
tikhomirov@518: repo.getChangelog().range(csetWithFileChange, csetWithFileChange, filterInsp);
tikhomirov@518: }
tikhomirov@126: }
tikhomirov@518: } else {
tikhomirov@518: // report complete file history (XXX may narrow range with [startRev, endRev], but need to go from file rev to link rev)
tikhomirov@518: int fileStartRev = 0; //fileNode.getChangesetRevisionIndex(0) >= startRev
tikhomirov@518: int fileEndRev = fileNode.getLastRevision();
tikhomirov@520: if (iterateDirection == IterateDirection.FromOldToNew) {
tikhomirov@520: fileNode.history(fileStartRev, fileEndRev, filterInsp);
tikhomirov@520: csetTransform.checkFailure();
tikhomirov@520: } else {
tikhomirov@520: assert iterateDirection == IterateDirection.FromNewToOld;
tikhomirov@520: BatchRangeHelper brh = new BatchRangeHelper(fileStartRev, fileEndRev, BATCH_SIZE, true);
tikhomirov@520: BatchChangesetInspector batchInspector = new BatchChangesetInspector(Math.min(fileEndRev-fileStartRev+1, BATCH_SIZE));
tikhomirov@520: filterInsp.delegateTo(batchInspector);
tikhomirov@520: while (brh.hasNext()) {
tikhomirov@520: brh.next();
tikhomirov@520: fileNode.history(brh.start(), brh.end(), filterInsp);
tikhomirov@520: for (BatchChangesetInspector.BatchRecord br : batchInspector.iterate(true /*iterateDirection == IterateDirection.FromNewToOld*/)) {
tikhomirov@520: csetTransform.next(br.csetIndex, br.csetRevision, br.cset);
tikhomirov@520: csetTransform.checkFailure();
tikhomirov@520: }
tikhomirov@520: batchInspector.reset();
tikhomirov@520: }
tikhomirov@520: }
tikhomirov@518: }
tikhomirov@518: if (followRenames && withCopyHandler != null && nameIndex + 1 < fileRenamesSize) {
tikhomirov@518: Pair nextRename = fileRenames.get(nameIndex+1);
tikhomirov@518: HgFileRevision src, dst;
tikhomirov@518: // A -> B
tikhomirov@518: if (iterateDirection == IterateDirection.FromOldToNew) {
tikhomirov@518: // curRename: A, nextRename: B
tikhomirov@518: src = new HgFileRevision(fileNode, curRename.second(), null);
tikhomirov@518: dst = new HgFileRevision(nextRename.first(), nextRename.first().getRevision(0), src.getPath());
tikhomirov@518: } else {
tikhomirov@518: assert iterateDirection == IterateDirection.FromNewToOld;
tikhomirov@518: // curRename: B, nextRename: A
tikhomirov@518: src = new HgFileRevision(nextRename.first(), nextRename.second(), null);
tikhomirov@518: dst = new HgFileRevision(fileNode, fileNode.getRevision(0), src.getPath());
tikhomirov@80: }
tikhomirov@518: withCopyHandler.copy(src, dst);
tikhomirov@518: }
tikhomirov@518: } // for renames
tikhomirov@518: } // file != null
tikhomirov@427: } catch (HgRuntimeException ex) {
tikhomirov@427: throw new HgLibraryFailureException(ex);
tikhomirov@64: } finally {
tikhomirov@193: csetTransform = null;
tikhomirov@215: progressHelper.done();
tikhomirov@64: }
tikhomirov@64: }
tikhomirov@328:
tikhomirov@520: private static class BatchChangesetInspector extends AdapterPlug implements HgChangelog.Inspector {
tikhomirov@520: private static class BatchRecord {
tikhomirov@520: public final int csetIndex;
tikhomirov@520: public final Nodeid csetRevision;
tikhomirov@520: public final RawChangeset cset;
tikhomirov@520:
tikhomirov@520: public BatchRecord(int index, Nodeid nodeid, RawChangeset changeset) {
tikhomirov@520: csetIndex = index;
tikhomirov@520: csetRevision = nodeid;
tikhomirov@520: cset = changeset;
tikhomirov@520: }
tikhomirov@520: }
tikhomirov@520: private final ArrayList batch;
tikhomirov@520:
tikhomirov@520: public BatchChangesetInspector(int batchSizeHint) {
tikhomirov@520: batch = new ArrayList(batchSizeHint);
tikhomirov@520: }
tikhomirov@520:
tikhomirov@520: public BatchChangesetInspector reset() {
tikhomirov@520: batch.clear();
tikhomirov@520: return this;
tikhomirov@520: }
tikhomirov@520:
tikhomirov@520: public void next(int revisionIndex, Nodeid nodeid, RawChangeset cset) {
tikhomirov@520: batch.add(new BatchRecord(revisionIndex, nodeid, cset.clone()));
tikhomirov@520: }
tikhomirov@520:
tikhomirov@520: public Iterable iterate(final boolean reverse) {
tikhomirov@520: return new Iterable() {
tikhomirov@520:
tikhomirov@520: public Iterator iterator() {
tikhomirov@520: return reverse ? new ReverseIterator(batch) : batch.iterator();
tikhomirov@520: }
tikhomirov@520: };
tikhomirov@520: }
tikhomirov@520:
tikhomirov@520: // alternative would be dispatch(HgChangelog.Inspector) and dispatchReverse()
tikhomirov@520: // methods, but progress and cancellation might get messy then
tikhomirov@520: }
tikhomirov@520:
tikhomirov@518: // public static void main(String[] args) {
tikhomirov@518: // int[] r = new int[] {17, 19, 21, 23, 25, 29};
tikhomirov@518: // System.out.println(Arrays.toString(narrowChangesetRange(r, 0, 45)));
tikhomirov@518: // System.out.println(Arrays.toString(narrowChangesetRange(r, 0, 25)));
tikhomirov@518: // System.out.println(Arrays.toString(narrowChangesetRange(r, 5, 26)));
tikhomirov@518: // System.out.println(Arrays.toString(narrowChangesetRange(r, 20, 26)));
tikhomirov@518: // System.out.println(Arrays.toString(narrowChangesetRange(r, 26, 28)));
tikhomirov@518: // }
tikhomirov@518:
tikhomirov@518: private static int[] narrowChangesetRange(int[] csetRange, int startCset, int endCset) {
tikhomirov@518: int lastInRange = csetRange[csetRange.length-1];
tikhomirov@518: assert csetRange.length < 2 || csetRange[0] < lastInRange; // sorted
tikhomirov@518: assert startCset >= 0 && startCset <= endCset;
tikhomirov@518: if (csetRange[0] >= startCset && lastInRange <= endCset) {
tikhomirov@518: // completely fits in
tikhomirov@518: return csetRange;
tikhomirov@518: }
tikhomirov@518: if (csetRange[0] > endCset || lastInRange < startCset) {
tikhomirov@518: return new int[0]; // trivial
tikhomirov@518: }
tikhomirov@518: int i = 0;
tikhomirov@518: while (i < csetRange.length && csetRange[i] < startCset) {
tikhomirov@518: i++;
tikhomirov@518: }
tikhomirov@518: int j = csetRange.length - 1;
tikhomirov@518: while (j > i && csetRange[j] > endCset) {
tikhomirov@518: j--;
tikhomirov@518: }
tikhomirov@518: if (i == j) {
tikhomirov@518: // no values in csetRange fit into [startCset, endCset]
tikhomirov@518: return new int[0];
tikhomirov@518: }
tikhomirov@518: int[] rv = new int[j-i+1];
tikhomirov@518: System.arraycopy(csetRange, i, rv, 0, rv.length);
tikhomirov@518: return rv;
tikhomirov@518: }
tikhomirov@518:
tikhomirov@370: /**
tikhomirov@515: * Tree-wise iteration of a file history, with handy access to parent-child relations between changesets.
tikhomirov@515: * When file history is being followed, handler may additionally implement {@link HgFileRenameHandlerMixin}
tikhomirov@515: * to get notified about switching between history chunks that belong to different names.
tikhomirov@402: *
tikhomirov@402: * @param handler callback to process changesets.
tikhomirov@515: * @see HgFileRenameHandlerMixin
tikhomirov@427: * @throws HgCallbackTargetException propagated exception from the handler
tikhomirov@427: * @throws HgException subclass thereof to indicate specific issue with the command arguments or repository state
tikhomirov@380: * @throws CancelledException if execution of the command was cancelled
tikhomirov@402: * @throws IllegalArgumentException if command is not satisfied with its arguments
tikhomirov@402: * @throws ConcurrentModificationException if this log command instance is already running
tikhomirov@370: */
tikhomirov@515: public void execute(final HgChangesetTreeHandler handler) throws HgCallbackTargetException, HgException, CancelledException {
tikhomirov@328: if (handler == null) {
tikhomirov@328: throw new IllegalArgumentException();
tikhomirov@328: }
tikhomirov@328: if (csetTransform != null) {
tikhomirov@328: throw new ConcurrentModificationException();
tikhomirov@328: }
tikhomirov@328: if (file == null) {
tikhomirov@328: throw new IllegalArgumentException("History tree is supported for files only (at least now), please specify file");
tikhomirov@328: }
tikhomirov@328: final ProgressSupport progressHelper = getProgressSupport(handler);
tikhomirov@328: final CancelSupport cancelHelper = getCancelSupport(handler, true);
tikhomirov@514: final HgFileRenameHandlerMixin renameHandler = Adaptable.Factory.getAdapter(handler, HgFileRenameHandlerMixin.class, null);
tikhomirov@507:
tikhomirov@509:
tikhomirov@517: // XXX rename. dispatcher is not a proper name (most of the job done - managing history chunk interconnection)
tikhomirov@516: final HandlerDispatcher dispatcher = new HandlerDispatcher() {
tikhomirov@515:
tikhomirov@516: @Override
tikhomirov@516: protected void once(HistoryNode n) throws HgCallbackTargetException, CancelledException {
tikhomirov@515: handler.treeElement(ei.init(n, currentFileNode));
tikhomirov@515: cancelHelper.checkCancelled();
tikhomirov@515: }
tikhomirov@515: };
tikhomirov@508:
tikhomirov@510: // renamed files in the queue are placed with respect to #iterateDirection
tikhomirov@510: // i.e. if we iterate from new to old, recent filenames come first
tikhomirov@514: List> fileRenamesQueue = buildFileRenamesQueue();
tikhomirov@517: // XXX perhaps, makes sense to look at selected file's revision when followAncestry is true
tikhomirov@517: // to ensure file we attempt to trace is in the WC's parent. Native hg aborts if not.
tikhomirov@507: progressHelper.start(4 * fileRenamesQueue.size());
tikhomirov@514: for (int namesIndex = 0, renamesQueueSize = fileRenamesQueue.size(); namesIndex < renamesQueueSize; namesIndex++) {
tikhomirov@510:
tikhomirov@514: final Pair renameInfo = fileRenamesQueue.get(namesIndex);
tikhomirov@516: dispatcher.prepare(progressHelper, renameInfo);
tikhomirov@423: cancelHelper.checkCancelled();
tikhomirov@516: if (namesIndex > 0) {
tikhomirov@516: dispatcher.connectWithLastJunctionPoint(renameInfo, fileRenamesQueue.get(namesIndex - 1), renameHandler);
tikhomirov@509: }
tikhomirov@514: if (namesIndex + 1 < renamesQueueSize) {
tikhomirov@516: // there's at least one more name we are going to look at
tikhomirov@516: dispatcher.updateJunctionPoint(renameInfo, fileRenamesQueue.get(namesIndex+1));
tikhomirov@509: } else {
tikhomirov@516: dispatcher.clearJunctionPoint();
tikhomirov@509: }
tikhomirov@516: dispatcher.dispatchAllChanges();
tikhomirov@514: } // for fileRenamesQueue;
tikhomirov@328: progressHelper.done();
tikhomirov@328: }
tikhomirov@328:
tikhomirov@520: /**
tikhomirov@520: * DO NOT USE THIS METHOD, DEBUG PURPOSES ONLY!!!
tikhomirov@520: */
tikhomirov@520: @Experimental(reason="Work in progress")
tikhomirov@520: public HgLogCommand debugSwitch1() {
tikhomirov@520: // FIXME can't expose iteration direction unless general iteration (changelog, not a file) supports it, too.
tikhomirov@520: // however, need to test the code already there, hence this debug switch
tikhomirov@520: if (iterateDirection == IterateDirection.FromOldToNew) {
tikhomirov@520: iterateDirection = IterateDirection.FromNewToOld;
tikhomirov@520: } else {
tikhomirov@520: iterateDirection = IterateDirection.FromOldToNew;
tikhomirov@520: }
tikhomirov@520: return this;
tikhomirov@520: }
tikhomirov@520:
tikhomirov@510: private IterateDirection iterateDirection = IterateDirection.FromOldToNew;
tikhomirov@510:
tikhomirov@510: private static class ReverseIterator implements Iterator {
tikhomirov@510: private final ListIterator listIterator;
tikhomirov@510:
tikhomirov@510: public ReverseIterator(List list) {
tikhomirov@510: listIterator = list.listIterator(list.size());
tikhomirov@510: }
tikhomirov@510:
tikhomirov@510: public boolean hasNext() {
tikhomirov@510: return listIterator.hasPrevious();
tikhomirov@510: }
tikhomirov@510: public E next() {
tikhomirov@510: return listIterator.previous();
tikhomirov@510: }
tikhomirov@510: public void remove() {
tikhomirov@510: listIterator.remove();
tikhomirov@510: }
tikhomirov@510: }
tikhomirov@510:
tikhomirov@507: /**
tikhomirov@508: * Follows file renames and build a list of all corresponding file nodes and revisions they were
tikhomirov@508: * copied/renamed/branched at (IOW, their latest revision to look at).
tikhomirov@508: *
tikhomirov@514: * If {@link #followRenames} is false
, the list contains one element only,
tikhomirov@508: * file node with the name of the file as it was specified by the user.
tikhomirov@508: *
tikhomirov@514: * For the most recent file revision depends on {@link #followAncestry}, and is file revision from working copy parent
tikhomirov@514: * in it's true. null
indicates file's TIP revision shall be used.
tikhomirov@508: *
tikhomirov@508: * TODO may use HgFileRevision (after some refactoring to accept HgDataFile and Nodeid) instead of Pair
tikhomirov@508: * and possibly reuse this functionality
tikhomirov@507: *
tikhomirov@510: * @return list of file renames, ordered with respect to {@link #iterateDirection}
tikhomirov@507: */
tikhomirov@518: private List> buildFileRenamesQueue() throws HgPathNotFoundException {
tikhomirov@508: LinkedList> rv = new LinkedList>();
tikhomirov@514: Nodeid startRev = null;
tikhomirov@514: HgDataFile fileNode = repo.getFileNode(file);
tikhomirov@518: if (!fileNode.exists()) {
tikhomirov@518: throw new HgPathNotFoundException(String.format("File %s not found in the repository", file), file);
tikhomirov@518: }
tikhomirov@514: if (followAncestry) {
tikhomirov@514: // TODO subject to dedicated method either in HgRepository (getWorkingCopyParentRevisionIndex)
tikhomirov@514: // or in the HgDataFile (getWorkingCopyOriginRevision)
tikhomirov@514: Nodeid wdParentChangeset = repo.getWorkingCopyParents().first();
tikhomirov@514: if (!wdParentChangeset.isNull()) {
tikhomirov@514: int wdParentRevIndex = repo.getChangelog().getRevisionIndex(wdParentChangeset);
tikhomirov@514: startRev = repo.getManifest().getFileRevision(wdParentRevIndex, fileNode.getPath());
tikhomirov@514: }
tikhomirov@514: // else fall-through, assume null (eventually, lastRevision()) is ok here
tikhomirov@514: }
tikhomirov@514: rv.add(new Pair(fileNode, startRev));
tikhomirov@514: if (!followRenames) {
tikhomirov@507: return rv;
tikhomirov@507: }
tikhomirov@514: while (fileNode.isCopy()) {
tikhomirov@514: Path fp = fileNode.getCopySourceName();
tikhomirov@514: Nodeid copyRev = fileNode.getCopySourceRevision();
tikhomirov@514: fileNode = repo.getFileNode(fp);
tikhomirov@510: Pair p = new Pair(fileNode, copyRev);
tikhomirov@510: if (iterateDirection == IterateDirection.FromOldToNew) {
tikhomirov@510: rv.addFirst(p);
tikhomirov@510: } else {
tikhomirov@510: assert iterateDirection == IterateDirection.FromNewToOld;
tikhomirov@510: rv.addLast(p);
tikhomirov@510: }
tikhomirov@514: };
tikhomirov@507: return rv;
tikhomirov@507: }
tikhomirov@508:
tikhomirov@508: private static class TreeBuildInspector implements HgChangelog.ParentInspector, HgChangelog.RevisionInspector {
tikhomirov@508: private final boolean followAncestry;
tikhomirov@508:
tikhomirov@508: private HistoryNode[] completeHistory;
tikhomirov@508: private int[] commitRevisions;
tikhomirov@509: private List resultHistory;
tikhomirov@508:
tikhomirov@508: TreeBuildInspector(boolean _followAncestry) {
tikhomirov@508: followAncestry = _followAncestry;
tikhomirov@508: }
tikhomirov@508:
tikhomirov@508: public void next(int revisionNumber, Nodeid revision, int linkedRevision) {
tikhomirov@508: commitRevisions[revisionNumber] = linkedRevision;
tikhomirov@508: }
tikhomirov@508:
tikhomirov@508: public void next(int revisionNumber, Nodeid revision, int parent1, int parent2, Nodeid nidParent1, Nodeid nidParent2) {
tikhomirov@508: HistoryNode p1 = null, p2 = null;
tikhomirov@516: // IMPORTANT: method #one(), below, doesn't expect this code expects reasonable values at parent indexes
tikhomirov@508: if (parent1 != -1) {
tikhomirov@508: p1 = completeHistory[parent1];
tikhomirov@508: }
tikhomirov@508: if (parent2!= -1) {
tikhomirov@508: p2 = completeHistory[parent2];
tikhomirov@508: }
tikhomirov@508: completeHistory[revisionNumber] = new HistoryNode(commitRevisions[revisionNumber], revision, p1, p2);
tikhomirov@508: }
tikhomirov@508:
tikhomirov@516: HistoryNode one(HgDataFile fileNode, Nodeid fileRevision) throws HgInvalidControlFileException {
tikhomirov@516: int fileRevIndexToVisit = fileNode.getRevisionIndex(fileRevision);
tikhomirov@516: return one(fileNode, fileRevIndexToVisit);
tikhomirov@516: }
tikhomirov@516:
tikhomirov@516: HistoryNode one(HgDataFile fileNode, int fileRevIndexToVisit) throws HgInvalidControlFileException {
tikhomirov@516: resultHistory = null;
tikhomirov@516: if (fileRevIndexToVisit == HgRepository.TIP) {
tikhomirov@516: fileRevIndexToVisit = fileNode.getLastRevision();
tikhomirov@516: }
tikhomirov@516: // still, allocate whole array, for #next to be able to get null parent values
tikhomirov@516: completeHistory = new HistoryNode[fileRevIndexToVisit+1];
tikhomirov@516: commitRevisions = new int[completeHistory.length];
tikhomirov@516: fileNode.indexWalk(fileRevIndexToVisit, fileRevIndexToVisit, this);
tikhomirov@516: // it's only single revision, no need to care about followAncestry
tikhomirov@516: // but won't hurt to keep resultHistory != null and commitRevisions initialized just in case
tikhomirov@516: HistoryNode rv = completeHistory[fileRevIndexToVisit];
tikhomirov@516: commitRevisions = new int[] { commitRevisions[fileRevIndexToVisit] };
tikhomirov@516: completeHistory = null; // no need to keep almost empty array in memory
tikhomirov@516: resultHistory = Collections.singletonList(rv);
tikhomirov@516: return rv;
tikhomirov@516: }
tikhomirov@516:
tikhomirov@508: /**
tikhomirov@508: * Builds history of file changes (in natural order, from oldest to newest) up to (and including) file revision specified.
tikhomirov@508: * If {@link TreeBuildInspector} follows ancestry, only elements that are on the line of ancestry of the revision at
tikhomirov@508: * lastRevisionIndex would be included.
tikhomirov@509: *
tikhomirov@509: * @return list of history elements, from oldest to newest. In case {@link #followAncestry} is true
, the list
tikhomirov@509: * is modifiable (to further augment with last/first elements of renamed file histories)
tikhomirov@508: */
tikhomirov@514: List go(HgDataFile fileNode, Nodeid fileLastRevisionToVisit) throws HgInvalidControlFileException {
tikhomirov@509: resultHistory = null;
tikhomirov@514: int fileLastRevIndexToVisit = fileLastRevisionToVisit == null ? fileNode.getLastRevision() : fileNode.getRevisionIndex(fileLastRevisionToVisit);
tikhomirov@514: completeHistory = new HistoryNode[fileLastRevIndexToVisit+1];
tikhomirov@508: commitRevisions = new int[completeHistory.length];
tikhomirov@514: fileNode.indexWalk(0, fileLastRevIndexToVisit, this);
tikhomirov@508: if (!followAncestry) {
tikhomirov@509: // in case when ancestor not followed, it's safe to return unmodifiable list
tikhomirov@509: resultHistory = Arrays.asList(completeHistory);
tikhomirov@509: completeHistory = null;
tikhomirov@509: // keep commitRevisions initialized, no need to recalculate them
tikhomirov@509: // as they correspond 1:1 to resultHistory
tikhomirov@509: return resultHistory;
tikhomirov@508: }
tikhomirov@508: /*
tikhomirov@509: * Changesets, newest at the top:
tikhomirov@508: * o <-- cset from working dir parent (as in dirstate), file not changed (file revision recorded points to that from A)
tikhomirov@508: * | x <-- revision with file changed (B')
tikhomirov@508: * x / <-- revision with file changed (A)
tikhomirov@508: * | x <-- revision with file changed (B)
tikhomirov@508: * |/
tikhomirov@508: * o <-- another changeset, where file wasn't changed
tikhomirov@508: * |
tikhomirov@508: * x <-- revision with file changed (C)
tikhomirov@508: *
tikhomirov@508: * File history: B', A, B, C
tikhomirov@508: *
tikhomirov@508: * When "follow", SHALL NOT report B and B', but A and C
tikhomirov@508: */
tikhomirov@508: // strippedHistory: only those HistoryNodes from completeHistory that are on the same
tikhomirov@508: // line of descendant, in order from older to newer
tikhomirov@508: LinkedList strippedHistoryList = new LinkedList();
tikhomirov@508: LinkedList queue = new LinkedList();
tikhomirov@508: // look for ancestors of the selected history node
tikhomirov@514: queue.add(completeHistory[fileLastRevIndexToVisit]);
tikhomirov@508: do {
tikhomirov@508: HistoryNode withFileChange = queue.removeFirst();
tikhomirov@511: if (strippedHistoryList.contains(withFileChange)) {
tikhomirov@511: // fork point for the change that was later merged (and we traced
tikhomirov@511: // both lines of development by now.
tikhomirov@511: continue;
tikhomirov@511: }
tikhomirov@508: if (withFileChange.children != null) {
tikhomirov@508: withFileChange.children.retainAll(strippedHistoryList);
tikhomirov@508: }
tikhomirov@508: strippedHistoryList.addFirst(withFileChange);
tikhomirov@508: if (withFileChange.parent1 != null) {
tikhomirov@508: queue.addLast(withFileChange.parent1);
tikhomirov@508: }
tikhomirov@508: if (withFileChange.parent2 != null) {
tikhomirov@508: queue.addLast(withFileChange.parent2);
tikhomirov@508: }
tikhomirov@508: } while (!queue.isEmpty());
tikhomirov@511: Collections.sort(strippedHistoryList, new Comparator() {
tikhomirov@511:
tikhomirov@511: public int compare(HistoryNode o1, HistoryNode o2) {
tikhomirov@511: return o1.changeset - o2.changeset;
tikhomirov@511: }
tikhomirov@511: });
tikhomirov@508: completeHistory = null;
tikhomirov@508: commitRevisions = null;
tikhomirov@508: // collected values are no longer valid - shall
tikhomirov@508: // strip off elements for missing HistoryNodes, but it's easier just to re-create the array
tikhomirov@509: // from resultHistory later, once (and if) needed
tikhomirov@509: return resultHistory = strippedHistoryList;
tikhomirov@508: }
tikhomirov@508:
tikhomirov@508: /**
tikhomirov@508: * handy access to all HistoryNode[i].changeset values
tikhomirov@508: */
tikhomirov@508: int[] getCommitRevisions() {
tikhomirov@509: if (commitRevisions == null) {
tikhomirov@509: commitRevisions = new int[resultHistory.size()];
tikhomirov@509: int i = 0;
tikhomirov@509: for (HistoryNode n : resultHistory) {
tikhomirov@509: commitRevisions[i++] = n.changeset;
tikhomirov@509: }
tikhomirov@509: }
tikhomirov@508: return commitRevisions;
tikhomirov@508: }
tikhomirov@508: };
tikhomirov@508:
tikhomirov@516: private abstract class HandlerDispatcher {
tikhomirov@516: private final int CACHE_CSET_IN_ADVANCE_THRESHOLD = 100; /* XXX is it really worth it? */
tikhomirov@516: // builds tree of nodes according to parents in file's revlog
tikhomirov@516: private final TreeBuildInspector treeBuildInspector = new TreeBuildInspector(followAncestry);
tikhomirov@516: private List changeHistory;
tikhomirov@516: protected ElementImpl ei = null;
tikhomirov@516: private ProgressSupport progress;
tikhomirov@516: protected HgDataFile currentFileNode;
tikhomirov@516: // node where current file history chunk intersects with same file under other name history
tikhomirov@516: // either mock of B(0) or A(k), depending on iteration order
tikhomirov@516: private HistoryNode junctionNode;
tikhomirov@516:
tikhomirov@516: // parentProgress shall be initialized with 4 XXX refactor all this stuff with parentProgress
tikhomirov@516: public void prepare(ProgressSupport parentProgress, Pair renameInfo) {
tikhomirov@516: // if we don't followAncestry, take complete history
tikhomirov@516: // XXX treeBuildInspector knows followAncestry, perhaps the logic
tikhomirov@516: // whether to take specific revision or the last one shall be there?
tikhomirov@516: changeHistory = treeBuildInspector.go(renameInfo.first(), followAncestry ? renameInfo.second() : null);
tikhomirov@516: assert changeHistory.size() > 0;
tikhomirov@516: parentProgress.worked(1);
tikhomirov@516: int historyNodeCount = changeHistory.size();
tikhomirov@516: if (ei == null) {
tikhomirov@516: // when follow is true, changeHistory.size() of the first revision might be quite short
tikhomirov@516: // (e.g. bad fname recognized soon), hence ensure at least cache size at once
tikhomirov@516: ei = new ElementImpl(Math.max(CACHE_CSET_IN_ADVANCE_THRESHOLD, historyNodeCount));
tikhomirov@516: }
tikhomirov@516: if (historyNodeCount < CACHE_CSET_IN_ADVANCE_THRESHOLD ) {
tikhomirov@516: int[] commitRevisions = treeBuildInspector.getCommitRevisions();
tikhomirov@516: assert commitRevisions.length == changeHistory.size();
tikhomirov@516: // read bunch of changesets at once and cache 'em
tikhomirov@516: ei.initTransform();
tikhomirov@516: repo.getChangelog().range(ei, commitRevisions);
tikhomirov@516: parentProgress.worked(1);
tikhomirov@516: progress = new ProgressSupport.Sub(parentProgress, 2);
tikhomirov@516: } else {
tikhomirov@516: progress = new ProgressSupport.Sub(parentProgress, 3);
tikhomirov@516: }
tikhomirov@516: progress.start(historyNodeCount);
tikhomirov@516: // switch to present chunk's file node
tikhomirov@516: switchTo(renameInfo.first());
tikhomirov@516: }
tikhomirov@516:
tikhomirov@516: public void updateJunctionPoint(Pair curRename, Pair nextRename) {
tikhomirov@516: // A (old) renamed to B(new). A(0..k..n) -> B(0..m). If followAncestry, k == n
tikhomirov@516: // curRename.second() points to A(k)
tikhomirov@516: if (iterateDirection == IterateDirection.FromOldToNew) {
tikhomirov@516: // looking at A chunk (curRename), nextRename points to B
tikhomirov@516: HistoryNode junctionSrc = findJunctionPointInCurrentChunk(curRename.second()); // A(k)
tikhomirov@516: HistoryNode junctionDestMock = treeBuildInspector.one(nextRename.first(), 0); // B(0)
tikhomirov@516: // junstionDestMock is mock object, once we iterate next rename, there'd be different HistoryNode
tikhomirov@516: // for B's first revision. This means we read it twice, but this seems to be reasonable
tikhomirov@516: // price for simplicity of the code (and opportunity to follow renames while not following ancestry)
tikhomirov@516: junctionSrc.bindChild(junctionDestMock);
tikhomirov@516: // Save mock A(k) 1) not to keep whole A history in memory 2) Don't need it's parent and children once get to B
tikhomirov@516: // moreover, children of original A(k) (junctionSrc) would list mock B(0) which is undesired once we iterate over real B
tikhomirov@516: junctionNode = new HistoryNode(junctionSrc.changeset, junctionSrc.fileRevision, null, null);
tikhomirov@516: } else {
tikhomirov@516: assert iterateDirection == IterateDirection.FromNewToOld;
tikhomirov@516: // looking at B chunk (curRename), nextRename points at A
tikhomirov@516: HistoryNode junctionDest = changeHistory.get(0); // B(0)
tikhomirov@516: // prepare mock A(k)
tikhomirov@516: HistoryNode junctionSrcMock = treeBuildInspector.one(nextRename.first(), nextRename.second()); // A(k)
tikhomirov@516: // B(0) to list A(k) as its parent
tikhomirov@516: // NOTE, A(k) would be different when we reach A chunk on the next iteration,
tikhomirov@516: // but we do not care as long as TreeElement needs only parent/child changesets
tikhomirov@516: // and not other TreeElements; so that it's enough to have mock parent node (just
tikhomirov@516: // for the sake of parent cset revisions). We have to, indeed, update real A(k),
tikhomirov@516: // once we get to iteration over A, with B(0) (junctionDest) as one more child.
tikhomirov@516: junctionSrcMock.bindChild(junctionDest);
tikhomirov@516: // Save mock B(0), for reasons see above for opposite direction
tikhomirov@516: junctionNode = new HistoryNode(junctionDest.changeset, junctionDest.fileRevision, null, null);
tikhomirov@516: }
tikhomirov@516: }
tikhomirov@516:
tikhomirov@516: public void clearJunctionPoint() {
tikhomirov@516: junctionNode = null;
tikhomirov@516: }
tikhomirov@516:
tikhomirov@516: public void connectWithLastJunctionPoint(Pair curRename, Pair prevRename, HgFileRenameHandlerMixin renameHandler) throws HgCallbackTargetException {
tikhomirov@516: assert junctionNode != null;
tikhomirov@516: // A renamed to B. A(0..k..n) -> B(0..m). If followAncestry: k == n
tikhomirov@516: if (iterateDirection == IterateDirection.FromOldToNew) {
tikhomirov@516: // forward, from old to new:
tikhomirov@516: // changeHistory points to B
tikhomirov@516: // Already reported: A(0)..A(n), A(k) is in junctionNode
tikhomirov@516: // Shall connect histories: A(k).bind(B(0))
tikhomirov@516: HistoryNode junctionDest = changeHistory.get(0); // B(0)
tikhomirov@516: // junctionNode is A(k)
tikhomirov@516: junctionNode.bindChild(junctionDest);
tikhomirov@516: if (renameHandler != null) { // shall report renames
tikhomirov@516: HgFileRevision copiedFrom = new HgFileRevision(prevRename.first(), junctionNode.fileRevision, null); // "A", A(k)
tikhomirov@516: HgFileRevision copiedTo = new HgFileRevision(curRename.first(), junctionDest.fileRevision, copiedFrom.getPath()); // "B", B(0)
tikhomirov@516: renameHandler.copy(copiedFrom, copiedTo);
tikhomirov@516: }
tikhomirov@516: } else {
tikhomirov@516: assert iterateDirection == IterateDirection.FromNewToOld;
tikhomirov@516: // changeHistory points to A
tikhomirov@516: // Already reported B(m), B(m-1)...B(0), B(0) is in junctionNode
tikhomirov@516: // Shall connect histories A(k).bind(B(0))
tikhomirov@516: // if followAncestry: A(k) is latest in changeHistory (k == n)
tikhomirov@516: HistoryNode junctionSrc = findJunctionPointInCurrentChunk(curRename.second()); // A(k)
tikhomirov@516: junctionSrc.bindChild(junctionNode);
tikhomirov@516: if (renameHandler != null) {
tikhomirov@516: HgFileRevision copiedFrom = new HgFileRevision(curRename.first(), junctionSrc.fileRevision, null); // "A", A(k)
tikhomirov@516: HgFileRevision copiedTo = new HgFileRevision(prevRename.first(), junctionNode.fileRevision, copiedFrom.getPath()); // "B", B(0)
tikhomirov@516: renameHandler.copy(copiedFrom, copiedTo);
tikhomirov@516: }
tikhomirov@516: }
tikhomirov@516: }
tikhomirov@516:
tikhomirov@516: private HistoryNode findJunctionPointInCurrentChunk(Nodeid fileRevision) {
tikhomirov@516: if (followAncestry) {
tikhomirov@516: // use the fact we don't go past junction point when followAncestry == true
tikhomirov@516: HistoryNode rv = changeHistory.get(changeHistory.size() - 1);
tikhomirov@516: assert rv.fileRevision.equals(fileRevision);
tikhomirov@516: return rv;
tikhomirov@516: }
tikhomirov@516: for (HistoryNode n : changeHistory) {
tikhomirov@516: if (n.fileRevision.equals(fileRevision)) {
tikhomirov@516: return n;
tikhomirov@516: }
tikhomirov@516: }
tikhomirov@516: int csetStart = changeHistory.get(0).changeset;
tikhomirov@516: int csetEnd = changeHistory.get(changeHistory.size() - 1).changeset;
tikhomirov@516: throw new HgInvalidStateException(String.format("For change history (cset[%d..%d]) could not find node for file change %s", csetStart, csetEnd, fileRevision.shortNotation()));
tikhomirov@516: }
tikhomirov@516:
tikhomirov@516: protected abstract void once(HistoryNode n) throws HgCallbackTargetException, CancelledException;
tikhomirov@516:
tikhomirov@516: public void dispatchAllChanges() throws HgCallbackTargetException, CancelledException {
tikhomirov@516: // XXX shall sort changeHistory according to changeset numbers?
tikhomirov@516: Iterator it;
tikhomirov@516: if (iterateDirection == IterateDirection.FromOldToNew) {
tikhomirov@516: it = changeHistory.listIterator();
tikhomirov@516: } else {
tikhomirov@516: assert iterateDirection == IterateDirection.FromNewToOld;
tikhomirov@516: it = new ReverseIterator(changeHistory);
tikhomirov@516: }
tikhomirov@516: while(it.hasNext()) {
tikhomirov@516: HistoryNode n = it.next();
tikhomirov@516: once(n);
tikhomirov@516: progress.worked(1);
tikhomirov@516: }
tikhomirov@516: changeHistory = null;
tikhomirov@516: }
tikhomirov@516:
tikhomirov@516: public void switchTo(HgDataFile df) {
tikhomirov@516: // from now on, use df in TreeElement
tikhomirov@516: currentFileNode = df;
tikhomirov@516: }
tikhomirov@516: }
tikhomirov@516:
tikhomirov@507:
tikhomirov@64: //
tikhomirov@64:
tikhomirov@520: private class FilteringInspector extends AdapterPlug implements HgChangelog.Inspector, Adaptable {
tikhomirov@518:
tikhomirov@518: private int firstCset = BAD_REVISION, lastCset = BAD_REVISION;
tikhomirov@520: private HgChangelog.Inspector delegate;
tikhomirov@520: // we use lifecycle to stop when limit is reached.
tikhomirov@520: // delegate, however, may use lifecycle, too, so give it a chance
tikhomirov@520: private LifecycleProxy lifecycleProxy;
tikhomirov@518:
tikhomirov@518: // limit to changesets in this range only
tikhomirov@518: public void changesets(int start, int end) {
tikhomirov@518: firstCset = start;
tikhomirov@518: lastCset = end;
tikhomirov@64: }
tikhomirov@520:
tikhomirov@520: public void delegateTo(HgChangelog.Inspector inspector) {
tikhomirov@520: delegate = inspector;
tikhomirov@520: // let delegate control life cycle, too
tikhomirov@520: if (lifecycleProxy == null) {
tikhomirov@520: super.attachAdapter(Lifecycle.class, lifecycleProxy = new LifecycleProxy(inspector));
tikhomirov@520: } else {
tikhomirov@520: lifecycleProxy.init(inspector);
tikhomirov@520: }
tikhomirov@520: }
tikhomirov@518:
tikhomirov@518: public void next(int revisionNumber, Nodeid nodeid, RawChangeset cset) {
tikhomirov@518: if (limit > 0 && count >= limit) {
tikhomirov@518: return;
tikhomirov@518: }
tikhomirov@518: // XXX may benefit from optional interface with #isInterested(int csetRev) - to avoid
tikhomirov@518: // RawChangeset instantiation
tikhomirov@518: if (firstCset != BAD_REVISION && revisionNumber < firstCset) {
tikhomirov@518: return;
tikhomirov@518: }
tikhomirov@518: if (lastCset != BAD_REVISION && revisionNumber > lastCset) {
tikhomirov@518: return;
tikhomirov@518: }
tikhomirov@518: if (branches != null && !branches.contains(cset.branch())) {
tikhomirov@518: return;
tikhomirov@518: }
tikhomirov@518: if (users != null) {
tikhomirov@518: String csetUser = cset.user().toLowerCase();
tikhomirov@518: boolean found = false;
tikhomirov@518: for (String u : users) {
tikhomirov@518: if (csetUser.indexOf(u) != -1) {
tikhomirov@518: found = true;
tikhomirov@518: break;
tikhomirov@518: }
tikhomirov@518: }
tikhomirov@518: if (!found) {
tikhomirov@518: return;
tikhomirov@64: }
tikhomirov@64: }
tikhomirov@518: if (date != null) {
tikhomirov@518: // TODO post-1.0 implement date support for log
tikhomirov@518: }
tikhomirov@520: delegate.next(revisionNumber, nodeid, cset);
tikhomirov@520: count++;
tikhomirov@520: if (limit > 0 && count >= limit) {
tikhomirov@520: lifecycleProxy.stop();
tikhomirov@64: }
tikhomirov@64: }
tikhomirov@520: }
tikhomirov@518:
tikhomirov@432: private HgParentChildMap getParentHelper(boolean create) throws HgInvalidControlFileException {
tikhomirov@328: if (parentHelper == null && create) {
tikhomirov@432: parentHelper = new HgParentChildMap(repo.getChangelog());
tikhomirov@195: parentHelper.init();
tikhomirov@195: }
tikhomirov@195: return parentHelper;
tikhomirov@195: }
tikhomirov@520:
tikhomirov@205: public static class CollectHandler implements HgChangesetHandler {
tikhomirov@129: private final List result = new LinkedList();
tikhomirov@64:
tikhomirov@129: public List getChanges() {
tikhomirov@64: return Collections.unmodifiableList(result);
tikhomirov@64: }
tikhomirov@64:
tikhomirov@427: public void cset(HgChangeset changeset) {
tikhomirov@64: result.add(changeset.clone());
tikhomirov@64: }
tikhomirov@64: }
tikhomirov@328:
tikhomirov@328: private static class HistoryNode {
tikhomirov@328: final int changeset;
tikhomirov@328: final Nodeid fileRevision;
tikhomirov@509: HistoryNode parent1; // there's special case when we can alter it, see #bindChild()
tikhomirov@509: final HistoryNode parent2;
tikhomirov@328: List children;
tikhomirov@328:
tikhomirov@328: HistoryNode(int cs, Nodeid revision, HistoryNode p1, HistoryNode p2) {
tikhomirov@328: changeset = cs;
tikhomirov@328: fileRevision = revision;
tikhomirov@328: parent1 = p1;
tikhomirov@328: parent2 = p2;
tikhomirov@328: if (p1 != null) {
tikhomirov@328: p1.addChild(this);
tikhomirov@328: }
tikhomirov@328: if (p2 != null) {
tikhomirov@328: p2.addChild(this);
tikhomirov@328: }
tikhomirov@328: }
tikhomirov@328:
tikhomirov@509: private void addChild(HistoryNode child) {
tikhomirov@328: if (children == null) {
tikhomirov@328: children = new ArrayList(2);
tikhomirov@328: }
tikhomirov@328: children.add(child);
tikhomirov@328: }
tikhomirov@509:
tikhomirov@509: /**
tikhomirov@509: * method to merge two history chunks for renamed file so that
tikhomirov@517: * this node's history continues (or forks, if we don't followAncestry)
tikhomirov@517: * with that of child
tikhomirov@509: * @param child
tikhomirov@509: */
tikhomirov@509: public void bindChild(HistoryNode child) {
tikhomirov@509: assert child.parent1 == null && child.parent2 == null;
tikhomirov@509: child.parent1 = this;
tikhomirov@509: addChild(child);
tikhomirov@509: }
tikhomirov@511:
tikhomirov@511: public String toString() {
tikhomirov@511: return String.format("", changeset, parent1 == null ? "-" : String.valueOf(parent1.changeset), parent2 == null ? "-" : String.valueOf(parent2.changeset));
tikhomirov@511: }
tikhomirov@328: }
tikhomirov@328:
tikhomirov@328: private class ElementImpl implements HgChangesetTreeHandler.TreeElement, HgChangelog.Inspector {
tikhomirov@328: private HistoryNode historyNode;
tikhomirov@515: private HgDataFile fileNode;
tikhomirov@328: private Pair parents;
tikhomirov@328: private List children;
tikhomirov@328: private IntMap cachedChangesets;
tikhomirov@328: private ChangesetTransformer.Transformation transform;
tikhomirov@328: private Nodeid changesetRevision;
tikhomirov@328: private Pair parentRevisions;
tikhomirov@328: private List childRevisions;
tikhomirov@328:
tikhomirov@328: public ElementImpl(int total) {
tikhomirov@328: cachedChangesets = new IntMap(total);
tikhomirov@328: }
tikhomirov@328:
tikhomirov@515: ElementImpl init(HistoryNode n, HgDataFile df) {
tikhomirov@328: historyNode = n;
tikhomirov@515: fileNode = df;
tikhomirov@328: parents = null;
tikhomirov@328: children = null;
tikhomirov@328: changesetRevision = null;
tikhomirov@328: parentRevisions = null;
tikhomirov@328: childRevisions = null;
tikhomirov@328: return this;
tikhomirov@328: }
tikhomirov@328:
tikhomirov@328: public Nodeid fileRevision() {
tikhomirov@328: return historyNode.fileRevision;
tikhomirov@328: }
tikhomirov@515:
tikhomirov@515: public HgDataFile file() {
tikhomirov@515: return fileNode;
tikhomirov@515: }
tikhomirov@328:
tikhomirov@423: public HgChangeset changeset() {
tikhomirov@328: return get(historyNode.changeset)[0];
tikhomirov@328: }
tikhomirov@328:
tikhomirov@423: public Pair parents() {
tikhomirov@328: if (parents != null) {
tikhomirov@328: return parents;
tikhomirov@328: }
tikhomirov@328: HistoryNode p;
tikhomirov@328: final int p1, p2;
tikhomirov@328: if ((p = historyNode.parent1) != null) {
tikhomirov@328: p1 = p.changeset;
tikhomirov@328: } else {
tikhomirov@328: p1 = -1;
tikhomirov@328: }
tikhomirov@328: if ((p = historyNode.parent2) != null) {
tikhomirov@328: p2 = p.changeset;
tikhomirov@328: } else {
tikhomirov@328: p2 = -1;
tikhomirov@328: }
tikhomirov@328: HgChangeset[] r = get(p1, p2);
tikhomirov@328: return parents = new Pair(r[0], r[1]);
tikhomirov@328: }
tikhomirov@328:
tikhomirov@423: public Collection children() {
tikhomirov@328: if (children != null) {
tikhomirov@328: return children;
tikhomirov@328: }
tikhomirov@328: if (historyNode.children == null) {
tikhomirov@328: children = Collections.emptyList();
tikhomirov@328: } else {
tikhomirov@328: int[] childrentChangesetNumbers = new int[historyNode.children.size()];
tikhomirov@328: int j = 0;
tikhomirov@328: for (HistoryNode hn : historyNode.children) {
tikhomirov@328: childrentChangesetNumbers[j++] = hn.changeset;
tikhomirov@328: }
tikhomirov@328: children = Arrays.asList(get(childrentChangesetNumbers));
tikhomirov@328: }
tikhomirov@328: return children;
tikhomirov@328: }
tikhomirov@328:
tikhomirov@328: void populate(HgChangeset cs) {
tikhomirov@403: cachedChangesets.put(cs.getRevisionIndex(), cs);
tikhomirov@328: }
tikhomirov@328:
tikhomirov@423: private HgChangeset[] get(int... changelogRevisionIndex) {
tikhomirov@403: HgChangeset[] rv = new HgChangeset[changelogRevisionIndex.length];
tikhomirov@403: IntVector misses = new IntVector(changelogRevisionIndex.length, -1);
tikhomirov@403: for (int i = 0; i < changelogRevisionIndex.length; i++) {
tikhomirov@403: if (changelogRevisionIndex[i] == -1) {
tikhomirov@328: rv[i] = null;
tikhomirov@328: continue;
tikhomirov@328: }
tikhomirov@403: HgChangeset cached = cachedChangesets.get(changelogRevisionIndex[i]);
tikhomirov@328: if (cached != null) {
tikhomirov@328: rv[i] = cached;
tikhomirov@328: } else {
tikhomirov@403: misses.add(changelogRevisionIndex[i]);
tikhomirov@328: }
tikhomirov@328: }
tikhomirov@328: if (misses.size() > 0) {
tikhomirov@328: final int[] changesets2read = misses.toArray();
tikhomirov@328: initTransform();
tikhomirov@328: repo.getChangelog().range(this, changesets2read);
tikhomirov@328: for (int changeset2read : changesets2read) {
tikhomirov@328: HgChangeset cs = cachedChangesets.get(changeset2read);
tikhomirov@403: if (cs == null) {
tikhomirov@423: HgInvalidStateException t = new HgInvalidStateException(String.format("Can't get changeset for revision %d", changeset2read));
tikhomirov@423: throw t.setRevisionIndex(changeset2read);
tikhomirov@403: }
tikhomirov@403: // HgChangelog.range may reorder changesets according to their order in the changelog
tikhomirov@403: // thus need to find original index
tikhomirov@403: boolean sanity = false;
tikhomirov@403: for (int i = 0; i < changelogRevisionIndex.length; i++) {
tikhomirov@403: if (changelogRevisionIndex[i] == cs.getRevisionIndex()) {
tikhomirov@403: rv[i] = cs;
tikhomirov@403: sanity = true;
tikhomirov@403: break;
tikhomirov@328: }
tikhomirov@403: }
tikhomirov@403: if (!sanity) {
tikhomirov@490: repo.getSessionContext().getLog().dump(getClass(), Error, "Index of revision %d:%s doesn't match any of requested", cs.getRevisionIndex(), cs.getNodeid().shortNotation());
tikhomirov@403: }
tikhomirov@403: assert sanity;
tikhomirov@328: }
tikhomirov@328: }
tikhomirov@328: return rv;
tikhomirov@328: }
tikhomirov@328:
tikhomirov@328: // init only when needed
tikhomirov@423: void initTransform() throws HgRuntimeException {
tikhomirov@328: if (transform == null) {
tikhomirov@328: transform = new ChangesetTransformer.Transformation(new HgStatusCollector(repo)/*XXX try to reuse from context?*/, getParentHelper(false));
tikhomirov@328: }
tikhomirov@328: }
tikhomirov@328:
tikhomirov@328: public void next(int revisionNumber, Nodeid nodeid, RawChangeset cset) {
tikhomirov@328: HgChangeset cs = transform.handle(revisionNumber, nodeid, cset);
tikhomirov@328: populate(cs.clone());
tikhomirov@328: }
tikhomirov@328:
tikhomirov@423: public Nodeid changesetRevision() {
tikhomirov@328: if (changesetRevision == null) {
tikhomirov@328: changesetRevision = getRevision(historyNode.changeset);
tikhomirov@328: }
tikhomirov@328: return changesetRevision;
tikhomirov@328: }
tikhomirov@328:
tikhomirov@423: public Pair parentRevisions() {
tikhomirov@328: if (parentRevisions == null) {
tikhomirov@328: HistoryNode p;
tikhomirov@328: final Nodeid p1, p2;
tikhomirov@328: if ((p = historyNode.parent1) != null) {
tikhomirov@328: p1 = getRevision(p.changeset);
tikhomirov@328: } else {
tikhomirov@328: p1 = Nodeid.NULL;;
tikhomirov@328: }
tikhomirov@328: if ((p = historyNode.parent2) != null) {
tikhomirov@328: p2 = getRevision(p.changeset);
tikhomirov@328: } else {
tikhomirov@328: p2 = Nodeid.NULL;
tikhomirov@328: }
tikhomirov@328: parentRevisions = new Pair(p1, p2);
tikhomirov@328: }
tikhomirov@328: return parentRevisions;
tikhomirov@328: }
tikhomirov@328:
tikhomirov@423: public Collection childRevisions() {
tikhomirov@328: if (childRevisions != null) {
tikhomirov@328: return childRevisions;
tikhomirov@328: }
tikhomirov@328: if (historyNode.children == null) {
tikhomirov@328: childRevisions = Collections.emptyList();
tikhomirov@328: } else {
tikhomirov@328: ArrayList rv = new ArrayList(historyNode.children.size());
tikhomirov@328: for (HistoryNode hn : historyNode.children) {
tikhomirov@328: rv.add(getRevision(hn.changeset));
tikhomirov@328: }
tikhomirov@328: childRevisions = Collections.unmodifiableList(rv);
tikhomirov@328: }
tikhomirov@328: return childRevisions;
tikhomirov@328: }
tikhomirov@328:
tikhomirov@328: // reading nodeid involves reading index only, guess, can afford not to optimize multiple reads
tikhomirov@423: private Nodeid getRevision(int changelogRevisionNumber) {
tikhomirov@423: // TODO post-1.0 pipe through pool
tikhomirov@328: HgChangeset cs = cachedChangesets.get(changelogRevisionNumber);
tikhomirov@328: if (cs != null) {
tikhomirov@328: return cs.getNodeid();
tikhomirov@328: } else {
tikhomirov@403: return repo.getChangelog().getRevision(changelogRevisionNumber);
tikhomirov@328: }
tikhomirov@328: }
tikhomirov@328: }
tikhomirov@510:
tikhomirov@510: private enum IterateDirection {
tikhomirov@510: FromOldToNew, FromNewToOld
tikhomirov@510: }
tikhomirov@64: }