# HG changeset patch # User Artem Tikhomirov # Date 1365617094 -7200 # Node ID c4fd1037bc6f6b0ed591ff293c6dea97e9a36de8 # Parent 8ed4f4f4f0a6c5f7f94e3ffe732de2f4e3acaea7 Support for copy/rename follow/no-follow for annotate diff -r 8ed4f4f4f0a6 -r c4fd1037bc6f cmdline/org/tmatesoft/hg/console/Annotate.java --- a/cmdline/org/tmatesoft/hg/console/Annotate.java Wed Apr 10 15:45:53 2013 +0200 +++ b/cmdline/org/tmatesoft/hg/console/Annotate.java Wed Apr 10 20:04:54 2013 +0200 @@ -16,14 +16,15 @@ */ package org.tmatesoft.hg.console; +import static java.util.Arrays.asList; import static org.tmatesoft.hg.repo.HgRepository.TIP; -import java.util.Collections; +import java.util.TreeSet; import org.tmatesoft.hg.core.HgAnnotateCommand; import org.tmatesoft.hg.core.HgAnnotateCommand.LineInfo; import org.tmatesoft.hg.core.HgRepoFacade; -import org.tmatesoft.hg.repo.HgDataFile; +import org.tmatesoft.hg.util.Path; /** * @@ -33,7 +34,8 @@ public class Annotate { public static void main(String[] args) throws Exception { - Options cmdLineOpts = Options.parse(args, Collections.emptySet()); + String[] boolOpts = new String[] {"-l", "--line-number" }; + Options cmdLineOpts = Options.parse(args, new TreeSet(asList(boolOpts))); HgRepoFacade repo = new HgRepoFacade(); if (!repo.init(cmdLineOpts.findRepository())) { System.err.printf("Can't find repository in: %s\n", repo.getRepository().getLocation()); @@ -44,8 +46,7 @@ AnnotateDumpInspector insp = new AnnotateDumpInspector(cmdLineOpts.getBoolean(false, "-l", "--line-number")); cmd.changeset(rev); for (String fname : cmdLineOpts.getList("")) { - HgDataFile fn = repo.getRepository().getFileNode(fname); - cmd.file(fn); + cmd.file(Path.create(fname)); cmd.execute(insp); } } @@ -59,9 +60,9 @@ public void next(LineInfo lineInfo) { if (lineNumbers) { - System.out.printf("%3d:%3d:%s", lineInfo.getChangesetIndex(), lineInfo.getLineNumber(), new String(lineInfo.getContent())); + System.out.printf("%3d:%3d: %s", lineInfo.getChangesetIndex(), lineInfo.getLineNumber(), new String(lineInfo.getContent())); } else { - System.out.printf("%3d:%s", lineInfo.getChangesetIndex(), new String(lineInfo.getContent())); + System.out.printf("%3d: %s", lineInfo.getChangesetIndex(), new String(lineInfo.getContent())); } } } diff -r 8ed4f4f4f0a6 -r c4fd1037bc6f src/org/tmatesoft/hg/core/HgAnnotateCommand.java --- a/src/org/tmatesoft/hg/core/HgAnnotateCommand.java Wed Apr 10 15:45:53 2013 +0200 +++ b/src/org/tmatesoft/hg/core/HgAnnotateCommand.java Wed Apr 10 20:04:54 2013 +0200 @@ -27,9 +27,11 @@ import org.tmatesoft.hg.internal.FileAnnotation.LineDescriptor; import org.tmatesoft.hg.internal.FileAnnotation.LineInspector; import org.tmatesoft.hg.repo.HgBlameFacility.BlockData; +import org.tmatesoft.hg.repo.HgBlameFacility; import org.tmatesoft.hg.repo.HgDataFile; import org.tmatesoft.hg.repo.HgRepository; import org.tmatesoft.hg.util.CancelledException; +import org.tmatesoft.hg.util.Path; /** * WORK IN PROGRESS. UNSTABLE API @@ -44,7 +46,8 @@ private final HgRepository repo; private final CsetParamKeeper annotateRevision; - private HgDataFile file; + private Path file; + private boolean followRename; public HgAnnotateCommand(HgRepository hgRepo) { repo = hgRepo; @@ -62,15 +65,31 @@ return this; } - public HgAnnotateCommand file(HgDataFile fileNode) { - file = fileNode; + /** + * Select file to annotate, origin of renamed/copied file would be followed, too. + * + * @param filePath path relative to repository root + * @return this for convenience + */ + public HgAnnotateCommand file(Path filePath) { + return file(filePath, true); + } + + /** + * Select file to annotate. + * + * @param filePath path relative to repository root + * @param followCopyRename true to follow copies/renames. + * @return this for convenience + */ + public HgAnnotateCommand file(Path filePath, boolean followCopyRename) { + file = filePath; + followRename = followCopyRename; return this; } // TODO [1.1] set encoding and provide String line content from LineInfo - // FIXME [1.1] follow and no-follow parameters - public void execute(Inspector inspector) throws HgException, HgCallbackTargetException, CancelledException { if (inspector == null) { throw new IllegalArgumentException(); @@ -78,8 +97,15 @@ if (file == null) { throw new HgBadArgumentException("Command needs file argument", null); } + HgDataFile df = repo.getFileNode(file); + if (!df.exists()) { + return; + } + final int changesetStart = followRename ? 0 : df.getChangesetRevisionIndex(0); Collector c = new Collector(); - FileAnnotation.annotate(file, annotateRevision.get(), c); + FileAnnotation fa = new FileAnnotation(c); + HgBlameFacility af = new HgBlameFacility(df); + af.annotate(changesetStart, annotateRevision.get(), fa, HgIterateDirection.NewToOld); LineImpl li = new LineImpl(); for (int i = 0; i < c.lineRevisions.length; i++) { li.init(i+1, c.lineRevisions[i], c.line(i)); @@ -92,9 +118,9 @@ */ @Callback public interface Inspector { - // start(FileDescriptor); - void next(LineInfo lineInfo); - // end(FileDescriptor); + // start(FileDescriptor) throws HgCallbackTargetException; + void next(LineInfo lineInfo) throws HgCallbackTargetException; + // end(FileDescriptor) throws HgCallbackTargetException; } /** diff -r 8ed4f4f4f0a6 -r c4fd1037bc6f src/org/tmatesoft/hg/internal/BlameHelper.java --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/src/org/tmatesoft/hg/internal/BlameHelper.java Wed Apr 10 20:04:54 2013 +0200 @@ -0,0 +1,757 @@ +/* + * Copyright (c) 2013 TMate Software Ltd + * + * This program is free software; you can redistribute it and/or modify + * it under the terms of the GNU General Public License as published by + * the Free Software Foundation; version 2 of the License. + * + * This program is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the + * GNU General Public License for more details. + * + * For information on how to redistribute this software under + * the terms of a license other than GNU General Public License + * contact TMate Software at support@hg4j.com + */ +package org.tmatesoft.hg.internal; + +import static org.tmatesoft.hg.repo.HgRepository.NO_REVISION; + +import java.util.LinkedList; +import java.util.ListIterator; + +import org.tmatesoft.hg.core.HgCallbackTargetException; +import org.tmatesoft.hg.internal.DiffHelper.LineSequence; +import org.tmatesoft.hg.internal.DiffHelper.LineSequence.ByteChain; +import org.tmatesoft.hg.repo.HgBlameFacility.Block; +import org.tmatesoft.hg.repo.HgBlameFacility.BlockData; +import org.tmatesoft.hg.repo.HgBlameFacility.ChangeBlock; +import org.tmatesoft.hg.repo.HgBlameFacility.EqualBlock; +import org.tmatesoft.hg.repo.HgBlameFacility.Inspector; +import org.tmatesoft.hg.repo.HgBlameFacility.RevisionDescriptor; +import org.tmatesoft.hg.repo.HgBlameFacility.RevisionDescriptor.Recipient; +import org.tmatesoft.hg.repo.HgBlameFacility; +import org.tmatesoft.hg.repo.HgDataFile; +import org.tmatesoft.hg.repo.HgInvalidStateException; +import org.tmatesoft.hg.util.Adaptable; +import org.tmatesoft.hg.util.CancelledException; +import org.tmatesoft.hg.util.Pair; + +/** + * Blame implementation + * @see HgBlameFacility + * @author Artem Tikhomirov + * @author TMate Software Ltd. + */ +public class BlameHelper { + + private final Inspector insp; + private FileLinesCache linesCache; + + // FIXME exposing internals (use of FileLinesCache through cons arg and #useFileUpTo) smells bad, refactor! + + public BlameHelper(Inspector inspector, int cacheHint) { + insp = inspector; + linesCache = new FileLinesCache(cacheHint); + } + + public void useFileUpTo(HgDataFile df, int clogRevIndex) { + linesCache.useFileUpTo(df, clogRevIndex); + } + + // NO_REVISION is not allowed as any argument + public void diff(int fileRevIndex1, int clogRevIndex1, int fileRevIndex2, int clogRevIndex2) throws HgCallbackTargetException { + HgDataFile targetFile = linesCache.getFile(clogRevIndex2); + LineSequence c1 = linesCache.lines(clogRevIndex1, fileRevIndex1); + LineSequence c2 = linesCache.lines(clogRevIndex2, fileRevIndex2); + DiffHelper pg = new DiffHelper(); + pg.init(c1, c2); + BlameBlockInspector bbi = new BlameBlockInspector(targetFile, fileRevIndex2, insp, clogRevIndex1, clogRevIndex2); + pg.findMatchingBlocks(bbi); + bbi.checkErrors(); + } + + public void annotateChange(int fileRevIndex, int csetRevIndex, int[] fileParentRevs, int[] fileParentClogRevs) throws HgCallbackTargetException { + HgDataFile targetFile = linesCache.getFile(csetRevIndex); + final LineSequence fileRevLines = linesCache.lines(csetRevIndex, fileRevIndex); + if (fileParentClogRevs[0] != NO_REVISION && fileParentClogRevs[1] != NO_REVISION) { + int p1ClogIndex = fileParentClogRevs[0]; + int p2ClogIndex = fileParentClogRevs[1]; + LineSequence p1Lines = linesCache.lines(p1ClogIndex, fileParentRevs[0]); + LineSequence p2Lines = linesCache.lines(p2ClogIndex, fileParentRevs[1]); + DiffHelper pg = new DiffHelper(); + pg.init(p2Lines, fileRevLines); + EqualBlocksCollector p2MergeCommon = new EqualBlocksCollector(); + pg.findMatchingBlocks(p2MergeCommon); + // + pg.init(p1Lines); + BlameBlockInspector bbi = new BlameBlockInspector(targetFile, fileRevIndex, insp, p1ClogIndex, csetRevIndex); + bbi.setMergeParent2(p2MergeCommon, p2ClogIndex); + pg.findMatchingBlocks(bbi); + bbi.checkErrors(); + } else if (fileParentClogRevs[0] == fileParentClogRevs[1]) { + // may be equal iff both are unset + assert fileParentClogRevs[0] == NO_REVISION; + // everything added + BlameBlockInspector bbi = new BlameBlockInspector(targetFile, fileRevIndex, insp, NO_REVISION, csetRevIndex); + bbi.begin(LineSequence.newlines(new byte[0]), fileRevLines); + bbi.match(0, fileRevLines.chunkCount()-1, 0); + bbi.end(); + bbi.checkErrors(); + } else { + int soleParentIndex = fileParentClogRevs[0] == NO_REVISION ? 1 : 0; + assert fileParentClogRevs[soleParentIndex] != NO_REVISION; + LineSequence parentLines = linesCache.lines(fileParentClogRevs[soleParentIndex], fileParentRevs[soleParentIndex]); + + DiffHelper pg = new DiffHelper(); + pg.init(parentLines, fileRevLines); + BlameBlockInspector bbi = new BlameBlockInspector(targetFile, fileRevIndex, insp, fileParentClogRevs[soleParentIndex], csetRevIndex); + pg.findMatchingBlocks(bbi); + bbi.checkErrors(); + } + } + + private static class FileLinesCache { + private final LinkedList> lruCache; + private final int limit; + private final LinkedList> files; // TODO in fact, need sparse array + + public FileLinesCache(int lruLimit) { + limit = lruLimit; + lruCache = new LinkedList>(); + files = new LinkedList>(); + } + + public void useFileUpTo(HgDataFile df, int clogRevIndex) { + Pair newEntry = new Pair(clogRevIndex, df); + for (ListIterator> it = files.listIterator(); it.hasNext();) { + Pair e = it.next(); + if (e.first() == clogRevIndex) { + assert e.second().getPath().equals(df.getPath()); + return; + } + if (e.first() > clogRevIndex) { + // insert new entry before current + it.previous(); + it.add(newEntry); + return; + } + } + files.add(newEntry); + } + + public HgDataFile getFile(int clogRevIndex) { + for (Pair e : files) { + if (e.first() >= clogRevIndex) { + return e.second(); + } + } + throw new HgInvalidStateException(String.format("Got %d file-changelog mappings, but no luck for revision %d.", files.size(), clogRevIndex)); + } + + public LineSequence lines(int clogRevIndex, int fileRevIndex) { + Pair cached = checkCache(clogRevIndex); + if (cached != null) { + return cached.second(); + } + HgDataFile df = getFile(clogRevIndex); + try { + ByteArrayChannel c; + df.content(fileRevIndex, c = new ByteArrayChannel()); + LineSequence rv = LineSequence.newlines(c.toArray()); + lruCache.addFirst(new Pair(clogRevIndex, rv)); + if (lruCache.size() > limit) { + lruCache.removeLast(); + } + return rv; + } catch (CancelledException ex) { + // TODO likely it was bad idea to throw cancelled exception from content() + // deprecate and provide alternative? + HgInvalidStateException ise = new HgInvalidStateException("ByteArrayChannel never throws CancelledException"); + ise.initCause(ex); + throw ise; + } + } + + private Pair checkCache(int fileRevIndex) { + Pair rv = null; + for (ListIterator> it = lruCache.listIterator(); it.hasNext(); ) { + Pair p = it.next(); + if (p.first() == fileRevIndex) { + rv = p; + it.remove(); + break; + } + } + if (rv != null) { + lruCache.addFirst(rv); + } + return rv; + } + } + + private static class BlameBlockInspector extends DiffHelper.DeltaInspector { + private final Inspector insp; + private final int csetOrigin; + private final int csetTarget; + private EqualBlocksCollector p2MergeCommon; + private int csetMergeParent; + private IntVector mergeRanges; + private final AnnotateRev annotatedRevision; + private HgCallbackTargetException error; + + public BlameBlockInspector(HgDataFile df, int fileRevIndex, Inspector inspector, int originCset, int targetCset) { + assert inspector != null; + insp = inspector; + annotatedRevision = new AnnotateRev(); + annotatedRevision.set(df, fileRevIndex); + csetOrigin = originCset; + csetTarget = targetCset; + } + + public void setMergeParent2(EqualBlocksCollector p2Merge, int parentCset2) { + p2MergeCommon = p2Merge; + csetMergeParent = parentCset2; + mergeRanges = new IntVector(3*10, 3*10); + } + + @Override + public void begin(LineSequence s1, LineSequence s2) { + super.begin(s1, s2); + if (shallStop()) { + return; + } + ContentBlock originContent = new ContentBlock(s1); + ContentBlock targetContent = new ContentBlock(s2); + annotatedRevision.set(originContent, targetContent); + annotatedRevision.set(csetOrigin, csetTarget, p2MergeCommon != null ? csetMergeParent : NO_REVISION); + Recipient curious = Adaptable.Factory.getAdapter(insp, Recipient.class, null); + if (curious != null) { + try { + curious.start(annotatedRevision); + } catch (HgCallbackTargetException ex) { + error = ex; + } + } + } + + @Override + public void end() { + super.end(); + if (shallStop()) { + return; + } + Recipient curious = Adaptable.Factory.getAdapter(insp, Recipient.class, null); + if (curious != null) { + try { + curious.done(annotatedRevision); + } catch (HgCallbackTargetException ex) { + error = ex; + } + } + p2MergeCommon = null; + } + + @Override + protected void changed(int s1From, int s1To, int s2From, int s2To) { + if (shallStop()) { + return; + } + try { + if (p2MergeCommon != null) { + mergeRanges.clear(); + p2MergeCommon.combineAndMarkRangesWithTarget(s2From, s2To - s2From, csetOrigin, csetMergeParent, mergeRanges); + + /* + * Usecases, how it USED TO BE initially: + * 3 lines changed to 10 lines. range of 10 lines breaks down to 2 from p2, 3 from p1, and 5 from p2. + * We report: 2 lines changed to 2(p2), then 1 line changed with 3(p1) and 5 lines added from p2. + * + * 10 lines changed to 3 lines, range of 3 lines breaks down to 2 line from p1 and 1 line from p2. + * We report: 2 lines changed to 2(p1) and 8 lines changed to 1(p2) + * + * NOW, lines from p2 are always reported as pure add (since we need their insertion point to be in p2, not in p1) + * and we try to consume p1 changes as soon as we see first p1's range + */ + int s1TotalLines = s1To - s1From, s1ConsumedLines = 0, s1Start = s1From; + + for (int i = 0; i < mergeRanges.size(); i += 3) { + final int rangeOrigin = mergeRanges.get(i); + final int rangeStart = mergeRanges.get(i+1); + final int rangeLen = mergeRanges.get(i+2); + final boolean lastRange = i+3 >= mergeRanges.size(); + final int s1LinesLeft = s1TotalLines - s1ConsumedLines; + // how many lines we may report as changed (don't use more than in range unless it's the very last range) + final int s1LinesToBorrow = lastRange ? s1LinesLeft : Math.min(s1LinesLeft, rangeLen); + if (rangeOrigin != csetMergeParent && s1LinesToBorrow > 0) { + ChangeBlockImpl block = getChangeBlock(s1Start, s1LinesToBorrow, rangeStart, rangeLen); + block.setOriginAndTarget(rangeOrigin, csetTarget); + insp.changed(block); + s1ConsumedLines += s1LinesToBorrow; + s1Start += s1LinesToBorrow; + } else { + int blockInsPoint = rangeOrigin != csetMergeParent ? s1Start : p2MergeCommon.reverseMapLine(rangeStart); + ChangeBlockImpl block = getAddBlock(rangeStart, rangeLen, blockInsPoint); + block.setOriginAndTarget(rangeOrigin, csetTarget); + insp.added(block); + } + } + if (s1ConsumedLines != s1TotalLines) { + assert s1ConsumedLines < s1TotalLines : String.format("Expected to process %d lines, but actually was %d", s1TotalLines, s1ConsumedLines); + // either there were no ranges from p1, whole s2From..s2To range came from p2, shall report as deleted + // or the ranges found were not enough to consume whole s2From..s2To + // The "deletion point" is shifted to the end of last csetOrigin->csetTarget change + int s2DeletePoint = s2From + s1ConsumedLines; + ChangeBlockImpl block = new ChangeBlockImpl(annotatedRevision.origin, null, s1Start, s1To - s1Start, -1, -1, -1, s2DeletePoint); + block.setOriginAndTarget(csetOrigin, csetTarget); + insp.deleted(block); + } + } else { + ChangeBlockImpl block = getChangeBlock(s1From, s1To - s1From, s2From, s2To - s2From); + block.setOriginAndTarget(csetOrigin, csetTarget); + insp.changed(block); + } + } catch (HgCallbackTargetException ex) { + error = ex; + } + } + + @Override + protected void added(int s1InsertPoint, int s2From, int s2To) { + if (shallStop()) { + return; + } + try { + if (p2MergeCommon != null) { + mergeRanges.clear(); + p2MergeCommon.combineAndMarkRangesWithTarget(s2From, s2To - s2From, csetOrigin, csetMergeParent, mergeRanges); + int insPoint = s1InsertPoint; // track changes to insertion point + for (int i = 0; i < mergeRanges.size(); i += 3) { + int rangeOrigin = mergeRanges.get(i); + int rangeStart = mergeRanges.get(i+1); + int rangeLen = mergeRanges.get(i+2); + ChangeBlockImpl block = getAddBlock(rangeStart, rangeLen, insPoint); + block.setOriginAndTarget(rangeOrigin, csetTarget); + insp.added(block); + // indicate insPoint moved down number of lines we just reported + insPoint += rangeLen; + } + } else { + ChangeBlockImpl block = getAddBlock(s2From, s2To - s2From, s1InsertPoint); + block.setOriginAndTarget(csetOrigin, csetTarget); + insp.added(block); + } + } catch (HgCallbackTargetException ex) { + error = ex; + } + } + + @Override + protected void deleted(int s2DeletePoint, int s1From, int s1To) { + if (shallStop()) { + return; + } + try { + ChangeBlockImpl block = new ChangeBlockImpl(annotatedRevision.origin, null, s1From, s1To - s1From, -1, -1, -1, s2DeletePoint); + block.setOriginAndTarget(csetOrigin, csetTarget); + insp.deleted(block); + } catch (HgCallbackTargetException ex) { + error = ex; + } + } + + @Override + protected void unchanged(int s1From, int s2From, int length) { + if (shallStop()) { + return; + } + try { + EqualBlockImpl block = new EqualBlockImpl(s1From, s2From, length, annotatedRevision.target); + block.setOriginAndTarget(csetOrigin, csetTarget); + insp.same(block); + } catch (HgCallbackTargetException ex) { + error = ex; + } + } + + void checkErrors() throws HgCallbackTargetException { + if (error != null) { + throw error; + } + } + + private boolean shallStop() { + return error != null; + } + + private ChangeBlockImpl getAddBlock(int start, int len, int insPoint) { + return new ChangeBlockImpl(null, annotatedRevision.target, -1, -1, start, len, insPoint, -1); + } + + private ChangeBlockImpl getChangeBlock(int start1, int len1, int start2, int len2) { + return new ChangeBlockImpl(annotatedRevision.origin, annotatedRevision.target, start1, len1, start2, len2, start1, start2); + } + } + + private static class BlockImpl implements Block { + private int originCset; + private int targetCset; + + void setOriginAndTarget(int originChangesetIndex, int targetChangesetIndex) { + // XXX perhaps, shall be part of Inspector API, rather than Block's + // as they don't change between blocks (although the moment about merged revisions) + // is not yet clear to me + originCset = originChangesetIndex; + targetCset = targetChangesetIndex; + } + + public int originChangesetIndex() { + return originCset; + } + + public int targetChangesetIndex() { + return targetCset; + } + } + + private static class EqualBlockImpl extends BlockImpl implements EqualBlock { + private final int start1, start2; + private final int length; + private final ContentBlock fullContent; + private FilterBlock myContent; + + EqualBlockImpl(int blockStartSeq1, int blockStartSeq2, int blockLength, ContentBlock targetContent) { + start1 = blockStartSeq1; + start2 = blockStartSeq2; + length = blockLength; + fullContent = targetContent; + } + + public int originStart() { + return start1; + } + + public int targetStart() { + return start2; + } + + public int length() { + return length; + } + + public BlockData content() { + if (myContent == null) { + myContent = new FilterBlock(fullContent, start2, length); + } + return myContent; + } + + @Override + public String toString() { + return String.format("@@ [%d..%d) == [%d..%d) @@", start1, start1+length, start2, start2+length); + } + } + + private static class ChangeBlockImpl extends BlockImpl implements ChangeBlock { + private final ContentBlock oldContent; + private final ContentBlock newContent; + private final int s1Start; + private final int s1Len; + private final int s2Start; + private final int s2Len; + private final int s1InsertPoint; + private final int s2DeletePoint; + private FilterBlock addedBlock, removedBlock; + + public ChangeBlockImpl(ContentBlock c1, ContentBlock c2, int s1Start, int s1Len, int s2Start, int s2Len, int s1InsertPoint, int s2DeletePoint) { + oldContent = c1; + newContent = c2; + this.s1Start = s1Start; + this.s1Len = s1Len; + this.s2Start = s2Start; + this.s2Len = s2Len; + this.s1InsertPoint = s1InsertPoint; + this.s2DeletePoint = s2DeletePoint; + } + + public int insertedAt() { + return s1InsertPoint; + } + + public int firstAddedLine() { + return s2Start; + } + + public int totalAddedLines() { + return s2Len; + } + + public BlockData addedLines() { + if (addedBlock == null) { + addedBlock = new FilterBlock(newContent, firstAddedLine(), totalAddedLines()); + } + return addedBlock; + } + + public int removedAt() { + return s2DeletePoint; + } + + public int firstRemovedLine() { + return s1Start; + } + + public int totalRemovedLines() { + return s1Len; + } + + public BlockData removedLines() { + if (removedBlock == null) { + removedBlock = new FilterBlock(oldContent, firstRemovedLine(), totalRemovedLines()); + } + return removedBlock; + } + + @Override + public String toString() { + if (s2DeletePoint == -1) { + return String.format("@@ -%d,0 +%d,%d @@", insertedAt(), firstAddedLine(), totalAddedLines()); + } else if (s1InsertPoint == -1) { + // delete only + return String.format("@@ -%d,%d +%d,0 @@", firstRemovedLine(), totalRemovedLines(), removedAt()); + } + return String.format("@@ -%d,%d +%d,%d @@", firstRemovedLine(), totalRemovedLines(), firstAddedLine(), totalAddedLines()); + } + } + + private static class SingleLine implements BlockData { + private final ByteChain line; + + public SingleLine(ByteChain lineContent) { + line = lineContent; + } + + public BlockData elementAt(int index) { + assert false; + return null; + } + + public int elementCount() { + return 0; + } + + public byte[] asArray() { + return line.data(); + } + } + + private static class ContentBlock implements BlockData { + private final LineSequence seq; + + public ContentBlock(LineSequence sequence) { + seq = sequence; + } + + public BlockData elementAt(int index) { + return new SingleLine(seq.chunk(index)); + } + + public int elementCount() { + return seq.chunkCount() - 1; + } + + public byte[] asArray() { + return seq.data(0, seq.chunkCount() - 1); + } + } + + private static class FilterBlock implements BlockData { + private final ContentBlock contentBlock; + private final int from; + private final int length; + + public FilterBlock(ContentBlock bd, int startFrom, int len) { + assert bd != null; + assert startFrom + len < bd.seq.chunkCount(); // there's one extra chunk in the end, so strict less is ok + contentBlock = bd; + from = startFrom; + length = len; + } + + public BlockData elementAt(int index) { + if (index < 0 || index >= length) { + throw new IllegalArgumentException(String.format("Expected value from [0..%d), got %d", length, index)); + } + return contentBlock.elementAt(from + index); + } + + public int elementCount() { + return length; + } + + public byte[] asArray() { + return contentBlock.seq.data(from, from + length); + } + } + + + private static class EqualBlocksCollector implements DiffHelper.MatchInspector { + private final RangeSeq matches = new RangeSeq(); + + public void begin(LineSequence s1, LineSequence s2) { + } + + public void match(int startSeq1, int startSeq2, int matchLength) { + matches.add(startSeq1, startSeq2, matchLength); + } + + public void end() { + } + + public int reverseMapLine(int ln) { + return matches.reverseMapLine(ln); + } + + public void intersectWithTarget(int start, int length, IntVector result) { + int s = start; + for (int l = start, x = start + length; l < x; l++) { + if (!matches.includesTargetLine(l)) { + if (l - s > 0) { + result.add(s); + result.add(l - s); + } + s = l+1; + } + } + if (s < start+length) { + result.add(s); + result.add((start + length) - s); + } + } + + /* + * intersects [start..start+length) with ranges of target lines, and based on the intersection + * breaks initial range into smaller ranges and records them into result, with marker to indicate + * whether the range is from initial range (markerSource) or is a result of the intersection with target + * (markerTarget) + */ + public void combineAndMarkRangesWithTarget(int start, int length, int markerSource, int markerTarget, IntVector result) { + int sourceStart = start, targetStart = start, sourceEnd = start + length; + for (int l = sourceStart; l < sourceEnd; l++) { + if (matches.includesTargetLine(l)) { + // l is from target + if (sourceStart < l) { + // few lines from source range were not in the target, report them + result.add(markerSource); + result.add(sourceStart); + result.add(l - sourceStart); + } + // indicate the earliest line from source range to use + sourceStart = l + 1; + } else { + // l is not in target + if (targetStart < l) { + // report lines from target range + result.add(markerTarget); + result.add(targetStart); + result.add(l - targetStart); + } + // next line *may* be from target + targetStart = l + 1; + } + } + // if source range end with line from target, sourceStart would be == sourceEnd, and we need to add range with markerTarget + // if source range doesn't end with target line, targetStart == sourceEnd, while sourceStart < sourceEnd + if (sourceStart < sourceEnd) { + assert targetStart == sourceEnd; + // something left from the source range + result.add(markerSource); + result.add(sourceStart); + result.add(sourceEnd - sourceStart); + } else if (targetStart < sourceEnd) { + assert sourceStart == sourceEnd; + result.add(markerTarget); + result.add(targetStart); + result.add(sourceEnd - targetStart); + } + } + } + + private static class AnnotateRev implements RevisionDescriptor { + public ContentBlock origin, target; + public int originCset, targetCset, mergeCset, fileRevIndex; + public HgDataFile df; + + public void set(HgDataFile file, int fileRev) { + df = file; + fileRevIndex = fileRev; + } + public void set(ContentBlock o, ContentBlock t) { + origin = o; + target = t; + } + public void set(int o, int t, int m) { + originCset = o; + targetCset = t; + mergeCset = m; + } + + public BlockData origin() { + return origin; + } + + public BlockData target() { + return target; + } + + public int originChangesetIndex() { + return originCset; + } + + public int targetChangesetIndex() { + return targetCset; + } + + public boolean isMerge() { + return mergeCset != NO_REVISION; + } + + public int mergeChangesetIndex() { + return mergeCset; + } + + public int fileRevisionIndex() { + return fileRevIndex; + } + public HgDataFile file() { + return df; + } + @Override + public String toString() { + if (isMerge()) { + return String.format("[%d,%d->%d]", originCset, mergeCset, targetCset); + } + return String.format("[%d->%d]", originCset, targetCset); + } + } + + public static void main(String[] args) { + EqualBlocksCollector bc = new EqualBlocksCollector(); + bc.match(-1, 5, 3); + bc.match(-1, 10, 2); + bc.match(-1, 15, 3); + bc.match(-1, 20, 3); + IntVector r = new IntVector(); + bc.intersectWithTarget(7, 10, r); + for (int i = 0; i < r.size(); i+=2) { + System.out.printf("[%d..%d) ", r.get(i), r.get(i) + r.get(i+1)); + } + System.out.println(); + r.clear(); + bc.combineAndMarkRangesWithTarget(0, 16, 508, 514, r); + for (int i = 0; i < r.size(); i+=3) { + System.out.printf("%d:[%d..%d) ", r.get(i), r.get(i+1), r.get(i+1) + r.get(i+2)); + } + } +} diff -r 8ed4f4f4f0a6 -r c4fd1037bc6f src/org/tmatesoft/hg/repo/HgBlameFacility.java --- a/src/org/tmatesoft/hg/repo/HgBlameFacility.java Wed Apr 10 15:45:53 2013 +0200 +++ b/src/org/tmatesoft/hg/repo/HgBlameFacility.java Wed Apr 10 20:04:54 2013 +0200 @@ -16,30 +16,24 @@ */ package org.tmatesoft.hg.repo; -import static org.tmatesoft.hg.repo.HgRepository.NO_REVISION; -import static org.tmatesoft.hg.repo.HgRepository.TIP; +import static org.tmatesoft.hg.core.HgIterateDirection.NewToOld; +import static org.tmatesoft.hg.core.HgIterateDirection.OldToNew; +import static org.tmatesoft.hg.repo.HgInternals.wrongRevisionIndex; +import static org.tmatesoft.hg.repo.HgRepository.*; +import java.util.Arrays; import java.util.BitSet; +import java.util.Collections; import java.util.LinkedList; -import java.util.ListIterator; import org.tmatesoft.hg.core.HgCallbackTargetException; import org.tmatesoft.hg.core.HgIterateDirection; import org.tmatesoft.hg.core.Nodeid; -import org.tmatesoft.hg.internal.ByteArrayChannel; +import org.tmatesoft.hg.internal.BlameHelper; import org.tmatesoft.hg.internal.Callback; -import org.tmatesoft.hg.internal.DiffHelper; import org.tmatesoft.hg.internal.Experimental; -import org.tmatesoft.hg.internal.IntMap; import org.tmatesoft.hg.internal.IntVector; -import org.tmatesoft.hg.internal.Internals; -import org.tmatesoft.hg.internal.DiffHelper.LineSequence; -import org.tmatesoft.hg.internal.DiffHelper.LineSequence.ByteChain; -import org.tmatesoft.hg.internal.RangeSeq; -import org.tmatesoft.hg.repo.HgBlameFacility.RevisionDescriptor.Recipient; import org.tmatesoft.hg.util.Adaptable; -import org.tmatesoft.hg.util.CancelledException; -import org.tmatesoft.hg.util.Pair; /** * Facility with diff/annotate functionality. @@ -62,37 +56,95 @@ * mimic 'hg diff -r clogRevIndex1 -r clogRevIndex2' */ public void diff(int clogRevIndex1, int clogRevIndex2, Inspector insp) throws HgCallbackTargetException { + // FIXME clogRevIndex1 and clogRevIndex2 may point to different files, need to decide whether to throw an exception + // or to attempt to look up correct file node (tricky) int fileRevIndex1 = fileRevIndex(df, clogRevIndex1); int fileRevIndex2 = fileRevIndex(df, clogRevIndex2); - FileLinesCache fileInfoCache = new FileLinesCache(df, 5); - LineSequence c1 = fileInfoCache.lines(fileRevIndex1); - LineSequence c2 = fileInfoCache.lines(fileRevIndex2); - DiffHelper pg = new DiffHelper(); - pg.init(c1, c2); - BlameBlockInspector bbi = new BlameBlockInspector(fileRevIndex2, insp, clogRevIndex1, clogRevIndex2); - pg.findMatchingBlocks(bbi); - bbi.checkErrors(); + BlameHelper bh = new BlameHelper(insp, 5); + bh.useFileUpTo(df, clogRevIndex2); + bh.diff(fileRevIndex1, clogRevIndex1, fileRevIndex2, clogRevIndex2); } /** * Walk file history up/down to revision at given changeset and report changes for each revision */ public void annotate(int changelogRevisionIndex, Inspector insp, HgIterateDirection iterateOrder) throws HgCallbackTargetException { + annotate(0, changelogRevisionIndex, insp, iterateOrder); + } + + /** + * Walk file history range and report changes for each revision + */ + public void annotate(int changelogRevIndexStart, int changelogRevIndexEnd, Inspector insp, HgIterateDirection iterateOrder) throws HgCallbackTargetException { + if (wrongRevisionIndex(changelogRevIndexStart) || wrongRevisionIndex(changelogRevIndexEnd)) { + throw new IllegalArgumentException(); + } + // Note, changelogRevisionIndex may be TIP, while the code below doesn't tolerate constants + // + int lastRevision = df.getRepo().getChangelog().getLastRevision(); + if (changelogRevIndexEnd == TIP) { + changelogRevIndexEnd = lastRevision; + } + HgInternals.checkRevlogRange(changelogRevIndexStart, changelogRevIndexEnd, lastRevision); if (!df.exists()) { return; } - // Note, changelogRevisionIndex may be TIP, while #implAnnotateChange doesn't tolerate constants - // - FileRevisionHistoryChunk fileHistory = new FileRevisionHistoryChunk(df); - fileHistory.init(changelogRevisionIndex); -// fileHistory.linkTo(null); FIXME + BlameHelper bh = new BlameHelper(insp, 10); + HgDataFile currentFile = df; + int fileLastClogRevIndex = changelogRevIndexEnd; + FileRevisionHistoryChunk nextChunk = null; + LinkedList fileCompleteHistory = new LinkedList(); + do { + FileRevisionHistoryChunk fileHistory = new FileRevisionHistoryChunk(currentFile); + fileHistory.init(fileLastClogRevIndex); + fileHistory.linkTo(nextChunk); + fileCompleteHistory.addFirst(fileHistory); // to get the list in old-to-new order + nextChunk = fileHistory; + bh.useFileUpTo(currentFile, fileLastClogRevIndex); + if (currentFile.isCopy()) { + // TODO SessionContext.getPathFactory() and replace all Path.create + HgRepository repo = currentFile.getRepo(); + currentFile = repo.getFileNode(currentFile.getCopySourceName()); + fileLastClogRevIndex = repo.getChangelog().getRevisionIndex(currentFile.getCopySourceRevision()); + // XXX perhaps, shall fail with meaningful exception if new file doesn't exist (.i/.d not found for whatever reason) + // or source revision is missing? + } else { + currentFile = null; // stop iterating + } + } while (currentFile != null && fileLastClogRevIndex >= changelogRevIndexStart); + // fileCompleteHistory is in (origin, intermediate target, ultimate target) order - int[] fileRevParents = new int[2]; - FileLinesCache fileInfoCache = new FileLinesCache(df, 10); - for (int fri : fileHistory.fileRevisions(iterateOrder)) { - int clogRevIndex = df.getChangesetRevisionIndex(fri); - fileHistory.getParents(fri, fileRevParents); - implAnnotateChange(fileInfoCache, clogRevIndex, fri, fileRevParents, insp); + int[] fileClogParentRevs = new int[2]; + int[] fileParentRevs = new int[2]; + if (iterateOrder == NewToOld) { + Collections.reverse(fileCompleteHistory); + } + boolean shallFilterStart = changelogRevIndexStart != 0; // no reason if complete history is walked + for (FileRevisionHistoryChunk fileHistory : fileCompleteHistory) { + for (int fri : fileHistory.fileRevisions(iterateOrder)) { + int clogRevIndex = fileHistory.changeset(fri); + if (shallFilterStart) { + if (iterateOrder == NewToOld) { + // clogRevIndex decreases + if (clogRevIndex < changelogRevIndexStart) { + break; + } + // fall-through, clogRevIndex is in the [start..end] range + } else { // old to new + // the way we built fileHistory ensures we won't walk past changelogRevIndexEnd + // here we ensure we start from the right one, the one indicated with changelogRevIndexStart + if (clogRevIndex < changelogRevIndexStart) { + continue; + } else { + shallFilterStart = false; // once boundary is crossed, no need to check + // fall-through + } + } + } + fileHistory.fillFileParents(fri, fileParentRevs); + fileHistory.fillCsetParents(fri, fileClogParentRevs); + bh.annotateChange(fri, clogRevIndex, fileParentRevs, fileClogParentRevs); + } } } @@ -109,181 +161,12 @@ if (changelogRevisionIndex == TIP) { changelogRevisionIndex = df.getChangesetRevisionIndex(fileRevIndex); } - implAnnotateChange(new FileLinesCache(df, 5), changelogRevisionIndex, fileRevIndex, fileRevParents, insp); - } - - private void implAnnotateChange(FileLinesCache fl, int csetRevIndex, int fileRevIndex, int[] fileParentRevs, Inspector insp) throws HgCallbackTargetException { - final LineSequence fileRevLines = fl.lines(fileRevIndex); - if (fileParentRevs[0] != NO_REVISION && fileParentRevs[1] != NO_REVISION) { - LineSequence p1Lines = fl.lines(fileParentRevs[0]); - LineSequence p2Lines = fl.lines(fileParentRevs[1]); - int p1ClogIndex = fl.getChangesetRevisionIndex(fileParentRevs[0]); - int p2ClogIndex = fl.getChangesetRevisionIndex(fileParentRevs[1]); - DiffHelper pg = new DiffHelper(); - pg.init(p2Lines, fileRevLines); - EqualBlocksCollector p2MergeCommon = new EqualBlocksCollector(); - pg.findMatchingBlocks(p2MergeCommon); - // - pg.init(p1Lines); - BlameBlockInspector bbi = new BlameBlockInspector(fileRevIndex, insp, p1ClogIndex, csetRevIndex); - bbi.setMergeParent2(p2MergeCommon, p2ClogIndex); - pg.findMatchingBlocks(bbi); - bbi.checkErrors(); - } else if (fileParentRevs[0] == fileParentRevs[1]) { - // may be equal iff both are unset - assert fileParentRevs[0] == NO_REVISION; - // everything added - BlameBlockInspector bbi = new BlameBlockInspector(fileRevIndex, insp, NO_REVISION, csetRevIndex); - bbi.begin(LineSequence.newlines(new byte[0]), fileRevLines); - bbi.match(0, fileRevLines.chunkCount()-1, 0); - bbi.end(); - bbi.checkErrors(); - } else { - int soleParent = fileParentRevs[0] == NO_REVISION ? fileParentRevs[1] : fileParentRevs[0]; - assert soleParent != NO_REVISION; - LineSequence parentLines = fl.lines(soleParent); - - int parentChangesetRevIndex = fl.getChangesetRevisionIndex(soleParent); - DiffHelper pg = new DiffHelper(); - pg.init(parentLines, fileRevLines); - BlameBlockInspector bbi = new BlameBlockInspector(fileRevIndex, insp, parentChangesetRevIndex, csetRevIndex); - pg.findMatchingBlocks(bbi); - bbi.checkErrors(); - } - } - - private static int fileRevIndex(HgDataFile df, int csetRevIndex) { - Nodeid fileRev = df.getRepo().getManifest().getFileRevision(csetRevIndex, df.getPath()); - return df.getRevisionIndex(fileRev); - } - - private static class FileRevisionHistoryChunk { - private final HgDataFile df; - private IntVector fileRevsToVisit; - private IntVector fileParentRevs; - - public FileRevisionHistoryChunk(HgDataFile file) { - df = file; - } - - public void getParents(int fileRevIndex, int[] fileRevParents) { - fileRevParents[0] = fileParentRevs.get(fileRevIndex * 2); - fileRevParents[1] = fileParentRevs.get(fileRevIndex * 2 + 1); - } - - public void init (int changelogRevisionIndex) { - // XXX df.indexWalk(0, fileRevIndex, ) might be more effective - int fileRevIndex = fileRevIndex(df, changelogRevisionIndex); - int[] fileRevParents = new int[2]; - fileParentRevs = new IntVector((fileRevIndex+1) * 2, 0); - fileParentRevs.add(NO_REVISION, NO_REVISION); // parents of fileRevIndex == 0 - for (int i = 1; i <= fileRevIndex; i++) { - df.parents(i, fileRevParents, null, null); - fileParentRevs.add(fileRevParents[0], fileRevParents[1]); - } - fileRevsToVisit = new IntVector(fileRevIndex + 1, 0); - LinkedList queue = new LinkedList(); - BitSet seen = new BitSet(fileRevIndex + 1); - queue.add(fileRevIndex); - do { - int x = queue.removeFirst(); - if (seen.get(x)) { - continue; - } - seen.set(x); - fileRevsToVisit.add(x); - int p1 = fileParentRevs.get(2*x); - int p2 = fileParentRevs.get(2*x + 1); - if (p1 != NO_REVISION) { - queue.addLast(p1); - } - if (p2 != NO_REVISION) { - queue.addLast(p2); - } - } while (!queue.isEmpty()); - // make sure no child is processed before we handled all (grand-)parents of the element - fileRevsToVisit.sort(false); - // now fileRevsToVisit keep file change ancestry from new to old - } - - public void linkTo(FileRevisionHistoryChunk origin) { - Internals.notImplemented(); - } - - public int[] fileRevisions(HgIterateDirection iterateOrder) { - // fileRevsToVisit is { r10, r7, r6, r5, r0 }, new to old - int[] rv = fileRevsToVisit.toArray(); - if (iterateOrder == HgIterateDirection.OldToNew) { - // reverse return value - for (int a = 0, b = rv.length-1; a < b; a++, b--) { - int t = rv[b]; - rv[b] = rv[a]; - rv[a] = t; - } - } - return rv; - } - } - - private static class FileLinesCache { - private final HgDataFile df; - private final LinkedList> lruCache; - private final int limit; - private IntMap fileToClogIndexMap = new IntMap(20); - - public FileLinesCache(HgDataFile file, int lruLimit) { - df = file; - limit = lruLimit; - lruCache = new LinkedList>(); - } - - public int getChangesetRevisionIndex(int fileRevIndex) { - Integer cached = fileToClogIndexMap.get(fileRevIndex); - if (cached == null) { - cached = df.getChangesetRevisionIndex(fileRevIndex); - fileToClogIndexMap.put(fileRevIndex, cached); - } - return cached.intValue(); - } - - public LineSequence lines(int fileRevIndex) { - Pair cached = checkCache(fileRevIndex); - if (cached != null) { - return cached.second(); - } - try { - ByteArrayChannel c; - df.content(fileRevIndex, c = new ByteArrayChannel()); - LineSequence rv = LineSequence.newlines(c.toArray()); - lruCache.addFirst(new Pair(fileRevIndex, rv)); - if (lruCache.size() > limit) { - lruCache.removeLast(); - } - return rv; - } catch (CancelledException ex) { - // TODO likely it was bad idea to throw cancelled exception from content() - // deprecate and provide alternative? - HgInvalidStateException ise = new HgInvalidStateException("ByteArrayChannel never throws CancelledException"); - ise.initCause(ex); - throw ise; - } - } - - private Pair checkCache(int fileRevIndex) { - Pair rv = null; - for (ListIterator> it = lruCache.listIterator(); it.hasNext(); ) { - Pair p = it.next(); - if (p.first() == fileRevIndex) { - rv = p; - it.remove(); - break; - } - } - if (rv != null) { - lruCache.addFirst(rv); - } - return rv; - } + BlameHelper bh = new BlameHelper(insp, 5); + bh.useFileUpTo(df, changelogRevisionIndex); + int[] fileClogParentRevs = new int[2]; + fileClogParentRevs[0] = fileRevParents[0] == NO_REVISION ? NO_REVISION : df.getChangesetRevisionIndex(fileRevParents[0]); + fileClogParentRevs[1] = fileRevParents[1] == NO_REVISION ? NO_REVISION : df.getChangesetRevisionIndex(fileRevParents[1]); + bh.annotateChange(fileRevIndex, changelogRevisionIndex, fileRevParents, fileClogParentRevs); } /** @@ -373,6 +256,11 @@ int fileRevisionIndex(); /** + * @return file object under blame (target file) + */ + HgDataFile file(); + + /** * Implement to indicate interest in {@link RevisionDescriptor}. * * Note, instance of {@link RevisionDescriptor} is the same for @@ -447,563 +335,118 @@ } public interface ChangeBlock extends AddBlock, DeleteBlock { } - - private static class BlameBlockInspector extends DiffHelper.DeltaInspector { - private final Inspector insp; - private final int csetOrigin; - private final int csetTarget; - private EqualBlocksCollector p2MergeCommon; - private int csetMergeParent; - private IntVector mergeRanges; - private final AnnotateRev annotatedRevision; - private HgCallbackTargetException error; - - public BlameBlockInspector(int fileRevIndex, Inspector inspector, int originCset, int targetCset) { - assert inspector != null; - insp = inspector; - annotatedRevision = new AnnotateRev(); - annotatedRevision.set(fileRevIndex); - csetOrigin = originCset; - csetTarget = targetCset; - } - - public void setMergeParent2(EqualBlocksCollector p2Merge, int parentCset2) { - p2MergeCommon = p2Merge; - csetMergeParent = parentCset2; - mergeRanges = new IntVector(3*10, 3*10); - } - - @Override - public void begin(LineSequence s1, LineSequence s2) { - super.begin(s1, s2); - if (shallStop()) { - return; - } - ContentBlock originContent = new ContentBlock(s1); - ContentBlock targetContent = new ContentBlock(s2); - annotatedRevision.set(originContent, targetContent); - annotatedRevision.set(csetOrigin, csetTarget, p2MergeCommon != null ? csetMergeParent : NO_REVISION); - Recipient curious = Adaptable.Factory.getAdapter(insp, Recipient.class, null); - if (curious != null) { - try { - curious.start(annotatedRevision); - } catch (HgCallbackTargetException ex) { - error = ex; - } - } - } - - @Override - public void end() { - super.end(); - if (shallStop()) { - return; - } - Recipient curious = Adaptable.Factory.getAdapter(insp, Recipient.class, null); - if (curious != null) { - try { - curious.done(annotatedRevision); - } catch (HgCallbackTargetException ex) { - error = ex; - } - } - p2MergeCommon = null; - } - @Override - protected void changed(int s1From, int s1To, int s2From, int s2To) { - if (shallStop()) { - return; - } - try { - if (p2MergeCommon != null) { - mergeRanges.clear(); - p2MergeCommon.combineAndMarkRangesWithTarget(s2From, s2To - s2From, csetOrigin, csetMergeParent, mergeRanges); - - /* - * Usecases, how it USED TO BE initially: - * 3 lines changed to 10 lines. range of 10 lines breaks down to 2 from p2, 3 from p1, and 5 from p2. - * We report: 2 lines changed to 2(p2), then 1 line changed with 3(p1) and 5 lines added from p2. - * - * 10 lines changed to 3 lines, range of 3 lines breaks down to 2 line from p1 and 1 line from p2. - * We report: 2 lines changed to 2(p1) and 8 lines changed to 1(p2) - * - * NOW, lines from p2 are always reported as pure add (since we need their insertion point to be in p2, not in p1) - * and we try to consume p1 changes as soon as we see first p1's range - */ - int s1TotalLines = s1To - s1From, s1ConsumedLines = 0, s1Start = s1From; - - for (int i = 0; i < mergeRanges.size(); i += 3) { - final int rangeOrigin = mergeRanges.get(i); - final int rangeStart = mergeRanges.get(i+1); - final int rangeLen = mergeRanges.get(i+2); - final boolean lastRange = i+3 >= mergeRanges.size(); - final int s1LinesLeft = s1TotalLines - s1ConsumedLines; - // how many lines we may report as changed (don't use more than in range unless it's the very last range) - final int s1LinesToBorrow = lastRange ? s1LinesLeft : Math.min(s1LinesLeft, rangeLen); - if (rangeOrigin != csetMergeParent && s1LinesToBorrow > 0) { - ChangeBlockImpl block = getChangeBlock(s1Start, s1LinesToBorrow, rangeStart, rangeLen); - block.setOriginAndTarget(rangeOrigin, csetTarget); - insp.changed(block); - s1ConsumedLines += s1LinesToBorrow; - s1Start += s1LinesToBorrow; - } else { - int blockInsPoint = rangeOrigin != csetMergeParent ? s1Start : p2MergeCommon.reverseMapLine(rangeStart); - ChangeBlockImpl block = getAddBlock(rangeStart, rangeLen, blockInsPoint); - block.setOriginAndTarget(rangeOrigin, csetTarget); - insp.added(block); - } - } - if (s1ConsumedLines != s1TotalLines) { - assert s1ConsumedLines < s1TotalLines : String.format("Expected to process %d lines, but actually was %d", s1TotalLines, s1ConsumedLines); - // either there were no ranges from p1, whole s2From..s2To range came from p2, shall report as deleted - // or the ranges found were not enough to consume whole s2From..s2To - // The "deletion point" is shifted to the end of last csetOrigin->csetTarget change - int s2DeletePoint = s2From + s1ConsumedLines; - ChangeBlockImpl block = new ChangeBlockImpl(annotatedRevision.origin, null, s1Start, s1To - s1Start, -1, -1, -1, s2DeletePoint); - block.setOriginAndTarget(csetOrigin, csetTarget); - insp.deleted(block); - } - } else { - ChangeBlockImpl block = getChangeBlock(s1From, s1To - s1From, s2From, s2To - s2From); - block.setOriginAndTarget(csetOrigin, csetTarget); - insp.changed(block); - } - } catch (HgCallbackTargetException ex) { - error = ex; - } - } - - @Override - protected void added(int s1InsertPoint, int s2From, int s2To) { - if (shallStop()) { - return; - } - try { - if (p2MergeCommon != null) { - mergeRanges.clear(); - p2MergeCommon.combineAndMarkRangesWithTarget(s2From, s2To - s2From, csetOrigin, csetMergeParent, mergeRanges); - int insPoint = s1InsertPoint; // track changes to insertion point - for (int i = 0; i < mergeRanges.size(); i += 3) { - int rangeOrigin = mergeRanges.get(i); - int rangeStart = mergeRanges.get(i+1); - int rangeLen = mergeRanges.get(i+2); - ChangeBlockImpl block = getAddBlock(rangeStart, rangeLen, insPoint); - block.setOriginAndTarget(rangeOrigin, csetTarget); - insp.added(block); - // indicate insPoint moved down number of lines we just reported - insPoint += rangeLen; - } - } else { - ChangeBlockImpl block = getAddBlock(s2From, s2To - s2From, s1InsertPoint); - block.setOriginAndTarget(csetOrigin, csetTarget); - insp.added(block); - } - } catch (HgCallbackTargetException ex) { - error = ex; - } - } - - @Override - protected void deleted(int s2DeletePoint, int s1From, int s1To) { - if (shallStop()) { - return; - } - try { - ChangeBlockImpl block = new ChangeBlockImpl(annotatedRevision.origin, null, s1From, s1To - s1From, -1, -1, -1, s2DeletePoint); - block.setOriginAndTarget(csetOrigin, csetTarget); - insp.deleted(block); - } catch (HgCallbackTargetException ex) { - error = ex; - } - } - @Override - protected void unchanged(int s1From, int s2From, int length) { - if (shallStop()) { - return; - } - try { - EqualBlockImpl block = new EqualBlockImpl(s1From, s2From, length, annotatedRevision.target); - block.setOriginAndTarget(csetOrigin, csetTarget); - insp.same(block); - } catch (HgCallbackTargetException ex) { - error = ex; - } - } - - void checkErrors() throws HgCallbackTargetException { - if (error != null) { - throw error; - } - } - - private boolean shallStop() { - return error != null; - } - - private ChangeBlockImpl getAddBlock(int start, int len, int insPoint) { - return new ChangeBlockImpl(null, annotatedRevision.target, -1, -1, start, len, insPoint, -1); - } - - private ChangeBlockImpl getChangeBlock(int start1, int len1, int start2, int len2) { - return new ChangeBlockImpl(annotatedRevision.origin, annotatedRevision.target, start1, len1, start2, len2, start1, start2); - } - } - - private static class BlockImpl implements Block { - private int originCset; - private int targetCset; - - void setOriginAndTarget(int originChangesetIndex, int targetChangesetIndex) { - // XXX perhaps, shall be part of Inspector API, rather than Block's - // as they don't change between blocks (although the moment about merged revisions) - // is not yet clear to me - originCset = originChangesetIndex; - targetCset = targetChangesetIndex; - } - - public int originChangesetIndex() { - return originCset; - } - - public int targetChangesetIndex() { - return targetCset; - } - } - - private static class EqualBlockImpl extends BlockImpl implements EqualBlock { - private final int start1, start2; - private final int length; - private final ContentBlock fullContent; - private FilterBlock myContent; - - EqualBlockImpl(int blockStartSeq1, int blockStartSeq2, int blockLength, ContentBlock targetContent) { - start1 = blockStartSeq1; - start2 = blockStartSeq2; - length = blockLength; - fullContent = targetContent; - } - - public int originStart() { - return start1; - } - - public int targetStart() { - return start2; - } - - public int length() { - return length; - } - - public BlockData content() { - if (myContent == null) { - myContent = new FilterBlock(fullContent, start2, length); - } - return myContent; - } - - @Override - public String toString() { - return String.format("@@ [%d..%d) == [%d..%d) @@", start1, start1+length, start2, start2+length); - } + private static int fileRevIndex(HgDataFile df, int csetRevIndex) { + Nodeid fileRev = df.getRepo().getManifest().getFileRevision(csetRevIndex, df.getPath()); + return df.getRevisionIndex(fileRev); } - private static class ChangeBlockImpl extends BlockImpl implements ChangeBlock { - private final ContentBlock oldContent; - private final ContentBlock newContent; - private final int s1Start; - private final int s1Len; - private final int s2Start; - private final int s2Len; - private final int s1InsertPoint; - private final int s2DeletePoint; - private FilterBlock addedBlock, removedBlock; - - public ChangeBlockImpl(ContentBlock c1, ContentBlock c2, int s1Start, int s1Len, int s2Start, int s2Len, int s1InsertPoint, int s2DeletePoint) { - oldContent = c1; - newContent = c2; - this.s1Start = s1Start; - this.s1Len = s1Len; - this.s2Start = s2Start; - this.s2Len = s2Len; - this.s1InsertPoint = s1InsertPoint; - this.s2DeletePoint = s2DeletePoint; - } - - public int insertedAt() { - return s1InsertPoint; - } + private static class FileRevisionHistoryChunk { + private final HgDataFile df; + // change ancestry, sequence of file revisions + private IntVector fileRevsToVisit; + // parent pairs of complete file history + private IntVector fileParentRevs; + // map file revision to changelog revision (sparse array, only file revisions to visit are set) + private int[] file2changelog; + private int originChangelogRev = BAD_REVISION, originFileRev = BAD_REVISION; - public int firstAddedLine() { - return s2Start; - } - - public int totalAddedLines() { - return s2Len; - } - - public BlockData addedLines() { - if (addedBlock == null) { - addedBlock = new FilterBlock(newContent, firstAddedLine(), totalAddedLines()); - } - return addedBlock; - } - - public int removedAt() { - return s2DeletePoint; - } - - public int firstRemovedLine() { - return s1Start; - } - - public int totalRemovedLines() { - return s1Len; - } - - public BlockData removedLines() { - if (removedBlock == null) { - removedBlock = new FilterBlock(oldContent, firstRemovedLine(), totalRemovedLines()); - } - return removedBlock; + public FileRevisionHistoryChunk(HgDataFile file) { + df = file; } - @Override - public String toString() { - if (s2DeletePoint == -1) { - return String.format("@@ -%d,0 +%d,%d @@", insertedAt(), firstAddedLine(), totalAddedLines()); - } else if (s1InsertPoint == -1) { - // delete only - return String.format("@@ -%d,%d +%d,0 @@", firstRemovedLine(), totalRemovedLines(), removedAt()); + public void init(int changelogRevisionIndex) { + // XXX df.indexWalk(0, fileRevIndex, ) might be more effective + int fileRevIndex = fileRevIndex(df, changelogRevisionIndex); + int[] fileRevParents = new int[2]; + fileParentRevs = new IntVector((fileRevIndex+1) * 2, 0); + fileParentRevs.add(NO_REVISION, NO_REVISION); // parents of fileRevIndex == 0 + for (int i = 1; i <= fileRevIndex; i++) { + df.parents(i, fileRevParents, null, null); + fileParentRevs.add(fileRevParents[0], fileRevParents[1]); } - return String.format("@@ -%d,%d +%d,%d @@", firstRemovedLine(), totalRemovedLines(), firstAddedLine(), totalAddedLines()); - } - } - - private static class SingleLine implements BlockData { - private final ByteChain line; - - public SingleLine(ByteChain lineContent) { - line = lineContent; - } - - public BlockData elementAt(int index) { - assert false; - return null; - } - - public int elementCount() { - return 0; - } - - public byte[] asArray() { - return line.data(); - } - } - - private static class ContentBlock implements BlockData { - private final LineSequence seq; - - public ContentBlock(LineSequence sequence) { - seq = sequence; - } - - public BlockData elementAt(int index) { - return new SingleLine(seq.chunk(index)); - } - - public int elementCount() { - return seq.chunkCount() - 1; - } - - public byte[] asArray() { - return seq.data(0, seq.chunkCount() - 1); - } - } - - private static class FilterBlock implements BlockData { - private final ContentBlock contentBlock; - private final int from; - private final int length; - - public FilterBlock(ContentBlock bd, int startFrom, int len) { - assert bd != null; - assert startFrom + len < bd.seq.chunkCount(); // there's one extra chunk in the end, so strict less is ok - contentBlock = bd; - from = startFrom; - length = len; - } - - public BlockData elementAt(int index) { - if (index < 0 || index >= length) { - throw new IllegalArgumentException(String.format("Expected value from [0..%d), got %d", length, index)); - } - return contentBlock.elementAt(from + index); - } - - public int elementCount() { - return length; - } - - public byte[] asArray() { - return contentBlock.seq.data(from, from + length); - } - } - - - private static class EqualBlocksCollector implements DiffHelper.MatchInspector { - private final RangeSeq matches = new RangeSeq(); - - public void begin(LineSequence s1, LineSequence s2) { - } - - public void match(int startSeq1, int startSeq2, int matchLength) { - matches.add(startSeq1, startSeq2, matchLength); - } - - public void end() { - } - - public int reverseMapLine(int ln) { - return matches.reverseMapLine(ln); - } - - public void intersectWithTarget(int start, int length, IntVector result) { - int s = start; - for (int l = start, x = start + length; l < x; l++) { - if (!matches.includesTargetLine(l)) { - if (l - s > 0) { - result.add(s); - result.add(l - s); - } - s = l+1; + // fileRevsToVisit keep file change ancestry from new to old + fileRevsToVisit = new IntVector(fileRevIndex + 1, 0); + // keep map of file revision to changelog revision + file2changelog = new int[fileRevIndex+1]; + // only elements worth visit would get mapped, so there would be unfilled areas in the file2changelog, + // prevent from error (make it explicit) by bad value + Arrays.fill(file2changelog, BAD_REVISION); + LinkedList queue = new LinkedList(); + BitSet seen = new BitSet(fileRevIndex + 1); + queue.add(fileRevIndex); + do { + int x = queue.removeFirst(); + if (seen.get(x)) { + continue; } - } - if (s < start+length) { - result.add(s); - result.add((start + length) - s); - } + seen.set(x); + fileRevsToVisit.add(x); + file2changelog[x] = df.getChangesetRevisionIndex(x); + int p1 = fileParentRevs.get(2*x); + int p2 = fileParentRevs.get(2*x + 1); + if (p1 != NO_REVISION) { + queue.addLast(p1); + } + if (p2 != NO_REVISION) { + queue.addLast(p2); + } + } while (!queue.isEmpty()); + // make sure no child is processed before we handled all (grand-)parents of the element + fileRevsToVisit.sort(false); } - /* - * intersects [start..start+length) with ranges of target lines, and based on the intersection - * breaks initial range into smaller ranges and records them into result, with marker to indicate - * whether the range is from initial range (markerSource) or is a result of the intersection with target - * (markerTarget) - */ - public void combineAndMarkRangesWithTarget(int start, int length, int markerSource, int markerTarget, IntVector result) { - int sourceStart = start, targetStart = start, sourceEnd = start + length; - for (int l = sourceStart; l < sourceEnd; l++) { - if (matches.includesTargetLine(l)) { - // l is from target - if (sourceStart < l) { - // few lines from source range were not in the target, report them - result.add(markerSource); - result.add(sourceStart); - result.add(l - sourceStart); - } - // indicate the earliest line from source range to use - sourceStart = l + 1; - } else { - // l is not in target - if (targetStart < l) { - // report lines from target range - result.add(markerTarget); - result.add(targetStart); - result.add(l - targetStart); - } - // next line *may* be from target - targetStart = l + 1; + public void linkTo(FileRevisionHistoryChunk target) { + // assume that target.init() has been called already + if (target == null) { + return; + } + target.originFileRev = fileRevsToVisit.get(0); // files to visit are new to old + target.originChangelogRev = changeset(target.originFileRev); + } + + public int[] fileRevisions(HgIterateDirection iterateOrder) { + // fileRevsToVisit is { r10, r7, r6, r5, r0 }, new to old + int[] rv = fileRevsToVisit.toArray(); + if (iterateOrder == OldToNew) { + // reverse return value + for (int a = 0, b = rv.length-1; a < b; a++, b--) { + int t = rv[b]; + rv[b] = rv[a]; + rv[a] = t; } } - // if source range end with line from target, sourceStart would be == sourceEnd, and we need to add range with markerTarget - // if source range doesn't end with target line, targetStart == sourceEnd, while sourceStart < sourceEnd - if (sourceStart < sourceEnd) { - assert targetStart == sourceEnd; - // something left from the source range - result.add(markerSource); - result.add(sourceStart); - result.add(sourceEnd - sourceStart); - } else if (targetStart < sourceEnd) { - assert sourceStart == sourceEnd; - result.add(markerTarget); - result.add(targetStart); - result.add(sourceEnd - targetStart); - } - } - } - - private static class AnnotateRev implements RevisionDescriptor { - public ContentBlock origin, target; - public int originCset, targetCset, mergeCset, fileRevIndex; - - public void set(int fileRev) { - fileRevIndex = fileRev; - } - public void set(ContentBlock o, ContentBlock t) { - origin = o; - target = t; - } - public void set(int o, int t, int m) { - originCset = o; - targetCset = t; - mergeCset = m; + return rv; } - public BlockData origin() { - return origin; - } - - public BlockData target() { - return target; - } - - public int originChangesetIndex() { - return originCset; - } - - public int targetChangesetIndex() { - return targetCset; - } - - public boolean isMerge() { - return mergeCset != NO_REVISION; - } - - public int mergeChangesetIndex() { - return mergeCset; + public int changeset(int fileRevIndex) { + return file2changelog[fileRevIndex]; } - - public int fileRevisionIndex() { - return fileRevIndex; - } - @Override - public String toString() { - if (isMerge()) { - return String.format("[%d,%d->%d]", originCset, mergeCset, targetCset); + + public void fillFileParents(int fileRevIndex, int[] fileParents) { + if (fileRevIndex == 0 && originFileRev != BAD_REVISION) { + // this chunk continues another file + assert originFileRev != NO_REVISION; + fileParents[0] = originFileRev; + fileParents[1] = NO_REVISION; + return; } - return String.format("[%d->%d]", originCset, targetCset); + fileParents[0] = fileParentRevs.get(fileRevIndex * 2); + fileParents[1] = fileParentRevs.get(fileRevIndex * 2 + 1); } - } - - public static void main(String[] args) { - EqualBlocksCollector bc = new EqualBlocksCollector(); - bc.match(-1, 5, 3); - bc.match(-1, 10, 2); - bc.match(-1, 15, 3); - bc.match(-1, 20, 3); - IntVector r = new IntVector(); - bc.intersectWithTarget(7, 10, r); - for (int i = 0; i < r.size(); i+=2) { - System.out.printf("[%d..%d) ", r.get(i), r.get(i) + r.get(i+1)); - } - System.out.println(); - r.clear(); - bc.combineAndMarkRangesWithTarget(0, 16, 508, 514, r); - for (int i = 0; i < r.size(); i+=3) { - System.out.printf("%d:[%d..%d) ", r.get(i), r.get(i+1), r.get(i+1) + r.get(i+2)); + + public void fillCsetParents(int fileRevIndex, int[] csetParents) { + if (fileRevIndex == 0 && originFileRev != BAD_REVISION) { + assert originFileRev != NO_REVISION; + csetParents[0] = originChangelogRev; + csetParents[1] = NO_REVISION; // I wonder if possible to start a copy with two parents? + return; + } + int fp1 = fileParentRevs.get(fileRevIndex * 2); + int fp2 = fileParentRevs.get(fileRevIndex * 2 + 1); + csetParents[0] = fp1 == NO_REVISION ? NO_REVISION : changeset(fp1); + csetParents[1] = fp2 == NO_REVISION ? NO_REVISION : changeset(fp2); } } }