123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334 |
- package plumbing
- import (
- "log"
- "sort"
- "unicode/utf8"
- "github.com/sergi/go-diff/diffmatchpatch"
- "gopkg.in/src-d/go-git.v4"
- "gopkg.in/src-d/go-git.v4/plumbing"
- "gopkg.in/src-d/go-git.v4/plumbing/object"
- "gopkg.in/src-d/go-git.v4/utils/merkletrie"
- "gopkg.in/src-d/hercules.v5/internal"
- "gopkg.in/src-d/hercules.v5/internal/core"
- )
- // RenameAnalysis improves TreeDiff's results by searching for changed blobs under different
- // paths which are likely to be the result of a rename with subsequent edits.
- // RenameAnalysis is a PipelineItem.
- type RenameAnalysis struct {
- core.NoopMerger
- // SimilarityThreshold adjusts the heuristic to determine file renames.
- // It has the same units as cgit's -X rename-threshold or -M. Better to
- // set it to the default value of 80 (80%).
- SimilarityThreshold int
- repository *git.Repository
- }
- const (
- // RenameAnalysisDefaultThreshold specifies the default percentage of common lines in a pair
- // of files to consider them linked. The exact code of the decision is sizesAreClose().
- // CGit's default is 50%. Ours is 80% because 50% can be too computationally expensive.
- RenameAnalysisDefaultThreshold = 80
- // ConfigRenameAnalysisSimilarityThreshold is the name of the configuration option
- // (RenameAnalysis.Configure()) which sets the similarity threshold.
- ConfigRenameAnalysisSimilarityThreshold = "RenameAnalysis.SimilarityThreshold"
- )
- // Name of this PipelineItem. Uniquely identifies the type, used for mapping keys, etc.
- func (ra *RenameAnalysis) Name() string {
- return "RenameAnalysis"
- }
- // Provides returns the list of names of entities which are produced by this PipelineItem.
- // Each produced entity will be inserted into `deps` of dependent Consume()-s according
- // to this list. Also used by core.Registry to build the global map of providers.
- func (ra *RenameAnalysis) Provides() []string {
- arr := [...]string{DependencyTreeChanges}
- return arr[:]
- }
- // Requires returns the list of names of entities which are needed by this PipelineItem.
- // Each requested entity will be inserted into `deps` of Consume(). In turn, those
- // entities are Provides() upstream.
- func (ra *RenameAnalysis) Requires() []string {
- arr := [...]string{DependencyBlobCache, DependencyTreeChanges}
- return arr[:]
- }
- // ListConfigurationOptions returns the list of changeable public properties of this PipelineItem.
- func (ra *RenameAnalysis) ListConfigurationOptions() []core.ConfigurationOption {
- options := [...]core.ConfigurationOption{{
- Name: ConfigRenameAnalysisSimilarityThreshold,
- Description: "The threshold on the similarity index used to detect renames.",
- Flag: "M",
- Type: core.IntConfigurationOption,
- Default: RenameAnalysisDefaultThreshold},
- }
- return options[:]
- }
- // Configure sets the properties previously published by ListConfigurationOptions().
- func (ra *RenameAnalysis) Configure(facts map[string]interface{}) {
- if val, exists := facts[ConfigRenameAnalysisSimilarityThreshold].(int); exists {
- ra.SimilarityThreshold = val
- }
- }
- // Initialize resets the temporary caches and prepares this PipelineItem for a series of Consume()
- // calls. The repository which is going to be analysed is supplied as an argument.
- func (ra *RenameAnalysis) Initialize(repository *git.Repository) {
- if ra.SimilarityThreshold < 0 || ra.SimilarityThreshold > 100 {
- log.Printf("Warning: adjusted the similarity threshold to %d\n",
- RenameAnalysisDefaultThreshold)
- ra.SimilarityThreshold = RenameAnalysisDefaultThreshold
- }
- ra.repository = repository
- }
- // Consume runs this PipelineItem on the next commit data.
- // `deps` contain all the results from upstream PipelineItem-s as requested by Requires().
- // Additionally, DependencyCommit is always present there and represents the analysed *object.Commit.
- // This function returns the mapping with analysis results. The keys must be the same as
- // in Provides(). If there was an error, nil is returned.
- func (ra *RenameAnalysis) Consume(deps map[string]interface{}) (map[string]interface{}, error) {
- changes := deps[DependencyTreeChanges].(object.Changes)
- cache := deps[DependencyBlobCache].(map[plumbing.Hash]*CachedBlob)
- reducedChanges := make(object.Changes, 0, changes.Len())
- // Stage 1 - find renames by matching the hashes
- // n log(n)
- // We sort additions and deletions by hash and then do the single scan along
- // both slices.
- deleted := make(sortableChanges, 0, changes.Len())
- added := make(sortableChanges, 0, changes.Len())
- for _, change := range changes {
- action, err := change.Action()
- if err != nil {
- return nil, err
- }
- switch action {
- case merkletrie.Insert:
- added = append(added, sortableChange{change, change.To.TreeEntry.Hash})
- case merkletrie.Delete:
- deleted = append(deleted, sortableChange{change, change.From.TreeEntry.Hash})
- case merkletrie.Modify:
- reducedChanges = append(reducedChanges, change)
- }
- }
- sort.Sort(deleted)
- sort.Sort(added)
- a := 0
- d := 0
- stillDeleted := make(object.Changes, 0, deleted.Len())
- stillAdded := make(object.Changes, 0, added.Len())
- for a < added.Len() && d < deleted.Len() {
- if added[a].hash == deleted[d].hash {
- reducedChanges = append(
- reducedChanges,
- &object.Change{From: deleted[d].change.From, To: added[a].change.To})
- a++
- d++
- } else if added[a].Less(&deleted[d]) {
- stillAdded = append(stillAdded, added[a].change)
- a++
- } else {
- stillDeleted = append(stillDeleted, deleted[d].change)
- d++
- }
- }
- for ; a < added.Len(); a++ {
- stillAdded = append(stillAdded, added[a].change)
- }
- for ; d < deleted.Len(); d++ {
- stillDeleted = append(stillDeleted, deleted[d].change)
- }
- // Stage 2 - apply the similarity threshold
- // n^2 but actually linear
- // We sort the blobs by size and do the single linear scan.
- addedBlobs := make(sortableBlobs, 0, stillAdded.Len())
- deletedBlobs := make(sortableBlobs, 0, stillDeleted.Len())
- for _, change := range stillAdded {
- blob := cache[change.To.TreeEntry.Hash]
- addedBlobs = append(
- addedBlobs, sortableBlob{change: change, size: blob.Size})
- }
- for _, change := range stillDeleted {
- blob := cache[change.From.TreeEntry.Hash]
- deletedBlobs = append(
- deletedBlobs, sortableBlob{change: change, size: blob.Size})
- }
- sort.Sort(addedBlobs)
- sort.Sort(deletedBlobs)
- dStart := 0
- for a = 0; a < addedBlobs.Len(); a++ {
- myBlob := cache[addedBlobs[a].change.To.TreeEntry.Hash]
- mySize := addedBlobs[a].size
- for d = dStart; d < deletedBlobs.Len() && !ra.sizesAreClose(mySize, deletedBlobs[d].size); d++ {
- }
- dStart = d
- foundMatch := false
- for d = dStart; d < deletedBlobs.Len() && ra.sizesAreClose(mySize, deletedBlobs[d].size); d++ {
- blobsAreClose, err := ra.blobsAreClose(
- myBlob, cache[deletedBlobs[d].change.From.TreeEntry.Hash])
- if err != nil {
- return nil, err
- }
- if blobsAreClose {
- foundMatch = true
- reducedChanges = append(
- reducedChanges,
- &object.Change{From: deletedBlobs[d].change.From,
- To: addedBlobs[a].change.To})
- break
- }
- }
- if foundMatch {
- addedBlobs = append(addedBlobs[:a], addedBlobs[a+1:]...)
- a--
- deletedBlobs = append(deletedBlobs[:d], deletedBlobs[d+1:]...)
- }
- }
- // Stage 3 - we give up, everything left are independent additions and deletions
- for _, blob := range addedBlobs {
- reducedChanges = append(reducedChanges, blob.change)
- }
- for _, blob := range deletedBlobs {
- reducedChanges = append(reducedChanges, blob.change)
- }
- return map[string]interface{}{DependencyTreeChanges: reducedChanges}, nil
- }
- // Fork clones this PipelineItem.
- func (ra *RenameAnalysis) Fork(n int) []core.PipelineItem {
- return core.ForkSamePipelineItem(ra, n)
- }
- func (ra *RenameAnalysis) sizesAreClose(size1 int64, size2 int64) bool {
- size := internal.Max64(1, internal.Max64(size1, size2))
- return (internal.Abs64(size1-size2)*100)/size <= int64(100-ra.SimilarityThreshold)
- }
- func (ra *RenameAnalysis) blobsAreClose(
- blob1 *CachedBlob, blob2 *CachedBlob) (bool, error) {
- src, dst := string(blob1.Data), string(blob2.Data)
- maxSize := internal.Max(1, internal.Max(utf8.RuneCountInString(src), utf8.RuneCountInString(dst)))
- // compute the line-by-line diff, then the char-level diffs of the del-ins blocks
- // yes, this algorithm is greedy and not exact
- dmp := diffmatchpatch.New()
- srcLines, dstLines, lines := dmp.DiffLinesToRunes(src, dst)
- diffs := dmp.DiffMainRunes(srcLines, dstLines, false)
- var common, posSrc, prevPosSrc, posDst int
- possibleDelInsBlock := false
- for _, edit := range diffs {
- switch edit.Type {
- case diffmatchpatch.DiffDelete:
- possibleDelInsBlock = true
- prevPosSrc = posSrc
- for _, lineno := range edit.Text {
- posSrc += len(lines[lineno])
- }
- case diffmatchpatch.DiffInsert:
- nextPosDst := posDst
- for _, lineno := range edit.Text {
- nextPosDst += len(lines[lineno])
- }
- if possibleDelInsBlock {
- possibleDelInsBlock = false
- localDmp := diffmatchpatch.New()
- localSrc := src[prevPosSrc:posSrc]
- localDst := dst[posDst:nextPosDst]
- localDiffs := localDmp.DiffMainRunes([]rune(localSrc), []rune(localDst), false)
- for _, localEdit := range localDiffs {
- if localEdit.Type == diffmatchpatch.DiffEqual {
- common += utf8.RuneCountInString(localEdit.Text)
- }
- }
- }
- posDst = nextPosDst
- case diffmatchpatch.DiffEqual:
- possibleDelInsBlock = false
- for _, lineno := range edit.Text {
- common += utf8.RuneCountInString(lines[lineno])
- step := len(lines[lineno])
- posSrc += step
- posDst += step
- }
- }
- // supposing that the rest of the lines are the same (they are not - too optimistic),
- // estimate the maximum similarity and exit the loop if it lower than our threshold
- maxCommon := common + internal.Min(
- utf8.RuneCountInString(src[posSrc:]),
- utf8.RuneCountInString(dst[posDst:]))
- similarity := (maxCommon * 100) / maxSize
- if similarity < ra.SimilarityThreshold {
- return false, nil
- }
- }
- // the very last "overly optimistic" estimate was actually precise, so since we are still here
- // the blobs are similar
- return true, nil
- }
- type sortableChange struct {
- change *object.Change
- hash plumbing.Hash
- }
- type sortableChanges []sortableChange
- func (change *sortableChange) Less(other *sortableChange) bool {
- for x := 0; x < 20; x++ {
- if change.hash[x] < other.hash[x] {
- return true
- }
- }
- return false
- }
- func (slice sortableChanges) Len() int {
- return len(slice)
- }
- func (slice sortableChanges) Less(i, j int) bool {
- return slice[i].Less(&slice[j])
- }
- func (slice sortableChanges) Swap(i, j int) {
- slice[i], slice[j] = slice[j], slice[i]
- }
- type sortableBlob struct {
- change *object.Change
- size int64
- }
- type sortableBlobs []sortableBlob
- func (change *sortableBlob) Less(other *sortableBlob) bool {
- return change.size < other.size
- }
- func (slice sortableBlobs) Len() int {
- return len(slice)
- }
- func (slice sortableBlobs) Less(i, j int) bool {
- return slice[i].Less(&slice[j])
- }
- func (slice sortableBlobs) Swap(i, j int) {
- slice[i], slice[j] = slice[j], slice[i]
- }
- func init() {
- core.Registry.Register(&RenameAnalysis{})
- }
|