Compute a digest of each report file, and nuke reports that have the same
digest as a file already added to the index.  This a workaround for the
analyzer analyzing the same file more than once, which can happen when building
multiple targets.


git-svn-id: https://llvm.org/svn/llvm-project/cfe/trunk@49903 91177308-0d34-0410-b5e6-96231b3b80d8
diff --git a/utils/scan-build b/utils/scan-build
index f6d894b..441f63f 100755
--- a/utils/scan-build
+++ b/utils/scan-build
@@ -116,15 +116,46 @@
 }
 
 ##----------------------------------------------------------------------------##
+# ComputeDigest - Compute a digest of the specified file.
+##----------------------------------------------------------------------------##
+
+sub ComputeDigest {
+  my $FName = shift;
+  die "Cannot read $FName" if (! -r $FName);  
+  my $Result = `sha1sum -b $FName`;
+  my @Output = split /\s+/,$Result;
+  die "Bad output from sha1sum" if (scalar(@Output) != 2);
+  return $Output[0];
+}
+
+##----------------------------------------------------------------------------##
 # ScanFile - Scan a report file for various identifying attributes.
 ##----------------------------------------------------------------------------##
 
+# Sometimes a source file is scanned more than once, and thus produces
+# multiple error reports.  We use a cache to solve this problem.
+
+my %AlreadyScanned;
+
 sub ScanFile {
   
   my $Index = shift;
   my $Dir = shift;
   my $FName = shift;
   
+  # Compute a digest for the report file.  Determine if we have already
+  # scanned a file that looks just like it.
+  
+  my $digest = ComputeDigest("$Dir/$FName");
+
+  if (defined($AlreadyScanned{$digest})) {
+    # Redundant file.  Remove it.
+    `rm -f $Dir/$FName`;
+    return;
+  }
+  
+  $AlreadyScanned{$digest} = 1;
+  
   open(IN, "$Dir/$FName") or die "$Prog: Cannot open '$Dir/$FName'\n";
 
   my $BugDesc = "";