//===- FuzzerMerge.cpp - merging corpora ----------------------------------===// // // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. // See https://llvm.org/LICENSE.txt for license information. // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception // //===----------------------------------------------------------------------===// // Merging corpora. //===----------------------------------------------------------------------===// #include "FuzzerCommand.h" #include "FuzzerMerge.h" #include "FuzzerIO.h" #include "FuzzerInternal.h" #include "FuzzerTracePC.h" #include "FuzzerUtil.h" #include #include #include #include #include namespace fuzzer { bool Merger::Parse(const std::string &Str, bool ParseCoverage) { std::istringstream SS(Str); return Parse(SS, ParseCoverage); } void Merger::ParseOrExit(std::istream &IS, bool ParseCoverage) { if (!Parse(IS, ParseCoverage)) { Printf("MERGE: failed to parse the control file (unexpected error)\n"); exit(1); } } // The control file example: // // 3 # The number of inputs // 1 # The number of inputs in the first corpus, <= the previous number // file0 // file1 // file2 # One file name per line. // STARTED 0 123 # FileID, file size // FT 0 1 4 6 8 # FileID COV1 COV2 ... // COV 0 7 8 9 # FileID COV1 COV1 // STARTED 1 456 # If FT is missing, the input crashed while processing. // STARTED 2 567 // FT 2 8 9 // COV 2 11 12 bool Merger::Parse(std::istream &IS, bool ParseCoverage) { LastFailure.clear(); std::string Line; // Parse NumFiles. if (!std::getline(IS, Line, '\n')) return false; std::istringstream L1(Line); size_t NumFiles = 0; L1 >> NumFiles; if (NumFiles == 0 || NumFiles > 10000000) return false; // Parse NumFilesInFirstCorpus. if (!std::getline(IS, Line, '\n')) return false; std::istringstream L2(Line); NumFilesInFirstCorpus = NumFiles + 1; L2 >> NumFilesInFirstCorpus; if (NumFilesInFirstCorpus > NumFiles) return false; // Parse file names. Files.resize(NumFiles); for (size_t i = 0; i < NumFiles; i++) if (!std::getline(IS, Files[i].Name, '\n')) return false; // Parse STARTED, FT, and COV lines. size_t ExpectedStartMarker = 0; const size_t kInvalidStartMarker = -1; size_t LastSeenStartMarker = kInvalidStartMarker; std::vector TmpFeatures; std::set PCs; while (std::getline(IS, Line, '\n')) { std::istringstream ISS1(Line); std::string Marker; uint32_t N; if (!(ISS1 >> Marker) || !(ISS1 >> N)) return false; if (Marker == "STARTED") { // STARTED FILE_ID FILE_SIZE if (ExpectedStartMarker != N) return false; ISS1 >> Files[ExpectedStartMarker].Size; LastSeenStartMarker = ExpectedStartMarker; assert(ExpectedStartMarker < Files.size()); ExpectedStartMarker++; } else if (Marker == "FT") { // FT FILE_ID COV1 COV2 COV3 ... size_t CurrentFileIdx = N; if (CurrentFileIdx != LastSeenStartMarker) return false; LastSeenStartMarker = kInvalidStartMarker; if (ParseCoverage) { TmpFeatures.clear(); // use a vector from outer scope to avoid resizes. while (ISS1 >> N) TmpFeatures.push_back(N); std::sort(TmpFeatures.begin(), TmpFeatures.end()); Files[CurrentFileIdx].Features = TmpFeatures; } } else if (Marker == "COV") { size_t CurrentFileIdx = N; if (ParseCoverage) while (ISS1 >> N) if (PCs.insert(N).second) Files[CurrentFileIdx].Cov.push_back(N); } else { return false; } } if (LastSeenStartMarker != kInvalidStartMarker) LastFailure = Files[LastSeenStartMarker].Name; FirstNotProcessedFile = ExpectedStartMarker; return true; } size_t Merger::ApproximateMemoryConsumption() const { size_t Res = 0; for (const auto &F: Files) Res += sizeof(F) + F.Features.size() * sizeof(F.Features[0]); return Res; } // Decides which files need to be merged (add those to NewFiles). // Returns the number of new features added. size_t Merger::Merge(const std::set &InitialFeatures, std::set *NewFeatures, const std::set &InitialCov, std::set *NewCov, std::vector *NewFiles) { NewFiles->clear(); NewFeatures->clear(); NewCov->clear(); assert(NumFilesInFirstCorpus <= Files.size()); std::set AllFeatures = InitialFeatures; // What features are in the initial corpus? for (size_t i = 0; i < NumFilesInFirstCorpus; i++) { auto &Cur = Files[i].Features; AllFeatures.insert(Cur.begin(), Cur.end()); } // Remove all features that we already know from all other inputs. for (size_t i = NumFilesInFirstCorpus; i < Files.size(); i++) { auto &Cur = Files[i].Features; std::vector Tmp; std::set_difference(Cur.begin(), Cur.end(), AllFeatures.begin(), AllFeatures.end(), std::inserter(Tmp, Tmp.begin())); Cur.swap(Tmp); } // Sort. Give preference to // * smaller files // * files with more features. std::sort(Files.begin() + NumFilesInFirstCorpus, Files.end(), [&](const MergeFileInfo &a, const MergeFileInfo &b) -> bool { if (a.Size != b.Size) return a.Size < b.Size; return a.Features.size() > b.Features.size(); }); // One greedy pass: add the file's features to AllFeatures. // If new features were added, add this file to NewFiles. for (size_t i = NumFilesInFirstCorpus; i < Files.size(); i++) { auto &Cur = Files[i].Features; // Printf("%s -> sz %zd ft %zd\n", Files[i].Name.c_str(), // Files[i].Size, Cur.size()); bool FoundNewFeatures = false; for (auto Fe: Cur) { if (AllFeatures.insert(Fe).second) { FoundNewFeatures = true; NewFeatures->insert(Fe); } } if (FoundNewFeatures) NewFiles->push_back(Files[i].Name); for (auto Cov : Files[i].Cov) if (InitialCov.find(Cov) == InitialCov.end()) NewCov->insert(Cov); } return NewFeatures->size(); } std::set Merger::AllFeatures() const { std::set S; for (auto &File : Files) S.insert(File.Features.begin(), File.Features.end()); return S; } // Inner process. May crash if the target crashes. void Fuzzer::CrashResistantMergeInternalStep(const std::string &CFPath, bool IsSetCoverMerge) { Printf("MERGE-INNER: using the control file '%s'\n", CFPath.c_str()); Merger M; std::ifstream IF(CFPath); M.ParseOrExit(IF, false); IF.close(); if (!M.LastFailure.empty()) Printf("MERGE-INNER: '%s' caused a failure at the previous merge step\n", M.LastFailure.c_str()); Printf("MERGE-INNER: %zd total files;" " %zd processed earlier; will process %zd files now\n", M.Files.size(), M.FirstNotProcessedFile, M.Files.size() - M.FirstNotProcessedFile); std::ofstream OF(CFPath, std::ofstream::out | std::ofstream::app); std::set AllFeatures; auto PrintStatsWrapper = [this, &AllFeatures](const char* Where) { this->PrintStats(Where, "\n", 0, AllFeatures.size()); }; std::set AllPCs; for (size_t i = M.FirstNotProcessedFile; i < M.Files.size(); i++) { Fuzzer::MaybeExitGracefully(); auto U = FileToVector(M.Files[i].Name); if (U.size() > MaxInputLen) { U.resize(MaxInputLen); U.shrink_to_fit(); } // Write the pre-run marker. OF << "STARTED " << i << " " << U.size() << "\n"; OF.flush(); // Flush is important since Command::Execute may crash. // Run. TPC.ResetMaps(); ExecuteCallback(U.data(), U.size()); // Collect coverage. We are iterating over the files in this order: // * First, files in the initial corpus ordered by size, smallest first. // * Then, all other files, smallest first. std::set Features; if (IsSetCoverMerge) TPC.CollectFeatures([&](size_t Feature) { Features.insert(Feature); }); else TPC.CollectFeatures([&](size_t Feature) { if (AllFeatures.insert(Feature).second) Features.insert(Feature); }); TPC.UpdateObservedPCs(); // Show stats. if (!(TotalNumberOfRuns & (TotalNumberOfRuns - 1))) PrintStatsWrapper("pulse "); if (TotalNumberOfRuns == M.NumFilesInFirstCorpus) PrintStatsWrapper("LOADED"); // Write the post-run marker and the coverage. OF << "FT " << i; for (size_t F : Features) OF << " " << F; OF << "\n"; OF << "COV " << i; TPC.ForEachObservedPC([&](const TracePC::PCTableEntry *TE) { if (AllPCs.insert(TE).second) OF << " " << TPC.PCTableEntryIdx(TE); }); OF << "\n"; OF.flush(); } PrintStatsWrapper("DONE "); } // Merges all corpora into the first corpus. A file is added into // the first corpus only if it adds new features. Unlike `Merger::Merge`, // this implementation calculates an approximation of the minimum set // of corpora files, that cover all known features (set cover problem). // Generally, this means that files with more features are preferred for // merge into the first corpus. When two files have the same number of // features, the smaller one is preferred. size_t Merger::SetCoverMerge(const std::set &InitialFeatures, std::set *NewFeatures, const std::set &InitialCov, std::set *NewCov, std::vector *NewFiles) { assert(NumFilesInFirstCorpus <= Files.size()); NewFiles->clear(); NewFeatures->clear(); NewCov->clear(); std::set AllFeatures; // 1 << 21 - 1 is the maximum feature index. // See 'kFeatureSetSize' in 'FuzzerCorpus.h'. const uint32_t kFeatureSetSize = 1 << 21; std::vector Covered(kFeatureSetSize, false); size_t NumCovered = 0; std::set ExistingFeatures = InitialFeatures; for (size_t i = 0; i < NumFilesInFirstCorpus; ++i) ExistingFeatures.insert(Files[i].Features.begin(), Files[i].Features.end()); // Mark the existing features as covered. for (const auto &F : ExistingFeatures) { if (!Covered[F % kFeatureSetSize]) { ++NumCovered; Covered[F % kFeatureSetSize] = true; } // Calculate an underestimation of the set of covered features // since the `Covered` bitvector is smaller than the feature range. AllFeatures.insert(F % kFeatureSetSize); } std::set RemainingFiles; for (size_t i = NumFilesInFirstCorpus; i < Files.size(); ++i) { // Construct an incremental sequence which represent the // indices to all files (excluding those in the initial corpus). // RemainingFiles = range(NumFilesInFirstCorpus..Files.size()). RemainingFiles.insert(i); // Insert this file's unique features to all features. for (const auto &F : Files[i].Features) AllFeatures.insert(F % kFeatureSetSize); } // Integrate files into Covered until set is complete. while (NumCovered != AllFeatures.size()) { // Index to file with largest number of unique features. size_t MaxFeaturesIndex = NumFilesInFirstCorpus; // Indices to remove from RemainingFiles. std::set RemoveIndices; // Running max unique feature count. // Updated upon finding a file with more features. size_t MaxNumFeatures = 0; // Iterate over all files not yet integrated into Covered, // to find the file which has the largest number of // features that are not already in Covered. for (const auto &i : RemainingFiles) { const auto &File = Files[i]; size_t CurrentUnique = 0; // Count number of features in this file // which are not yet in Covered. for (const auto &F : File.Features) if (!Covered[F % kFeatureSetSize]) ++CurrentUnique; if (CurrentUnique == 0) { // All features in this file are already in Covered: skip next time. RemoveIndices.insert(i); } else if (CurrentUnique > MaxNumFeatures || (CurrentUnique == MaxNumFeatures && File.Size < Files[MaxFeaturesIndex].Size)) { // Update the max features file based on unique features // Break ties by selecting smaller files. MaxNumFeatures = CurrentUnique; MaxFeaturesIndex = i; } } // Must be a valid index/ assert(MaxFeaturesIndex < Files.size()); // Remove any feature-less files found. for (const auto &i : RemoveIndices) RemainingFiles.erase(i); if (MaxNumFeatures == 0) { // Did not find a file that adds unique features. // This means that we should have no remaining files. assert(RemainingFiles.size() == 0); assert(NumCovered == AllFeatures.size()); break; } // MaxFeaturesIndex must be an element of Remaining. assert(RemainingFiles.find(MaxFeaturesIndex) != RemainingFiles.end()); // Remove the file with the most features from Remaining. RemainingFiles.erase(MaxFeaturesIndex); const auto &MaxFeatureFile = Files[MaxFeaturesIndex]; // Add the features of the max feature file to Covered. for (const auto &F : MaxFeatureFile.Features) { if (!Covered[F % kFeatureSetSize]) { ++NumCovered; Covered[F % kFeatureSetSize] = true; NewFeatures->insert(F); } } // Add the index to this file to the result. NewFiles->push_back(MaxFeatureFile.Name); // Update NewCov with the additional coverage // that MaxFeatureFile provides. for (const auto &C : MaxFeatureFile.Cov) if (InitialCov.find(C) == InitialCov.end()) NewCov->insert(C); } return NewFeatures->size(); } static size_t WriteNewControlFile(const std::string &CFPath, const std::vector &OldCorpus, const std::vector &NewCorpus, const std::vector &KnownFiles) { std::unordered_set FilesToSkip; for (auto &SF: KnownFiles) FilesToSkip.insert(SF.Name); std::vector FilesToUse; auto MaybeUseFile = [=, &FilesToUse](std::string Name) { if (FilesToSkip.find(Name) == FilesToSkip.end()) FilesToUse.push_back(Name); }; for (auto &SF: OldCorpus) MaybeUseFile(SF.File); auto FilesToUseFromOldCorpus = FilesToUse.size(); for (auto &SF: NewCorpus) MaybeUseFile(SF.File); RemoveFile(CFPath); std::ofstream ControlFile(CFPath); ControlFile << FilesToUse.size() << "\n"; ControlFile << FilesToUseFromOldCorpus << "\n"; for (auto &FN: FilesToUse) ControlFile << FN << "\n"; if (!ControlFile) { Printf("MERGE-OUTER: failed to write to the control file: %s\n", CFPath.c_str()); exit(1); } return FilesToUse.size(); } // Outer process. Does not call the target code and thus should not fail. void CrashResistantMerge(const std::vector &Args, const std::vector &OldCorpus, const std::vector &NewCorpus, std::vector *NewFiles, const std::set &InitialFeatures, std::set *NewFeatures, const std::set &InitialCov, std::set *NewCov, const std::string &CFPath, bool V, /*Verbose*/ bool IsSetCoverMerge) { if (NewCorpus.empty() && OldCorpus.empty()) return; // Nothing to merge. size_t NumAttempts = 0; std::vector KnownFiles; if (FileSize(CFPath)) { VPrintf(V, "MERGE-OUTER: non-empty control file provided: '%s'\n", CFPath.c_str()); Merger M; std::ifstream IF(CFPath); if (M.Parse(IF, /*ParseCoverage=*/true)) { VPrintf(V, "MERGE-OUTER: control file ok, %zd files total," " first not processed file %zd\n", M.Files.size(), M.FirstNotProcessedFile); if (!M.LastFailure.empty()) VPrintf(V, "MERGE-OUTER: '%s' will be skipped as unlucky " "(merge has stumbled on it the last time)\n", M.LastFailure.c_str()); if (M.FirstNotProcessedFile >= M.Files.size()) { // Merge has already been completed with the given merge control file. if (M.Files.size() == OldCorpus.size() + NewCorpus.size()) { VPrintf( V, "MERGE-OUTER: nothing to do, merge has been completed before\n"); exit(0); } // Number of input files likely changed, start merge from scratch, but // reuse coverage information from the given merge control file. VPrintf( V, "MERGE-OUTER: starting merge from scratch, but reusing coverage " "information from the given control file\n"); KnownFiles = M.Files; } else { // There is a merge in progress, continue. NumAttempts = M.Files.size() - M.FirstNotProcessedFile; } } else { VPrintf(V, "MERGE-OUTER: bad control file, will overwrite it\n"); } } if (!NumAttempts) { // The supplied control file is empty or bad, create a fresh one. VPrintf(V, "MERGE-OUTER: " "%zd files, %zd in the initial corpus, %zd processed earlier\n", OldCorpus.size() + NewCorpus.size(), OldCorpus.size(), KnownFiles.size()); NumAttempts = WriteNewControlFile(CFPath, OldCorpus, NewCorpus, KnownFiles); } // Execute the inner process until it passes. // Every inner process should execute at least one input. Command BaseCmd(Args); BaseCmd.removeFlag("merge"); BaseCmd.removeFlag("set_cover_merge"); BaseCmd.removeFlag("fork"); BaseCmd.removeFlag("collect_data_flow"); for (size_t Attempt = 1; Attempt <= NumAttempts; Attempt++) { Fuzzer::MaybeExitGracefully(); VPrintf(V, "MERGE-OUTER: attempt %zd\n", Attempt); Command Cmd(BaseCmd); Cmd.addFlag("merge_control_file", CFPath); // If we are going to use the set cover implementation for // minimization add the merge_inner=2 internal flag. Cmd.addFlag("merge_inner", IsSetCoverMerge ? "2" : "1"); if (!V) { Cmd.setOutputFile(getDevNull()); Cmd.combineOutAndErr(); } auto ExitCode = ExecuteCommand(Cmd); if (!ExitCode) { VPrintf(V, "MERGE-OUTER: successful in %zd attempt(s)\n", Attempt); break; } } // Read the control file and do the merge. Merger M; std::ifstream IF(CFPath); IF.seekg(0, IF.end); VPrintf(V, "MERGE-OUTER: the control file has %zd bytes\n", (size_t)IF.tellg()); IF.seekg(0, IF.beg); M.ParseOrExit(IF, true); IF.close(); VPrintf(V, "MERGE-OUTER: consumed %zdMb (%zdMb rss) to parse the control file\n", M.ApproximateMemoryConsumption() >> 20, GetPeakRSSMb()); M.Files.insert(M.Files.end(), KnownFiles.begin(), KnownFiles.end()); if (IsSetCoverMerge) M.SetCoverMerge(InitialFeatures, NewFeatures, InitialCov, NewCov, NewFiles); else M.Merge(InitialFeatures, NewFeatures, InitialCov, NewCov, NewFiles); VPrintf(V, "MERGE-OUTER: %zd new files with %zd new features added; " "%zd new coverage edges\n", NewFiles->size(), NewFeatures->size(), NewCov->size()); } } // namespace fuzzer