FAQ (Fuzzily Asked Questions) #
The fuzzer is showing crashes, but when I run the SUT on the test cases outside of the fuzzer, then no crash is shown. What is happening? #
There are several potential reasons:
- The SUT is behaving nondeterministically.
- The SUT depends on global state (e.g., it may read or write to disk or use in-memory singletons).
- You are experiencing crashes because your system is running out of memory and killing processes. This is typically signaled by signal 6 on Linux. You may also want to check the output of
dmesg
and see if you have entries about out-of-memory issues.
When should I stop fuzzing? #
This question is very hard to answer. The code coverage is essentially the only metric available in today’s fuzzers. Code coverage is a proxy for measuring a fuzzer’s efficiency in finding bugs.
However, we have acknowledged that code coverage is the only tool we have nowadays. So the practical answer to the question is: Stop fuzzing when the fuzzer does not find new test cases for several hours. This is equivalent to saying that the fuzzer is not increasing its coverage.
Even though most academic papers about fuzzing run campaigns for 24 hours (see Evaluating Fuzz Testing, Table 1 and section 6), there is no fixed recommendation for a specific duration. Running the fuzzer for longer can be beneficial. Experiment with longer durations than 24 hours and observe the fuzzing campaign.
Definitely keep in mind that the total count of executions is a more important metric than time. If your SUT is slow to execute, fuzz longer. Do not blindly stick to a specific time.
Also, based on our experience at Trail of Bits, we recommend running multiple fuzzing campaigns, periodically resetting the corpus, and verifying if the coverage has changed. Doing so supports the fuzzer when it encounters difficulties in exploring additional program states, attributed to the absence of sufficiently informative feedback or feedback that adequately guides the fuzzer’s exploration process. Restarting a fuzzing campaign could result in randomly exploring different parts of the program than in previous campaigns.
Should I keep the corpus confidential or make it public? #
By default, you should keep the corpus confidential. The corpus is a valuable resource, and generating one takes a lot of effort. An attacker who is provided a corpus may find vulnerabilities more easily.
However, a number of projects, such as OpenSSL, decided to make their corpora public. The probable reason for this is to support security researchers. Also, it simplifies the setup of using the corpus for fuzzing in CI.
In conclusion, this is a risk assessment you have to make for your project. We recommend starting by keeping the corpus private and then eventually making it public.
My fuzzer is not finding anything. What are indicators that there is a bug in the fuzzing setup? #
Take a look at the code coverage you achieve when fuzzing. In this fuzzing chapter, we have guides on how to check the coverage for every discussed fuzzer.
A proxy for code coverage can be the executions per second:
- If you are seeing too many (>10⁶ executions per second), then the critical code is probably not executing and the code coverage is likely low. The SUT may be hitting an error early and returning gracefully.
- If you are seeing not enough executions per second (<100), your target may be executing too slowly and the fuzzer is not progressing quickly enough to discover interesting test cases.
My corpus has grown quite large. How do I deal with these thousands of small files? #
There is a technique called “corpus minimization”. Stay tuned for an update of the testing handbook to find out how to maintain and manage corpora! Subscribe to our newsletter here to get notified.
My fuzzer found inputs that crash the SUT. However, it is very large and complex. Is there a way to simplify finding the root cause of the crash? #
A technique called “test case minimization” reduces the size of inputs while preserving the crash.
Stay tuned for an update to the testing handbook that will cover how to maintain and manage corpora! Subscribe the newsletter here to get notified.
How can I collect core dumps from crashes during or after a fuzzing campaign? #
On Linux, two main settings affect core dumps: 1) the “core”
resource limit that enables core dumps per process, and 2) the sysctl
setting
kernel.core-pattern
, which determines where the core dump is stored or sent to.
For the Bash shell, you can use ulimit -c
to query the current limits on core dumps; use ulimit -c 0
to disable them and ulimit -c unlimited
to enable them. The default for resource limits differs between environments and operating systems (i.e., server vs. desktop operating systems). In order to store core dumps in the current working directory in a file called core
, you can configure the kernel using:
sysctl -w kernel.core_pattern=core
Depending on how you execute the crashing program, you may still not get a core dump file; fuzzers can interfere with the creation of core dumps either by setting the resource limits for the current process or by avoiding the creation of core dumps by overwriting signal handlers. Fuzzers disable core dumps because they may take a long time to create. Ideally, fuzzers want to continue fuzzing quickly after a crash is encountered.
Consult the documentation of your fuzzer to verify how to re-enable core dumps if you need them for debugging. Make sure to configure resource limits and the kernel.core_pattern
.
For libFuzzer, AFL++, and Hongfuzz, follow this guidance:
- libFuzzer. By default, libFuzzer disables core dumps by registering signal handlers. To enable writing core dumps, you must
enable ASan and then set the environment variable
ASAN_OPTIONS=abort_on_error=1:disable_coredump=0
. Note that this works only for crashes that are handled by ASan. A call toabort
does not cause the creation of a core dump with this method. - AFL++. Set the
AFL_DEBUG
environment variable to1
(see the source code for this option here). - Hongfuzz. Set the flag
--rlimit_core
to a high value.
I have a fuzzing setup. How often and where should I run my harness? #
Ideally, you fuzz continuously on dedicated servers and in the continuous integration service you are already using.
Stay tuned for an update to the testing handbook to find out how to set up continuous fuzzing! Subscribe to the newsletter here to get notified.
My fuzzer has found multiple crashes. In fact, I have hundreds of crash files. How can I find the corresponding bugs? #
The process of pinpointing specific bugs is called “bug triaging.”
Stay tuned for an update to the testing handbook to find out how to triage bugs! Subscribe to the newsletter here to get notified.
How do I fuzz my Go, Python, Java, or JavaScript project? #
Stay tuned for an update to the testing handbook to find out how to use the right tools for your technology stack! Subscribe to the newsletter here to get notified and message @trailofbits.
I’m using Bazel, Buck, or some other build system. How can I integrate fuzzing into my project? #
Message @trailofbits and tell us about your project setup. We’d be happy to check if there is an easy way to integrate the preferred fuzzer.
My program runs only on Windows. How can I fuzz it? #
Stay tuned for an update to the testing handbook to find out how to fuzz on Windows. Subscribe to the newsletter here to get notified and message @trailofbits to let us know you want to fuzz on Windows.
What is the best approach if I have an OSS-Fuzz fuzzing harness ready, but my project is not eligible for continuous fuzzing by Google infrastructure?
When you already have harnesses, but your project is not eligible to be fuzzed continuously by Google infrastructure, it’s important to fuzz the project regularly and for extended periods. It’s best to configure fuzzing so it runs on an updated codebase automatically.
One way of doing this is using CIFuzz (with ClusterFuzzLite if your project is not enrolled in the OSS-Fuzz project) to perform short fuzzing as a post-commit (or pre-commit) CI job. Because CIFuzz tests code from every commit, you can easily see which commit introduced the problem. This method also simplifies adding regression testing, as you can automatically add problematic inputs to corpora.
Additionally, if your project supports code-coverage calculations, CIFuzz can run only harnesses that touch modified code and not all of them. Stay tuned for an upcoming update to the testing handbook with a robust Continuous Fuzzing chapter!