aboutsummaryrefslogtreecommitdiff
path: root/instrumentation
diff options
context:
space:
mode:
authorvan Hauser <vh@thc.org>2021-01-05 11:52:55 +0100
committervan Hauser <vh@thc.org>2021-01-05 11:52:55 +0100
commitcaf1fbd6323f4069bed20e386d425e529be48a27 (patch)
tree93dd03000295490e88e99d0d8ca7e804d53fd220 /instrumentation
parenta561de6e974f18e193507578fa376ad4cb815974 (diff)
downloadAFLplusplus-caf1fbd6323f4069bed20e386d425e529be48a27.tar.gz
selective_coverage implementation in afl-compiler-rt
Diffstat (limited to 'instrumentation')
-rw-r--r--instrumentation/afl-compiler-rt.o.c149
1 files changed, 123 insertions, 26 deletions
diff --git a/instrumentation/afl-compiler-rt.o.c b/instrumentation/afl-compiler-rt.o.c
index e7cdcf4a..6a7c3dd0 100644
--- a/instrumentation/afl-compiler-rt.o.c
+++ b/instrumentation/afl-compiler-rt.o.c
@@ -76,7 +76,9 @@
#endif
u8 __afl_area_initial[MAP_INITIAL_SIZE];
+u8 * __afl_area_ptr_dummy = __afl_area_initial;
u8 * __afl_area_ptr = __afl_area_initial;
+u8 * __afl_area_ptr_backup = __afl_area_initial;
u8 * __afl_dictionary;
u8 * __afl_fuzz_ptr;
u32 __afl_fuzz_len_dummy;
@@ -87,6 +89,11 @@ u32 __afl_map_size = MAP_SIZE;
u32 __afl_dictionary_len;
u64 __afl_map_addr;
+// for the __AFL_COVERAGE_ON/__AFL_COVERAGE_OFF features to work:
+int __afl_selective_coverage __attribute__((weak));
+int __afl_selective_coverage_start_off __attribute__((weak));
+int __afl_selective_coverage_temp = 1;
+
#ifdef __ANDROID__
PREV_LOC_T __afl_prev_loc[NGRAM_SIZE_MAX];
u32 __afl_prev_ctx;
@@ -100,6 +107,7 @@ __thread u32 __afl_cmp_counter;
int __afl_sharedmem_fuzzing __attribute__((weak));
struct cmp_map *__afl_cmp_map;
+struct cmp_map *__afl_cmp_map_backup;
/* Child pid? */
@@ -230,7 +238,7 @@ static void __afl_map_shm_fuzz() {
static void __afl_map_shm(void) {
// if we are not running in afl ensure the map exists
- if (!__afl_area_ptr) { __afl_area_ptr = __afl_area_initial; }
+ if (!__afl_area_ptr) { __afl_area_ptr = __afl_area_ptr_dummy; }
char *id_str = getenv(SHM_ENV_VAR);
@@ -295,11 +303,17 @@ static void __afl_map_shm(void) {
if (__afl_area_ptr && __afl_area_ptr != __afl_area_initial) {
- if (__afl_map_addr)
+ if (__afl_map_addr) {
+
munmap((void *)__afl_map_addr, __afl_final_loc);
- else
+
+ } else {
+
free(__afl_area_ptr);
- __afl_area_ptr = __afl_area_initial;
+
+ }
+
+ __afl_area_ptr = __afl_area_ptr_dummy;
}
@@ -396,9 +410,42 @@ static void __afl_map_shm(void) {
free(__afl_area_ptr);
__afl_area_ptr = NULL;
- if (__afl_final_loc > MAP_INITIAL_SIZE)
+
+ if (__afl_final_loc > MAP_INITIAL_SIZE) {
+
__afl_area_ptr = malloc(__afl_final_loc);
- if (!__afl_area_ptr) __afl_area_ptr = __afl_area_initial;
+
+ }
+
+ if (!__afl_area_ptr) { __afl_area_ptr = __afl_area_ptr_dummy; }
+
+ }
+
+ __afl_area_ptr_backup = __afl_area_ptr;
+
+ if (__afl_selective_coverage) {
+
+ if (__afl_map_size > MAP_INITIAL_SIZE) {
+
+ __afl_area_ptr_dummy = malloc(__afl_map_size);
+
+ if (__afl_area_ptr_dummy) {
+
+ if (__afl_selective_coverage_start_off) {
+
+ __afl_area_ptr = __afl_area_ptr_dummy;
+
+ }
+
+ } else {
+
+ fprintf(stderr, "Error: __afl_selective_coverage failed!\n");
+ __afl_selective_coverage = 0;
+ // continue;
+
+ }
+
+ }
}
@@ -449,6 +496,8 @@ static void __afl_map_shm(void) {
__afl_cmp_map = shmat(shm_id, NULL, 0);
#endif
+ __afl_cmp_map_backup = __afl_cmp_map;
+
if (!__afl_cmp_map || __afl_cmp_map == (void *)-1) {
perror("shmat for cmplog");
@@ -903,6 +952,8 @@ int __afl_persistent_loop(unsigned int max_cnt) {
cycle_cnt = max_cnt;
first_pass = 0;
+ __afl_selective_coverage_temp = 1;
+
return 1;
}
@@ -915,6 +966,7 @@ int __afl_persistent_loop(unsigned int max_cnt) {
__afl_area_ptr[0] = 1;
memset(__afl_prev_loc, 0, NGRAM_SIZE_MAX * sizeof(PREV_LOC_T));
+ __afl_selective_coverage_temp = 1;
return 1;
@@ -924,7 +976,7 @@ int __afl_persistent_loop(unsigned int max_cnt) {
follows the loop is not traced. We do that by pivoting back to the
dummy output region. */
- __afl_area_ptr = __afl_area_initial;
+ __afl_area_ptr = __afl_area_ptr_dummy;
}
@@ -946,7 +998,7 @@ void __afl_manual_init(void) {
init_done = 1;
is_persistent = 0;
__afl_sharedmem_fuzzing = 0;
- if (__afl_area_ptr == NULL) __afl_area_ptr = __afl_area_initial;
+ if (__afl_area_ptr == NULL) __afl_area_ptr = __afl_area_ptr_dummy;
if (getenv("AFL_DEBUG"))
fprintf(stderr,
@@ -1007,7 +1059,12 @@ __attribute__((constructor(1))) void __afl_auto_second(void) {
else
ptr = (u8 *)malloc(__afl_final_loc);
- if (ptr && (ssize_t)ptr != -1) __afl_area_ptr = ptr;
+ if (ptr && (ssize_t)ptr != -1) {
+
+ __afl_area_ptr = ptr;
+ __afl_area_ptr_backup = __afl_area_ptr;
+
+ }
}
@@ -1023,7 +1080,12 @@ __attribute__((constructor(0))) void __afl_auto_first(void) {
ptr = (u8 *)malloc(1024000);
- if (ptr && (ssize_t)ptr != -1) __afl_area_ptr = ptr;
+ if (ptr && (ssize_t)ptr != -1) {
+
+ __afl_area_ptr = ptr;
+ __afl_area_ptr_backup = __afl_area_ptr;
+
+ }
}
@@ -1315,6 +1377,35 @@ void __cmplog_rtn_hook(u8 *ptr1, u8 *ptr2) {
/* COVERAGE manipulation features */
+// this variable is then used in the shm setup to create an additional map
+// if __afl_map_size > MAP_SIZE or cmplog is used.
+// Especially with cmplog this would result in a ~260MB mem increase per
+// target run.
+
+// disable coverage from this point onwards until turned on again
+void __afl_coverage_off() {
+
+ if (likely(__afl_selective_coverage)) {
+
+ __afl_area_ptr = __afl_area_ptr_dummy;
+ __afl_cmp_map = NULL;
+
+ }
+
+}
+
+// enable coverage
+void __afl_coverage_on() {
+
+ if (likely(__afl_selective_coverage && __afl_selective_coverage_temp)) {
+
+ __afl_area_ptr = __afl_area_ptr_backup;
+ __afl_cmp_map = __afl_cmp_map_backup;
+
+ }
+
+}
+
// discard all coverage up to this point
void __afl_coverage_discard() {
@@ -1329,30 +1420,36 @@ void __afl_coverage_discard() {
void __afl_coverage_abort() {
__afl_coverage_discard();
- exit(0);
-}
+ if (likely(is_persistent && __afl_selective_coverage)) {
-// For the following two functions to work there needs to be a global define,
-// eg. __AFL_COVERAGE(); after the headers which translates to
-// int __afl_selective_coverage = 1; by a -D from afl-cc
-int __afl_selective_coverage __attribute__((weak));
-// this variable is then used in the shm setup to create an additional map
-// if __afl_map_size > MAP_SIZE or cmplog is used.
-// Especially with cmplog this would result in a ~260MB mem increase per
-// target run.
+ __afl_coverage_off();
+ __afl_selective_coverage_temp = 0;
-// disable coverage from this point onwards until turned on again
-void __afl_coverage_off() {
+ } else {
- // switch __afl_area_ptr and __afl_cmp_map to (the same) dummy pointer
+ exit(0);
+
+ }
}
-// enable coverage
-void __afl_coverage_on() {
+// mark this area as especially interesting
+void __afl_coverage_interesting(u32 id, u8 val) {
+
+ if (val) {
+
+ __afl_area_ptr[id] = val;
+
+ } else {
+
+ do {
+
+ __afl_area_ptr[id] = (u8)rand();
- // switch __afl_area_ptr and __afl_cmp_map to the real map
+ } while (!__afl_area_ptr[id]);
+
+ }
}