Remove leading blank lines from function bodies.

This resolves #535.
diff --git a/include/jemalloc/internal/arena_inlines_a.h b/include/jemalloc/internal/arena_inlines_a.h
index 743727b..d241b8a 100644
--- a/include/jemalloc/internal/arena_inlines_a.h
+++ b/include/jemalloc/internal/arena_inlines_a.h
@@ -16,35 +16,30 @@
 JEMALLOC_INLINE unsigned
 arena_ind_get(const arena_t *arena)
 {
-
 	return (base_ind_get(arena->base));
 }
 
 JEMALLOC_INLINE void
 arena_internal_add(arena_t *arena, size_t size)
 {
-
 	atomic_add_zu(&arena->stats.internal, size);
 }
 
 JEMALLOC_INLINE void
 arena_internal_sub(arena_t *arena, size_t size)
 {
-
 	atomic_sub_zu(&arena->stats.internal, size);
 }
 
 JEMALLOC_INLINE size_t
 arena_internal_get(arena_t *arena)
 {
-
 	return (atomic_read_zu(&arena->stats.internal));
 }
 
 JEMALLOC_INLINE bool
 arena_prof_accum_impl(arena_t *arena, uint64_t accumbytes)
 {
-
 	cassert(config_prof);
 	assert(prof_interval != 0);
 
@@ -59,7 +54,6 @@
 JEMALLOC_INLINE bool
 arena_prof_accum_locked(arena_t *arena, uint64_t accumbytes)
 {
-
 	cassert(config_prof);
 
 	if (likely(prof_interval == 0))
@@ -70,7 +64,6 @@
 JEMALLOC_INLINE bool
 arena_prof_accum(tsdn_t *tsdn, arena_t *arena, uint64_t accumbytes)
 {
-
 	cassert(config_prof);
 
 	if (likely(prof_interval == 0))
diff --git a/include/jemalloc/internal/arena_inlines_b.h b/include/jemalloc/internal/arena_inlines_b.h
index 9068cf4..9461466 100644
--- a/include/jemalloc/internal/arena_inlines_b.h
+++ b/include/jemalloc/internal/arena_inlines_b.h
@@ -33,7 +33,6 @@
 JEMALLOC_INLINE prof_tctx_t *
 arena_prof_tctx_get(tsdn_t *tsdn, const extent_t *extent, const void *ptr)
 {
-
 	cassert(config_prof);
 	assert(ptr != NULL);
 
@@ -46,7 +45,6 @@
 arena_prof_tctx_set(tsdn_t *tsdn, extent_t *extent, const void *ptr,
     size_t usize, prof_tctx_t *tctx)
 {
-
 	cassert(config_prof);
 	assert(ptr != NULL);
 
@@ -58,7 +56,6 @@
 arena_prof_tctx_reset(tsdn_t *tsdn, extent_t *extent, const void *ptr,
     prof_tctx_t *tctx)
 {
-
 	cassert(config_prof);
 	assert(ptr != NULL);
 	assert(!extent_slab_get(extent));
@@ -85,7 +82,6 @@
 JEMALLOC_ALWAYS_INLINE void
 arena_decay_tick(tsdn_t *tsdn, arena_t *arena)
 {
-
 	malloc_mutex_assert_not_owner(tsdn, &arena->lock);
 
 	arena_decay_ticks(tsdn, arena, 1);
@@ -95,7 +91,6 @@
 arena_malloc(tsdn_t *tsdn, arena_t *arena, size_t size, szind_t ind, bool zero,
     tcache_t *tcache, bool slow_path)
 {
-
 	assert(!tsdn_null(tsdn) || tcache == NULL);
 	assert(size != 0);
 
@@ -118,7 +113,6 @@
 JEMALLOC_ALWAYS_INLINE arena_t *
 arena_aalloc(tsdn_t *tsdn, const void *ptr)
 {
-
 	return (extent_arena_get(iealloc(tsdn, ptr)));
 }
 
@@ -142,7 +136,6 @@
 arena_dalloc(tsdn_t *tsdn, extent_t *extent, void *ptr, tcache_t *tcache,
     bool slow_path)
 {
-
 	assert(!tsdn_null(tsdn) || tcache == NULL);
 	assert(ptr != NULL);
 
@@ -176,7 +169,6 @@
 arena_sdalloc(tsdn_t *tsdn, extent_t *extent, void *ptr, size_t size,
     tcache_t *tcache, bool slow_path)
 {
-
 	assert(!tsdn_null(tsdn) || tcache == NULL);
 	assert(ptr != NULL);
 
diff --git a/include/jemalloc/internal/atomic_inlines.h b/include/jemalloc/internal/atomic_inlines.h
index de0ac6a..89d1b35 100644
--- a/include/jemalloc/internal/atomic_inlines.h
+++ b/include/jemalloc/internal/atomic_inlines.h
@@ -101,7 +101,6 @@
 JEMALLOC_INLINE void
 atomic_write_u64(uint64_t *p, uint64_t x)
 {
-
 	asm volatile (
 	    "xchgq %1, %0;" /* Lock is implied by xchgq. */
 	    : "=m" (*p), "+r" (x) /* Outputs. */
@@ -141,7 +140,6 @@
 JEMALLOC_INLINE uint64_t
 atomic_add_u64(uint64_t *p, uint64_t x)
 {
-
 	/*
 	 * atomic_fetchadd_64() doesn't exist, but we only ever use this
 	 * function on LP64 systems, so atomic_fetchadd_long() will do.
@@ -154,7 +152,6 @@
 JEMALLOC_INLINE uint64_t
 atomic_sub_u64(uint64_t *p, uint64_t x)
 {
-
 	assert(sizeof(uint64_t) == sizeof(unsigned long));
 
 	return (atomic_fetchadd_long(p, (unsigned long)(-(long)x)) - x);
@@ -163,7 +160,6 @@
 JEMALLOC_INLINE bool
 atomic_cas_u64(uint64_t *p, uint64_t c, uint64_t s)
 {
-
 	assert(sizeof(uint64_t) == sizeof(unsigned long));
 
 	return (!atomic_cmpset_long(p, (unsigned long)c, (unsigned long)s));
@@ -172,7 +168,6 @@
 JEMALLOC_INLINE void
 atomic_write_u64(uint64_t *p, uint64_t x)
 {
-
 	assert(sizeof(uint64_t) == sizeof(unsigned long));
 
 	atomic_store_rel_long(p, x);
@@ -181,21 +176,18 @@
 JEMALLOC_INLINE uint64_t
 atomic_add_u64(uint64_t *p, uint64_t x)
 {
-
 	return (OSAtomicAdd64((int64_t)x, (int64_t *)p));
 }
 
 JEMALLOC_INLINE uint64_t
 atomic_sub_u64(uint64_t *p, uint64_t x)
 {
-
 	return (OSAtomicAdd64(-((int64_t)x), (int64_t *)p));
 }
 
 JEMALLOC_INLINE bool
 atomic_cas_u64(uint64_t *p, uint64_t c, uint64_t s)
 {
-
 	return (!OSAtomicCompareAndSwap64(c, s, (int64_t *)p));
 }
 
@@ -213,14 +205,12 @@
 JEMALLOC_INLINE uint64_t
 atomic_add_u64(uint64_t *p, uint64_t x)
 {
-
 	return (InterlockedExchangeAdd64(p, x) + x);
 }
 
 JEMALLOC_INLINE uint64_t
 atomic_sub_u64(uint64_t *p, uint64_t x)
 {
-
 	return (InterlockedExchangeAdd64(p, -((int64_t)x)) - x);
 }
 
@@ -236,7 +226,6 @@
 JEMALLOC_INLINE void
 atomic_write_u64(uint64_t *p, uint64_t x)
 {
-
 	InterlockedExchange64(p, x);
 }
 #  elif (defined(__GCC_HAVE_SYNC_COMPARE_AND_SWAP_8) || \
@@ -244,28 +233,24 @@
 JEMALLOC_INLINE uint64_t
 atomic_add_u64(uint64_t *p, uint64_t x)
 {
-
 	return (__sync_add_and_fetch(p, x));
 }
 
 JEMALLOC_INLINE uint64_t
 atomic_sub_u64(uint64_t *p, uint64_t x)
 {
-
 	return (__sync_sub_and_fetch(p, x));
 }
 
 JEMALLOC_INLINE bool
 atomic_cas_u64(uint64_t *p, uint64_t c, uint64_t s)
 {
-
 	return (!__sync_bool_compare_and_swap(p, c, s));
 }
 
 JEMALLOC_INLINE void
 atomic_write_u64(uint64_t *p, uint64_t x)
 {
-
 	__sync_lock_test_and_set(p, x);
 }
 #  else
@@ -325,7 +310,6 @@
 JEMALLOC_INLINE void
 atomic_write_u32(uint32_t *p, uint32_t x)
 {
-
 	asm volatile (
 	    "xchgl %1, %0;" /* Lock is implied by xchgl. */
 	    : "=m" (*p), "+r" (x) /* Outputs. */
@@ -365,49 +349,42 @@
 JEMALLOC_INLINE uint32_t
 atomic_add_u32(uint32_t *p, uint32_t x)
 {
-
 	return (atomic_fetchadd_32(p, x) + x);
 }
 
 JEMALLOC_INLINE uint32_t
 atomic_sub_u32(uint32_t *p, uint32_t x)
 {
-
 	return (atomic_fetchadd_32(p, (uint32_t)(-(int32_t)x)) - x);
 }
 
 JEMALLOC_INLINE bool
 atomic_cas_u32(uint32_t *p, uint32_t c, uint32_t s)
 {
-
 	return (!atomic_cmpset_32(p, c, s));
 }
 
 JEMALLOC_INLINE void
 atomic_write_u32(uint32_t *p, uint32_t x)
 {
-
 	atomic_store_rel_32(p, x);
 }
 #elif (defined(JEMALLOC_OSATOMIC))
 JEMALLOC_INLINE uint32_t
 atomic_add_u32(uint32_t *p, uint32_t x)
 {
-
 	return (OSAtomicAdd32((int32_t)x, (int32_t *)p));
 }
 
 JEMALLOC_INLINE uint32_t
 atomic_sub_u32(uint32_t *p, uint32_t x)
 {
-
 	return (OSAtomicAdd32(-((int32_t)x), (int32_t *)p));
 }
 
 JEMALLOC_INLINE bool
 atomic_cas_u32(uint32_t *p, uint32_t c, uint32_t s)
 {
-
 	return (!OSAtomicCompareAndSwap32(c, s, (int32_t *)p));
 }
 
@@ -425,14 +402,12 @@
 JEMALLOC_INLINE uint32_t
 atomic_add_u32(uint32_t *p, uint32_t x)
 {
-
 	return (InterlockedExchangeAdd(p, x) + x);
 }
 
 JEMALLOC_INLINE uint32_t
 atomic_sub_u32(uint32_t *p, uint32_t x)
 {
-
 	return (InterlockedExchangeAdd(p, -((int32_t)x)) - x);
 }
 
@@ -448,7 +423,6 @@
 JEMALLOC_INLINE void
 atomic_write_u32(uint32_t *p, uint32_t x)
 {
-
 	InterlockedExchange(p, x);
 }
 #elif (defined(__GCC_HAVE_SYNC_COMPARE_AND_SWAP_4) || \
@@ -456,28 +430,24 @@
 JEMALLOC_INLINE uint32_t
 atomic_add_u32(uint32_t *p, uint32_t x)
 {
-
 	return (__sync_add_and_fetch(p, x));
 }
 
 JEMALLOC_INLINE uint32_t
 atomic_sub_u32(uint32_t *p, uint32_t x)
 {
-
 	return (__sync_sub_and_fetch(p, x));
 }
 
 JEMALLOC_INLINE bool
 atomic_cas_u32(uint32_t *p, uint32_t c, uint32_t s)
 {
-
 	return (!__sync_bool_compare_and_swap(p, c, s));
 }
 
 JEMALLOC_INLINE void
 atomic_write_u32(uint32_t *p, uint32_t x)
 {
-
 	__sync_lock_test_and_set(p, x);
 }
 #else
@@ -489,7 +459,6 @@
 JEMALLOC_INLINE void *
 atomic_add_p(void **p, void *x)
 {
-
 #if (LG_SIZEOF_PTR == 3)
 	return ((void *)atomic_add_u64((uint64_t *)p, (uint64_t)x));
 #elif (LG_SIZEOF_PTR == 2)
@@ -500,7 +469,6 @@
 JEMALLOC_INLINE void *
 atomic_sub_p(void **p, void *x)
 {
-
 #if (LG_SIZEOF_PTR == 3)
 	return ((void *)atomic_add_u64((uint64_t *)p, (uint64_t)-((int64_t)x)));
 #elif (LG_SIZEOF_PTR == 2)
@@ -511,7 +479,6 @@
 JEMALLOC_INLINE bool
 atomic_cas_p(void **p, void *c, void *s)
 {
-
 #if (LG_SIZEOF_PTR == 3)
 	return (atomic_cas_u64((uint64_t *)p, (uint64_t)c, (uint64_t)s));
 #elif (LG_SIZEOF_PTR == 2)
@@ -522,7 +489,6 @@
 JEMALLOC_INLINE void
 atomic_write_p(void **p, const void *x)
 {
-
 #if (LG_SIZEOF_PTR == 3)
 	atomic_write_u64((uint64_t *)p, (uint64_t)x);
 #elif (LG_SIZEOF_PTR == 2)
@@ -535,7 +501,6 @@
 JEMALLOC_INLINE size_t
 atomic_add_zu(size_t *p, size_t x)
 {
-
 #if (LG_SIZEOF_PTR == 3)
 	return ((size_t)atomic_add_u64((uint64_t *)p, (uint64_t)x));
 #elif (LG_SIZEOF_PTR == 2)
@@ -546,7 +511,6 @@
 JEMALLOC_INLINE size_t
 atomic_sub_zu(size_t *p, size_t x)
 {
-
 #if (LG_SIZEOF_PTR == 3)
 	return ((size_t)atomic_add_u64((uint64_t *)p, (uint64_t)-((int64_t)x)));
 #elif (LG_SIZEOF_PTR == 2)
@@ -557,7 +521,6 @@
 JEMALLOC_INLINE bool
 atomic_cas_zu(size_t *p, size_t c, size_t s)
 {
-
 #if (LG_SIZEOF_PTR == 3)
 	return (atomic_cas_u64((uint64_t *)p, (uint64_t)c, (uint64_t)s));
 #elif (LG_SIZEOF_PTR == 2)
@@ -568,7 +531,6 @@
 JEMALLOC_INLINE void
 atomic_write_zu(size_t *p, size_t x)
 {
-
 #if (LG_SIZEOF_PTR == 3)
 	atomic_write_u64((uint64_t *)p, (uint64_t)x);
 #elif (LG_SIZEOF_PTR == 2)
@@ -581,7 +543,6 @@
 JEMALLOC_INLINE unsigned
 atomic_add_u(unsigned *p, unsigned x)
 {
-
 #if (LG_SIZEOF_INT == 3)
 	return ((unsigned)atomic_add_u64((uint64_t *)p, (uint64_t)x));
 #elif (LG_SIZEOF_INT == 2)
@@ -592,7 +553,6 @@
 JEMALLOC_INLINE unsigned
 atomic_sub_u(unsigned *p, unsigned x)
 {
-
 #if (LG_SIZEOF_INT == 3)
 	return ((unsigned)atomic_add_u64((uint64_t *)p,
 	    (uint64_t)-((int64_t)x)));
@@ -605,7 +565,6 @@
 JEMALLOC_INLINE bool
 atomic_cas_u(unsigned *p, unsigned c, unsigned s)
 {
-
 #if (LG_SIZEOF_INT == 3)
 	return (atomic_cas_u64((uint64_t *)p, (uint64_t)c, (uint64_t)s));
 #elif (LG_SIZEOF_INT == 2)
@@ -616,7 +575,6 @@
 JEMALLOC_INLINE void
 atomic_write_u(unsigned *p, unsigned x)
 {
-
 #if (LG_SIZEOF_INT == 3)
 	atomic_write_u64((uint64_t *)p, (uint64_t)x);
 #elif (LG_SIZEOF_INT == 2)
diff --git a/include/jemalloc/internal/base_inlines.h b/include/jemalloc/internal/base_inlines.h
index f882bcd..63547d6 100644
--- a/include/jemalloc/internal/base_inlines.h
+++ b/include/jemalloc/internal/base_inlines.h
@@ -9,7 +9,6 @@
 JEMALLOC_INLINE unsigned
 base_ind_get(const base_t *base)
 {
-
 	return (base->ind);
 }
 #endif
diff --git a/include/jemalloc/internal/extent_inlines.h b/include/jemalloc/internal/extent_inlines.h
index e48af92..87e0bcd 100644
--- a/include/jemalloc/internal/extent_inlines.h
+++ b/include/jemalloc/internal/extent_inlines.h
@@ -55,14 +55,12 @@
 JEMALLOC_INLINE arena_t *
 extent_arena_get(const extent_t *extent)
 {
-
 	return (extent->e_arena);
 }
 
 JEMALLOC_INLINE void *
 extent_base_get(const extent_t *extent)
 {
-
 	assert(extent->e_addr == PAGE_ADDR2BASE(extent->e_addr) ||
 	    !extent->e_slab);
 	return (PAGE_ADDR2BASE(extent->e_addr));
@@ -71,7 +69,6 @@
 JEMALLOC_INLINE void *
 extent_addr_get(const extent_t *extent)
 {
-
 	assert(extent->e_addr == PAGE_ADDR2BASE(extent->e_addr) ||
 	    !extent->e_slab);
 	return (extent->e_addr);
@@ -80,14 +77,12 @@
 JEMALLOC_INLINE size_t
 extent_size_get(const extent_t *extent)
 {
-
 	return (extent->e_size);
 }
 
 JEMALLOC_INLINE size_t
 extent_usize_get(const extent_t *extent)
 {
-
 	assert(!extent->e_slab);
 	return (extent->e_usize);
 }
@@ -95,14 +90,12 @@
 JEMALLOC_INLINE void *
 extent_before_get(const extent_t *extent)
 {
-
 	return ((void *)((uintptr_t)extent_base_get(extent) - PAGE));
 }
 
 JEMALLOC_INLINE void *
 extent_last_get(const extent_t *extent)
 {
-
 	return ((void *)((uintptr_t)extent_base_get(extent) +
 	    extent_size_get(extent) - PAGE));
 }
@@ -110,7 +103,6 @@
 JEMALLOC_INLINE void *
 extent_past_get(const extent_t *extent)
 {
-
 	return ((void *)((uintptr_t)extent_base_get(extent) +
 	    extent_size_get(extent)));
 }
@@ -118,49 +110,42 @@
 JEMALLOC_INLINE size_t
 extent_sn_get(const extent_t *extent)
 {
-
 	return (extent->e_sn);
 }
 
 JEMALLOC_INLINE bool
 extent_active_get(const extent_t *extent)
 {
-
 	return (extent->e_active);
 }
 
 JEMALLOC_INLINE bool
 extent_retained_get(const extent_t *extent)
 {
-
 	return (qr_next(extent, qr_link) == extent);
 }
 
 JEMALLOC_INLINE bool
 extent_zeroed_get(const extent_t *extent)
 {
-
 	return (extent->e_zeroed);
 }
 
 JEMALLOC_INLINE bool
 extent_committed_get(const extent_t *extent)
 {
-
 	return (extent->e_committed);
 }
 
 JEMALLOC_INLINE bool
 extent_slab_get(const extent_t *extent)
 {
-
 	return (extent->e_slab);
 }
 
 JEMALLOC_INLINE arena_slab_data_t *
 extent_slab_data_get(extent_t *extent)
 {
-
 	assert(extent->e_slab);
 	return (&extent->e_slab_data);
 }
@@ -168,7 +153,6 @@
 JEMALLOC_INLINE const arena_slab_data_t *
 extent_slab_data_get_const(const extent_t *extent)
 {
-
 	assert(extent->e_slab);
 	return (&extent->e_slab_data);
 }
@@ -176,7 +160,6 @@
 JEMALLOC_INLINE prof_tctx_t *
 extent_prof_tctx_get(const extent_t *extent)
 {
-
 	return ((prof_tctx_t *)atomic_read_p(
 	    &((extent_t *)extent)->e_prof_tctx_pun));
 }
@@ -184,21 +167,18 @@
 JEMALLOC_INLINE void
 extent_arena_set(extent_t *extent, arena_t *arena)
 {
-
 	extent->e_arena = arena;
 }
 
 JEMALLOC_INLINE void
 extent_addr_set(extent_t *extent, void *addr)
 {
-
 	extent->e_addr = addr;
 }
 
 JEMALLOC_INLINE void
 extent_addr_randomize(tsdn_t *tsdn, extent_t *extent, size_t alignment)
 {
-
 	assert(extent_base_get(extent) == extent_addr_get(extent));
 
 	if (alignment < PAGE) {
@@ -219,56 +199,48 @@
 JEMALLOC_INLINE void
 extent_size_set(extent_t *extent, size_t size)
 {
-
 	extent->e_size = size;
 }
 
 JEMALLOC_INLINE void
 extent_usize_set(extent_t *extent, size_t usize)
 {
-
 	extent->e_usize = usize;
 }
 
 JEMALLOC_INLINE void
 extent_sn_set(extent_t *extent, size_t sn)
 {
-
 	extent->e_sn = sn;
 }
 
 JEMALLOC_INLINE void
 extent_active_set(extent_t *extent, bool active)
 {
-
 	extent->e_active = active;
 }
 
 JEMALLOC_INLINE void
 extent_zeroed_set(extent_t *extent, bool zeroed)
 {
-
 	extent->e_zeroed = zeroed;
 }
 
 JEMALLOC_INLINE void
 extent_committed_set(extent_t *extent, bool committed)
 {
-
 	extent->e_committed = committed;
 }
 
 JEMALLOC_INLINE void
 extent_slab_set(extent_t *extent, bool slab)
 {
-
 	extent->e_slab = slab;
 }
 
 JEMALLOC_INLINE void
 extent_prof_tctx_set(extent_t *extent, prof_tctx_t *tctx)
 {
-
 	atomic_write_p(&extent->e_prof_tctx_pun, tctx);
 }
 
@@ -277,7 +249,6 @@
     size_t usize, size_t sn, bool active, bool zeroed, bool committed,
     bool slab)
 {
-
 	assert(addr == PAGE_ADDR2BASE(addr) || !slab);
 
 	extent_arena_set(extent, arena);
@@ -297,14 +268,12 @@
 JEMALLOC_INLINE void
 extent_ring_insert(extent_t *sentinel, extent_t *extent)
 {
-
 	qr_meld(sentinel, extent, extent_t, qr_link);
 }
 
 JEMALLOC_INLINE void
 extent_ring_remove(extent_t *extent)
 {
-
 	qr_remove(extent, qr_link);
 }
 
diff --git a/include/jemalloc/internal/hash_inlines.h b/include/jemalloc/internal/hash_inlines.h
index 0340418..4bb7850 100644
--- a/include/jemalloc/internal/hash_inlines.h
+++ b/include/jemalloc/internal/hash_inlines.h
@@ -23,21 +23,18 @@
 JEMALLOC_INLINE uint32_t
 hash_rotl_32(uint32_t x, int8_t r)
 {
-
 	return ((x << r) | (x >> (32 - r)));
 }
 
 JEMALLOC_INLINE uint64_t
 hash_rotl_64(uint64_t x, int8_t r)
 {
-
 	return ((x << r) | (x >> (64 - r)));
 }
 
 JEMALLOC_INLINE uint32_t
 hash_get_block_32(const uint32_t *p, int i)
 {
-
 	/* Handle unaligned read. */
 	if (unlikely((uintptr_t)p & (sizeof(uint32_t)-1)) != 0) {
 		uint32_t ret;
@@ -52,7 +49,6 @@
 JEMALLOC_INLINE uint64_t
 hash_get_block_64(const uint64_t *p, int i)
 {
-
 	/* Handle unaligned read. */
 	if (unlikely((uintptr_t)p & (sizeof(uint64_t)-1)) != 0) {
 		uint64_t ret;
@@ -67,7 +63,6 @@
 JEMALLOC_INLINE uint32_t
 hash_fmix_32(uint32_t h)
 {
-
 	h ^= h >> 16;
 	h *= 0x85ebca6b;
 	h ^= h >> 13;
@@ -80,7 +75,6 @@
 JEMALLOC_INLINE uint64_t
 hash_fmix_64(uint64_t k)
 {
-
 	k ^= k >> 33;
 	k *= KQU(0xff51afd7ed558ccd);
 	k ^= k >> 33;
@@ -326,7 +320,6 @@
 JEMALLOC_INLINE void
 hash(const void *key, size_t len, const uint32_t seed, size_t r_hash[2])
 {
-
 	assert(len <= INT_MAX); /* Unfortunate implementation limitation. */
 
 #if (LG_SIZEOF_PTR == 3 && !defined(JEMALLOC_BIG_ENDIAN))
diff --git a/include/jemalloc/internal/jemalloc_internal.h.in b/include/jemalloc/internal/jemalloc_internal.h.in
index dfbb4b6..00dce68 100644
--- a/include/jemalloc/internal/jemalloc_internal.h.in
+++ b/include/jemalloc/internal/jemalloc_internal.h.in
@@ -553,7 +553,6 @@
 JEMALLOC_ALWAYS_INLINE pszind_t
 psz2ind(size_t psz)
 {
-
 	if (unlikely(psz > LARGE_MAXCLASS))
 		return (NPSIZES);
 	{
@@ -577,7 +576,6 @@
 JEMALLOC_INLINE size_t
 pind2sz_compute(pszind_t pind)
 {
-
 	if (unlikely(pind == NPSIZES))
 		return (LARGE_MAXCLASS + PAGE);
 	{
@@ -608,7 +606,6 @@
 JEMALLOC_INLINE size_t
 pind2sz(pszind_t pind)
 {
-
 	assert(pind < NPSIZES+1);
 	return (pind2sz_lookup(pind));
 }
@@ -616,7 +613,6 @@
 JEMALLOC_INLINE size_t
 psz2u(size_t psz)
 {
-
 	if (unlikely(psz > LARGE_MAXCLASS))
 		return (LARGE_MAXCLASS + PAGE);
 	{
@@ -633,7 +629,6 @@
 JEMALLOC_INLINE szind_t
 size2index_compute(size_t size)
 {
-
 	if (unlikely(size > LARGE_MAXCLASS))
 		return (NSIZES);
 #if (NTBINS != 0)
@@ -664,7 +659,6 @@
 JEMALLOC_ALWAYS_INLINE szind_t
 size2index_lookup(size_t size)
 {
-
 	assert(size <= LOOKUP_MAXCLASS);
 	{
 		szind_t ret = (size2index_tab[(size-1) >> LG_TINY_MIN]);
@@ -676,7 +670,6 @@
 JEMALLOC_ALWAYS_INLINE szind_t
 size2index(size_t size)
 {
-
 	assert(size > 0);
 	if (likely(size <= LOOKUP_MAXCLASS))
 		return (size2index_lookup(size));
@@ -686,7 +679,6 @@
 JEMALLOC_INLINE size_t
 index2size_compute(szind_t index)
 {
-
 #if (NTBINS > 0)
 	if (index < NTBINS)
 		return (ZU(1) << (LG_TINY_MAXCLASS - NTBINS + 1 + index));
@@ -721,7 +713,6 @@
 JEMALLOC_ALWAYS_INLINE size_t
 index2size(szind_t index)
 {
-
 	assert(index < NSIZES);
 	return (index2size_lookup(index));
 }
@@ -729,7 +720,6 @@
 JEMALLOC_ALWAYS_INLINE size_t
 s2u_compute(size_t size)
 {
-
 	if (unlikely(size > LARGE_MAXCLASS))
 		return (0);
 #if (NTBINS > 0)
@@ -767,7 +757,6 @@
 JEMALLOC_ALWAYS_INLINE size_t
 s2u(size_t size)
 {
-
 	assert(size > 0);
 	if (likely(size <= LOOKUP_MAXCLASS))
 		return (s2u_lookup(size));
@@ -852,14 +841,12 @@
 JEMALLOC_INLINE arena_t *
 arena_choose(tsd_t *tsd, arena_t *arena)
 {
-
 	return (arena_choose_impl(tsd, arena, false));
 }
 
 JEMALLOC_INLINE arena_t *
 arena_ichoose(tsd_t *tsd, arena_t *arena)
 {
-
 	return (arena_choose_impl(tsd, arena, true));
 }
 
@@ -933,7 +920,6 @@
 JEMALLOC_ALWAYS_INLINE extent_t *
 iealloc(tsdn_t *tsdn, const void *ptr)
 {
-
 	return (extent_lookup(tsdn, ptr, true));
 }
 #endif
@@ -975,7 +961,6 @@
 JEMALLOC_ALWAYS_INLINE arena_t *
 iaalloc(tsdn_t *tsdn, const void *ptr)
 {
-
 	assert(ptr != NULL);
 
 	return (arena_aalloc(tsdn, ptr));
@@ -991,7 +976,6 @@
 JEMALLOC_ALWAYS_INLINE size_t
 isalloc(tsdn_t *tsdn, const extent_t *extent, const void *ptr)
 {
-
 	assert(ptr != NULL);
 
 	return (arena_salloc(tsdn, extent, ptr));
@@ -1019,7 +1003,6 @@
 JEMALLOC_ALWAYS_INLINE void *
 ialloc(tsd_t *tsd, size_t size, szind_t ind, bool zero, bool slow_path)
 {
-
 	return (iallocztm(tsd_tsdn(tsd), size, ind, zero, tcache_get(tsd, true),
 	    false, NULL, slow_path));
 }
@@ -1049,14 +1032,12 @@
 ipalloct(tsdn_t *tsdn, size_t usize, size_t alignment, bool zero,
     tcache_t *tcache, arena_t *arena)
 {
-
 	return (ipallocztm(tsdn, usize, alignment, zero, tcache, false, arena));
 }
 
 JEMALLOC_ALWAYS_INLINE void *
 ipalloc(tsd_t *tsd, size_t usize, size_t alignment, bool zero)
 {
-
 	return (ipallocztm(tsd_tsdn(tsd), usize, alignment, zero,
 	    tcache_get(tsd, true), false, NULL));
 }
@@ -1088,7 +1069,6 @@
 idalloctm(tsdn_t *tsdn, extent_t *extent, void *ptr, tcache_t *tcache,
     bool is_internal, bool slow_path)
 {
-
 	assert(ptr != NULL);
 	assert(!is_internal || tcache == NULL);
 	assert(!is_internal || arena_ind_get(iaalloc(tsdn, ptr)) <
@@ -1104,7 +1084,6 @@
 JEMALLOC_ALWAYS_INLINE void
 idalloc(tsd_t *tsd, extent_t *extent, void *ptr)
 {
-
 	idalloctm(tsd_tsdn(tsd), extent, ptr, tcache_get(tsd, false), false,
 	    true);
 }
@@ -1113,7 +1092,6 @@
 isdalloct(tsdn_t *tsdn, extent_t *extent, void *ptr, size_t size,
     tcache_t *tcache, bool slow_path)
 {
-
 	arena_sdalloc(tsdn, extent, ptr, size, tcache, slow_path);
 }
 
@@ -1154,7 +1132,6 @@
 iralloct(tsdn_t *tsdn, extent_t *extent, void *ptr, size_t oldsize, size_t size,
     size_t alignment, bool zero, tcache_t *tcache, arena_t *arena)
 {
-
 	assert(ptr != NULL);
 	assert(size != 0);
 
@@ -1176,7 +1153,6 @@
 iralloc(tsd_t *tsd, extent_t *extent, void *ptr, size_t oldsize, size_t size,
     size_t alignment, bool zero)
 {
-
 	return (iralloct(tsd_tsdn(tsd), extent, ptr, oldsize, size, alignment,
 	    zero, tcache_get(tsd, true), NULL));
 }
@@ -1185,7 +1161,6 @@
 ixalloc(tsdn_t *tsdn, extent_t *extent, void *ptr, size_t oldsize, size_t size,
     size_t extra, size_t alignment, bool zero)
 {
-
 	assert(ptr != NULL);
 	assert(size != 0);
 
diff --git a/include/jemalloc/internal/jemalloc_internal_decls.h b/include/jemalloc/internal/jemalloc_internal_decls.h
index c907d91..277027f 100644
--- a/include/jemalloc/internal/jemalloc_internal_decls.h
+++ b/include/jemalloc/internal/jemalloc_internal_decls.h
@@ -63,7 +63,6 @@
 static int
 isblank(int c)
 {
-
 	return (c == '\t' || c == ' ');
 }
 #endif
diff --git a/include/jemalloc/internal/mutex_inlines.h b/include/jemalloc/internal/mutex_inlines.h
index b769f0c..d65fa13 100644
--- a/include/jemalloc/internal/mutex_inlines.h
+++ b/include/jemalloc/internal/mutex_inlines.h
@@ -12,7 +12,6 @@
 JEMALLOC_INLINE void
 malloc_mutex_lock(tsdn_t *tsdn, malloc_mutex_t *mutex)
 {
-
 	if (isthreaded) {
 		witness_assert_not_owner(tsdn, &mutex->witness);
 #ifdef _WIN32
@@ -35,7 +34,6 @@
 JEMALLOC_INLINE void
 malloc_mutex_unlock(tsdn_t *tsdn, malloc_mutex_t *mutex)
 {
-
 	if (isthreaded) {
 		witness_unlock(tsdn, &mutex->witness);
 #ifdef _WIN32
@@ -57,7 +55,6 @@
 JEMALLOC_INLINE void
 malloc_mutex_assert_owner(tsdn_t *tsdn, malloc_mutex_t *mutex)
 {
-
 	if (isthreaded)
 		witness_assert_owner(tsdn, &mutex->witness);
 }
@@ -65,7 +62,6 @@
 JEMALLOC_INLINE void
 malloc_mutex_assert_not_owner(tsdn_t *tsdn, malloc_mutex_t *mutex)
 {
-
 	if (isthreaded)
 		witness_assert_not_owner(tsdn, &mutex->witness);
 }
diff --git a/include/jemalloc/internal/ph.h b/include/jemalloc/internal/ph.h
index 4f91c33..9efb7b7 100644
--- a/include/jemalloc/internal/ph.h
+++ b/include/jemalloc/internal/ph.h
@@ -207,19 +207,16 @@
 a_attr void								\
 a_prefix##new(a_ph_type *ph)						\
 {									\
-									\
 	memset(ph, 0, sizeof(ph(a_type)));				\
 }									\
 a_attr bool								\
 a_prefix##empty(a_ph_type *ph)						\
 {									\
-									\
 	return (ph->ph_root == NULL);					\
 }									\
 a_attr a_type *								\
 a_prefix##first(a_ph_type *ph)						\
 {									\
-									\
 	if (ph->ph_root == NULL)					\
 		return (NULL);						\
 	ph_merge_aux(a_type, a_field, ph, a_cmp);			\
@@ -228,7 +225,6 @@
 a_attr void								\
 a_prefix##insert(a_ph_type *ph, a_type *phn)				\
 {									\
-									\
 	memset(&phn->a_field, 0, sizeof(phn(a_type)));			\
 									\
 	/*								\
diff --git a/include/jemalloc/internal/prng_inlines.h b/include/jemalloc/internal/prng_inlines.h
index b82a662..8cc19ce 100644
--- a/include/jemalloc/internal/prng_inlines.h
+++ b/include/jemalloc/internal/prng_inlines.h
@@ -20,21 +20,18 @@
 JEMALLOC_ALWAYS_INLINE uint32_t
 prng_state_next_u32(uint32_t state)
 {
-
 	return ((state * PRNG_A_32) + PRNG_C_32);
 }
 
 JEMALLOC_ALWAYS_INLINE uint64_t
 prng_state_next_u64(uint64_t state)
 {
-
 	return ((state * PRNG_A_64) + PRNG_C_64);
 }
 
 JEMALLOC_ALWAYS_INLINE size_t
 prng_state_next_zu(size_t state)
 {
-
 #if LG_SIZEOF_PTR == 2
 	return ((state * PRNG_A_32) + PRNG_C_32);
 #elif LG_SIZEOF_PTR == 3
diff --git a/include/jemalloc/internal/prof_inlines.h b/include/jemalloc/internal/prof_inlines.h
index 0b58042..a1ea7a3 100644
--- a/include/jemalloc/internal/prof_inlines.h
+++ b/include/jemalloc/internal/prof_inlines.h
@@ -29,7 +29,6 @@
 JEMALLOC_ALWAYS_INLINE bool
 prof_active_get_unlocked(void)
 {
-
 	/*
 	 * Even if opt_prof is true, sampling can be temporarily disabled by
 	 * setting prof_active to false.  No locking is used when reading
@@ -42,7 +41,6 @@
 JEMALLOC_ALWAYS_INLINE bool
 prof_gdump_get_unlocked(void)
 {
-
 	/*
 	 * No locking is used when reading prof_gdump_val in the fast path, so
 	 * there are no guarantees regarding how long it will take for all
@@ -78,7 +76,6 @@
 JEMALLOC_ALWAYS_INLINE prof_tctx_t *
 prof_tctx_get(tsdn_t *tsdn, const extent_t *extent, const void *ptr)
 {
-
 	cassert(config_prof);
 	assert(ptr != NULL);
 
@@ -89,7 +86,6 @@
 prof_tctx_set(tsdn_t *tsdn, extent_t *extent, const void *ptr, size_t usize,
     prof_tctx_t *tctx)
 {
-
 	cassert(config_prof);
 	assert(ptr != NULL);
 
@@ -100,7 +96,6 @@
 prof_tctx_reset(tsdn_t *tsdn, extent_t *extent, const void *ptr,
     prof_tctx_t *tctx)
 {
-
 	cassert(config_prof);
 	assert(ptr != NULL);
 
@@ -162,7 +157,6 @@
 prof_malloc(tsdn_t *tsdn, extent_t *extent, const void *ptr, size_t usize,
     prof_tctx_t *tctx)
 {
-
 	cassert(config_prof);
 	assert(ptr != NULL);
 	assert(usize == isalloc(tsdn, extent, ptr));
diff --git a/include/jemalloc/internal/rtree_inlines.h b/include/jemalloc/internal/rtree_inlines.h
index 7efba54..7e79a6a 100644
--- a/include/jemalloc/internal/rtree_inlines.h
+++ b/include/jemalloc/internal/rtree_inlines.h
@@ -74,7 +74,6 @@
 JEMALLOC_ALWAYS_INLINE uintptr_t
 rtree_subkey(rtree_t *rtree, uintptr_t key, unsigned level)
 {
-
 	return ((key >> ((ZU(1) << (LG_SIZEOF_PTR+3)) -
 	    rtree->levels[level].cumbits)) & ((ZU(1) <<
 	    rtree->levels[level].bits) - 1));
@@ -83,7 +82,6 @@
 JEMALLOC_ALWAYS_INLINE bool
 rtree_node_valid(rtree_elm_t *node)
 {
-
 	return ((uintptr_t)node != (uintptr_t)0);
 }
 
@@ -144,7 +142,6 @@
 JEMALLOC_INLINE void
 rtree_elm_write(rtree_elm_t *elm, const extent_t *extent)
 {
-
 	atomic_write_p(&elm->pun, extent);
 }
 
@@ -408,7 +405,6 @@
 rtree_elm_write_acquired(tsdn_t *tsdn, const rtree_t *rtree, rtree_elm_t *elm,
     const extent_t *extent)
 {
-
 	assert(((uintptr_t)extent & (uintptr_t)0x1) == (uintptr_t)0x0);
 	assert(((uintptr_t)elm->pun & (uintptr_t)0x1) == (uintptr_t)0x1);
 
@@ -422,7 +418,6 @@
 JEMALLOC_INLINE void
 rtree_elm_release(tsdn_t *tsdn, const rtree_t *rtree, rtree_elm_t *elm)
 {
-
 	rtree_elm_write(elm, rtree_elm_read_acquired(tsdn, rtree, elm));
 	if (config_debug)
 		rtree_elm_witness_release(tsdn, rtree, elm);
diff --git a/include/jemalloc/internal/spin_inlines.h b/include/jemalloc/internal/spin_inlines.h
index b10f67e..b4e779f 100644
--- a/include/jemalloc/internal/spin_inlines.h
+++ b/include/jemalloc/internal/spin_inlines.h
@@ -10,7 +10,6 @@
 JEMALLOC_INLINE void
 spin_init(spin_t *spin)
 {
-
 	spin->iteration = 0;
 }
 
diff --git a/include/jemalloc/internal/tcache_inlines.h b/include/jemalloc/internal/tcache_inlines.h
index e522d9e..2762b0e 100644
--- a/include/jemalloc/internal/tcache_inlines.h
+++ b/include/jemalloc/internal/tcache_inlines.h
@@ -88,7 +88,6 @@
 JEMALLOC_ALWAYS_INLINE void
 tcache_event(tsd_t *tsd, tcache_t *tcache)
 {
-
 	if (TCACHE_GC_INCR == 0)
 		return;
 
diff --git a/include/jemalloc/internal/ticker_inlines.h b/include/jemalloc/internal/ticker_inlines.h
index 42f37eb..1a4395f 100644
--- a/include/jemalloc/internal/ticker_inlines.h
+++ b/include/jemalloc/internal/ticker_inlines.h
@@ -13,7 +13,6 @@
 JEMALLOC_INLINE void
 ticker_init(ticker_t *ticker, int32_t nticks)
 {
-
 	ticker->tick = nticks;
 	ticker->nticks = nticks;
 }
@@ -21,21 +20,18 @@
 JEMALLOC_INLINE void
 ticker_copy(ticker_t *ticker, const ticker_t *other)
 {
-
 	*ticker = *other;
 }
 
 JEMALLOC_INLINE int32_t
 ticker_read(const ticker_t *ticker)
 {
-
 	return (ticker->tick);
 }
 
 JEMALLOC_INLINE bool
 ticker_ticks(ticker_t *ticker, int32_t nticks)
 {
-
 	if (unlikely(ticker->tick < nticks)) {
 		ticker->tick = ticker->nticks;
 		return (true);
@@ -47,7 +43,6 @@
 JEMALLOC_INLINE bool
 ticker_tick(ticker_t *ticker)
 {
-
 	return (ticker_ticks(ticker, 1));
 }
 #endif
diff --git a/include/jemalloc/internal/tsd_inlines.h b/include/jemalloc/internal/tsd_inlines.h
index ad915d1..0df21ad 100644
--- a/include/jemalloc/internal/tsd_inlines.h
+++ b/include/jemalloc/internal/tsd_inlines.h
@@ -51,21 +51,18 @@
 JEMALLOC_ALWAYS_INLINE tsd_t *
 tsd_fetch(void)
 {
-
 	return (tsd_fetch_impl(true));
 }
 
 JEMALLOC_ALWAYS_INLINE tsdn_t *
 tsd_tsdn(tsd_t *tsd)
 {
-
 	return ((tsdn_t *)tsd);
 }
 
 JEMALLOC_INLINE bool
 tsd_nominal(tsd_t *tsd)
 {
-
 	return (tsd->state == tsd_state_nominal);
 }
 
@@ -73,21 +70,18 @@
 JEMALLOC_ALWAYS_INLINE t *						\
 tsd_##n##p_get(tsd_t *tsd)						\
 {									\
-									\
 	return (&tsd->n);						\
 }									\
 									\
 JEMALLOC_ALWAYS_INLINE t						\
 tsd_##n##_get(tsd_t *tsd)						\
 {									\
-									\
 	return (*tsd_##n##p_get(tsd));					\
 }									\
 									\
 JEMALLOC_ALWAYS_INLINE void						\
 tsd_##n##_set(tsd_t *tsd, t n)						\
 {									\
-									\
 	assert(tsd->state == tsd_state_nominal);			\
 	tsd->n = n;							\
 }
@@ -97,7 +91,6 @@
 JEMALLOC_ALWAYS_INLINE tsdn_t *
 tsdn_fetch(void)
 {
-
 	if (!tsd_booted_get())
 		return (NULL);
 
@@ -107,14 +100,12 @@
 JEMALLOC_ALWAYS_INLINE bool
 tsdn_null(const tsdn_t *tsdn)
 {
-
 	return (tsdn == NULL);
 }
 
 JEMALLOC_ALWAYS_INLINE tsd_t *
 tsdn_tsd(tsdn_t *tsdn)
 {
-
 	assert(!tsdn_null(tsdn));
 
 	return (&tsdn->tsd);
@@ -123,7 +114,6 @@
 JEMALLOC_ALWAYS_INLINE rtree_ctx_t *
 tsdn_rtree_ctx(tsdn_t *tsdn, rtree_ctx_t *fallback)
 {
-
 	/*
 	 * If tsd cannot be accessed, initialize the fallback rtree_ctx and
 	 * return a pointer to it.
diff --git a/include/jemalloc/internal/tsd_types.h b/include/jemalloc/internal/tsd_types.h
index b48eaec..17e3da9 100644
--- a/include/jemalloc/internal/tsd_types.h
+++ b/include/jemalloc/internal/tsd_types.h
@@ -177,7 +177,6 @@
 a_attr bool								\
 a_name##tsd_cleanup_wrapper(void)					\
 {									\
-									\
 	if (a_name##tsd_initialized) {					\
 		a_name##tsd_initialized = false;			\
 		a_cleanup(&a_name##tsd_tls);				\
@@ -187,7 +186,6 @@
 a_attr bool								\
 a_name##tsd_boot0(void)							\
 {									\
-									\
 	if (a_cleanup != malloc_tsd_no_cleanup) {			\
 		malloc_tsd_cleanup_register(				\
 		    &a_name##tsd_cleanup_wrapper);			\
@@ -198,39 +196,33 @@
 a_attr void								\
 a_name##tsd_boot1(void)							\
 {									\
-									\
 	/* Do nothing. */						\
 }									\
 a_attr bool								\
 a_name##tsd_boot(void)							\
 {									\
-									\
 	return (a_name##tsd_boot0());					\
 }									\
 a_attr bool								\
 a_name##tsd_booted_get(void)						\
 {									\
-									\
 	return (a_name##tsd_booted);					\
 }									\
 a_attr bool								\
 a_name##tsd_get_allocates(void)						\
 {									\
-									\
 	return (false);							\
 }									\
 /* Get/set. */								\
 a_attr a_type *								\
 a_name##tsd_get(bool init)						\
 {									\
-									\
 	assert(a_name##tsd_booted);					\
 	return (&a_name##tsd_tls);					\
 }									\
 a_attr void								\
 a_name##tsd_set(a_type *val)						\
 {									\
-									\
 	assert(a_name##tsd_booted);					\
 	if (likely(&a_name##tsd_tls != val))				\
 		a_name##tsd_tls = (*val);				\
@@ -244,7 +236,6 @@
 a_attr bool								\
 a_name##tsd_boot0(void)							\
 {									\
-									\
 	if (a_cleanup != malloc_tsd_no_cleanup) {			\
 		if (pthread_key_create(&a_name##tsd_tsd, a_cleanup) !=	\
 		    0)							\
@@ -256,39 +247,33 @@
 a_attr void								\
 a_name##tsd_boot1(void)							\
 {									\
-									\
 	/* Do nothing. */						\
 }									\
 a_attr bool								\
 a_name##tsd_boot(void)							\
 {									\
-									\
 	return (a_name##tsd_boot0());					\
 }									\
 a_attr bool								\
 a_name##tsd_booted_get(void)						\
 {									\
-									\
 	return (a_name##tsd_booted);					\
 }									\
 a_attr bool								\
 a_name##tsd_get_allocates(void)						\
 {									\
-									\
 	return (false);							\
 }									\
 /* Get/set. */								\
 a_attr a_type *								\
 a_name##tsd_get(bool init)						\
 {									\
-									\
 	assert(a_name##tsd_booted);					\
 	return (&a_name##tsd_tls);					\
 }									\
 a_attr void								\
 a_name##tsd_set(a_type *val)						\
 {									\
-									\
 	assert(a_name##tsd_booted);					\
 	if (likely(&a_name##tsd_tls != val))				\
 		a_name##tsd_tls = (*val);				\
@@ -331,7 +316,6 @@
 a_attr void								\
 a_name##tsd_wrapper_set(a_name##tsd_wrapper_t *wrapper)			\
 {									\
-									\
 	if (!TlsSetValue(a_name##tsd_tsd, (void *)wrapper)) {		\
 		malloc_write("<jemalloc>: Error setting"		\
 		    " TSD for "#a_name"\n");				\
@@ -364,7 +348,6 @@
 a_attr bool								\
 a_name##tsd_boot0(void)							\
 {									\
-									\
 	a_name##tsd_tsd = TlsAlloc();					\
 	if (a_name##tsd_tsd == TLS_OUT_OF_INDEXES)			\
 		return (true);						\
@@ -394,7 +377,6 @@
 a_attr bool								\
 a_name##tsd_boot(void)							\
 {									\
-									\
 	if (a_name##tsd_boot0())					\
 		return (true);						\
 	a_name##tsd_boot1();						\
@@ -403,13 +385,11 @@
 a_attr bool								\
 a_name##tsd_booted_get(void)						\
 {									\
-									\
 	return (a_name##tsd_booted);					\
 }									\
 a_attr bool								\
 a_name##tsd_get_allocates(void)						\
 {									\
-									\
 	return (true);							\
 }									\
 /* Get/set. */								\
@@ -466,7 +446,6 @@
 a_attr void								\
 a_name##tsd_wrapper_set(a_name##tsd_wrapper_t *wrapper)			\
 {									\
-									\
 	if (pthread_setspecific(a_name##tsd_tsd,			\
 	    (void *)wrapper)) {						\
 		malloc_write("<jemalloc>: Error setting"		\
@@ -506,7 +485,6 @@
 a_attr bool								\
 a_name##tsd_boot0(void)							\
 {									\
-									\
 	if (pthread_key_create(&a_name##tsd_tsd,			\
 	    a_name##tsd_cleanup_wrapper) != 0)				\
 		return (true);						\
@@ -532,7 +510,6 @@
 a_attr bool								\
 a_name##tsd_boot(void)							\
 {									\
-									\
 	if (a_name##tsd_boot0())					\
 		return (true);						\
 	a_name##tsd_boot1();						\
@@ -541,13 +518,11 @@
 a_attr bool								\
 a_name##tsd_booted_get(void)						\
 {									\
-									\
 	return (a_name##tsd_booted);					\
 }									\
 a_attr bool								\
 a_name##tsd_get_allocates(void)						\
 {									\
-									\
 	return (true);							\
 }									\
 /* Get/set. */								\
diff --git a/include/jemalloc/internal/util_inlines.h b/include/jemalloc/internal/util_inlines.h
index 93f5b1d..4ceed06 100644
--- a/include/jemalloc/internal/util_inlines.h
+++ b/include/jemalloc/internal/util_inlines.h
@@ -27,28 +27,24 @@
 JEMALLOC_ALWAYS_INLINE unsigned
 ffs_llu(unsigned long long bitmap)
 {
-
 	return (JEMALLOC_INTERNAL_FFSLL(bitmap));
 }
 
 JEMALLOC_ALWAYS_INLINE unsigned
 ffs_lu(unsigned long bitmap)
 {
-
 	return (JEMALLOC_INTERNAL_FFSL(bitmap));
 }
 
 JEMALLOC_ALWAYS_INLINE unsigned
 ffs_u(unsigned bitmap)
 {
-
 	return (JEMALLOC_INTERNAL_FFS(bitmap));
 }
 
 JEMALLOC_ALWAYS_INLINE unsigned
 ffs_zu(size_t bitmap)
 {
-
 #if LG_SIZEOF_PTR == LG_SIZEOF_INT
 	return (ffs_u(bitmap));
 #elif LG_SIZEOF_PTR == LG_SIZEOF_LONG
@@ -63,7 +59,6 @@
 JEMALLOC_ALWAYS_INLINE unsigned
 ffs_u64(uint64_t bitmap)
 {
-
 #if LG_SIZEOF_LONG == 3
 	return (ffs_lu(bitmap));
 #elif LG_SIZEOF_LONG_LONG == 3
@@ -76,7 +71,6 @@
 JEMALLOC_ALWAYS_INLINE unsigned
 ffs_u32(uint32_t bitmap)
 {
-
 #if LG_SIZEOF_INT == 2
 	return (ffs_u(bitmap));
 #else
@@ -88,7 +82,6 @@
 JEMALLOC_INLINE uint64_t
 pow2_ceil_u64(uint64_t x)
 {
-
 	x--;
 	x |= x >> 1;
 	x |= x >> 2;
@@ -103,7 +96,6 @@
 JEMALLOC_INLINE uint32_t
 pow2_ceil_u32(uint32_t x)
 {
-
 	x--;
 	x |= x >> 1;
 	x |= x >> 2;
@@ -118,7 +110,6 @@
 JEMALLOC_INLINE size_t
 pow2_ceil_zu(size_t x)
 {
-
 #if (LG_SIZEOF_PTR == 3)
 	return (pow2_ceil_u64(x));
 #else
@@ -163,7 +154,6 @@
 JEMALLOC_INLINE unsigned
 lg_floor(size_t x)
 {
-
 	assert(x != 0);
 
 #if (LG_SIZEOF_PTR == LG_SIZEOF_INT)
@@ -178,7 +168,6 @@
 JEMALLOC_INLINE unsigned
 lg_floor(size_t x)
 {
-
 	assert(x != 0);
 
 	x |= (x >> 1);
@@ -200,7 +189,6 @@
 JEMALLOC_INLINE void
 set_errno(int errnum)
 {
-
 #ifdef _WIN32
 	SetLastError(errnum);
 #else
@@ -212,7 +200,6 @@
 JEMALLOC_INLINE int
 get_errno(void)
 {
-
 #ifdef _WIN32
 	return (GetLastError());
 #else
diff --git a/include/msvc_compat/strings.h b/include/msvc_compat/strings.h
index a3ee250..47998be 100644
--- a/include/msvc_compat/strings.h
+++ b/include/msvc_compat/strings.h
@@ -17,7 +17,6 @@
 
 static __forceinline int ffs(int x)
 {
-
 	return (ffsl(x));
 }
 
diff --git a/src/arena.c b/src/arena.c
index 1f0c4df..80af3f9 100644
--- a/src/arena.c
+++ b/src/arena.c
@@ -40,7 +40,6 @@
 static size_t
 arena_extent_dirty_npages(const extent_t *extent)
 {
-
 	return (extent_size_get(extent) >> LG_PAGE);
 }
 
@@ -76,7 +75,6 @@
 arena_extent_cache_dalloc_locked(tsdn_t *tsdn, arena_t *arena,
     extent_hooks_t **r_extent_hooks, extent_t *extent)
 {
-
 	malloc_mutex_assert_owner(tsdn, &arena->lock);
 
 	extent_dalloc_cache(tsdn, arena, r_extent_hooks, extent);
@@ -87,7 +85,6 @@
 arena_extent_cache_dalloc(tsdn_t *tsdn, arena_t *arena,
     extent_hooks_t **r_extent_hooks, extent_t *extent)
 {
-
 	malloc_mutex_lock(tsdn, &arena->lock);
 	arena_extent_cache_dalloc_locked(tsdn, arena, r_extent_hooks, extent);
 	malloc_mutex_unlock(tsdn, &arena->lock);
@@ -97,7 +94,6 @@
 arena_extent_cache_maybe_insert(tsdn_t *tsdn, arena_t *arena, extent_t *extent,
     bool cache)
 {
-
 	malloc_mutex_assert_owner(tsdn, &arena->extents_mtx);
 
 	if (cache) {
@@ -110,7 +106,6 @@
 arena_extent_cache_maybe_remove(tsdn_t *tsdn, arena_t *arena, extent_t *extent,
     bool dirty)
 {
-
 	malloc_mutex_assert_owner(tsdn, &arena->extents_mtx);
 
 	if (dirty) {
@@ -196,14 +191,12 @@
 static void
 arena_nactive_add(arena_t *arena, size_t add_pages)
 {
-
 	arena->nactive += add_pages;
 }
 
 static void
 arena_nactive_sub(arena_t *arena, size_t sub_pages)
 {
-
 	assert(arena->nactive >= sub_pages);
 	arena->nactive -= sub_pages;
 }
@@ -279,7 +272,6 @@
 static void
 arena_large_ralloc_stats_update(arena_t *arena, size_t oldusize, size_t usize)
 {
-
 	arena_large_dalloc_stats_update(arena, oldusize);
 	arena_large_malloc_stats_update(arena, usize);
 }
@@ -391,7 +383,6 @@
 static void
 arena_decay_deadline_init(arena_t *arena)
 {
-
 	/*
 	 * Generate a new deadline that is uniformly random within the next
 	 * epoch after the current one.
@@ -410,7 +401,6 @@
 static bool
 arena_decay_deadline_reached(const arena_t *arena, const nstime_t *time)
 {
-
 	return (nstime_compare(&arena->decay.deadline, time) <= 0);
 }
 
@@ -451,7 +441,6 @@
 static void
 arena_decay_backlog_update(arena_t *arena, uint64_t nadvance_u64)
 {
-
 	if (nadvance_u64 >= SMOOTHSTEP_NSTEPS) {
 		memset(arena->decay.backlog, 0, (SMOOTHSTEP_NSTEPS-1) *
 		    sizeof(size_t));
@@ -509,7 +498,6 @@
 static void
 arena_decay_epoch_advance(tsdn_t *tsdn, arena_t *arena, const nstime_t *time)
 {
-
 	arena_decay_epoch_advance_helper(arena, time);
 	arena_decay_epoch_advance_purge(tsdn, arena);
 }
@@ -517,7 +505,6 @@
 static void
 arena_decay_init(arena_t *arena, ssize_t decay_time)
 {
-
 	arena->decay.time = decay_time;
 	if (decay_time > 0) {
 		nstime_init2(&arena->decay.interval, decay_time, 0);
@@ -535,7 +522,6 @@
 static bool
 arena_decay_time_valid(ssize_t decay_time)
 {
-
 	if (decay_time < -1)
 		return (false);
 	if (decay_time == -1 || (uint64_t)decay_time <= NSTIME_SEC_MAX)
@@ -558,7 +544,6 @@
 bool
 arena_decay_time_set(tsdn_t *tsdn, arena_t *arena, ssize_t decay_time)
 {
-
 	if (!arena_decay_time_valid(decay_time))
 		return (true);
 
@@ -623,7 +608,6 @@
 void
 arena_maybe_purge(tsdn_t *tsdn, arena_t *arena)
 {
-
 	malloc_mutex_assert_owner(tsdn, &arena->lock);
 
 	/* Don't recursively purge. */
@@ -762,7 +746,6 @@
 void
 arena_purge(tsdn_t *tsdn, arena_t *arena, bool all)
 {
-
 	malloc_mutex_lock(tsdn, &arena->lock);
 	if (all)
 		arena_purge_to_limit(tsdn, arena, 0);
@@ -783,7 +766,6 @@
 static void
 arena_bin_slabs_nonfull_insert(arena_bin_t *bin, extent_t *slab)
 {
-
 	assert(extent_slab_data_get(slab)->nfree > 0);
 	extent_heap_insert(&bin->slabs_nonfull, slab);
 }
@@ -791,7 +773,6 @@
 static void
 arena_bin_slabs_nonfull_remove(arena_bin_t *bin, extent_t *slab)
 {
-
 	extent_heap_remove(&bin->slabs_nonfull, slab);
 }
 
@@ -809,7 +790,6 @@
 static void
 arena_bin_slabs_full_insert(arena_bin_t *bin, extent_t *slab)
 {
-
 	assert(extent_slab_data_get(slab)->nfree == 0);
 	extent_ring_insert(&bin->slabs_full, slab);
 }
@@ -817,7 +797,6 @@
 static void
 arena_bin_slabs_full_remove(extent_t *slab)
 {
-
 	extent_ring_remove(slab);
 }
 
@@ -935,7 +914,6 @@
 void
 arena_destroy(tsd_t *tsd, arena_t *arena)
 {
-
 	assert(base_ind_get(arena->base) >= narenas_auto);
 	assert(arena_nthreads_get(arena, false) == 0);
 	assert(arena_nthreads_get(arena, true) == 0);
@@ -1176,7 +1154,6 @@
 void
 arena_alloc_junk_small(void *ptr, const arena_bin_info_t *bin_info, bool zero)
 {
-
 	if (!zero)
 		memset(ptr, JEMALLOC_ALLOC_JUNK, bin_info->reg_size);
 }
@@ -1188,7 +1165,6 @@
 void
 arena_dalloc_junk_small(void *ptr, const arena_bin_info_t *bin_info)
 {
-
 	memset(ptr, JEMALLOC_FREE_JUNK, bin_info->reg_size);
 }
 #ifdef JEMALLOC_JET
@@ -1255,7 +1231,6 @@
 arena_malloc_hard(tsdn_t *tsdn, arena_t *arena, size_t size, szind_t ind,
     bool zero)
 {
-
 	assert(!tsdn_null(tsdn) || arena != NULL);
 
 	if (likely(!tsdn_null(tsdn)))
@@ -1320,7 +1295,6 @@
 static size_t
 arena_prof_demote(tsdn_t *tsdn, extent_t *extent, const void *ptr)
 {
-
 	cassert(config_prof);
 	assert(ptr != NULL);
 
@@ -1351,7 +1325,6 @@
 static void
 arena_dissociate_bin_slab(extent_t *slab, arena_bin_t *bin)
 {
-
 	/* Dissociate slab from bin. */
 	if (slab == bin->slabcur)
 		bin->slabcur = NULL;
@@ -1375,7 +1348,6 @@
 arena_dalloc_bin_slab(tsdn_t *tsdn, arena_t *arena, extent_t *slab,
     arena_bin_t *bin)
 {
-
 	assert(slab != bin->slabcur);
 
 	malloc_mutex_unlock(tsdn, &bin->lock);
@@ -1393,7 +1365,6 @@
 arena_bin_lower_slab(tsdn_t *tsdn, arena_t *arena, extent_t *slab,
     arena_bin_t *bin)
 {
-
 	assert(extent_slab_data_get(slab)->nfree > 0);
 
 	/*
@@ -1446,7 +1417,6 @@
 arena_dalloc_bin_junked_locked(tsdn_t *tsdn, arena_t *arena, extent_t *extent,
     void *ptr)
 {
-
 	arena_dalloc_bin_locked_impl(tsdn, arena, extent, ptr, true);
 }
 
@@ -1463,7 +1433,6 @@
 void
 arena_dalloc_small(tsdn_t *tsdn, arena_t *arena, extent_t *extent, void *ptr)
 {
-
 	arena_dalloc_bin(tsdn, arena, extent, ptr);
 	arena_decay_tick(tsdn, arena);
 }
@@ -1508,7 +1477,6 @@
 arena_ralloc_move_helper(tsdn_t *tsdn, arena_t *arena, size_t usize,
     size_t alignment, bool zero, tcache_t *tcache)
 {
-
 	if (alignment == 0)
 		return (arena_malloc(tsdn, arena, usize, size2index(usize),
 		    zero, tcache, true));
@@ -1575,7 +1543,6 @@
 bool
 arena_dss_prec_set(tsdn_t *tsdn, arena_t *arena, dss_prec_t dss_prec)
 {
-
 	if (!have_dss)
 		return (dss_prec != dss_prec_disabled);
 	malloc_mutex_lock(tsdn, &arena->lock);
@@ -1587,14 +1554,12 @@
 ssize_t
 arena_decay_time_default_get(void)
 {
-
 	return ((ssize_t)atomic_read_zu((size_t *)&decay_time_default));
 }
 
 bool
 arena_decay_time_default_set(ssize_t decay_time)
 {
-
 	if (!arena_decay_time_valid(decay_time))
 		return (true);
 	atomic_write_zu((size_t *)&decay_time_default, (size_t)decay_time);
@@ -1605,7 +1570,6 @@
 arena_basic_stats_merge_locked(arena_t *arena, unsigned *nthreads,
     const char **dss, ssize_t *decay_time, size_t *nactive, size_t *ndirty)
 {
-
 	*nthreads += arena_nthreads_get(arena, false);
 	*dss = dss_prec_names[arena->dss_prec];
 	*decay_time = arena->decay.time;
@@ -1617,7 +1581,6 @@
 arena_basic_stats_merge(tsdn_t *tsdn, arena_t *arena, unsigned *nthreads,
     const char **dss, ssize_t *decay_time, size_t *nactive, size_t *ndirty)
 {
-
 	malloc_mutex_lock(tsdn, &arena->lock);
 	arena_basic_stats_merge_locked(arena, nthreads, dss, decay_time,
 	    nactive, ndirty);
@@ -1686,28 +1649,24 @@
 unsigned
 arena_nthreads_get(arena_t *arena, bool internal)
 {
-
 	return (atomic_read_u(&arena->nthreads[internal]));
 }
 
 void
 arena_nthreads_inc(arena_t *arena, bool internal)
 {
-
 	atomic_add_u(&arena->nthreads[internal], 1);
 }
 
 void
 arena_nthreads_dec(arena_t *arena, bool internal)
 {
-
 	atomic_sub_u(&arena->nthreads[internal], 1);
 }
 
 size_t
 arena_extent_sn_next(arena_t *arena)
 {
-
 	return (atomic_add_zu(&arena->extent_sn_next, 1) - 1);
 }
 
@@ -1813,28 +1772,24 @@
 void
 arena_boot(void)
 {
-
 	arena_decay_time_default_set(opt_decay_time);
 }
 
 void
 arena_prefork0(tsdn_t *tsdn, arena_t *arena)
 {
-
 	malloc_mutex_prefork(tsdn, &arena->lock);
 }
 
 void
 arena_prefork1(tsdn_t *tsdn, arena_t *arena)
 {
-
 	malloc_mutex_prefork(tsdn, &arena->extents_mtx);
 }
 
 void
 arena_prefork2(tsdn_t *tsdn, arena_t *arena)
 {
-
 	malloc_mutex_prefork(tsdn, &arena->extent_cache_mtx);
 }
 
diff --git a/src/base.c b/src/base.c
index 5eab7cd..7c0ef2c 100644
--- a/src/base.c
+++ b/src/base.c
@@ -30,7 +30,6 @@
 static void
 base_unmap(extent_hooks_t *extent_hooks, unsigned ind, void *addr, size_t size)
 {
-
 	/*
 	 * Cascade through dalloc, decommit, purge_lazy, and purge_forced,
 	 * stopping at first success.  This cascade is performed for consistency
@@ -107,7 +106,6 @@
 base_extent_bump_alloc_post(tsdn_t *tsdn, base_t *base, extent_t *extent,
     size_t gap_size, void *addr, size_t size)
 {
-
 	if (extent_size_get(extent) > 0) {
 		/*
 		 * Compute the index for the largest size class that does not
@@ -202,7 +200,6 @@
 base_t *
 b0get(void)
 {
-
 	return (b0);
 }
 
@@ -263,7 +260,6 @@
 extent_hooks_t *
 base_extent_hooks_get(base_t *base)
 {
-
 	return ((extent_hooks_t *)atomic_read_p(&base->extent_hooks_pun));
 }
 
@@ -330,7 +326,6 @@
 base_stats_get(tsdn_t *tsdn, base_t *base, size_t *allocated, size_t *resident,
     size_t *mapped)
 {
-
 	cassert(config_stats);
 
 	malloc_mutex_lock(tsdn, &base->mtx);
@@ -345,28 +340,24 @@
 void
 base_prefork(tsdn_t *tsdn, base_t *base)
 {
-
 	malloc_mutex_prefork(tsdn, &base->mtx);
 }
 
 void
 base_postfork_parent(tsdn_t *tsdn, base_t *base)
 {
-
 	malloc_mutex_postfork_parent(tsdn, &base->mtx);
 }
 
 void
 base_postfork_child(tsdn_t *tsdn, base_t *base)
 {
-
 	malloc_mutex_postfork_child(tsdn, &base->mtx);
 }
 
 bool
 base_boot(tsdn_t *tsdn)
 {
-
 	b0 = base_new(tsdn, 0, (extent_hooks_t *)&extent_hooks_default);
 	return (b0 == NULL);
 }
diff --git a/src/bitmap.c b/src/bitmap.c
index 6655445..3d27f05 100644
--- a/src/bitmap.c
+++ b/src/bitmap.c
@@ -37,7 +37,6 @@
 static size_t
 bitmap_info_ngroups(const bitmap_info_t *binfo)
 {
-
 	return (binfo->levels[binfo->nlevels].group_offset);
 }
 
@@ -74,7 +73,6 @@
 void
 bitmap_info_init(bitmap_info_t *binfo, size_t nbits)
 {
-
 	assert(nbits > 0);
 	assert(nbits <= (ZU(1) << LG_BITMAP_MAXBITS));
 
@@ -85,7 +83,6 @@
 static size_t
 bitmap_info_ngroups(const bitmap_info_t *binfo)
 {
-
 	return (binfo->ngroups);
 }
 
@@ -106,6 +103,5 @@
 size_t
 bitmap_size(const bitmap_info_t *binfo)
 {
-
 	return (bitmap_info_ngroups(binfo) << LG_SIZEOF_BITMAP);
 }
diff --git a/src/ckh.c b/src/ckh.c
index 6f0f1e4..fe79862 100644
--- a/src/ckh.c
+++ b/src/ckh.c
@@ -411,7 +411,6 @@
 void
 ckh_delete(tsd_t *tsd, ckh_t *ckh)
 {
-
 	assert(ckh != NULL);
 
 #ifdef CKH_VERBOSE
@@ -435,7 +434,6 @@
 size_t
 ckh_count(ckh_t *ckh)
 {
-
 	assert(ckh != NULL);
 
 	return (ckh->count);
@@ -539,14 +537,12 @@
 void
 ckh_string_hash(const void *key, size_t r_hash[2])
 {
-
 	hash(key, strlen((const char *)key), 0x94122f33U, r_hash);
 }
 
 bool
 ckh_string_keycomp(const void *k1, const void *k2)
 {
-
 	assert(k1 != NULL);
 	assert(k2 != NULL);
 
@@ -569,6 +565,5 @@
 bool
 ckh_pointer_keycomp(const void *k1, const void *k2)
 {
-
 	return ((k1 == k2) ? true : false);
 }
diff --git a/src/ctl.c b/src/ctl.c
index 76fbce4..8484ba8 100644
--- a/src/ctl.c
+++ b/src/ctl.c
@@ -18,7 +18,6 @@
 JEMALLOC_INLINE_C const ctl_named_node_t *
 ctl_named_node(const ctl_node_t *node)
 {
-
 	return ((node->named) ? (const ctl_named_node_t *)node : NULL);
 }
 
@@ -33,7 +32,6 @@
 JEMALLOC_INLINE_C const ctl_indexed_node_t *
 ctl_indexed_node(const ctl_node_t *node)
 {
-
 	return (!node->named ? (const ctl_indexed_node_t *)node : NULL);
 }
 
@@ -475,7 +473,6 @@
 static unsigned
 stats_arenas_i2a(size_t i)
 {
-
 	return (stats_arenas_i2a_impl(i, true, false));
 }
 
@@ -513,7 +510,6 @@
 static void
 ctl_arena_clear(ctl_arena_stats_t *astats)
 {
-
 	astats->nthreads = 0;
 	astats->dss = dss_prec_names[dss_prec_limit];
 	astats->decay_time = -1;
@@ -985,7 +981,6 @@
 bool
 ctl_boot(void)
 {
-
 	if (malloc_mutex_init(&ctl_mtx, "ctl", WITNESS_RANK_CTL))
 		return (true);
 
@@ -997,21 +992,18 @@
 void
 ctl_prefork(tsdn_t *tsdn)
 {
-
 	malloc_mutex_prefork(tsdn, &ctl_mtx);
 }
 
 void
 ctl_postfork_parent(tsdn_t *tsdn)
 {
-
 	malloc_mutex_postfork_parent(tsdn, &ctl_mtx);
 }
 
 void
 ctl_postfork_child(tsdn_t *tsdn)
 {
-
 	malloc_mutex_postfork_child(tsdn, &ctl_mtx);
 }
 
@@ -1540,7 +1532,6 @@
 static void
 arena_i_purge(tsdn_t *tsdn, unsigned arena_ind, bool all)
 {
-
 	malloc_mutex_lock(tsdn, &ctl_mtx);
 	{
 		unsigned narenas = ctl_stats->narenas;
@@ -1918,7 +1909,6 @@
 static const ctl_named_node_t *
 arenas_bin_i_index(tsdn_t *tsdn, const size_t *mib, size_t miblen, size_t i)
 {
-
 	if (i > NBINS)
 		return (NULL);
 	return (super_arenas_bin_i_node);
@@ -1929,7 +1919,6 @@
 static const ctl_named_node_t *
 arenas_lextent_i_index(tsdn_t *tsdn, const size_t *mib, size_t miblen, size_t i)
 {
-
 	if (i > NSIZES - NBINS)
 		return (NULL);
 	return (super_arenas_lextent_i_node);
@@ -2159,7 +2148,6 @@
 stats_arenas_i_bins_j_index(tsdn_t *tsdn, const size_t *mib, size_t miblen,
     size_t j)
 {
-
 	if (j > NBINS)
 		return (NULL);
 	return (super_stats_arenas_i_bins_j_node);
@@ -2178,7 +2166,6 @@
 stats_arenas_i_lextents_j_index(tsdn_t *tsdn, const size_t *mib, size_t miblen,
     size_t j)
 {
-
 	if (j > NSIZES - NBINS)
 		return (NULL);
 	return (super_stats_arenas_i_lextents_j_node);
diff --git a/src/extent.c b/src/extent.c
index 7eb4970..73f79c1 100644
--- a/src/extent.c
+++ b/src/extent.c
@@ -94,7 +94,6 @@
 void
 extent_dalloc(tsdn_t *tsdn, arena_t *arena, extent_t *extent)
 {
-
 	malloc_mutex_lock(tsdn, &arena->extent_cache_mtx);
 	ql_elm_new(extent, ql_link);
 	ql_tail_insert(&arena->extent_cache, extent, ql_link);
@@ -104,21 +103,18 @@
 extent_hooks_t *
 extent_hooks_get(arena_t *arena)
 {
-
 	return (base_extent_hooks_get(arena->base));
 }
 
 extent_hooks_t *
 extent_hooks_set(arena_t *arena, extent_hooks_t *extent_hooks)
 {
-
 	return (base_extent_hooks_set(arena->base, extent_hooks));
 }
 
 static void
 extent_hooks_assure_initialized(arena_t *arena, extent_hooks_t **r_extent_hooks)
 {
-
 	if (*r_extent_hooks == EXTENT_HOOKS_INITIALIZER)
 		*r_extent_hooks = extent_hooks_get(arena);
 }
@@ -226,7 +222,6 @@
     const extent_t *extent, bool dependent, bool init_missing,
     rtree_elm_t **r_elm_a, rtree_elm_t **r_elm_b)
 {
-
 	*r_elm_a = rtree_elm_acquire(tsdn, &extents_rtree, rtree_ctx,
 	    (uintptr_t)extent_base_get(extent), dependent, init_missing);
 	if (!dependent && *r_elm_a == NULL)
@@ -252,7 +247,6 @@
 extent_rtree_write_acquired(tsdn_t *tsdn, rtree_elm_t *elm_a,
     rtree_elm_t *elm_b, const extent_t *extent)
 {
-
 	rtree_elm_write_acquired(tsdn, &extents_rtree, elm_a, extent);
 	if (elm_b != NULL)
 		rtree_elm_write_acquired(tsdn, &extents_rtree, elm_b, extent);
@@ -261,7 +255,6 @@
 static void
 extent_rtree_release(tsdn_t *tsdn, rtree_elm_t *elm_a, rtree_elm_t *elm_b)
 {
-
 	rtree_elm_release(tsdn, &extents_rtree, elm_a);
 	if (elm_b != NULL)
 		rtree_elm_release(tsdn, &extents_rtree, elm_b);
@@ -285,7 +278,6 @@
 static void
 extent_gprof_add(tsdn_t *tsdn, const extent_t *extent)
 {
-
 	cassert(config_prof);
 
 	if (opt_prof && extent_active_get(extent)) {
@@ -307,7 +299,6 @@
 static void
 extent_gprof_sub(tsdn_t *tsdn, const extent_t *extent)
 {
-
 	cassert(config_prof);
 
 	if (opt_prof && extent_active_get(extent)) {
@@ -406,7 +397,6 @@
 extent_leak(tsdn_t *tsdn, arena_t *arena, extent_hooks_t **r_extent_hooks,
     bool cache, extent_t *extent)
 {
-
 	/*
 	 * Leak extent after making sure its pages have already been purged, so
 	 * that this is only a virtual memory leak.
@@ -641,7 +631,6 @@
     extent_hooks_t **r_extent_hooks, void *new_addr, size_t usize, size_t pad,
     size_t alignment, bool *zero, bool *commit, bool slab)
 {
-
 	malloc_mutex_assert_owner(tsdn, &arena->extents_mtx);
 
 	return (extent_alloc_cache_impl(tsdn, arena, r_extent_hooks, true,
@@ -653,7 +642,6 @@
     extent_hooks_t **r_extent_hooks, void *new_addr, size_t usize, size_t pad,
     size_t alignment, bool *zero, bool *commit, bool slab)
 {
-
 	return (extent_alloc_cache_impl(tsdn, arena, r_extent_hooks, false,
 	    new_addr, usize, pad, alignment, zero, commit, slab));
 }
@@ -694,7 +682,6 @@
 extent_retain(tsdn_t *tsdn, arena_t *arena, extent_hooks_t **r_extent_hooks,
     extent_t *extent)
 {
-
 	if (config_stats)
 		arena->stats.retained += extent_size_get(extent);
 	extent_record(tsdn, arena, r_extent_hooks, arena->extents_retained,
@@ -906,7 +893,6 @@
 static bool
 extent_can_coalesce(const extent_t *a, const extent_t *b)
 {
-
 	if (extent_arena_get(a) != extent_arena_get(b))
 		return (false);
 	if (extent_active_get(a) != extent_active_get(b))
@@ -924,7 +910,6 @@
     extent_hooks_t **r_extent_hooks, extent_t *a, extent_t *b,
     extent_heap_t extent_heaps[NPSIZES+1], bool cache)
 {
-
 	if (!extent_can_coalesce(a, b))
 		return;
 
@@ -1008,7 +993,6 @@
 extent_dalloc_cache(tsdn_t *tsdn, arena_t *arena,
     extent_hooks_t **r_extent_hooks, extent_t *extent)
 {
-
 	assert(extent_base_get(extent) != NULL);
 	assert(extent_size_get(extent) != 0);
 
@@ -1022,7 +1006,6 @@
 static bool
 extent_dalloc_default_impl(void *addr, size_t size)
 {
-
 	if (!have_dss || !extent_in_dss(addr))
 		return (extent_dalloc_mmap(addr, size));
 	return (true);
@@ -1033,7 +1016,6 @@
 extent_dalloc_default(extent_hooks_t *extent_hooks, void *addr, size_t size,
     bool committed, unsigned arena_ind)
 {
-
 	assert(extent_hooks == &extent_hooks_default);
 
 	return (extent_dalloc_default_impl(addr, size));
@@ -1116,7 +1098,6 @@
 extent_commit_default(extent_hooks_t *extent_hooks, void *addr, size_t size,
     size_t offset, size_t length, unsigned arena_ind)
 {
-
 	assert(extent_hooks == &extent_hooks_default);
 
 	return (pages_commit((void *)((uintptr_t)addr + (uintptr_t)offset),
@@ -1142,7 +1123,6 @@
 extent_decommit_default(extent_hooks_t *extent_hooks, void *addr, size_t size,
     size_t offset, size_t length, unsigned arena_ind)
 {
-
 	assert(extent_hooks == &extent_hooks_default);
 
 	return (pages_decommit((void *)((uintptr_t)addr + (uintptr_t)offset),
@@ -1171,7 +1151,6 @@
 extent_purge_lazy_default(extent_hooks_t *extent_hooks, void *addr, size_t size,
     size_t offset, size_t length, unsigned arena_ind)
 {
-
 	assert(extent_hooks == &extent_hooks_default);
 	assert(addr != NULL);
 	assert((offset & PAGE_MASK) == 0);
@@ -1188,7 +1167,6 @@
     extent_hooks_t **r_extent_hooks, extent_t *extent, size_t offset,
     size_t length)
 {
-
 	extent_hooks_assure_initialized(arena, r_extent_hooks);
 	return ((*r_extent_hooks)->purge_lazy == NULL ||
 	    (*r_extent_hooks)->purge_lazy(*r_extent_hooks,
@@ -1201,7 +1179,6 @@
 extent_purge_forced_default(extent_hooks_t *extent_hooks, void *addr,
     size_t size, size_t offset, size_t length, unsigned arena_ind)
 {
-
 	assert(extent_hooks == &extent_hooks_default);
 	assert(addr != NULL);
 	assert((offset & PAGE_MASK) == 0);
@@ -1218,7 +1195,6 @@
     extent_hooks_t **r_extent_hooks, extent_t *extent, size_t offset,
     size_t length)
 {
-
 	extent_hooks_assure_initialized(arena, r_extent_hooks);
 	return ((*r_extent_hooks)->purge_forced == NULL ||
 	    (*r_extent_hooks)->purge_forced(*r_extent_hooks,
@@ -1231,7 +1207,6 @@
 extent_split_default(extent_hooks_t *extent_hooks, void *addr, size_t size,
     size_t size_a, size_t size_b, bool committed, unsigned arena_ind)
 {
-
 	assert(extent_hooks == &extent_hooks_default);
 
 	if (!maps_coalesce)
@@ -1310,7 +1285,6 @@
 static bool
 extent_merge_default_impl(void *addr_a, void *addr_b)
 {
-
 	if (!maps_coalesce)
 		return (true);
 	if (have_dss && !extent_dss_mergeable(addr_a, addr_b))
@@ -1324,7 +1298,6 @@
 extent_merge_default(extent_hooks_t *extent_hooks, void *addr_a, size_t size_a,
     void *addr_b, size_t size_b, bool committed, unsigned arena_ind)
 {
-
 	assert(extent_hooks == &extent_hooks_default);
 
 	return (extent_merge_default_impl(addr_a, addr_b));
@@ -1396,7 +1369,6 @@
 bool
 extent_boot(void)
 {
-
 	if (rtree_new(&extents_rtree, (unsigned)((ZU(1) << (LG_SIZEOF_PTR+3)) -
 	    LG_PAGE)))
 		return (true);
diff --git a/src/extent_dss.c b/src/extent_dss.c
index 0f0c689..5aa95b1 100644
--- a/src/extent_dss.c
+++ b/src/extent_dss.c
@@ -32,7 +32,6 @@
 static void *
 extent_dss_sbrk(intptr_t increment)
 {
-
 #ifdef JEMALLOC_DSS
 	return (sbrk(increment));
 #else
@@ -55,7 +54,6 @@
 bool
 extent_dss_prec_set(dss_prec_t dss_prec)
 {
-
 	if (!have_dss)
 		return (dss_prec != dss_prec_disabled);
 	atomic_write_u(&dss_prec_default, (unsigned)dss_prec);
@@ -208,7 +206,6 @@
 static bool
 extent_in_dss_helper(void *addr, void *max)
 {
-
 	return ((uintptr_t)addr >= (uintptr_t)dss_base && (uintptr_t)addr <
 	    (uintptr_t)max);
 }
@@ -216,7 +213,6 @@
 bool
 extent_in_dss(void *addr)
 {
-
 	cassert(have_dss);
 
 	return (extent_in_dss_helper(addr, atomic_read_p(&dss_max)));
@@ -241,7 +237,6 @@
 void
 extent_dss_boot(void)
 {
-
 	cassert(have_dss);
 
 	dss_base = extent_dss_sbrk(0);
diff --git a/src/extent_mmap.c b/src/extent_mmap.c
index 23dd4f8..e685a45 100644
--- a/src/extent_mmap.c
+++ b/src/extent_mmap.c
@@ -69,7 +69,6 @@
 bool
 extent_dalloc_mmap(void *addr, size_t size)
 {
-
 	if (config_munmap)
 		pages_unmap(addr, size);
 	return (!config_munmap);
diff --git a/src/jemalloc.c b/src/jemalloc.c
index a053983..1dc9183 100644
--- a/src/jemalloc.c
+++ b/src/jemalloc.c
@@ -211,7 +211,6 @@
 static void WINAPI
 _init_init_lock(void)
 {
-
 	/*
 	 * If another constructor in the same binary is using mallctl to e.g.
 	 * set up extent hooks, it may end up running before this one, and
@@ -276,14 +275,12 @@
 JEMALLOC_ALWAYS_INLINE_C bool
 malloc_initialized(void)
 {
-
 	return (malloc_init_state == malloc_init_initialized);
 }
 
 JEMALLOC_ALWAYS_INLINE_C bool
 malloc_init_a0(void)
 {
-
 	if (unlikely(malloc_init_state == malloc_init_uninitialized))
 		return (malloc_init_hard_a0());
 	return (false);
@@ -292,7 +289,6 @@
 JEMALLOC_ALWAYS_INLINE_C bool
 malloc_init(void)
 {
-
 	if (unlikely(!malloc_initialized()) && malloc_init_hard())
 		return (true);
 	return (false);
@@ -306,7 +302,6 @@
 static void *
 a0ialloc(size_t size, bool zero, bool is_internal)
 {
-
 	if (unlikely(malloc_init_a0()))
 		return (NULL);
 
@@ -317,21 +312,18 @@
 static void
 a0idalloc(extent_t *extent, void *ptr, bool is_internal)
 {
-
 	idalloctm(TSDN_NULL, extent, ptr, false, is_internal, true);
 }
 
 void *
 a0malloc(size_t size)
 {
-
 	return (a0ialloc(size, false, true));
 }
 
 void
 a0dalloc(void *ptr)
 {
-
 	a0idalloc(iealloc(NULL, ptr), ptr, true);
 }
 
@@ -344,7 +336,6 @@
 void *
 bootstrap_malloc(size_t size)
 {
-
 	if (unlikely(size == 0))
 		size = 1;
 
@@ -368,7 +359,6 @@
 void
 bootstrap_free(void *ptr)
 {
-
 	if (unlikely(ptr == NULL))
 		return;
 
@@ -378,28 +368,24 @@
 void
 arena_set(unsigned ind, arena_t *arena)
 {
-
 	atomic_write_p((void **)&arenas[ind], arena);
 }
 
 static void
 narenas_total_set(unsigned narenas)
 {
-
 	atomic_write_u(&narenas_total, narenas);
 }
 
 static void
 narenas_total_inc(void)
 {
-
 	atomic_add_u(&narenas_total, 1);
 }
 
 unsigned
 narenas_total_get(void)
 {
-
 	return (atomic_read_u(&narenas_total));
 }
 
@@ -689,7 +675,6 @@
 static void
 stats_print_atexit(void)
 {
-
 	if (config_tcache && config_stats) {
 		tsdn_t *tsdn;
 		unsigned narenas, i;
@@ -737,7 +722,6 @@
 static char *
 secure_getenv(const char *name)
 {
-
 #  ifdef JEMALLOC_HAVE_ISSETUGID
 	if (issetugid() != 0)
 		return (NULL);
@@ -855,7 +839,6 @@
 malloc_conf_error(const char *msg, const char *k, size_t klen, const char *v,
     size_t vlen)
 {
-
 	malloc_printf("<jemalloc>: %s: %.*s:%.*s\n", msg, (int)klen, k,
 	    (int)vlen, v);
 }
@@ -1167,7 +1150,6 @@
 static bool
 malloc_init_hard_needed(void)
 {
-
 	if (malloc_initialized() || (IS_INITIALIZER && malloc_init_state ==
 	    malloc_init_recursible)) {
 		/*
@@ -1197,7 +1179,6 @@
 static bool
 malloc_init_hard_a0_locked()
 {
-
 	malloc_initializer = INITIALIZER;
 
 	if (config_prof)
@@ -1261,7 +1242,6 @@
 static bool
 malloc_init_hard_recursible(void)
 {
-
 	malloc_init_state = malloc_init_recursible;
 
 	ncpus = malloc_ncpus();
@@ -1285,7 +1265,6 @@
 static bool
 malloc_init_hard_finish(tsdn_t *tsdn)
 {
-
 	if (malloc_mutex_boot())
 		return (true);
 
@@ -1458,7 +1437,6 @@
 ialloc_post_check(void *ret, tsdn_t *tsdn, size_t usize, const char *func,
     bool update_errno, bool slow_path)
 {
-
 	assert(!tsdn_null(tsdn) || ret == NULL);
 
 	if (unlikely(ret == NULL)) {
@@ -1617,7 +1595,6 @@
 JEMALLOC_ATTR(nonnull(1))
 je_posix_memalign(void **memptr, size_t alignment, size_t size)
 {
-
 	return (imemalign(memptr, alignment, size, sizeof(void *)));
 }
 
@@ -1754,7 +1731,6 @@
 isfree(tsd_t *tsd, extent_t *extent, void *ptr, size_t usize, tcache_t *tcache,
     bool slow_path)
 {
-
 	witness_assert_lockless(tsd_tsdn(tsd));
 
 	assert(ptr != NULL);
@@ -1850,7 +1826,6 @@
 JEMALLOC_EXPORT void JEMALLOC_NOTHROW
 je_free(void *ptr)
 {
-
 	UTRACE(ptr, 0, 0);
 	if (likely(ptr != NULL)) {
 		tsd_t *tsd = tsd_fetch();
@@ -1959,7 +1934,6 @@
 imallocx_flags_decode(tsd_t *tsd, size_t size, int flags, size_t *usize,
     size_t *alignment, bool *zero, tcache_t **tcache, arena_t **arena)
 {
-
 	if ((flags & MALLOCX_LG_ALIGN_MASK) == 0) {
 		*alignment = 0;
 		*usize = s2u(size);
@@ -2641,7 +2615,6 @@
 static void
 jemalloc_constructor(void)
 {
-
 	malloc_init();
 }
 #endif
diff --git a/src/jemalloc_cpp.cpp b/src/jemalloc_cpp.cpp
index 4d88f99..84d47ae 100644
--- a/src/jemalloc_cpp.cpp
+++ b/src/jemalloc_cpp.cpp
@@ -70,55 +70,47 @@
 void *
 operator new(std::size_t size)
 {
-
 	return (newImpl<false>(size));
 }
 
 void *
 operator new[](std::size_t size)
 {
-
 	return (newImpl<false>(size));
 }
 
 void *
 operator new(std::size_t size, const std::nothrow_t&) noexcept
 {
-
 	return (newImpl<true>(size));
 }
 
 void *
 operator new[](std::size_t size, const std::nothrow_t&) noexcept
 {
-
 	return (newImpl<true>(size));
 }
 
 void
 operator delete(void* ptr) noexcept
 {
-
 	je_free(ptr);
 }
 
 void
 operator delete[](void* ptr) noexcept
 {
-
 	je_free(ptr);
 }
 
 void
 operator delete(void* ptr, const std::nothrow_t&) noexcept
 {
-
 	je_free(ptr);
 }
 
 void operator delete[](void* ptr, const std::nothrow_t&) noexcept
 {
-
 	je_free(ptr);
 }
 
@@ -127,13 +119,11 @@
 void
 operator delete(void* ptr, std::size_t size) noexcept
 {
-
 	je_sdallocx(ptr, size, /*flags=*/0);
 }
 
 void operator delete[](void* ptr, std::size_t size) noexcept
 {
-
 	je_sdallocx(ptr, size, /*flags=*/0);
 }
 
diff --git a/src/large.c b/src/large.c
index ec22e64..9936b23 100644
--- a/src/large.c
+++ b/src/large.c
@@ -6,7 +6,6 @@
 void *
 large_malloc(tsdn_t *tsdn, arena_t *arena, size_t usize, bool zero)
 {
-
 	assert(usize == s2u(usize));
 
 	return (large_palloc(tsdn, arena, usize, CACHELINE, zero));
@@ -67,7 +66,6 @@
 void
 large_dalloc_junk(void *ptr, size_t usize)
 {
-
 	memset(ptr, JEMALLOC_FREE_JUNK, usize);
 }
 #ifdef JEMALLOC_JET
@@ -83,7 +81,6 @@
 void
 large_dalloc_maybe_junk(void *ptr, size_t usize)
 {
-
 	if (config_fill && have_dss && unlikely(opt_junk_free)) {
 		/*
 		 * Only bother junk filling if the extent isn't about to be
@@ -198,7 +195,6 @@
 large_ralloc_no_move(tsdn_t *tsdn, extent_t *extent, size_t usize_min,
     size_t usize_max, bool zero)
 {
-
 	assert(s2u(extent_usize_get(extent)) == extent_usize_get(extent));
 	/* The following should have been caught by callers. */
 	assert(usize_min > 0 && usize_max <= LARGE_MAXCLASS);
@@ -247,7 +243,6 @@
 large_ralloc_move_helper(tsdn_t *tsdn, arena_t *arena, size_t usize,
     size_t alignment, bool zero)
 {
-
 	if (alignment <= CACHELINE)
 		return (large_malloc(tsdn, arena, usize, zero));
 	return (large_palloc(tsdn, arena, usize, alignment, zero));
@@ -314,41 +309,35 @@
 void
 large_dalloc_junked_locked(tsdn_t *tsdn, extent_t *extent)
 {
-
 	large_dalloc_impl(tsdn, extent, true);
 }
 
 void
 large_dalloc(tsdn_t *tsdn, extent_t *extent)
 {
-
 	large_dalloc_impl(tsdn, extent, false);
 }
 
 size_t
 large_salloc(tsdn_t *tsdn, const extent_t *extent)
 {
-
 	return (extent_usize_get(extent));
 }
 
 prof_tctx_t *
 large_prof_tctx_get(tsdn_t *tsdn, const extent_t *extent)
 {
-
 	return (extent_prof_tctx_get(extent));
 }
 
 void
 large_prof_tctx_set(tsdn_t *tsdn, extent_t *extent, prof_tctx_t *tctx)
 {
-
 	extent_prof_tctx_set(extent, tctx);
 }
 
 void
 large_prof_tctx_reset(tsdn_t *tsdn, extent_t *extent)
 {
-
 	large_prof_tctx_set(tsdn, extent, (prof_tctx_t *)(uintptr_t)1U);
 }
diff --git a/src/mutex.c b/src/mutex.c
index b757ba8..bde536d 100644
--- a/src/mutex.c
+++ b/src/mutex.c
@@ -37,7 +37,6 @@
 static void
 pthread_create_once(void)
 {
-
 	pthread_create_fptr = dlsym(RTLD_NEXT, "pthread_create");
 	if (pthread_create_fptr == NULL) {
 		malloc_write("<jemalloc>: Error in dlsym(RTLD_NEXT, "
@@ -71,7 +70,6 @@
 bool
 malloc_mutex_init(malloc_mutex_t *mutex, const char *name, witness_rank_t rank)
 {
-
 #ifdef _WIN32
 #  if _WIN32_WINNT >= 0x0600
 	InitializeSRWLock(&mutex->lock);
@@ -113,21 +111,18 @@
 void
 malloc_mutex_prefork(tsdn_t *tsdn, malloc_mutex_t *mutex)
 {
-
 	malloc_mutex_lock(tsdn, mutex);
 }
 
 void
 malloc_mutex_postfork_parent(tsdn_t *tsdn, malloc_mutex_t *mutex)
 {
-
 	malloc_mutex_unlock(tsdn, mutex);
 }
 
 void
 malloc_mutex_postfork_child(tsdn_t *tsdn, malloc_mutex_t *mutex)
 {
-
 #ifdef JEMALLOC_MUTEX_INIT_CB
 	malloc_mutex_unlock(tsdn, mutex);
 #else
@@ -144,7 +139,6 @@
 bool
 malloc_mutex_boot(void)
 {
-
 #ifdef JEMALLOC_MUTEX_INIT_CB
 	postpone_init = false;
 	while (postponed_mutexes != NULL) {
diff --git a/src/nstime.c b/src/nstime.c
index 0948e29..57ebf2e 100644
--- a/src/nstime.c
+++ b/src/nstime.c
@@ -5,56 +5,48 @@
 void
 nstime_init(nstime_t *time, uint64_t ns)
 {
-
 	time->ns = ns;
 }
 
 void
 nstime_init2(nstime_t *time, uint64_t sec, uint64_t nsec)
 {
-
 	time->ns = sec * BILLION + nsec;
 }
 
 uint64_t
 nstime_ns(const nstime_t *time)
 {
-
 	return (time->ns);
 }
 
 uint64_t
 nstime_sec(const nstime_t *time)
 {
-
 	return (time->ns / BILLION);
 }
 
 uint64_t
 nstime_nsec(const nstime_t *time)
 {
-
 	return (time->ns % BILLION);
 }
 
 void
 nstime_copy(nstime_t *time, const nstime_t *source)
 {
-
 	*time = *source;
 }
 
 int
 nstime_compare(const nstime_t *a, const nstime_t *b)
 {
-
 	return ((a->ns > b->ns) - (a->ns < b->ns));
 }
 
 void
 nstime_add(nstime_t *time, const nstime_t *addend)
 {
-
 	assert(UINT64_MAX - time->ns >= addend->ns);
 
 	time->ns += addend->ns;
@@ -63,7 +55,6 @@
 void
 nstime_subtract(nstime_t *time, const nstime_t *subtrahend)
 {
-
 	assert(nstime_compare(time, subtrahend) >= 0);
 
 	time->ns -= subtrahend->ns;
@@ -72,7 +63,6 @@
 void
 nstime_imultiply(nstime_t *time, uint64_t multiplier)
 {
-
 	assert((((time->ns | multiplier) & (UINT64_MAX << (sizeof(uint64_t) <<
 	    2))) == 0) || ((time->ns * multiplier) / multiplier == time->ns));
 
@@ -82,7 +72,6 @@
 void
 nstime_idivide(nstime_t *time, uint64_t divisor)
 {
-
 	assert(divisor != 0);
 
 	time->ns /= divisor;
@@ -91,7 +80,6 @@
 uint64_t
 nstime_divide(const nstime_t *time, const nstime_t *divisor)
 {
-
 	assert(divisor->ns != 0);
 
 	return (time->ns / divisor->ns);
@@ -135,7 +123,6 @@
 static void
 nstime_get(nstime_t *time)
 {
-
 	nstime_init(time, mach_absolute_time());
 }
 #else
@@ -157,7 +144,6 @@
 bool
 nstime_monotonic(void)
 {
-
 	return (NSTIME_MONOTONIC);
 #undef NSTIME_MONOTONIC
 }
diff --git a/src/pages.c b/src/pages.c
index d5a0a21..7c26a28 100644
--- a/src/pages.c
+++ b/src/pages.c
@@ -64,7 +64,6 @@
 void
 pages_unmap(void *addr, size_t size)
 {
-
 #ifdef _WIN32
 	if (VirtualFree(addr, 0, MEM_RELEASE) == 0)
 #else
@@ -121,7 +120,6 @@
 static bool
 pages_commit_impl(void *addr, size_t size, bool commit)
 {
-
 	if (os_overcommits)
 		return (true);
 
@@ -151,21 +149,18 @@
 bool
 pages_commit(void *addr, size_t size)
 {
-
 	return (pages_commit_impl(addr, size, true));
 }
 
 bool
 pages_decommit(void *addr, size_t size)
 {
-
 	return (pages_commit_impl(addr, size, false));
 }
 
 bool
 pages_purge_lazy(void *addr, size_t size)
 {
-
 	if (!pages_can_purge_lazy)
 		return (true);
 
@@ -182,7 +177,6 @@
 bool
 pages_purge_forced(void *addr, size_t size)
 {
-
 	if (!pages_can_purge_forced)
 		return (true);
 
@@ -196,7 +190,6 @@
 bool
 pages_huge(void *addr, size_t size)
 {
-
 	assert(HUGEPAGE_ADDR2BASE(addr) == addr);
 	assert(HUGEPAGE_CEILING(size) == size);
 
@@ -210,7 +203,6 @@
 bool
 pages_nohuge(void *addr, size_t size)
 {
-
 	assert(HUGEPAGE_ADDR2BASE(addr) == addr);
 	assert(HUGEPAGE_CEILING(size) == size);
 
@@ -284,7 +276,6 @@
 void
 pages_boot(void)
 {
-
 #ifndef _WIN32
 	mmap_flags = MAP_PRIVATE | MAP_ANON;
 #endif
diff --git a/src/prof.c b/src/prof.c
index b9a9d65..237cbb5 100644
--- a/src/prof.c
+++ b/src/prof.c
@@ -226,7 +226,6 @@
 prof_malloc_sample_object(tsdn_t *tsdn, extent_t *extent, const void *ptr,
     size_t usize, prof_tctx_t *tctx)
 {
-
 	prof_tctx_set(tsdn, extent, ptr, usize, tctx);
 
 	malloc_mutex_lock(tsdn, tctx->tdata->lock);
@@ -243,7 +242,6 @@
 void
 prof_free_sampled_object(tsd_t *tsd, size_t usize, prof_tctx_t *tctx)
 {
-
 	malloc_mutex_lock(tsd_tsdn(tsd), tctx->tdata->lock);
 	assert(tctx->cnts.curobjs > 0);
 	assert(tctx->cnts.curbytes >= usize);
@@ -259,7 +257,6 @@
 void
 bt_init(prof_bt_t *bt, void **vec)
 {
-
 	cassert(config_prof);
 
 	bt->vec = vec;
@@ -269,7 +266,6 @@
 JEMALLOC_INLINE_C void
 prof_enter(tsd_t *tsd, prof_tdata_t *tdata)
 {
-
 	cassert(config_prof);
 	assert(tdata == prof_tdata_get(tsd, false));
 
@@ -284,7 +280,6 @@
 JEMALLOC_INLINE_C void
 prof_leave(tsd_t *tsd, prof_tdata_t *tdata)
 {
-
 	cassert(config_prof);
 	assert(tdata == prof_tdata_get(tsd, false));
 
@@ -326,7 +321,6 @@
 static _Unwind_Reason_Code
 prof_unwind_init_callback(struct _Unwind_Context *context, void *arg)
 {
-
 	cassert(config_prof);
 
 	return (_URC_NO_REASON);
@@ -525,7 +519,6 @@
 void
 prof_backtrace(prof_bt_t *bt)
 {
-
 	cassert(config_prof);
 	not_reached();
 }
@@ -542,7 +535,6 @@
 static malloc_mutex_t *
 prof_tdata_mutex_choose(uint64_t thr_uid)
 {
-
 	return (&tdata_locks[thr_uid % PROF_NTDATA_LOCKS]);
 }
 
@@ -576,7 +568,6 @@
 prof_gctx_try_destroy(tsd_t *tsd, prof_tdata_t *tdata_self, prof_gctx_t *gctx,
     prof_tdata_t *tdata)
 {
-
 	cassert(config_prof);
 
 	/*
@@ -612,7 +603,6 @@
 static bool
 prof_tctx_should_destroy(tsdn_t *tsdn, prof_tctx_t *tctx)
 {
-
 	malloc_mutex_assert_owner(tsdn, tctx->tdata->lock);
 
 	if (opt_prof_accum)
@@ -627,7 +617,6 @@
 static bool
 prof_gctx_should_destroy(prof_gctx_t *gctx)
 {
-
 	if (opt_prof_accum)
 		return (false);
 	if (!tctx_tree_empty(&gctx->tctxs))
@@ -1044,7 +1033,6 @@
 static void
 prof_tctx_merge_tdata(tsdn_t *tsdn, prof_tctx_t *tctx, prof_tdata_t *tdata)
 {
-
 	malloc_mutex_assert_owner(tsdn, tctx->tdata->lock);
 
 	malloc_mutex_lock(tsdn, tctx->gctx->lock);
@@ -1077,7 +1065,6 @@
 static void
 prof_tctx_merge_gctx(tsdn_t *tsdn, prof_tctx_t *tctx, prof_gctx_t *gctx)
 {
-
 	malloc_mutex_assert_owner(tsdn, gctx->lock);
 
 	gctx->cnt_summed.curobjs += tctx->dump_cnts.curobjs;
@@ -1173,7 +1160,6 @@
 static void
 prof_dump_gctx_prep(tsdn_t *tsdn, prof_gctx_t *gctx, prof_gctx_tree_t *gctxs)
 {
-
 	cassert(config_prof);
 
 	malloc_mutex_lock(tsdn, gctx->lock);
@@ -1421,7 +1407,6 @@
 static int
 prof_getpid(void)
 {
-
 #ifdef _WIN32
 	return (GetCurrentProcessId());
 #else
@@ -1491,7 +1476,6 @@
 prof_leakcheck(const prof_cnt_t *cnt_all, size_t leak_ngctx,
     const char *filename)
 {
-
 #ifdef JEMALLOC_PROF
 	/*
 	 * Scaling is equivalent AdjustSamples() in jeprof, but the result may
@@ -1640,7 +1624,6 @@
 static void
 prof_dump_filename(char *filename, char v, uint64_t vseq)
 {
-
 	cassert(config_prof);
 
 	if (vseq != VSEQ_INVALID) {
@@ -1844,7 +1827,6 @@
 prof_tdata_t *
 prof_tdata_init(tsd_t *tsd)
 {
-
 	return (prof_tdata_init_impl(tsd, prof_thr_uid_alloc(tsd_tsdn(tsd)), 0,
 	    NULL, prof_thread_active_init_get(tsd_tsdn(tsd))));
 }
@@ -1852,7 +1834,6 @@
 static bool
 prof_tdata_should_destroy_unlocked(prof_tdata_t *tdata, bool even_if_attached)
 {
-
 	if (tdata->attached && !even_if_attached)
 		return (false);
 	if (ckh_count(&tdata->bt2tctx) != 0)
@@ -1864,7 +1845,6 @@
 prof_tdata_should_destroy(tsdn_t *tsdn, prof_tdata_t *tdata,
     bool even_if_attached)
 {
-
 	malloc_mutex_assert_owner(tsdn, tdata->lock);
 
 	return (prof_tdata_should_destroy_unlocked(tdata, even_if_attached));
@@ -1874,7 +1854,6 @@
 prof_tdata_destroy_locked(tsd_t *tsd, prof_tdata_t *tdata,
     bool even_if_attached)
 {
-
 	malloc_mutex_assert_owner(tsd_tsdn(tsd), &tdatas_mtx);
 
 	tdata_tree_remove(&tdatas, tdata);
@@ -1893,7 +1872,6 @@
 static void
 prof_tdata_destroy(tsd_t *tsd, prof_tdata_t *tdata, bool even_if_attached)
 {
-
 	malloc_mutex_lock(tsd_tsdn(tsd), &tdatas_mtx);
 	prof_tdata_destroy_locked(tsd, tdata, even_if_attached);
 	malloc_mutex_unlock(tsd_tsdn(tsd), &tdatas_mtx);
@@ -2162,7 +2140,6 @@
 void
 prof_boot0(void)
 {
-
 	cassert(config_prof);
 
 	memcpy(opt_prof_prefix, PROF_PREFIX_DEFAULT,
@@ -2172,7 +2149,6 @@
 void
 prof_boot1(void)
 {
-
 	cassert(config_prof);
 
 	/*
@@ -2198,7 +2174,6 @@
 bool
 prof_boot2(tsd_t *tsd)
 {
-
 	cassert(config_prof);
 
 	if (opt_prof) {
@@ -2292,7 +2267,6 @@
 void
 prof_prefork0(tsdn_t *tsdn)
 {
-
 	if (opt_prof) {
 		unsigned i;
 
@@ -2309,7 +2283,6 @@
 void
 prof_prefork1(tsdn_t *tsdn)
 {
-
 	if (opt_prof) {
 		malloc_mutex_prefork(tsdn, &prof_active_mtx);
 		malloc_mutex_prefork(tsdn, &prof_dump_seq_mtx);
@@ -2322,7 +2295,6 @@
 void
 prof_postfork_parent(tsdn_t *tsdn)
 {
-
 	if (opt_prof) {
 		unsigned i;
 
@@ -2345,7 +2317,6 @@
 void
 prof_postfork_child(tsdn_t *tsdn)
 {
-
 	if (opt_prof) {
 		unsigned i;
 
diff --git a/src/rtree.c b/src/rtree.c
index fd5e85d..43f2165 100644
--- a/src/rtree.c
+++ b/src/rtree.c
@@ -4,7 +4,6 @@
 static unsigned
 hmin(unsigned ha, unsigned hb)
 {
-
 	return (ha < hb ? ha : hb);
 }
 
@@ -71,7 +70,6 @@
 static rtree_elm_t *
 rtree_node_alloc(tsdn_t *tsdn, rtree_t *rtree, size_t nelms)
 {
-
 	return ((rtree_elm_t *)base_alloc(tsdn, b0get(), nelms *
 	    sizeof(rtree_elm_t), CACHELINE));
 }
@@ -88,7 +86,6 @@
 UNUSED static void
 rtree_node_dalloc(tsdn_t *tsdn, rtree_t *rtree, rtree_elm_t *node)
 {
-
 	/* Nodes are never deleted during normal operation. */
 	not_reached();
 }
@@ -103,7 +100,6 @@
 rtree_delete_subtree(tsdn_t *tsdn, rtree_t *rtree, rtree_elm_t *node,
     unsigned level)
 {
-
 	if (level + 1 < rtree->height) {
 		size_t nchildren, i;
 
@@ -157,7 +153,6 @@
 rtree_elm_t *
 rtree_subtree_read_hard(tsdn_t *tsdn, rtree_t *rtree, unsigned level)
 {
-
 	return (rtree_node_init(tsdn, rtree, level,
 	    &rtree->levels[level].subtree));
 }
@@ -166,7 +161,6 @@
 rtree_child_read_hard(tsdn_t *tsdn, rtree_t *rtree, rtree_elm_t *elm,
     unsigned level)
 {
-
 	return (rtree_node_init(tsdn, rtree, level+1, &elm->child));
 }
 
diff --git a/src/tcache.c b/src/tcache.c
index fad5277..66e255d 100644
--- a/src/tcache.c
+++ b/src/tcache.c
@@ -26,7 +26,6 @@
 size_t
 tcache_salloc(tsdn_t *tsdn, const void *ptr)
 {
-
 	return (arena_salloc(tsdn, iealloc(tsdn, ptr), ptr));
 }
 
@@ -249,7 +248,6 @@
 static void
 tcache_arena_associate(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena)
 {
-
 	if (config_stats) {
 		/* Link into list of extant tcaches. */
 		malloc_mutex_lock(tsdn, &arena->lock);
@@ -262,7 +260,6 @@
 static void
 tcache_arena_dissociate(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena)
 {
-
 	if (config_stats) {
 		/* Unlink from list of extant tcaches. */
 		malloc_mutex_lock(tsdn, &arena->lock);
@@ -287,7 +284,6 @@
 tcache_arena_reassociate(tsdn_t *tsdn, tcache_t *tcache, arena_t *oldarena,
     arena_t *newarena)
 {
-
 	tcache_arena_dissociate(tsdn, tcache, oldarena);
 	tcache_arena_associate(tsdn, tcache, newarena);
 }
@@ -473,7 +469,6 @@
 static void
 tcaches_elm_flush(tsd_t *tsd, tcaches_t *elm)
 {
-
 	if (elm->tcache == NULL)
 		return;
 	tcache_destroy(tsd, elm->tcache);
@@ -483,7 +478,6 @@
 void
 tcaches_flush(tsd_t *tsd, unsigned ind)
 {
-
 	tcaches_elm_flush(tsd, &tcaches[ind]);
 }
 
diff --git a/src/tsd.c b/src/tsd.c
index 5d9fc9f..b4d7e79 100644
--- a/src/tsd.c
+++ b/src/tsd.c
@@ -14,21 +14,18 @@
 void *
 malloc_tsd_malloc(size_t size)
 {
-
 	return (a0malloc(CACHELINE_CEILING(size)));
 }
 
 void
 malloc_tsd_dalloc(void *wrapper)
 {
-
 	a0dalloc(wrapper);
 }
 
 void
 malloc_tsd_no_cleanup(void *arg)
 {
-
 	not_reached();
 }
 
@@ -61,7 +58,6 @@
 void
 malloc_tsd_cleanup_register(bool (*f)(void))
 {
-
 	assert(ncleanups < MALLOC_TSD_CLEANUPS_MAX);
 	cleanups[ncleanups] = f;
 	ncleanups++;
@@ -127,7 +123,6 @@
 void
 malloc_tsd_boot1(void)
 {
-
 	tsd_boot1();
 	*tsd_arenas_tdata_bypassp_get(tsd_fetch()) = false;
 }
@@ -136,7 +131,6 @@
 static BOOL WINAPI
 _tls_callback(HINSTANCE hinstDLL, DWORD fdwReason, LPVOID lpvReserved)
 {
-
 	switch (fdwReason) {
 #ifdef JEMALLOC_LAZY_LOCK
 	case DLL_THREAD_ATTACH:
@@ -194,7 +188,6 @@
 void
 tsd_init_finish(tsd_init_head_t *head, tsd_init_block_t *block)
 {
-
 	malloc_mutex_lock(TSDN_NULL, &head->lock);
 	ql_remove(&head->blocks, block, link);
 	malloc_mutex_unlock(TSDN_NULL, &head->lock);
diff --git a/src/util.c b/src/util.c
index dd8c236..c6ac4e1 100644
--- a/src/util.c
+++ b/src/util.c
@@ -48,7 +48,6 @@
 static void
 wrtmessage(void *cbopaque, const char *s)
 {
-
 #if defined(JEMALLOC_USE_SYSCALL) && defined(SYS_write)
 	/*
 	 * Use syscall(2) rather than write(2) when possible in order to avoid
@@ -74,7 +73,6 @@
 void
 malloc_write(const char *s)
 {
-
 	if (je_malloc_message != NULL)
 		je_malloc_message(NULL, s);
 	else
@@ -88,7 +86,6 @@
 int
 buferror(int err, char *buf, size_t buflen)
 {
-
 #ifdef _WIN32
 	FormatMessageA(FORMAT_MESSAGE_FROM_SYSTEM, NULL, err, 0,
 	    (LPSTR)buf, (DWORD)buflen, NULL);
@@ -292,7 +289,6 @@
 static char *
 o2s(uintmax_t x, bool alt_form, char *s, size_t *slen_p)
 {
-
 	s = u2s(x, 8, false, s, slen_p);
 	if (alt_form && *s != '0') {
 		s--;
@@ -305,7 +301,6 @@
 static char *
 x2s(uintmax_t x, bool alt_form, bool uppercase, char *s, size_t *slen_p)
 {
-
 	s = u2s(x, 16, uppercase, s, slen_p);
 	if (alt_form) {
 		s -= 2;
diff --git a/src/witness.c b/src/witness.c
index 0f5c0d7..ffc7e24 100644
--- a/src/witness.c
+++ b/src/witness.c
@@ -5,7 +5,6 @@
 witness_init(witness_t *witness, const char *name, witness_rank_t rank,
     witness_comp_t *comp, void *opaque)
 {
-
 	witness->name = name;
 	witness->rank = rank;
 	witness->comp = comp;
@@ -41,7 +40,6 @@
 void
 witness_owner_error(const witness_t *witness)
 {
-
 	malloc_printf("<jemalloc>: Should own %s(%u)\n", witness->name,
 	    witness->rank);
 	abort();
@@ -59,7 +57,6 @@
 void
 witness_not_owner_error(const witness_t *witness)
 {
-
 	malloc_printf("<jemalloc>: Should not own %s(%u)\n", witness->name,
 	    witness->rank);
 	abort();
@@ -97,7 +94,6 @@
 void
 witnesses_cleanup(tsd_t *tsd)
 {
-
 	witness_assert_lockless(tsd_tsdn(tsd));
 
 	/* Do nothing. */
@@ -106,14 +102,12 @@
 void
 witness_prefork(tsd_t *tsd)
 {
-
 	tsd_witness_fork_set(tsd, true);
 }
 
 void
 witness_postfork_parent(tsd_t *tsd)
 {
-
 	tsd_witness_fork_set(tsd, false);
 }
 
diff --git a/src/zone.c b/src/zone.c
index 66ba02b..1fcff64 100644
--- a/src/zone.c
+++ b/src/zone.c
@@ -47,7 +47,6 @@
 static size_t
 zone_size(malloc_zone_t *zone, void *ptr)
 {
-
 	/*
 	 * There appear to be places within Darwin (such as setenv(3)) that
 	 * cause calls to this function with pointers that *no* zone owns.  If
@@ -63,14 +62,12 @@
 static void *
 zone_malloc(malloc_zone_t *zone, size_t size)
 {
-
 	return (je_malloc(size));
 }
 
 static void *
 zone_calloc(malloc_zone_t *zone, size_t num, size_t size)
 {
-
 	return (je_calloc(num, size));
 }
 
@@ -87,7 +84,6 @@
 static void
 zone_free(malloc_zone_t *zone, void *ptr)
 {
-
 	if (ivsalloc(tsdn_fetch(), ptr) != 0) {
 		je_free(ptr);
 		return;
@@ -99,7 +95,6 @@
 static void *
 zone_realloc(malloc_zone_t *zone, void *ptr, size_t size)
 {
-
 	if (ivsalloc(tsdn_fetch(), ptr) != 0)
 		return (je_realloc(ptr, size));
 
@@ -138,7 +133,6 @@
 static void *
 zone_destroy(malloc_zone_t *zone)
 {
-
 	/* This function should never be called. */
 	not_reached();
 	return (NULL);
@@ -147,7 +141,6 @@
 static size_t
 zone_good_size(malloc_zone_t *zone, size_t size)
 {
-
 	if (size == 0)
 		size = 1;
 	return (s2u(size));
@@ -156,7 +149,6 @@
 static void
 zone_force_lock(malloc_zone_t *zone)
 {
-
 	if (isthreaded)
 		jemalloc_prefork();
 }
@@ -164,7 +156,6 @@
 static void
 zone_force_unlock(malloc_zone_t *zone)
 {
-
 	/*
 	 * Call jemalloc_postfork_child() rather than
 	 * jemalloc_postfork_parent(), because this function is executed by both
@@ -179,7 +170,6 @@
 static void
 zone_init(void)
 {
-
 	jemalloc_zone.size = (void *)zone_size;
 	jemalloc_zone.malloc = (void *)zone_malloc;
 	jemalloc_zone.calloc = (void *)zone_calloc;
@@ -297,7 +287,6 @@
 void
 zone_register(void)
 {
-
 	/*
 	 * If something else replaced the system default zone allocator, don't
 	 * register jemalloc's.
diff --git a/test/include/test/math.h b/test/include/test/math.h
index b057b29..1728d60 100644
--- a/test/include/test/math.h
+++ b/test/include/test/math.h
@@ -305,7 +305,6 @@
 JEMALLOC_INLINE double
 pt_gamma(double p, double shape, double scale, double ln_gamma_shape)
 {
-
 	return (pt_chi2(p, shape * 2.0, ln_gamma_shape) * 0.5 * scale);
 }
 #endif
diff --git a/test/include/test/mq.h b/test/include/test/mq.h
index 7c4df49..a974eb9 100644
--- a/test/include/test/mq.h
+++ b/test/include/test/mq.h
@@ -46,7 +46,6 @@
 a_attr void								\
 a_prefix##fini(a_mq_type *mq)						\
 {									\
-									\
 	mtx_fini(&mq->lock);						\
 }									\
 a_attr unsigned								\
diff --git a/test/integration/MALLOCX_ARENA.c b/test/integration/MALLOCX_ARENA.c
index 58032da..1d9e423 100644
--- a/test/integration/MALLOCX_ARENA.c
+++ b/test/integration/MALLOCX_ARENA.c
@@ -63,7 +63,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_MALLOCX_ARENA));
 }
diff --git a/test/integration/aligned_alloc.c b/test/integration/aligned_alloc.c
index 36fb699..52b69ac 100644
--- a/test/integration/aligned_alloc.c
+++ b/test/integration/aligned_alloc.c
@@ -10,7 +10,6 @@
 static void
 purge(void)
 {
-
 	assert_d_eq(mallctl("arena.0.purge", NULL, NULL, NULL, 0), 0,
 	    "Unexpected mallctl error");
 }
@@ -130,7 +129,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_alignment_errors,
 	    test_oom_errors,
diff --git a/test/integration/allocated.c b/test/integration/allocated.c
index 6ce145b..7570c52 100644
--- a/test/integration/allocated.c
+++ b/test/integration/allocated.c
@@ -98,7 +98,6 @@
 
 TEST_BEGIN(test_main_thread)
 {
-
 	thd_start(NULL);
 }
 TEST_END
@@ -115,7 +114,6 @@
 int
 main(void)
 {
-
 	/* Run tests multiple times to check for bad interactions. */
 	return (test(
 	    test_main_thread,
diff --git a/test/integration/cpp/basic.cpp b/test/integration/cpp/basic.cpp
index eeb93c4..4a87a3b 100644
--- a/test/integration/cpp/basic.cpp
+++ b/test/integration/cpp/basic.cpp
@@ -12,7 +12,6 @@
 int
 main()
 {
-
 	return (test(
 	    test_basic));
 }
diff --git a/test/integration/extent.c b/test/integration/extent.c
index e347b66..30849b0 100644
--- a/test/integration/extent.c
+++ b/test/integration/extent.c
@@ -176,7 +176,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_extent_manual_hook,
 	    test_extent_auto_hook));
diff --git a/test/integration/mallocx.c b/test/integration/mallocx.c
index 2298f72..7617b1b 100644
--- a/test/integration/mallocx.c
+++ b/test/integration/mallocx.c
@@ -20,7 +20,6 @@
 static unsigned
 get_nlarge(void)
 {
-
 	return (get_nsizes_impl("arenas.nlextents"));
 }
 
@@ -46,7 +45,6 @@
 static size_t
 get_large_size(size_t ind)
 {
-
 	return (get_size_impl("arenas.lextent.0.size", ind));
 }
 
@@ -58,7 +56,6 @@
 static void
 purge(void)
 {
-
 	assert_d_eq(mallctl("arena.0.purge", NULL, NULL, NULL, 0), 0,
 	    "Unexpected mallctl error");
 }
@@ -225,7 +222,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_overflow,
 	    test_oom,
diff --git a/test/integration/overflow.c b/test/integration/overflow.c
index 3e1e15f..ad867e7 100644
--- a/test/integration/overflow.c
+++ b/test/integration/overflow.c
@@ -43,7 +43,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_overflow));
 }
diff --git a/test/integration/posix_memalign.c b/test/integration/posix_memalign.c
index 9f3156a..dace10f 100644
--- a/test/integration/posix_memalign.c
+++ b/test/integration/posix_memalign.c
@@ -10,7 +10,6 @@
 static void
 purge(void)
 {
-
 	assert_d_eq(mallctl("arena.0.purge", NULL, NULL, NULL, 0), 0,
 	    "Unexpected mallctl error");
 }
@@ -124,7 +123,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_alignment_errors,
 	    test_oom_errors,
diff --git a/test/integration/rallocx.c b/test/integration/rallocx.c
index dd89e8c..0a8b50c 100644
--- a/test/integration/rallocx.c
+++ b/test/integration/rallocx.c
@@ -16,7 +16,6 @@
 static unsigned
 get_nlarge(void)
 {
-
 	return (get_nsizes_impl("arenas.nlextents"));
 }
 
@@ -42,7 +41,6 @@
 static size_t
 get_large_size(size_t ind)
 {
-
 	return (get_size_impl("arenas.lextent.0.size", ind));
 }
 
@@ -249,7 +247,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_grow_and_shrink,
 	    test_zero,
diff --git a/test/integration/sdallocx.c b/test/integration/sdallocx.c
index f92e058..5d0a8f8 100644
--- a/test/integration/sdallocx.c
+++ b/test/integration/sdallocx.c
@@ -50,7 +50,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_basic,
 	    test_alignment_and_size));
diff --git a/test/integration/thread_arena.c b/test/integration/thread_arena.c
index 7a35a63..cf8240d 100644
--- a/test/integration/thread_arena.c
+++ b/test/integration/thread_arena.c
@@ -75,7 +75,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_thread_arena));
 }
diff --git a/test/integration/thread_tcache_enabled.c b/test/integration/thread_tcache_enabled.c
index 2c2825e..1394371 100644
--- a/test/integration/thread_tcache_enabled.c
+++ b/test/integration/thread_tcache_enabled.c
@@ -86,7 +86,6 @@
 
 TEST_BEGIN(test_main_thread)
 {
-
 	thd_start(NULL);
 }
 TEST_END
@@ -103,7 +102,6 @@
 int
 main(void)
 {
-
 	/* Run tests multiple times to check for bad interactions. */
 	return (test(
 	    test_main_thread,
diff --git a/test/integration/xallocx.c b/test/integration/xallocx.c
index d35ca39..647404a 100644
--- a/test/integration/xallocx.c
+++ b/test/integration/xallocx.c
@@ -87,14 +87,12 @@
 static unsigned
 get_nsmall(void)
 {
-
 	return (get_nsizes_impl("arenas.nbins"));
 }
 
 static unsigned
 get_nlarge(void)
 {
-
 	return (get_nsizes_impl("arenas.nlextents"));
 }
 
@@ -120,14 +118,12 @@
 static size_t
 get_small_size(size_t ind)
 {
-
 	return (get_size_impl("arenas.bin.0.size", ind));
 }
 
 static size_t
 get_large_size(size_t ind)
 {
-
 	return (get_size_impl("arenas.lextent.0.size", ind));
 }
 
@@ -397,7 +393,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_same_size,
 	    test_extra_no_move,
diff --git a/test/src/btalloc.c b/test/src/btalloc.c
index 9a253d9..a78cb89 100644
--- a/test/src/btalloc.c
+++ b/test/src/btalloc.c
@@ -3,6 +3,5 @@
 void *
 btalloc(size_t size, unsigned bits)
 {
-
 	return (btalloc_0(size, bits));
 }
diff --git a/test/src/mq.c b/test/src/mq.c
index 40b31c1..47f362c 100644
--- a/test/src/mq.c
+++ b/test/src/mq.c
@@ -7,7 +7,6 @@
 void
 mq_nanosleep(unsigned ns)
 {
-
 	assert(ns <= 1000*1000*1000);
 
 #ifdef _WIN32
diff --git a/test/src/mtx.c b/test/src/mtx.c
index 8a5dfdd..bbfec4a 100644
--- a/test/src/mtx.c
+++ b/test/src/mtx.c
@@ -7,7 +7,6 @@
 bool
 mtx_init(mtx_t *mtx)
 {
-
 #ifdef _WIN32
 	if (!InitializeCriticalSectionAndSpinCount(&mtx->lock, _CRT_SPINCOUNT))
 		return (true);
@@ -33,7 +32,6 @@
 void
 mtx_fini(mtx_t *mtx)
 {
-
 #ifdef _WIN32
 #elif (defined(JEMALLOC_OS_UNFAIR_LOCK))
 #elif (defined(JEMALLOC_OSSPIN))
@@ -45,7 +43,6 @@
 void
 mtx_lock(mtx_t *mtx)
 {
-
 #ifdef _WIN32
 	EnterCriticalSection(&mtx->lock);
 #elif (defined(JEMALLOC_OS_UNFAIR_LOCK))
@@ -60,7 +57,6 @@
 void
 mtx_unlock(mtx_t *mtx)
 {
-
 #ifdef _WIN32
 	LeaveCriticalSection(&mtx->lock);
 #elif (defined(JEMALLOC_OS_UNFAIR_LOCK))
diff --git a/test/src/test.c b/test/src/test.c
index d70cc75..345cc1c 100644
--- a/test/src/test.c
+++ b/test/src/test.c
@@ -34,7 +34,6 @@
 static const char *
 test_status_string(test_status_t test_status)
 {
-
 	switch (test_status) {
 	case test_status_pass: return "pass";
 	case test_status_skip: return "skip";
@@ -46,7 +45,6 @@
 void
 p_test_init(const char *name)
 {
-
 	test_count++;
 	test_status = test_status_pass;
 	test_name = name;
@@ -55,7 +53,6 @@
 void
 p_test_fini(void)
 {
-
 	test_counts[test_status]++;
 	malloc_printf("%s: %s\n", test_name, test_status_string(test_status));
 }
@@ -127,7 +124,6 @@
 void
 p_test_fail(const char *prefix, const char *message)
 {
-
 	malloc_cprintf(NULL, NULL, "%s%s\n", prefix, message);
 	test_status = test_status_fail;
 }
diff --git a/test/src/thd.c b/test/src/thd.c
index c9d0065..e316708 100644
--- a/test/src/thd.c
+++ b/test/src/thd.c
@@ -13,7 +13,6 @@
 void
 thd_join(thd_t thd, void **ret)
 {
-
 	if (WaitForSingleObject(thd, INFINITE) == WAIT_OBJECT_0 && ret) {
 		DWORD exit_code;
 		GetExitCodeThread(thd, (LPDWORD) &exit_code);
@@ -25,7 +24,6 @@
 void
 thd_create(thd_t *thd, void *(*proc)(void *), void *arg)
 {
-
 	if (pthread_create(thd, NULL, proc, arg) != 0)
 		test_fail("Error in pthread_create()\n");
 }
@@ -33,7 +31,6 @@
 void
 thd_join(thd_t thd, void **ret)
 {
-
 	pthread_join(thd, ret);
 }
 #endif
diff --git a/test/src/timer.c b/test/src/timer.c
index 3c7e63a..82f69d0 100644
--- a/test/src/timer.c
+++ b/test/src/timer.c
@@ -3,7 +3,6 @@
 void
 timer_start(timedelta_t *timer)
 {
-
 	nstime_init(&timer->t0, 0);
 	nstime_update(&timer->t0);
 }
@@ -11,7 +10,6 @@
 void
 timer_stop(timedelta_t *timer)
 {
-
 	nstime_copy(&timer->t1, &timer->t0);
 	nstime_update(&timer->t1);
 }
diff --git a/test/stress/microbench.c b/test/stress/microbench.c
index 7dc45f8..c599d9d 100644
--- a/test/stress/microbench.c
+++ b/test/stress/microbench.c
@@ -65,7 +65,6 @@
 
 TEST_BEGIN(test_malloc_vs_mallocx)
 {
-
 	compare_funcs(10*1000*1000, 100*1000*1000, "malloc",
 	    malloc_free, "mallocx", mallocx_free);
 }
@@ -95,7 +94,6 @@
 
 TEST_BEGIN(test_free_vs_dallocx)
 {
-
 	compare_funcs(10*1000*1000, 100*1000*1000, "free", malloc_free,
 	    "dallocx", malloc_dallocx);
 }
@@ -103,7 +101,6 @@
 
 TEST_BEGIN(test_dallocx_vs_sdallocx)
 {
-
 	compare_funcs(10*1000*1000, 100*1000*1000, "dallocx", malloc_dallocx,
 	    "sdallocx", malloc_sdallocx);
 }
@@ -140,7 +137,6 @@
 
 TEST_BEGIN(test_mus_vs_sallocx)
 {
-
 	compare_funcs(10*1000*1000, 100*1000*1000, "malloc_usable_size",
 	    malloc_mus_free, "sallocx", malloc_sallocx_free);
 }
@@ -163,7 +159,6 @@
 
 TEST_BEGIN(test_sallocx_vs_nallocx)
 {
-
 	compare_funcs(10*1000*1000, 100*1000*1000, "sallocx",
 	    malloc_sallocx_free, "nallocx", malloc_nallocx_free);
 }
@@ -172,7 +167,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_malloc_vs_mallocx,
 	    test_free_vs_dallocx,
diff --git a/test/unit/SFMT.c b/test/unit/SFMT.c
index ba4be87..cf52670 100644
--- a/test/unit/SFMT.c
+++ b/test/unit/SFMT.c
@@ -1596,7 +1596,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_gen_rand_32,
 	    test_by_array_32,
diff --git a/test/unit/a0.c b/test/unit/a0.c
index b9ba45a..87f7e52 100644
--- a/test/unit/a0.c
+++ b/test/unit/a0.c
@@ -13,7 +13,6 @@
 int
 main(void)
 {
-
 	return (test_no_malloc_init(
 	    test_a0));
 }
diff --git a/test/unit/arena_reset.c b/test/unit/arena_reset.c
index 65ff103..257f972 100644
--- a/test/unit/arena_reset.c
+++ b/test/unit/arena_reset.c
@@ -20,14 +20,12 @@
 static unsigned
 get_nsmall(void)
 {
-
 	return (get_nsizes_impl("arenas.nbins"));
 }
 
 static unsigned
 get_nlarge(void)
 {
-
 	return (get_nsizes_impl("arenas.nlextents"));
 }
 
@@ -53,14 +51,12 @@
 static size_t
 get_small_size(size_t ind)
 {
-
 	return (get_size_impl("arenas.bin.0.size", ind));
 }
 
 static size_t
 get_large_size(size_t ind)
 {
-
 	return (get_size_impl("arenas.lextent.0.size", ind));
 }
 
@@ -164,14 +160,12 @@
 static void
 do_arena_reset(unsigned arena_ind)
 {
-
 	do_arena_reset_destroy("arena.0.reset", arena_ind);
 }
 
 static void
 do_arena_destroy(unsigned arena_ind)
 {
-
 	do_arena_reset_destroy("arena.0.destroy", arena_ind);
 }
 
@@ -214,7 +208,6 @@
 
 TEST_BEGIN(test_arena_destroy_initial)
 {
-
 	assert_false(arena_i_initialized(MALLCTL_ARENAS_DESTROYED, false),
 	    "Destroyed arena stats should not be initialized");
 }
@@ -269,7 +262,6 @@
 extent_dalloc_unmap(extent_hooks_t *extent_hooks, void *addr, size_t size,
     bool committed, unsigned arena_ind)
 {
-
 	TRACE_HOOK("%s(extent_hooks=%p, addr=%p, size=%zu, committed=%s, "
 	    "arena_ind=%u)\n", __func__, extent_hooks, addr, size, committed ?
 	    "true" : "false", arena_ind);
@@ -338,7 +330,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_arena_reset,
 	    test_arena_destroy_initial,
diff --git a/test/unit/atomic.c b/test/unit/atomic.c
index b8933a6..1d14368 100644
--- a/test/unit/atomic.c
+++ b/test/unit/atomic.c
@@ -68,7 +68,6 @@
 TEST_STRUCT(u64, uint64_t)
 TEST_BEGIN(test_atomic_u64)
 {
-
 #if !(LG_SIZEOF_PTR == 3 || LG_SIZEOF_INT == 3)
 	test_skip("64-bit atomic operations not supported");
 #else
@@ -80,7 +79,6 @@
 TEST_STRUCT(u32, uint32_t)
 TEST_BEGIN(test_atomic_u32)
 {
-
 	TEST_BODY(u32, uint32_t, uint32_t, u32, "#"FMTx32);
 }
 TEST_END
@@ -88,7 +86,6 @@
 TEST_STRUCT(p, void *)
 TEST_BEGIN(test_atomic_p)
 {
-
 	TEST_BODY(p, void *, uintptr_t, ptr, "p");
 }
 TEST_END
@@ -96,7 +93,6 @@
 TEST_STRUCT(zu, size_t)
 TEST_BEGIN(test_atomic_zu)
 {
-
 	TEST_BODY(zu, size_t, size_t, zu, "#zx");
 }
 TEST_END
@@ -104,7 +100,6 @@
 TEST_STRUCT(u, unsigned)
 TEST_BEGIN(test_atomic_u)
 {
-
 	TEST_BODY(u, unsigned, unsigned, u, "#x");
 }
 TEST_END
@@ -112,7 +107,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_atomic_u64,
 	    test_atomic_u32,
diff --git a/test/unit/base.c b/test/unit/base.c
index 8f97e8b..9aa43ea 100644
--- a/test/unit/base.c
+++ b/test/unit/base.c
@@ -208,7 +208,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_base_hooks_default,
 	    test_base_hooks_null,
diff --git a/test/unit/bitmap.c b/test/unit/bitmap.c
index 10d47c7..b502bfe 100644
--- a/test/unit/bitmap.c
+++ b/test/unit/bitmap.c
@@ -126,7 +126,6 @@
 
 TEST_BEGIN(test_bitmap_initializer)
 {
-
 #define	NB(nbits) {							\
 		if (nbits <= BITMAP_MAXBITS) {				\
 			bitmap_info_t binfo =				\
@@ -339,7 +338,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_bitmap_initializer,
 	    test_bitmap_size,
diff --git a/test/unit/ckh.c b/test/unit/ckh.c
index 2cbc226..1f57668 100644
--- a/test/unit/ckh.c
+++ b/test/unit/ckh.c
@@ -206,7 +206,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_new_delete,
 	    test_count_insert_search_remove,
diff --git a/test/unit/decay.c b/test/unit/decay.c
index 7efecf0..b3b1dd9 100644
--- a/test/unit/decay.c
+++ b/test/unit/decay.c
@@ -12,14 +12,12 @@
 static bool
 nstime_monotonic_mock(void)
 {
-
 	return (monotonic_mock);
 }
 
 static bool
 nstime_update_mock(nstime_t *time)
 {
-
 	nupdates_mock++;
 	if (monotonic_mock)
 		nstime_copy(time, &time_mock);
@@ -357,7 +355,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_decay_ticks,
 	    test_decay_ticker,
diff --git a/test/unit/extent_quantize.c b/test/unit/extent_quantize.c
index 43fa360..a5c1b7a 100644
--- a/test/unit/extent_quantize.c
+++ b/test/unit/extent_quantize.c
@@ -138,7 +138,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_small_extent_size,
 	    test_large_extent_size,
diff --git a/test/unit/fork.c b/test/unit/fork.c
index c530797..58091c6 100644
--- a/test/unit/fork.c
+++ b/test/unit/fork.c
@@ -58,7 +58,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_fork));
 }
diff --git a/test/unit/hash.c b/test/unit/hash.c
index 010c9d7..ff23777 100644
--- a/test/unit/hash.c
+++ b/test/unit/hash.c
@@ -38,7 +38,6 @@
 static int
 hash_variant_bits(hash_variant_t variant)
 {
-
 	switch (variant) {
 	case hash_variant_x86_32: return (32);
 	case hash_variant_x86_128: return (128);
@@ -50,7 +49,6 @@
 static const char *
 hash_variant_string(hash_variant_t variant)
 {
-
 	switch (variant) {
 	case hash_variant_x86_32: return ("hash_x86_32");
 	case hash_variant_x86_128: return ("hash_x86_128");
@@ -155,21 +153,18 @@
 
 TEST_BEGIN(test_hash_x86_32)
 {
-
 	hash_variant_verify(hash_variant_x86_32);
 }
 TEST_END
 
 TEST_BEGIN(test_hash_x86_128)
 {
-
 	hash_variant_verify(hash_variant_x86_128);
 }
 TEST_END
 
 TEST_BEGIN(test_hash_x64_128)
 {
-
 	hash_variant_verify(hash_variant_x64_128);
 }
 TEST_END
@@ -177,7 +172,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_hash_x86_32,
 	    test_hash_x86_128,
diff --git a/test/unit/junk.c b/test/unit/junk.c
index 680f0d2..5f34d05 100644
--- a/test/unit/junk.c
+++ b/test/unit/junk.c
@@ -17,7 +17,6 @@
 static void
 watch_junking(void *p)
 {
-
 	watch_for_junking = p;
 	saw_junking = false;
 }
@@ -55,7 +54,6 @@
 static void
 large_dalloc_maybe_junk_intercept(void *ptr, size_t usize)
 {
-
 	large_dalloc_maybe_junk_orig(ptr, usize);
 	if (ptr == watch_for_junking)
 		saw_junking = true;
@@ -130,7 +128,6 @@
 
 TEST_BEGIN(test_junk_small)
 {
-
 	test_skip_if(!config_fill);
 	test_junk(1, SMALL_MAXCLASS-1);
 }
@@ -138,7 +135,6 @@
 
 TEST_BEGIN(test_junk_large)
 {
-
 	test_skip_if(!config_fill);
 	test_junk(SMALL_MAXCLASS+1, (1U << (LG_LARGE_MINCLASS+1)));
 }
@@ -147,7 +143,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_junk_small,
 	    test_junk_large));
diff --git a/test/unit/mallctl.c b/test/unit/mallctl.c
index fbe76cb..5b734e1 100644
--- a/test/unit/mallctl.c
+++ b/test/unit/mallctl.c
@@ -121,7 +121,6 @@
 
 TEST_BEGIN(test_mallctl_config)
 {
-
 #define	TEST_MALLCTL_CONFIG(config, t) do {				\
 	t oldval;							\
 	size_t sz = sizeof(oldval);					\
@@ -551,7 +550,6 @@
 
 TEST_BEGIN(test_arenas_constants)
 {
-
 #define	TEST_ARENAS_CONSTANT(t, name, expected) do {			\
 	t name;								\
 	size_t sz = sizeof(t);						\
@@ -571,7 +569,6 @@
 
 TEST_BEGIN(test_arenas_bin_constants)
 {
-
 #define	TEST_ARENAS_BIN_CONSTANT(t, name, expected) do {		\
 	t name;								\
 	size_t sz = sizeof(t);						\
@@ -591,7 +588,6 @@
 
 TEST_BEGIN(test_arenas_lextent_constants)
 {
-
 #define	TEST_ARENAS_LEXTENT_CONSTANT(t, name, expected) do {		\
 	t name;								\
 	size_t sz = sizeof(t);						\
@@ -626,7 +622,6 @@
 
 TEST_BEGIN(test_stats_arenas)
 {
-
 #define	TEST_STATS_ARENAS(t, name) do {					\
 	t name;								\
 	size_t sz = sizeof(t);						\
@@ -647,7 +642,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_mallctl_errors,
 	    test_mallctlnametomib_errors,
diff --git a/test/unit/math.c b/test/unit/math.c
index adb72be..8e5ec61 100644
--- a/test/unit/math.c
+++ b/test/unit/math.c
@@ -387,7 +387,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_ln_gamma_factorial,
 	    test_ln_gamma_misc,
diff --git a/test/unit/mq.c b/test/unit/mq.c
index bde2a48..bd289c5 100644
--- a/test/unit/mq.c
+++ b/test/unit/mq.c
@@ -85,7 +85,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_mq_basic,
 	    test_mq_threaded));
diff --git a/test/unit/mtx.c b/test/unit/mtx.c
index 96ff694..2eccc98 100644
--- a/test/unit/mtx.c
+++ b/test/unit/mtx.c
@@ -53,7 +53,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_mtx_basic,
 	    test_mtx_race));
diff --git a/test/unit/nstime.c b/test/unit/nstime.c
index 0368bc2..6548ba2 100644
--- a/test/unit/nstime.c
+++ b/test/unit/nstime.c
@@ -178,7 +178,6 @@
 
 TEST_BEGIN(test_nstime_monotonic)
 {
-
 	nstime_monotonic();
 }
 TEST_END
@@ -211,7 +210,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_nstime_init,
 	    test_nstime_init2,
diff --git a/test/unit/pack.c b/test/unit/pack.c
index 81ded4e..316b6df 100644
--- a/test/unit/pack.c
+++ b/test/unit/pack.c
@@ -161,7 +161,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_pack));
 }
diff --git a/test/unit/pages.c b/test/unit/pages.c
index f297215..1e6add9 100644
--- a/test/unit/pages.c
+++ b/test/unit/pages.c
@@ -24,7 +24,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_pages_huge));
 }
diff --git a/test/unit/ph.c b/test/unit/ph.c
index da442f0..10bf99e 100644
--- a/test/unit/ph.c
+++ b/test/unit/ph.c
@@ -148,7 +148,6 @@
 static void
 node_remove(heap_t *heap, node_t *node)
 {
-
 	heap_remove(heap, node);
 
 	node->magic = 0;
@@ -283,7 +282,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_ph_empty,
 	    test_ph_random));
diff --git a/test/unit/prng.c b/test/unit/prng.c
index 80c9d73..f32d82a 100644
--- a/test/unit/prng.c
+++ b/test/unit/prng.c
@@ -114,35 +114,30 @@
 
 TEST_BEGIN(test_prng_lg_range_u32_nonatomic)
 {
-
 	test_prng_lg_range_u32(false);
 }
 TEST_END
 
 TEST_BEGIN(test_prng_lg_range_u32_atomic)
 {
-
 	test_prng_lg_range_u32(true);
 }
 TEST_END
 
 TEST_BEGIN(test_prng_lg_range_u64_nonatomic)
 {
-
 	test_prng_lg_range_u64();
 }
 TEST_END
 
 TEST_BEGIN(test_prng_lg_range_zu_nonatomic)
 {
-
 	test_prng_lg_range_zu(false);
 }
 TEST_END
 
 TEST_BEGIN(test_prng_lg_range_zu_atomic)
 {
-
 	test_prng_lg_range_zu(true);
 }
 TEST_END
@@ -212,35 +207,30 @@
 
 TEST_BEGIN(test_prng_range_u32_nonatomic)
 {
-
 	test_prng_range_u32(false);
 }
 TEST_END
 
 TEST_BEGIN(test_prng_range_u32_atomic)
 {
-
 	test_prng_range_u32(true);
 }
 TEST_END
 
 TEST_BEGIN(test_prng_range_u64_nonatomic)
 {
-
 	test_prng_range_u64();
 }
 TEST_END
 
 TEST_BEGIN(test_prng_range_zu_nonatomic)
 {
-
 	test_prng_range_zu(false);
 }
 TEST_END
 
 TEST_BEGIN(test_prng_range_zu_atomic)
 {
-
 	test_prng_range_zu(true);
 }
 TEST_END
@@ -248,7 +238,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_prng_lg_range_u32_nonatomic,
 	    test_prng_lg_range_u32_atomic,
diff --git a/test/unit/prof_accum.c b/test/unit/prof_accum.c
index d941b5b..41ebeea 100644
--- a/test/unit/prof_accum.c
+++ b/test/unit/prof_accum.c
@@ -24,7 +24,6 @@
 static void *
 alloc_from_permuted_backtrace(unsigned thd_ind, unsigned iteration)
 {
-
 	return (btalloc(1, thd_ind*NALLOCS_PER_THREAD + iteration));
 }
 
@@ -86,7 +85,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_idump));
 }
diff --git a/test/unit/prof_active.c b/test/unit/prof_active.c
index d00943a..d3b341d 100644
--- a/test/unit/prof_active.c
+++ b/test/unit/prof_active.c
@@ -38,7 +38,6 @@
 mallctl_prof_active_get_impl(bool prof_active_old_expected, const char *func,
     int line)
 {
-
 	mallctl_bool_get("prof.active", prof_active_old_expected, func, line);
 }
 #define	mallctl_prof_active_get(a)					\
@@ -48,7 +47,6 @@
 mallctl_prof_active_set_impl(bool prof_active_old_expected,
     bool prof_active_new, const char *func, int line)
 {
-
 	mallctl_bool_set("prof.active", prof_active_old_expected,
 	    prof_active_new, func, line);
 }
@@ -59,7 +57,6 @@
 mallctl_thread_prof_active_get_impl(bool thread_prof_active_old_expected,
     const char *func, int line)
 {
-
 	mallctl_bool_get("thread.prof.active", thread_prof_active_old_expected,
 	    func, line);
 }
@@ -70,7 +67,6 @@
 mallctl_thread_prof_active_set_impl(bool thread_prof_active_old_expected,
     bool thread_prof_active_new, const char *func, int line)
 {
-
 	mallctl_bool_set("thread.prof.active", thread_prof_active_old_expected,
 	    thread_prof_active_new, func, line);
 }
@@ -96,7 +92,6 @@
 
 TEST_BEGIN(test_prof_active)
 {
-
 	test_skip_if(!config_prof);
 
 	mallctl_prof_active_get(true);
@@ -131,7 +126,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_prof_active));
 }
diff --git a/test/unit/prof_gdump.c b/test/unit/prof_gdump.c
index cb99acd..53f7cad 100644
--- a/test/unit/prof_gdump.c
+++ b/test/unit/prof_gdump.c
@@ -76,7 +76,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_gdump));
 }
diff --git a/test/unit/prof_idump.c b/test/unit/prof_idump.c
index c293350..43824c6 100644
--- a/test/unit/prof_idump.c
+++ b/test/unit/prof_idump.c
@@ -53,7 +53,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_idump));
 }
diff --git a/test/unit/prof_reset.c b/test/unit/prof_reset.c
index 59d7079..cc13e37 100644
--- a/test/unit/prof_reset.c
+++ b/test/unit/prof_reset.c
@@ -19,7 +19,6 @@
 static void
 set_prof_active(bool active)
 {
-
 	assert_d_eq(mallctl("prof.active", NULL, NULL, (void *)&active,
 	    sizeof(active)), 0, "Unexpected mallctl failure");
 }
@@ -98,7 +97,6 @@
 prof_dump_header_intercept(tsdn_t *tsdn, bool propagate_err,
     const prof_cnt_t *cnt_all)
 {
-
 	prof_dump_header_intercepted = true;
 	memcpy(&cnt_all_copy, cnt_all, sizeof(prof_cnt_t));
 
@@ -292,7 +290,6 @@
 int
 main(void)
 {
-
 	/* Intercept dumping prior to running any tests. */
 	prof_dump_open = prof_dump_open_intercept;
 
diff --git a/test/unit/prof_thread_name.c b/test/unit/prof_thread_name.c
index 9ec5497..8699936 100644
--- a/test/unit/prof_thread_name.c
+++ b/test/unit/prof_thread_name.c
@@ -26,7 +26,6 @@
 mallctl_thread_name_set_impl(const char *thread_name, const char *func,
     int line)
 {
-
 	assert_d_eq(mallctl("thread.prof.name", NULL, NULL,
 	    (void *)&thread_name, sizeof(thread_name)), 0,
 	    "%s():%d: Unexpected mallctl failure reading thread.prof.name",
@@ -124,7 +123,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_prof_thread_name_validation,
 	    test_prof_thread_name_threaded));
diff --git a/test/unit/ql.c b/test/unit/ql.c
index 05fad45..2ebb450 100644
--- a/test/unit/ql.c
+++ b/test/unit/ql.c
@@ -198,7 +198,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_ql_empty,
 	    test_ql_tail_insert,
diff --git a/test/unit/qr.c b/test/unit/qr.c
index 8b764e1..7c9c102 100644
--- a/test/unit/qr.c
+++ b/test/unit/qr.c
@@ -238,7 +238,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_qr_one,
 	    test_qr_after_insert,
diff --git a/test/unit/rb.c b/test/unit/rb.c
index cf3d3a7..56e0021 100644
--- a/test/unit/rb.c
+++ b/test/unit/rb.c
@@ -347,7 +347,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_rb_empty,
 	    test_rb_random));
diff --git a/test/unit/rtree.c b/test/unit/rtree.c
index 03f4e26..d2f3705 100644
--- a/test/unit/rtree.c
+++ b/test/unit/rtree.c
@@ -24,7 +24,6 @@
 static void
 rtree_node_dalloc_intercept(tsdn_t *tsdn, rtree_t *rtree, rtree_elm_t *node)
 {
-
 	if (rtree != test_rtree) {
 		rtree_node_dalloc_orig(tsdn, rtree, node);
 		return;
@@ -283,7 +282,6 @@
 int
 main(void)
 {
-
 	rtree_node_alloc_orig = rtree_node_alloc;
 	rtree_node_alloc = rtree_node_alloc_intercept;
 	rtree_node_dalloc_orig = rtree_node_dalloc;
diff --git a/test/unit/size_classes.c b/test/unit/size_classes.c
index 9b47b20..f7c14bc 100644
--- a/test/unit/size_classes.c
+++ b/test/unit/size_classes.c
@@ -178,7 +178,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_size_classes,
 	    test_psize_classes,
diff --git a/test/unit/slab.c b/test/unit/slab.c
index 42e82a8..7e6a62f 100644
--- a/test/unit/slab.c
+++ b/test/unit/slab.c
@@ -29,7 +29,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_arena_slab_regind));
 }
diff --git a/test/unit/smoothstep.c b/test/unit/smoothstep.c
index 4cfb213..071aede 100644
--- a/test/unit/smoothstep.c
+++ b/test/unit/smoothstep.c
@@ -98,7 +98,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_smoothstep_integral,
 	    test_smoothstep_monotonic,
diff --git a/test/unit/stats.c b/test/unit/stats.c
index a99a88f..18856f1 100644
--- a/test/unit/stats.c
+++ b/test/unit/stats.c
@@ -120,7 +120,6 @@
 void *
 thd_start(void *arg)
 {
-
 	return (NULL);
 }
 
@@ -350,7 +349,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_stats_summary,
 	    test_stats_large,
diff --git a/test/unit/ticker.c b/test/unit/ticker.c
index e737020..b8af46c 100644
--- a/test/unit/ticker.c
+++ b/test/unit/ticker.c
@@ -68,7 +68,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_ticker_tick,
 	    test_ticker_ticks,
diff --git a/test/unit/tsd.c b/test/unit/tsd.c
index d5f96ac..5313ef8 100644
--- a/test/unit/tsd.c
+++ b/test/unit/tsd.c
@@ -78,7 +78,6 @@
 
 TEST_BEGIN(test_tsd_main_thread)
 {
-
 	thd_start((void *)(uintptr_t)0xa5f3e329);
 }
 TEST_END
@@ -98,7 +97,6 @@
 int
 main(void)
 {
-
 	/* Core tsd bootstrapping must happen prior to data_tsd_boot(). */
 	if (nallocx(1, 0) == 0) {
 		malloc_printf("Initialization error");
diff --git a/test/unit/util.c b/test/unit/util.c
index b1f9abd..b891a19 100644
--- a/test/unit/util.c
+++ b/test/unit/util.c
@@ -33,21 +33,18 @@
 
 TEST_BEGIN(test_pow2_ceil_u64)
 {
-
 	TEST_POW2_CEIL(uint64_t, u64, FMTu64);
 }
 TEST_END
 
 TEST_BEGIN(test_pow2_ceil_u32)
 {
-
 	TEST_POW2_CEIL(uint32_t, u32, FMTu32);
 }
 TEST_END
 
 TEST_BEGIN(test_pow2_ceil_zu)
 {
-
 	TEST_POW2_CEIL(size_t, zu, "zu");
 }
 TEST_END
@@ -307,7 +304,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_pow2_ceil_u64,
 	    test_pow2_ceil_u32,
diff --git a/test/unit/witness.c b/test/unit/witness.c
index 2b01203..1359398 100644
--- a/test/unit/witness.c
+++ b/test/unit/witness.c
@@ -14,35 +14,30 @@
 witness_lock_error_intercept(const witness_list_t *witnesses,
     const witness_t *witness)
 {
-
 	saw_lock_error = true;
 }
 
 static void
 witness_owner_error_intercept(const witness_t *witness)
 {
-
 	saw_owner_error = true;
 }
 
 static void
 witness_not_owner_error_intercept(const witness_t *witness)
 {
-
 	saw_not_owner_error = true;
 }
 
 static void
 witness_lockless_error_intercept(const witness_list_t *witnesses)
 {
-
 	saw_lockless_error = true;
 }
 
 static int
 witness_comp(const witness_t *a, void *oa, const witness_t *b, void *ob)
 {
-
 	assert_u_eq(a->rank, b->rank, "Witnesses should have equal rank");
 
 	assert(oa == (void *)a);
@@ -54,7 +49,6 @@
 static int
 witness_comp_reverse(const witness_t *a, void *oa, const witness_t *b, void *ob)
 {
-
 	assert_u_eq(a->rank, b->rank, "Witnesses should have equal rank");
 
 	assert(oa == (void *)a);
@@ -273,7 +267,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_witness,
 	    test_witness_comp,
diff --git a/test/unit/zero.c b/test/unit/zero.c
index c025c83..c752954 100644
--- a/test/unit/zero.c
+++ b/test/unit/zero.c
@@ -47,7 +47,6 @@
 
 TEST_BEGIN(test_zero_small)
 {
-
 	test_skip_if(!config_fill);
 	test_zero(1, SMALL_MAXCLASS-1);
 }
@@ -55,7 +54,6 @@
 
 TEST_BEGIN(test_zero_large)
 {
-
 	test_skip_if(!config_fill);
 	test_zero(SMALL_MAXCLASS+1, (1U << (LG_LARGE_MINCLASS+1)));
 }
@@ -64,7 +62,6 @@
 int
 main(void)
 {
-
 	return (test(
 	    test_zero_small,
 	    test_zero_large));