blob: b21ab23a7b4c311a16a8edf48e63b8e321476e32 [file] [log] [blame]
Dmitry Vyukov9fc0df82013-01-11 08:07:43 +00001//===-- sanitizer_quarantine.h ----------------------------------*- C++ -*-===//
2//
3// The LLVM Compiler Infrastructure
4//
5// This file is distributed under the University of Illinois Open Source
6// License. See LICENSE.TXT for details.
7//
8//===----------------------------------------------------------------------===//
9//
10// Memory quarantine for AddressSanitizer and potentially other tools.
11// Quarantine caches some specified amount of memory in per-thread caches,
12// then evicts to global FIFO queue. When the queue reaches specified threshold,
13// oldest memory is recycled.
14//
15//===----------------------------------------------------------------------===//
16
17#ifndef SANITIZER_QUARANTINE_H
18#define SANITIZER_QUARANTINE_H
19
20#include "sanitizer_internal_defs.h"
21#include "sanitizer_mutex.h"
Dmitry Vyukova61ec812013-01-11 11:03:35 +000022#include "sanitizer_list.h"
Dmitry Vyukov9fc0df82013-01-11 08:07:43 +000023
24namespace __sanitizer {
25
26template<typename Node> class QuarantineCache;
27
Dmitry Vyukova61ec812013-01-11 11:03:35 +000028struct QuarantineBatch {
29 static const uptr kSize = 1024;
30 QuarantineBatch *next;
31 uptr size;
32 uptr count;
33 void *batch[kSize];
34};
35
Dmitry Vyukov9fc0df82013-01-11 08:07:43 +000036// The callback interface is:
Dmitry Vyukova61ec812013-01-11 11:03:35 +000037// void Callback::Recycle(Node *ptr);
38// void *cb.Allocate(uptr size);
39// void cb.Deallocate(void *ptr);
Dmitry Vyukov9fc0df82013-01-11 08:07:43 +000040template<typename Callback, typename Node>
41class Quarantine {
42 public:
Dmitry Vyukova61ec812013-01-11 11:03:35 +000043 typedef QuarantineCache<Callback> Cache;
Dmitry Vyukov9fc0df82013-01-11 08:07:43 +000044
45 explicit Quarantine(LinkerInitialized)
46 : cache_(LINKER_INITIALIZED) {
47 }
48
49 void Init(uptr size, uptr cache_size) {
50 max_size_ = size;
Dmitry Vyukova61ec812013-01-11 11:03:35 +000051 min_size_ = size / 10 * 9; // 90% of max size.
Dmitry Vyukov9fc0df82013-01-11 08:07:43 +000052 max_cache_size_ = cache_size;
53 }
54
Dmitry Vyukova61ec812013-01-11 11:03:35 +000055 void Put(Cache *c, Callback cb, Node *ptr, uptr size) {
56 c->Enqueue(cb, ptr, size);
Dmitry Vyukov9fc0df82013-01-11 08:07:43 +000057 if (c->Size() > max_cache_size_)
58 Drain(c, cb);
59 }
60
Dmitry Vyukova61ec812013-01-11 11:03:35 +000061 void NOINLINE Drain(Cache *c, Callback cb) {
62 {
63 SpinMutexLock l(&cache_mutex_);
64 cache_.Transfer(c);
65 }
Dmitry Vyukovf7941952013-01-11 11:39:59 +000066 if (cache_.Size() > max_size_ && recycle_mutex_.TryLock())
67 Recycle(cb);
Dmitry Vyukov9fc0df82013-01-11 08:07:43 +000068 }
69
70 private:
Dmitry Vyukova61ec812013-01-11 11:03:35 +000071 // Read-only data.
72 char pad0_[kCacheLineSize];
Dmitry Vyukov9fc0df82013-01-11 08:07:43 +000073 uptr max_size_;
Dmitry Vyukova61ec812013-01-11 11:03:35 +000074 uptr min_size_;
Dmitry Vyukov9fc0df82013-01-11 08:07:43 +000075 uptr max_cache_size_;
Dmitry Vyukova61ec812013-01-11 11:03:35 +000076 char pad1_[kCacheLineSize];
77 SpinMutex cache_mutex_;
78 SpinMutex recycle_mutex_;
Dmitry Vyukov9fc0df82013-01-11 08:07:43 +000079 Cache cache_;
Dmitry Vyukova61ec812013-01-11 11:03:35 +000080 char pad2_[kCacheLineSize];
Dmitry Vyukovf7941952013-01-11 11:39:59 +000081
82 void NOINLINE Recycle(Callback cb) {
83 Cache tmp;
84 {
85 SpinMutexLock l(&cache_mutex_);
86 while (cache_.Size() > min_size_) {
87 QuarantineBatch *b = cache_.DequeueBatch();
88 tmp.EnqueueBatch(b);
89 }
90 }
91 recycle_mutex_.Unlock();
Dmitry Vyukovf99b94e2013-01-11 16:40:01 +000092 DoRecycle(&tmp, cb);
93 }
94
95 void NOINLINE DoRecycle(Cache *c, Callback cb) {
96 while (QuarantineBatch *b = c->DequeueBatch()) {
97 const uptr kPrefetch = 16;
98 for (uptr i = 0; i < kPrefetch; i++)
99 PREFETCH(b->batch[i]);
100 for (uptr i = 0; i < b->count; i++) {
101 PREFETCH(b->batch[i + kPrefetch]);
Dmitry Vyukovf7941952013-01-11 11:39:59 +0000102 cb.Recycle((Node*)b->batch[i]);
Dmitry Vyukovf99b94e2013-01-11 16:40:01 +0000103 }
Dmitry Vyukovf7941952013-01-11 11:39:59 +0000104 cb.Deallocate(b);
105 }
106 }
Dmitry Vyukov9fc0df82013-01-11 08:07:43 +0000107};
108
Dmitry Vyukova61ec812013-01-11 11:03:35 +0000109// Per-thread cache of memory blocks.
110template<typename Callback>
Dmitry Vyukov9fc0df82013-01-11 08:07:43 +0000111class QuarantineCache {
112 public:
113 explicit QuarantineCache(LinkerInitialized) {
114 }
115
Dmitry Vyukova61ec812013-01-11 11:03:35 +0000116 QuarantineCache()
117 : size_() {
118 list_.clear();
119 }
120
Dmitry Vyukov9fc0df82013-01-11 08:07:43 +0000121 uptr Size() const {
Dmitry Vyukova61ec812013-01-11 11:03:35 +0000122 return atomic_load(&size_, memory_order_relaxed);
Dmitry Vyukov9fc0df82013-01-11 08:07:43 +0000123 }
124
Dmitry Vyukova61ec812013-01-11 11:03:35 +0000125 void Enqueue(Callback cb, void *ptr, uptr size) {
126 if (list_.empty() || list_.back()->count == QuarantineBatch::kSize)
127 AllocBatch(cb);
128 QuarantineBatch *b = list_.back();
129 b->batch[b->count++] = ptr;
130 b->size += size;
131 SizeAdd(size);
Dmitry Vyukov9fc0df82013-01-11 08:07:43 +0000132 }
133
Dmitry Vyukova61ec812013-01-11 11:03:35 +0000134 void Transfer(QuarantineCache *c) {
135 list_.append_back(&c->list_);
136 SizeAdd(c->Size());
137 atomic_store(&c->size_, 0, memory_order_relaxed);
138 }
139
140 void EnqueueBatch(QuarantineBatch *b) {
141 list_.push_back(b);
142 SizeAdd(b->size);
143 }
144
145 QuarantineBatch *DequeueBatch() {
146 if (list_.empty())
Dmitry Vyukov9fc0df82013-01-11 08:07:43 +0000147 return 0;
Dmitry Vyukova61ec812013-01-11 11:03:35 +0000148 QuarantineBatch *b = list_.front();
149 list_.pop_front();
Timur Iskhodzhanove174cd12013-05-29 12:03:49 +0000150 // FIXME: should probably add SizeSub method?
151 // See https://code.google.com/p/thread-sanitizer/issues/detail?id=20
152 SizeAdd(0 - b->size);
Dmitry Vyukova61ec812013-01-11 11:03:35 +0000153 return b;
Dmitry Vyukov9fc0df82013-01-11 08:07:43 +0000154 }
155
156 private:
Dmitry Vyukova61ec812013-01-11 11:03:35 +0000157 IntrusiveList<QuarantineBatch> list_;
158 atomic_uintptr_t size_;
159
160 void SizeAdd(uptr add) {
161 atomic_store(&size_, Size() + add, memory_order_relaxed);
162 }
163
Timur Iskhodzhanov2b10d392013-02-08 12:02:00 +0000164 NOINLINE QuarantineBatch* AllocBatch(Callback cb) {
Dmitry Vyukova61ec812013-01-11 11:03:35 +0000165 QuarantineBatch *b = (QuarantineBatch *)cb.Allocate(sizeof(*b));
166 b->count = 0;
167 b->size = 0;
168 list_.push_back(b);
169 return b;
170 }
Dmitry Vyukov9fc0df82013-01-11 08:07:43 +0000171};
Alexey Samsonovba5e9962013-01-30 07:45:58 +0000172} // namespace __sanitizer
Dmitry Vyukov9fc0df82013-01-11 08:07:43 +0000173
174#endif // #ifndef SANITIZER_QUARANTINE_H