blob: 78efa0a197a9ec66467a1c1e666768039842c4e5 [file] [log] [blame]
Li Zefand0b6e042009-07-13 10:33:21 +08001#undef TRACE_SYSTEM
2#define TRACE_SYSTEM kmem
3
Steven Rostedtea20d922009-04-10 08:54:16 -04004#if !defined(_TRACE_KMEM_H) || defined(TRACE_HEADER_MULTI_READ)
Zhaolei02af61b2009-04-10 14:26:18 +08005#define _TRACE_KMEM_H
Eduard - Gabriel Munteanub9ce08c2008-08-10 20:14:03 +03006
7#include <linux/types.h>
Zhaoleifc182a42009-04-10 14:27:38 +08008#include <linux/tracepoint.h>
David Howellsa1ce3922012-10-02 18:01:25 +01009#include <trace/events/gfpflags.h>
Steven Rostedt62ba1802009-05-15 16:16:30 -040010
Li Zefan53d04222009-11-26 15:04:10 +080011DECLARE_EVENT_CLASS(kmem_alloc,
Steven Rostedtea20d922009-04-10 08:54:16 -040012
13 TP_PROTO(unsigned long call_site,
14 const void *ptr,
15 size_t bytes_req,
16 size_t bytes_alloc,
17 gfp_t gfp_flags),
18
19 TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags),
20
21 TP_STRUCT__entry(
22 __field( unsigned long, call_site )
23 __field( const void *, ptr )
24 __field( size_t, bytes_req )
25 __field( size_t, bytes_alloc )
26 __field( gfp_t, gfp_flags )
27 ),
28
29 TP_fast_assign(
30 __entry->call_site = call_site;
31 __entry->ptr = ptr;
32 __entry->bytes_req = bytes_req;
33 __entry->bytes_alloc = bytes_alloc;
34 __entry->gfp_flags = gfp_flags;
35 ),
36
Steven Rostedt62ba1802009-05-15 16:16:30 -040037 TP_printk("call_site=%lx ptr=%p bytes_req=%zu bytes_alloc=%zu gfp_flags=%s",
Steven Rostedtea20d922009-04-10 08:54:16 -040038 __entry->call_site,
39 __entry->ptr,
40 __entry->bytes_req,
41 __entry->bytes_alloc,
Steven Rostedt62ba1802009-05-15 16:16:30 -040042 show_gfp_flags(__entry->gfp_flags))
Steven Rostedtea20d922009-04-10 08:54:16 -040043);
44
Li Zefan53d04222009-11-26 15:04:10 +080045DEFINE_EVENT(kmem_alloc, kmalloc,
Steven Rostedtea20d922009-04-10 08:54:16 -040046
Li Zefan53d04222009-11-26 15:04:10 +080047 TP_PROTO(unsigned long call_site, const void *ptr,
48 size_t bytes_req, size_t bytes_alloc, gfp_t gfp_flags),
Steven Rostedtea20d922009-04-10 08:54:16 -040049
Li Zefan53d04222009-11-26 15:04:10 +080050 TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags)
Steven Rostedtea20d922009-04-10 08:54:16 -040051);
52
Li Zefan53d04222009-11-26 15:04:10 +080053DEFINE_EVENT(kmem_alloc, kmem_cache_alloc,
54
55 TP_PROTO(unsigned long call_site, const void *ptr,
56 size_t bytes_req, size_t bytes_alloc, gfp_t gfp_flags),
57
58 TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags)
59);
60
61DECLARE_EVENT_CLASS(kmem_alloc_node,
Steven Rostedtea20d922009-04-10 08:54:16 -040062
63 TP_PROTO(unsigned long call_site,
64 const void *ptr,
65 size_t bytes_req,
66 size_t bytes_alloc,
67 gfp_t gfp_flags,
68 int node),
69
70 TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags, node),
71
72 TP_STRUCT__entry(
73 __field( unsigned long, call_site )
74 __field( const void *, ptr )
75 __field( size_t, bytes_req )
76 __field( size_t, bytes_alloc )
77 __field( gfp_t, gfp_flags )
78 __field( int, node )
79 ),
80
81 TP_fast_assign(
82 __entry->call_site = call_site;
83 __entry->ptr = ptr;
84 __entry->bytes_req = bytes_req;
85 __entry->bytes_alloc = bytes_alloc;
86 __entry->gfp_flags = gfp_flags;
87 __entry->node = node;
88 ),
89
Steven Rostedt62ba1802009-05-15 16:16:30 -040090 TP_printk("call_site=%lx ptr=%p bytes_req=%zu bytes_alloc=%zu gfp_flags=%s node=%d",
Steven Rostedtea20d922009-04-10 08:54:16 -040091 __entry->call_site,
92 __entry->ptr,
93 __entry->bytes_req,
94 __entry->bytes_alloc,
Steven Rostedt62ba1802009-05-15 16:16:30 -040095 show_gfp_flags(__entry->gfp_flags),
Steven Rostedtea20d922009-04-10 08:54:16 -040096 __entry->node)
97);
98
Li Zefan53d04222009-11-26 15:04:10 +080099DEFINE_EVENT(kmem_alloc_node, kmalloc_node,
Steven Rostedtea20d922009-04-10 08:54:16 -0400100
Li Zefan53d04222009-11-26 15:04:10 +0800101 TP_PROTO(unsigned long call_site, const void *ptr,
102 size_t bytes_req, size_t bytes_alloc,
103 gfp_t gfp_flags, int node),
Steven Rostedtea20d922009-04-10 08:54:16 -0400104
Li Zefan53d04222009-11-26 15:04:10 +0800105 TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags, node)
Steven Rostedtea20d922009-04-10 08:54:16 -0400106);
107
Li Zefan53d04222009-11-26 15:04:10 +0800108DEFINE_EVENT(kmem_alloc_node, kmem_cache_alloc_node,
109
110 TP_PROTO(unsigned long call_site, const void *ptr,
111 size_t bytes_req, size_t bytes_alloc,
112 gfp_t gfp_flags, int node),
113
114 TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags, node)
115);
116
117DECLARE_EVENT_CLASS(kmem_free,
Steven Rostedtea20d922009-04-10 08:54:16 -0400118
119 TP_PROTO(unsigned long call_site, const void *ptr),
120
121 TP_ARGS(call_site, ptr),
122
123 TP_STRUCT__entry(
124 __field( unsigned long, call_site )
125 __field( const void *, ptr )
126 ),
127
128 TP_fast_assign(
129 __entry->call_site = call_site;
130 __entry->ptr = ptr;
131 ),
132
133 TP_printk("call_site=%lx ptr=%p", __entry->call_site, __entry->ptr)
134);
135
Li Zefan53d04222009-11-26 15:04:10 +0800136DEFINE_EVENT(kmem_free, kfree,
Steven Rostedtea20d922009-04-10 08:54:16 -0400137
138 TP_PROTO(unsigned long call_site, const void *ptr),
139
Li Zefan53d04222009-11-26 15:04:10 +0800140 TP_ARGS(call_site, ptr)
141);
Steven Rostedtea20d922009-04-10 08:54:16 -0400142
Shreyas B. Prabhue5feb1e2015-05-28 15:44:16 -0700143DEFINE_EVENT_CONDITION(kmem_free, kmem_cache_free,
Steven Rostedtea20d922009-04-10 08:54:16 -0400144
Li Zefan53d04222009-11-26 15:04:10 +0800145 TP_PROTO(unsigned long call_site, const void *ptr),
Steven Rostedtea20d922009-04-10 08:54:16 -0400146
Shreyas B. Prabhue5feb1e2015-05-28 15:44:16 -0700147 TP_ARGS(call_site, ptr),
148
149 /*
150 * This trace can be potentially called from an offlined cpu.
151 * Since trace points use RCU and RCU should not be used from
152 * offline cpus, filter such calls out.
153 * While this trace can be called from a preemptable section,
154 * it has no impact on the condition since tasks can migrate
155 * only from online cpus to other online cpus. Thus its safe
156 * to use raw_smp_processor_id.
157 */
158 TP_CONDITION(cpu_online(raw_smp_processor_id()))
Steven Rostedtea20d922009-04-10 08:54:16 -0400159);
Mel Gorman4b4f2782009-09-21 17:02:41 -0700160
Konstantin Khlebnikovb413d482012-01-10 15:07:09 -0800161TRACE_EVENT(mm_page_free,
Mel Gorman4b4f2782009-09-21 17:02:41 -0700162
163 TP_PROTO(struct page *page, unsigned int order),
164
165 TP_ARGS(page, order),
166
167 TP_STRUCT__entry(
Namhyung Kim9fdd8a82015-04-06 14:36:09 +0900168 __field( unsigned long, pfn )
Mel Gorman4b4f2782009-09-21 17:02:41 -0700169 __field( unsigned int, order )
170 ),
171
172 TP_fast_assign(
Namhyung Kim9fdd8a82015-04-06 14:36:09 +0900173 __entry->pfn = page_to_pfn(page);
Mel Gorman4b4f2782009-09-21 17:02:41 -0700174 __entry->order = order;
175 ),
176
177 TP_printk("page=%p pfn=%lu order=%d",
Namhyung Kim9fdd8a82015-04-06 14:36:09 +0900178 pfn_to_page(__entry->pfn),
179 __entry->pfn,
Mel Gorman4b4f2782009-09-21 17:02:41 -0700180 __entry->order)
181);
182
Konstantin Khlebnikovb413d482012-01-10 15:07:09 -0800183TRACE_EVENT(mm_page_free_batched,
Mel Gorman4b4f2782009-09-21 17:02:41 -0700184
185 TP_PROTO(struct page *page, int cold),
186
187 TP_ARGS(page, cold),
188
189 TP_STRUCT__entry(
Namhyung Kim9fdd8a82015-04-06 14:36:09 +0900190 __field( unsigned long, pfn )
Mel Gorman4b4f2782009-09-21 17:02:41 -0700191 __field( int, cold )
192 ),
193
194 TP_fast_assign(
Namhyung Kim9fdd8a82015-04-06 14:36:09 +0900195 __entry->pfn = page_to_pfn(page);
Mel Gorman4b4f2782009-09-21 17:02:41 -0700196 __entry->cold = cold;
197 ),
198
199 TP_printk("page=%p pfn=%lu order=0 cold=%d",
Namhyung Kim9fdd8a82015-04-06 14:36:09 +0900200 pfn_to_page(__entry->pfn),
201 __entry->pfn,
Mel Gorman4b4f2782009-09-21 17:02:41 -0700202 __entry->cold)
203);
204
205TRACE_EVENT(mm_page_alloc,
206
207 TP_PROTO(struct page *page, unsigned int order,
208 gfp_t gfp_flags, int migratetype),
209
210 TP_ARGS(page, order, gfp_flags, migratetype),
211
212 TP_STRUCT__entry(
Namhyung Kim9fdd8a82015-04-06 14:36:09 +0900213 __field( unsigned long, pfn )
Mel Gorman4b4f2782009-09-21 17:02:41 -0700214 __field( unsigned int, order )
215 __field( gfp_t, gfp_flags )
216 __field( int, migratetype )
217 ),
218
219 TP_fast_assign(
Namhyung Kim9fdd8a82015-04-06 14:36:09 +0900220 __entry->pfn = page ? page_to_pfn(page) : -1UL;
Mel Gorman4b4f2782009-09-21 17:02:41 -0700221 __entry->order = order;
222 __entry->gfp_flags = gfp_flags;
223 __entry->migratetype = migratetype;
224 ),
225
226 TP_printk("page=%p pfn=%lu order=%d migratetype=%d gfp_flags=%s",
Namhyung Kim9fdd8a82015-04-06 14:36:09 +0900227 __entry->pfn != -1UL ? pfn_to_page(__entry->pfn) : NULL,
228 __entry->pfn != -1UL ? __entry->pfn : 0,
Mel Gorman4b4f2782009-09-21 17:02:41 -0700229 __entry->order,
230 __entry->migratetype,
231 show_gfp_flags(__entry->gfp_flags))
232);
233
Li Zefan53d04222009-11-26 15:04:10 +0800234DECLARE_EVENT_CLASS(mm_page,
Mel Gorman0d3d0622009-09-21 17:02:44 -0700235
236 TP_PROTO(struct page *page, unsigned int order, int migratetype),
237
238 TP_ARGS(page, order, migratetype),
239
240 TP_STRUCT__entry(
Namhyung Kim9fdd8a82015-04-06 14:36:09 +0900241 __field( unsigned long, pfn )
Mel Gorman0d3d0622009-09-21 17:02:44 -0700242 __field( unsigned int, order )
243 __field( int, migratetype )
244 ),
245
246 TP_fast_assign(
Namhyung Kim9fdd8a82015-04-06 14:36:09 +0900247 __entry->pfn = page ? page_to_pfn(page) : -1UL;
Mel Gorman0d3d0622009-09-21 17:02:44 -0700248 __entry->order = order;
249 __entry->migratetype = migratetype;
250 ),
251
252 TP_printk("page=%p pfn=%lu order=%u migratetype=%d percpu_refill=%d",
Namhyung Kim9fdd8a82015-04-06 14:36:09 +0900253 __entry->pfn != -1UL ? pfn_to_page(__entry->pfn) : NULL,
254 __entry->pfn != -1UL ? __entry->pfn : 0,
Mel Gorman0d3d0622009-09-21 17:02:44 -0700255 __entry->order,
256 __entry->migratetype,
257 __entry->order == 0)
258);
259
Li Zefan53d04222009-11-26 15:04:10 +0800260DEFINE_EVENT(mm_page, mm_page_alloc_zone_locked,
Mel Gorman0d3d0622009-09-21 17:02:44 -0700261
Li Zefan53d04222009-11-26 15:04:10 +0800262 TP_PROTO(struct page *page, unsigned int order, int migratetype),
263
264 TP_ARGS(page, order, migratetype)
265);
266
267DEFINE_EVENT_PRINT(mm_page, mm_page_pcpu_drain,
268
269 TP_PROTO(struct page *page, unsigned int order, int migratetype),
Mel Gorman0d3d0622009-09-21 17:02:44 -0700270
271 TP_ARGS(page, order, migratetype),
272
Mel Gorman0d3d0622009-09-21 17:02:44 -0700273 TP_printk("page=%p pfn=%lu order=%d migratetype=%d",
Namhyung Kim9fdd8a82015-04-06 14:36:09 +0900274 pfn_to_page(__entry->pfn), __entry->pfn,
Li Zefan53d04222009-11-26 15:04:10 +0800275 __entry->order, __entry->migratetype)
Mel Gorman0d3d0622009-09-21 17:02:44 -0700276);
277
Mel Gormane0fff1b2009-09-21 17:02:42 -0700278TRACE_EVENT(mm_page_alloc_extfrag,
279
280 TP_PROTO(struct page *page,
KOSAKI Motohiro52c8f6a2013-11-12 15:08:19 -0800281 int alloc_order, int fallback_order,
Vlastimil Babka99592d52015-02-11 15:28:15 -0800282 int alloc_migratetype, int fallback_migratetype),
Mel Gormane0fff1b2009-09-21 17:02:42 -0700283
284 TP_ARGS(page,
285 alloc_order, fallback_order,
Vlastimil Babka99592d52015-02-11 15:28:15 -0800286 alloc_migratetype, fallback_migratetype),
Mel Gormane0fff1b2009-09-21 17:02:42 -0700287
288 TP_STRUCT__entry(
Namhyung Kim9fdd8a82015-04-06 14:36:09 +0900289 __field( unsigned long, pfn )
Mel Gormane0fff1b2009-09-21 17:02:42 -0700290 __field( int, alloc_order )
291 __field( int, fallback_order )
292 __field( int, alloc_migratetype )
293 __field( int, fallback_migratetype )
Srivatsa S. Bhatf92310c2013-09-11 14:20:36 -0700294 __field( int, change_ownership )
Mel Gormane0fff1b2009-09-21 17:02:42 -0700295 ),
296
297 TP_fast_assign(
Namhyung Kim9fdd8a82015-04-06 14:36:09 +0900298 __entry->pfn = page_to_pfn(page);
Mel Gormane0fff1b2009-09-21 17:02:42 -0700299 __entry->alloc_order = alloc_order;
300 __entry->fallback_order = fallback_order;
301 __entry->alloc_migratetype = alloc_migratetype;
302 __entry->fallback_migratetype = fallback_migratetype;
Vlastimil Babka99592d52015-02-11 15:28:15 -0800303 __entry->change_ownership = (alloc_migratetype ==
304 get_pageblock_migratetype(page));
Mel Gormane0fff1b2009-09-21 17:02:42 -0700305 ),
306
307 TP_printk("page=%p pfn=%lu alloc_order=%d fallback_order=%d pageblock_order=%d alloc_migratetype=%d fallback_migratetype=%d fragmenting=%d change_ownership=%d",
Namhyung Kim9fdd8a82015-04-06 14:36:09 +0900308 pfn_to_page(__entry->pfn),
309 __entry->pfn,
Mel Gormane0fff1b2009-09-21 17:02:42 -0700310 __entry->alloc_order,
311 __entry->fallback_order,
312 pageblock_order,
313 __entry->alloc_migratetype,
314 __entry->fallback_migratetype,
315 __entry->fallback_order < pageblock_order,
Srivatsa S. Bhatf92310c2013-09-11 14:20:36 -0700316 __entry->change_ownership)
Mel Gormane0fff1b2009-09-21 17:02:42 -0700317);
318
Steven Rostedta8d154b2009-04-10 09:36:00 -0400319#endif /* _TRACE_KMEM_H */
Steven Rostedtea20d922009-04-10 08:54:16 -0400320
Steven Rostedta8d154b2009-04-10 09:36:00 -0400321/* This part must be outside protection */
322#include <trace/define_trace.h>