~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

TOMOYO Linux Cross Reference
Linux/include/trace/events/kmem.h

Version: ~ [ linux-5.6 ] ~ [ linux-5.5.13 ] ~ [ linux-5.4.28 ] ~ [ linux-5.3.18 ] ~ [ linux-5.2.21 ] ~ [ linux-5.1.21 ] ~ [ linux-5.0.21 ] ~ [ linux-4.20.17 ] ~ [ linux-4.19.113 ] ~ [ linux-4.18.20 ] ~ [ linux-4.17.19 ] ~ [ linux-4.16.18 ] ~ [ linux-4.15.18 ] ~ [ linux-4.14.174 ] ~ [ linux-4.13.16 ] ~ [ linux-4.12.14 ] ~ [ linux-4.11.12 ] ~ [ linux-4.10.17 ] ~ [ linux-4.9.217 ] ~ [ linux-4.8.17 ] ~ [ linux-4.7.10 ] ~ [ linux-4.6.7 ] ~ [ linux-4.5.7 ] ~ [ linux-4.4.217 ] ~ [ linux-4.3.6 ] ~ [ linux-4.2.8 ] ~ [ linux-4.1.52 ] ~ [ linux-4.0.9 ] ~ [ linux-3.19.8 ] ~ [ linux-3.18.140 ] ~ [ linux-3.17.8 ] ~ [ linux-3.16.82 ] ~ [ linux-3.15.10 ] ~ [ linux-3.14.79 ] ~ [ linux-3.13.11 ] ~ [ linux-3.12.74 ] ~ [ linux-3.11.10 ] ~ [ linux-3.10.108 ] ~ [ linux-3.9.11 ] ~ [ linux-3.8.13 ] ~ [ linux-3.7.10 ] ~ [ linux-3.6.11 ] ~ [ linux-3.5.7 ] ~ [ linux-3.4.113 ] ~ [ linux-3.3.8 ] ~ [ linux-3.2.102 ] ~ [ linux-3.1.10 ] ~ [ linux-3.0.101 ] ~ [ linux-2.6.32.71 ] ~ [ linux-2.6.0 ] ~ [ linux-2.4.37.11 ] ~ [ unix-v6-master ] ~ [ ccs-tools-1.8.5 ] ~ [ policy-sample ] ~
Architecture: ~ [ i386 ] ~ [ alpha ] ~ [ m68k ] ~ [ mips ] ~ [ ppc ] ~ [ sparc ] ~ [ sparc64 ] ~

  1 #undef TRACE_SYSTEM
  2 #define TRACE_SYSTEM kmem
  3 
  4 #if !defined(_TRACE_KMEM_H) || defined(TRACE_HEADER_MULTI_READ)
  5 #define _TRACE_KMEM_H
  6 
  7 #include <linux/types.h>
  8 #include <linux/tracepoint.h>
  9 #include <trace/events/gfpflags.h>
 10 
 11 DECLARE_EVENT_CLASS(kmem_alloc,
 12 
 13         TP_PROTO(unsigned long call_site,
 14                  const void *ptr,
 15                  size_t bytes_req,
 16                  size_t bytes_alloc,
 17                  gfp_t gfp_flags),
 18 
 19         TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags),
 20 
 21         TP_STRUCT__entry(
 22                 __field(        unsigned long,  call_site       )
 23                 __field(        const void *,   ptr             )
 24                 __field(        size_t,         bytes_req       )
 25                 __field(        size_t,         bytes_alloc     )
 26                 __field(        gfp_t,          gfp_flags       )
 27         ),
 28 
 29         TP_fast_assign(
 30                 __entry->call_site      = call_site;
 31                 __entry->ptr            = ptr;
 32                 __entry->bytes_req      = bytes_req;
 33                 __entry->bytes_alloc    = bytes_alloc;
 34                 __entry->gfp_flags      = gfp_flags;
 35         ),
 36 
 37         TP_printk("call_site=%lx ptr=%p bytes_req=%zu bytes_alloc=%zu gfp_flags=%s",
 38                 __entry->call_site,
 39                 __entry->ptr,
 40                 __entry->bytes_req,
 41                 __entry->bytes_alloc,
 42                 show_gfp_flags(__entry->gfp_flags))
 43 );
 44 
 45 DEFINE_EVENT(kmem_alloc, kmalloc,
 46 
 47         TP_PROTO(unsigned long call_site, const void *ptr,
 48                  size_t bytes_req, size_t bytes_alloc, gfp_t gfp_flags),
 49 
 50         TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags)
 51 );
 52 
 53 DEFINE_EVENT(kmem_alloc, kmem_cache_alloc,
 54 
 55         TP_PROTO(unsigned long call_site, const void *ptr,
 56                  size_t bytes_req, size_t bytes_alloc, gfp_t gfp_flags),
 57 
 58         TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags)
 59 );
 60 
 61 DECLARE_EVENT_CLASS(kmem_alloc_node,
 62 
 63         TP_PROTO(unsigned long call_site,
 64                  const void *ptr,
 65                  size_t bytes_req,
 66                  size_t bytes_alloc,
 67                  gfp_t gfp_flags,
 68                  int node),
 69 
 70         TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags, node),
 71 
 72         TP_STRUCT__entry(
 73                 __field(        unsigned long,  call_site       )
 74                 __field(        const void *,   ptr             )
 75                 __field(        size_t,         bytes_req       )
 76                 __field(        size_t,         bytes_alloc     )
 77                 __field(        gfp_t,          gfp_flags       )
 78                 __field(        int,            node            )
 79         ),
 80 
 81         TP_fast_assign(
 82                 __entry->call_site      = call_site;
 83                 __entry->ptr            = ptr;
 84                 __entry->bytes_req      = bytes_req;
 85                 __entry->bytes_alloc    = bytes_alloc;
 86                 __entry->gfp_flags      = gfp_flags;
 87                 __entry->node           = node;
 88         ),
 89 
 90         TP_printk("call_site=%lx ptr=%p bytes_req=%zu bytes_alloc=%zu gfp_flags=%s node=%d",
 91                 __entry->call_site,
 92                 __entry->ptr,
 93                 __entry->bytes_req,
 94                 __entry->bytes_alloc,
 95                 show_gfp_flags(__entry->gfp_flags),
 96                 __entry->node)
 97 );
 98 
 99 DEFINE_EVENT(kmem_alloc_node, kmalloc_node,
100 
101         TP_PROTO(unsigned long call_site, const void *ptr,
102                  size_t bytes_req, size_t bytes_alloc,
103                  gfp_t gfp_flags, int node),
104 
105         TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags, node)
106 );
107 
108 DEFINE_EVENT(kmem_alloc_node, kmem_cache_alloc_node,
109 
110         TP_PROTO(unsigned long call_site, const void *ptr,
111                  size_t bytes_req, size_t bytes_alloc,
112                  gfp_t gfp_flags, int node),
113 
114         TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags, node)
115 );
116 
117 DECLARE_EVENT_CLASS(kmem_free,
118 
119         TP_PROTO(unsigned long call_site, const void *ptr),
120 
121         TP_ARGS(call_site, ptr),
122 
123         TP_STRUCT__entry(
124                 __field(        unsigned long,  call_site       )
125                 __field(        const void *,   ptr             )
126         ),
127 
128         TP_fast_assign(
129                 __entry->call_site      = call_site;
130                 __entry->ptr            = ptr;
131         ),
132 
133         TP_printk("call_site=%lx ptr=%p", __entry->call_site, __entry->ptr)
134 );
135 
136 DEFINE_EVENT(kmem_free, kfree,
137 
138         TP_PROTO(unsigned long call_site, const void *ptr),
139 
140         TP_ARGS(call_site, ptr)
141 );
142 
143 DEFINE_EVENT_CONDITION(kmem_free, kmem_cache_free,
144 
145         TP_PROTO(unsigned long call_site, const void *ptr),
146 
147         TP_ARGS(call_site, ptr),
148 
149         /*
150          * This trace can be potentially called from an offlined cpu.
151          * Since trace points use RCU and RCU should not be used from
152          * offline cpus, filter such calls out.
153          * While this trace can be called from a preemptable section,
154          * it has no impact on the condition since tasks can migrate
155          * only from online cpus to other online cpus. Thus its safe
156          * to use raw_smp_processor_id.
157          */
158         TP_CONDITION(cpu_online(raw_smp_processor_id()))
159 );
160 
161 TRACE_EVENT_CONDITION(mm_page_free,
162 
163         TP_PROTO(struct page *page, unsigned int order),
164 
165         TP_ARGS(page, order),
166 
167 
168         /*
169          * This trace can be potentially called from an offlined cpu.
170          * Since trace points use RCU and RCU should not be used from
171          * offline cpus, filter such calls out.
172          * While this trace can be called from a preemptable section,
173          * it has no impact on the condition since tasks can migrate
174          * only from online cpus to other online cpus. Thus its safe
175          * to use raw_smp_processor_id.
176          */
177         TP_CONDITION(cpu_online(raw_smp_processor_id())),
178 
179         TP_STRUCT__entry(
180                 __field(        unsigned long,  pfn             )
181                 __field(        unsigned int,   order           )
182         ),
183 
184         TP_fast_assign(
185                 __entry->pfn            = page_to_pfn(page);
186                 __entry->order          = order;
187         ),
188 
189         TP_printk("page=%p pfn=%lu order=%d",
190                         pfn_to_page(__entry->pfn),
191                         __entry->pfn,
192                         __entry->order)
193 );
194 
195 TRACE_EVENT(mm_page_free_batched,
196 
197         TP_PROTO(struct page *page, int cold),
198 
199         TP_ARGS(page, cold),
200 
201         TP_STRUCT__entry(
202                 __field(        unsigned long,  pfn             )
203                 __field(        int,            cold            )
204         ),
205 
206         TP_fast_assign(
207                 __entry->pfn            = page_to_pfn(page);
208                 __entry->cold           = cold;
209         ),
210 
211         TP_printk("page=%p pfn=%lu order=0 cold=%d",
212                         pfn_to_page(__entry->pfn),
213                         __entry->pfn,
214                         __entry->cold)
215 );
216 
217 TRACE_EVENT(mm_page_alloc,
218 
219         TP_PROTO(struct page *page, unsigned int order,
220                         gfp_t gfp_flags, int migratetype),
221 
222         TP_ARGS(page, order, gfp_flags, migratetype),
223 
224         TP_STRUCT__entry(
225                 __field(        unsigned long,  pfn             )
226                 __field(        unsigned int,   order           )
227                 __field(        gfp_t,          gfp_flags       )
228                 __field(        int,            migratetype     )
229         ),
230 
231         TP_fast_assign(
232                 __entry->pfn            = page ? page_to_pfn(page) : -1UL;
233                 __entry->order          = order;
234                 __entry->gfp_flags      = gfp_flags;
235                 __entry->migratetype    = migratetype;
236         ),
237 
238         TP_printk("page=%p pfn=%lu order=%d migratetype=%d gfp_flags=%s",
239                 __entry->pfn != -1UL ? pfn_to_page(__entry->pfn) : NULL,
240                 __entry->pfn != -1UL ? __entry->pfn : 0,
241                 __entry->order,
242                 __entry->migratetype,
243                 show_gfp_flags(__entry->gfp_flags))
244 );
245 
246 DECLARE_EVENT_CLASS(mm_page,
247 
248         TP_PROTO(struct page *page, unsigned int order, int migratetype),
249 
250         TP_ARGS(page, order, migratetype),
251 
252         TP_STRUCT__entry(
253                 __field(        unsigned long,  pfn             )
254                 __field(        unsigned int,   order           )
255                 __field(        int,            migratetype     )
256         ),
257 
258         TP_fast_assign(
259                 __entry->pfn            = page ? page_to_pfn(page) : -1UL;
260                 __entry->order          = order;
261                 __entry->migratetype    = migratetype;
262         ),
263 
264         TP_printk("page=%p pfn=%lu order=%u migratetype=%d percpu_refill=%d",
265                 __entry->pfn != -1UL ? pfn_to_page(__entry->pfn) : NULL,
266                 __entry->pfn != -1UL ? __entry->pfn : 0,
267                 __entry->order,
268                 __entry->migratetype,
269                 __entry->order == 0)
270 );
271 
272 DEFINE_EVENT(mm_page, mm_page_alloc_zone_locked,
273 
274         TP_PROTO(struct page *page, unsigned int order, int migratetype),
275 
276         TP_ARGS(page, order, migratetype)
277 );
278 
279 TRACE_EVENT_CONDITION(mm_page_pcpu_drain,
280 
281         TP_PROTO(struct page *page, unsigned int order, int migratetype),
282 
283         TP_ARGS(page, order, migratetype),
284 
285         /*
286          * This trace can be potentially called from an offlined cpu.
287          * Since trace points use RCU and RCU should not be used from
288          * offline cpus, filter such calls out.
289          * While this trace can be called from a preemptable section,
290          * it has no impact on the condition since tasks can migrate
291          * only from online cpus to other online cpus. Thus its safe
292          * to use raw_smp_processor_id.
293          */
294         TP_CONDITION(cpu_online(raw_smp_processor_id())),
295 
296         TP_STRUCT__entry(
297                 __field(        unsigned long,  pfn             )
298                 __field(        unsigned int,   order           )
299                 __field(        int,            migratetype     )
300         ),
301 
302         TP_fast_assign(
303                 __entry->pfn            = page ? page_to_pfn(page) : -1UL;
304                 __entry->order          = order;
305                 __entry->migratetype    = migratetype;
306         ),
307 
308         TP_printk("page=%p pfn=%lu order=%d migratetype=%d",
309                 pfn_to_page(__entry->pfn), __entry->pfn,
310                 __entry->order, __entry->migratetype)
311 );
312 
313 TRACE_EVENT(mm_page_alloc_extfrag,
314 
315         TP_PROTO(struct page *page,
316                 int alloc_order, int fallback_order,
317                 int alloc_migratetype, int fallback_migratetype),
318 
319         TP_ARGS(page,
320                 alloc_order, fallback_order,
321                 alloc_migratetype, fallback_migratetype),
322 
323         TP_STRUCT__entry(
324                 __field(        unsigned long,  pfn                     )
325                 __field(        int,            alloc_order             )
326                 __field(        int,            fallback_order          )
327                 __field(        int,            alloc_migratetype       )
328                 __field(        int,            fallback_migratetype    )
329                 __field(        int,            change_ownership        )
330         ),
331 
332         TP_fast_assign(
333                 __entry->pfn                    = page_to_pfn(page);
334                 __entry->alloc_order            = alloc_order;
335                 __entry->fallback_order         = fallback_order;
336                 __entry->alloc_migratetype      = alloc_migratetype;
337                 __entry->fallback_migratetype   = fallback_migratetype;
338                 __entry->change_ownership       = (alloc_migratetype ==
339                                         get_pageblock_migratetype(page));
340         ),
341 
342         TP_printk("page=%p pfn=%lu alloc_order=%d fallback_order=%d pageblock_order=%d alloc_migratetype=%d fallback_migratetype=%d fragmenting=%d change_ownership=%d",
343                 pfn_to_page(__entry->pfn),
344                 __entry->pfn,
345                 __entry->alloc_order,
346                 __entry->fallback_order,
347                 pageblock_order,
348                 __entry->alloc_migratetype,
349                 __entry->fallback_migratetype,
350                 __entry->fallback_order < pageblock_order,
351                 __entry->change_ownership)
352 );
353 
354 #endif /* _TRACE_KMEM_H */
355 
356 /* This part must be outside protection */
357 #include <trace/define_trace.h>
358 

~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

kernel.org | git.kernel.org | LWN.net | Project Home | Wiki (Japanese) | Wiki (English) | SVN repository | Mail admin

Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.

osdn.jp