1/* SPDX-License-Identifier: GPL-2.0 */
2#undef TRACE_SYSTEM
3#define TRACE_SYSTEM writeback
4
5#if !defined(_TRACE_WRITEBACK_H) || defined(TRACE_HEADER_MULTI_READ)
6#define _TRACE_WRITEBACK_H
7
8#include <linux/tracepoint.h>
9#include <linux/backing-dev.h>
10#include <linux/writeback.h>
11
12#define show_inode_state(state)					\
13	__print_flags(state, "|",				\
14		{I_DIRTY_SYNC,		"I_DIRTY_SYNC"},	\
15		{I_DIRTY_DATASYNC,	"I_DIRTY_DATASYNC"},	\
16		{I_DIRTY_PAGES,		"I_DIRTY_PAGES"},	\
17		{I_NEW,			"I_NEW"},		\
18		{I_WILL_FREE,		"I_WILL_FREE"},		\
19		{I_FREEING,		"I_FREEING"},		\
20		{I_CLEAR,		"I_CLEAR"},		\
21		{I_SYNC,		"I_SYNC"},		\
22		{I_DIRTY_TIME,		"I_DIRTY_TIME"},	\
23		{I_REFERENCED,		"I_REFERENCED"}		\
24	)
25
26/* enums need to be exported to user space */
27#undef EM
28#undef EMe
29#define EM(a,b) 	TRACE_DEFINE_ENUM(a);
30#define EMe(a,b)	TRACE_DEFINE_ENUM(a);
31
32#define WB_WORK_REASON							\
33	EM( WB_REASON_BACKGROUND,		"background")		\
34	EM( WB_REASON_VMSCAN,			"vmscan")		\
35	EM( WB_REASON_SYNC,			"sync")			\
36	EM( WB_REASON_PERIODIC,			"periodic")		\
37	EM( WB_REASON_LAPTOP_TIMER,		"laptop_timer")		\
38	EM( WB_REASON_FS_FREE_SPACE,		"fs_free_space")	\
39	EM( WB_REASON_FORKER_THREAD,		"forker_thread")	\
40	EMe(WB_REASON_FOREIGN_FLUSH,		"foreign_flush")
41
42WB_WORK_REASON
43
44/*
45 * Now redefine the EM() and EMe() macros to map the enums to the strings
46 * that will be printed in the output.
47 */
48#undef EM
49#undef EMe
50#define EM(a,b)		{ a, b },
51#define EMe(a,b)	{ a, b }
52
53struct wb_writeback_work;
54
55DECLARE_EVENT_CLASS(writeback_folio_template,
56
57	TP_PROTO(struct folio *folio, struct address_space *mapping),
58
59	TP_ARGS(folio, mapping),
60
61	TP_STRUCT__entry (
62		__array(char, name, 32)
63		__field(ino_t, ino)
64		__field(pgoff_t, index)
65	),
66
67	TP_fast_assign(
68		strscpy_pad(__entry->name,
69			    bdi_dev_name(mapping ? inode_to_bdi(mapping->host) :
70					 NULL), 32);
71		__entry->ino = (mapping && mapping->host) ? mapping->host->i_ino : 0;
72		__entry->index = folio->index;
73	),
74
75	TP_printk("bdi %s: ino=%lu index=%lu",
76		__entry->name,
77		(unsigned long)__entry->ino,
78		__entry->index
79	)
80);
81
82DEFINE_EVENT(writeback_folio_template, writeback_dirty_folio,
83
84	TP_PROTO(struct folio *folio, struct address_space *mapping),
85
86	TP_ARGS(folio, mapping)
87);
88
89DEFINE_EVENT(writeback_folio_template, folio_wait_writeback,
90
91	TP_PROTO(struct folio *folio, struct address_space *mapping),
92
93	TP_ARGS(folio, mapping)
94);
95
96DECLARE_EVENT_CLASS(writeback_dirty_inode_template,
97
98	TP_PROTO(struct inode *inode, int flags),
99
100	TP_ARGS(inode, flags),
101
102	TP_STRUCT__entry (
103		__array(char, name, 32)
104		__field(ino_t, ino)
105		__field(unsigned long, state)
106		__field(unsigned long, flags)
107	),
108
109	TP_fast_assign(
110		struct backing_dev_info *bdi = inode_to_bdi(inode);
111
112		/* may be called for files on pseudo FSes w/ unregistered bdi */
113		strscpy_pad(__entry->name, bdi_dev_name(bdi), 32);
114		__entry->ino		= inode->i_ino;
115		__entry->state		= inode->i_state;
116		__entry->flags		= flags;
117	),
118
119	TP_printk("bdi %s: ino=%lu state=%s flags=%s",
120		__entry->name,
121		(unsigned long)__entry->ino,
122		show_inode_state(__entry->state),
123		show_inode_state(__entry->flags)
124	)
125);
126
127DEFINE_EVENT(writeback_dirty_inode_template, writeback_mark_inode_dirty,
128
129	TP_PROTO(struct inode *inode, int flags),
130
131	TP_ARGS(inode, flags)
132);
133
134DEFINE_EVENT(writeback_dirty_inode_template, writeback_dirty_inode_start,
135
136	TP_PROTO(struct inode *inode, int flags),
137
138	TP_ARGS(inode, flags)
139);
140
141DEFINE_EVENT(writeback_dirty_inode_template, writeback_dirty_inode,
142
143	TP_PROTO(struct inode *inode, int flags),
144
145	TP_ARGS(inode, flags)
146);
147
148#ifdef CREATE_TRACE_POINTS
149#ifdef CONFIG_CGROUP_WRITEBACK
150
151static inline ino_t __trace_wb_assign_cgroup(struct bdi_writeback *wb)
152{
153	return cgroup_ino(wb->memcg_css->cgroup);
154}
155
156static inline ino_t __trace_wbc_assign_cgroup(struct writeback_control *wbc)
157{
158	if (wbc->wb)
159		return __trace_wb_assign_cgroup(wbc->wb);
160	else
161		return 1;
162}
163#else	/* CONFIG_CGROUP_WRITEBACK */
164
165static inline ino_t __trace_wb_assign_cgroup(struct bdi_writeback *wb)
166{
167	return 1;
168}
169
170static inline ino_t __trace_wbc_assign_cgroup(struct writeback_control *wbc)
171{
172	return 1;
173}
174
175#endif	/* CONFIG_CGROUP_WRITEBACK */
176#endif	/* CREATE_TRACE_POINTS */
177
178#ifdef CONFIG_CGROUP_WRITEBACK
179TRACE_EVENT(inode_foreign_history,
180
181	TP_PROTO(struct inode *inode, struct writeback_control *wbc,
182		 unsigned int history),
183
184	TP_ARGS(inode, wbc, history),
185
186	TP_STRUCT__entry(
187		__array(char,		name, 32)
188		__field(ino_t,		ino)
189		__field(ino_t,		cgroup_ino)
190		__field(unsigned int,	history)
191	),
192
193	TP_fast_assign(
194		strscpy_pad(__entry->name, bdi_dev_name(inode_to_bdi(inode)), 32);
195		__entry->ino		= inode->i_ino;
196		__entry->cgroup_ino	= __trace_wbc_assign_cgroup(wbc);
197		__entry->history	= history;
198	),
199
200	TP_printk("bdi %s: ino=%lu cgroup_ino=%lu history=0x%x",
201		__entry->name,
202		(unsigned long)__entry->ino,
203		(unsigned long)__entry->cgroup_ino,
204		__entry->history
205	)
206);
207
208TRACE_EVENT(inode_switch_wbs,
209
210	TP_PROTO(struct inode *inode, struct bdi_writeback *old_wb,
211		 struct bdi_writeback *new_wb),
212
213	TP_ARGS(inode, old_wb, new_wb),
214
215	TP_STRUCT__entry(
216		__array(char,		name, 32)
217		__field(ino_t,		ino)
218		__field(ino_t,		old_cgroup_ino)
219		__field(ino_t,		new_cgroup_ino)
220	),
221
222	TP_fast_assign(
223		strscpy_pad(__entry->name, bdi_dev_name(old_wb->bdi), 32);
224		__entry->ino		= inode->i_ino;
225		__entry->old_cgroup_ino	= __trace_wb_assign_cgroup(old_wb);
226		__entry->new_cgroup_ino	= __trace_wb_assign_cgroup(new_wb);
227	),
228
229	TP_printk("bdi %s: ino=%lu old_cgroup_ino=%lu new_cgroup_ino=%lu",
230		__entry->name,
231		(unsigned long)__entry->ino,
232		(unsigned long)__entry->old_cgroup_ino,
233		(unsigned long)__entry->new_cgroup_ino
234	)
235);
236
237TRACE_EVENT(track_foreign_dirty,
238
239	TP_PROTO(struct folio *folio, struct bdi_writeback *wb),
240
241	TP_ARGS(folio, wb),
242
243	TP_STRUCT__entry(
244		__array(char,		name, 32)
245		__field(u64,		bdi_id)
246		__field(ino_t,		ino)
247		__field(unsigned int,	memcg_id)
248		__field(ino_t,		cgroup_ino)
249		__field(ino_t,		page_cgroup_ino)
250	),
251
252	TP_fast_assign(
253		struct address_space *mapping = folio_mapping(folio);
254		struct inode *inode = mapping ? mapping->host : NULL;
255
256		strscpy_pad(__entry->name, bdi_dev_name(wb->bdi), 32);
257		__entry->bdi_id		= wb->bdi->id;
258		__entry->ino		= inode ? inode->i_ino : 0;
259		__entry->memcg_id	= wb->memcg_css->id;
260		__entry->cgroup_ino	= __trace_wb_assign_cgroup(wb);
261		__entry->page_cgroup_ino = cgroup_ino(folio_memcg(folio)->css.cgroup);
262	),
263
264	TP_printk("bdi %s[%llu]: ino=%lu memcg_id=%u cgroup_ino=%lu page_cgroup_ino=%lu",
265		__entry->name,
266		__entry->bdi_id,
267		(unsigned long)__entry->ino,
268		__entry->memcg_id,
269		(unsigned long)__entry->cgroup_ino,
270		(unsigned long)__entry->page_cgroup_ino
271	)
272);
273
274TRACE_EVENT(flush_foreign,
275
276	TP_PROTO(struct bdi_writeback *wb, unsigned int frn_bdi_id,
277		 unsigned int frn_memcg_id),
278
279	TP_ARGS(wb, frn_bdi_id, frn_memcg_id),
280
281	TP_STRUCT__entry(
282		__array(char,		name, 32)
283		__field(ino_t,		cgroup_ino)
284		__field(unsigned int,	frn_bdi_id)
285		__field(unsigned int,	frn_memcg_id)
286	),
287
288	TP_fast_assign(
289		strscpy_pad(__entry->name, bdi_dev_name(wb->bdi), 32);
290		__entry->cgroup_ino	= __trace_wb_assign_cgroup(wb);
291		__entry->frn_bdi_id	= frn_bdi_id;
292		__entry->frn_memcg_id	= frn_memcg_id;
293	),
294
295	TP_printk("bdi %s: cgroup_ino=%lu frn_bdi_id=%u frn_memcg_id=%u",
296		__entry->name,
297		(unsigned long)__entry->cgroup_ino,
298		__entry->frn_bdi_id,
299		__entry->frn_memcg_id
300	)
301);
302#endif
303
304DECLARE_EVENT_CLASS(writeback_write_inode_template,
305
306	TP_PROTO(struct inode *inode, struct writeback_control *wbc),
307
308	TP_ARGS(inode, wbc),
309
310	TP_STRUCT__entry (
311		__array(char, name, 32)
312		__field(ino_t, ino)
313		__field(int, sync_mode)
314		__field(ino_t, cgroup_ino)
315	),
316
317	TP_fast_assign(
318		strscpy_pad(__entry->name,
319			    bdi_dev_name(inode_to_bdi(inode)), 32);
320		__entry->ino		= inode->i_ino;
321		__entry->sync_mode	= wbc->sync_mode;
322		__entry->cgroup_ino	= __trace_wbc_assign_cgroup(wbc);
323	),
324
325	TP_printk("bdi %s: ino=%lu sync_mode=%d cgroup_ino=%lu",
326		__entry->name,
327		(unsigned long)__entry->ino,
328		__entry->sync_mode,
329		(unsigned long)__entry->cgroup_ino
330	)
331);
332
333DEFINE_EVENT(writeback_write_inode_template, writeback_write_inode_start,
334
335	TP_PROTO(struct inode *inode, struct writeback_control *wbc),
336
337	TP_ARGS(inode, wbc)
338);
339
340DEFINE_EVENT(writeback_write_inode_template, writeback_write_inode,
341
342	TP_PROTO(struct inode *inode, struct writeback_control *wbc),
343
344	TP_ARGS(inode, wbc)
345);
346
347DECLARE_EVENT_CLASS(writeback_work_class,
348	TP_PROTO(struct bdi_writeback *wb, struct wb_writeback_work *work),
349	TP_ARGS(wb, work),
350	TP_STRUCT__entry(
351		__array(char, name, 32)
352		__field(long, nr_pages)
353		__field(dev_t, sb_dev)
354		__field(int, sync_mode)
355		__field(int, for_kupdate)
356		__field(int, range_cyclic)
357		__field(int, for_background)
358		__field(int, reason)
359		__field(ino_t, cgroup_ino)
360	),
361	TP_fast_assign(
362		strscpy_pad(__entry->name, bdi_dev_name(wb->bdi), 32);
363		__entry->nr_pages = work->nr_pages;
364		__entry->sb_dev = work->sb ? work->sb->s_dev : 0;
365		__entry->sync_mode = work->sync_mode;
366		__entry->for_kupdate = work->for_kupdate;
367		__entry->range_cyclic = work->range_cyclic;
368		__entry->for_background	= work->for_background;
369		__entry->reason = work->reason;
370		__entry->cgroup_ino = __trace_wb_assign_cgroup(wb);
371	),
372	TP_printk("bdi %s: sb_dev %d:%d nr_pages=%ld sync_mode=%d "
373		  "kupdate=%d range_cyclic=%d background=%d reason=%s cgroup_ino=%lu",
374		  __entry->name,
375		  MAJOR(__entry->sb_dev), MINOR(__entry->sb_dev),
376		  __entry->nr_pages,
377		  __entry->sync_mode,
378		  __entry->for_kupdate,
379		  __entry->range_cyclic,
380		  __entry->for_background,
381		  __print_symbolic(__entry->reason, WB_WORK_REASON),
382		  (unsigned long)__entry->cgroup_ino
383	)
384);
385#define DEFINE_WRITEBACK_WORK_EVENT(name) \
386DEFINE_EVENT(writeback_work_class, name, \
387	TP_PROTO(struct bdi_writeback *wb, struct wb_writeback_work *work), \
388	TP_ARGS(wb, work))
389DEFINE_WRITEBACK_WORK_EVENT(writeback_queue);
390DEFINE_WRITEBACK_WORK_EVENT(writeback_exec);
391DEFINE_WRITEBACK_WORK_EVENT(writeback_start);
392DEFINE_WRITEBACK_WORK_EVENT(writeback_written);
393DEFINE_WRITEBACK_WORK_EVENT(writeback_wait);
394
395TRACE_EVENT(writeback_pages_written,
396	TP_PROTO(long pages_written),
397	TP_ARGS(pages_written),
398	TP_STRUCT__entry(
399		__field(long,		pages)
400	),
401	TP_fast_assign(
402		__entry->pages		= pages_written;
403	),
404	TP_printk("%ld", __entry->pages)
405);
406
407DECLARE_EVENT_CLASS(writeback_class,
408	TP_PROTO(struct bdi_writeback *wb),
409	TP_ARGS(wb),
410	TP_STRUCT__entry(
411		__array(char, name, 32)
412		__field(ino_t, cgroup_ino)
413	),
414	TP_fast_assign(
415		strscpy_pad(__entry->name, bdi_dev_name(wb->bdi), 32);
416		__entry->cgroup_ino = __trace_wb_assign_cgroup(wb);
417	),
418	TP_printk("bdi %s: cgroup_ino=%lu",
419		  __entry->name,
420		  (unsigned long)__entry->cgroup_ino
421	)
422);
423#define DEFINE_WRITEBACK_EVENT(name) \
424DEFINE_EVENT(writeback_class, name, \
425	TP_PROTO(struct bdi_writeback *wb), \
426	TP_ARGS(wb))
427
428DEFINE_WRITEBACK_EVENT(writeback_wake_background);
429
430TRACE_EVENT(writeback_bdi_register,
431	TP_PROTO(struct backing_dev_info *bdi),
432	TP_ARGS(bdi),
433	TP_STRUCT__entry(
434		__array(char, name, 32)
435	),
436	TP_fast_assign(
437		strscpy_pad(__entry->name, bdi_dev_name(bdi), 32);
438	),
439	TP_printk("bdi %s",
440		__entry->name
441	)
442);
443
444DECLARE_EVENT_CLASS(wbc_class,
445	TP_PROTO(struct writeback_control *wbc, struct backing_dev_info *bdi),
446	TP_ARGS(wbc, bdi),
447	TP_STRUCT__entry(
448		__array(char, name, 32)
449		__field(long, nr_to_write)
450		__field(long, pages_skipped)
451		__field(int, sync_mode)
452		__field(int, for_kupdate)
453		__field(int, for_background)
454		__field(int, for_reclaim)
455		__field(int, range_cyclic)
456		__field(long, range_start)
457		__field(long, range_end)
458		__field(ino_t, cgroup_ino)
459	),
460
461	TP_fast_assign(
462		strscpy_pad(__entry->name, bdi_dev_name(bdi), 32);
463		__entry->nr_to_write	= wbc->nr_to_write;
464		__entry->pages_skipped	= wbc->pages_skipped;
465		__entry->sync_mode	= wbc->sync_mode;
466		__entry->for_kupdate	= wbc->for_kupdate;
467		__entry->for_background	= wbc->for_background;
468		__entry->for_reclaim	= wbc->for_reclaim;
469		__entry->range_cyclic	= wbc->range_cyclic;
470		__entry->range_start	= (long)wbc->range_start;
471		__entry->range_end	= (long)wbc->range_end;
472		__entry->cgroup_ino	= __trace_wbc_assign_cgroup(wbc);
473	),
474
475	TP_printk("bdi %s: towrt=%ld skip=%ld mode=%d kupd=%d "
476		"bgrd=%d reclm=%d cyclic=%d "
477		"start=0x%lx end=0x%lx cgroup_ino=%lu",
478		__entry->name,
479		__entry->nr_to_write,
480		__entry->pages_skipped,
481		__entry->sync_mode,
482		__entry->for_kupdate,
483		__entry->for_background,
484		__entry->for_reclaim,
485		__entry->range_cyclic,
486		__entry->range_start,
487		__entry->range_end,
488		(unsigned long)__entry->cgroup_ino
489	)
490)
491
492#define DEFINE_WBC_EVENT(name) \
493DEFINE_EVENT(wbc_class, name, \
494	TP_PROTO(struct writeback_control *wbc, struct backing_dev_info *bdi), \
495	TP_ARGS(wbc, bdi))
496DEFINE_WBC_EVENT(wbc_writepage);
497
498TRACE_EVENT(writeback_queue_io,
499	TP_PROTO(struct bdi_writeback *wb,
500		 struct wb_writeback_work *work,
501		 unsigned long dirtied_before,
502		 int moved),
503	TP_ARGS(wb, work, dirtied_before, moved),
504	TP_STRUCT__entry(
505		__array(char,		name, 32)
506		__field(unsigned long,	older)
507		__field(long,		age)
508		__field(int,		moved)
509		__field(int,		reason)
510		__field(ino_t,		cgroup_ino)
511	),
512	TP_fast_assign(
513		strscpy_pad(__entry->name, bdi_dev_name(wb->bdi), 32);
514		__entry->older	= dirtied_before;
515		__entry->age	= (jiffies - dirtied_before) * 1000 / HZ;
516		__entry->moved	= moved;
517		__entry->reason	= work->reason;
518		__entry->cgroup_ino	= __trace_wb_assign_cgroup(wb);
519	),
520	TP_printk("bdi %s: older=%lu age=%ld enqueue=%d reason=%s cgroup_ino=%lu",
521		__entry->name,
522		__entry->older,	/* dirtied_before in jiffies */
523		__entry->age,	/* dirtied_before in relative milliseconds */
524		__entry->moved,
525		__print_symbolic(__entry->reason, WB_WORK_REASON),
526		(unsigned long)__entry->cgroup_ino
527	)
528);
529
530TRACE_EVENT(global_dirty_state,
531
532	TP_PROTO(unsigned long background_thresh,
533		 unsigned long dirty_thresh
534	),
535
536	TP_ARGS(background_thresh,
537		dirty_thresh
538	),
539
540	TP_STRUCT__entry(
541		__field(unsigned long,	nr_dirty)
542		__field(unsigned long,	nr_writeback)
543		__field(unsigned long,	background_thresh)
544		__field(unsigned long,	dirty_thresh)
545		__field(unsigned long,	dirty_limit)
546		__field(unsigned long,	nr_dirtied)
547		__field(unsigned long,	nr_written)
548	),
549
550	TP_fast_assign(
551		__entry->nr_dirty	= global_node_page_state(NR_FILE_DIRTY);
552		__entry->nr_writeback	= global_node_page_state(NR_WRITEBACK);
553		__entry->nr_dirtied	= global_node_page_state(NR_DIRTIED);
554		__entry->nr_written	= global_node_page_state(NR_WRITTEN);
555		__entry->background_thresh = background_thresh;
556		__entry->dirty_thresh	= dirty_thresh;
557		__entry->dirty_limit	= global_wb_domain.dirty_limit;
558	),
559
560	TP_printk("dirty=%lu writeback=%lu "
561		  "bg_thresh=%lu thresh=%lu limit=%lu "
562		  "dirtied=%lu written=%lu",
563		  __entry->nr_dirty,
564		  __entry->nr_writeback,
565		  __entry->background_thresh,
566		  __entry->dirty_thresh,
567		  __entry->dirty_limit,
568		  __entry->nr_dirtied,
569		  __entry->nr_written
570	)
571);
572
573#define KBps(x)			((x) << (PAGE_SHIFT - 10))
574
575TRACE_EVENT(bdi_dirty_ratelimit,
576
577	TP_PROTO(struct bdi_writeback *wb,
578		 unsigned long dirty_rate,
579		 unsigned long task_ratelimit),
580
581	TP_ARGS(wb, dirty_rate, task_ratelimit),
582
583	TP_STRUCT__entry(
584		__array(char,		bdi, 32)
585		__field(unsigned long,	write_bw)
586		__field(unsigned long,	avg_write_bw)
587		__field(unsigned long,	dirty_rate)
588		__field(unsigned long,	dirty_ratelimit)
589		__field(unsigned long,	task_ratelimit)
590		__field(unsigned long,	balanced_dirty_ratelimit)
591		__field(ino_t,		cgroup_ino)
592	),
593
594	TP_fast_assign(
595		strscpy_pad(__entry->bdi, bdi_dev_name(wb->bdi), 32);
596		__entry->write_bw	= KBps(wb->write_bandwidth);
597		__entry->avg_write_bw	= KBps(wb->avg_write_bandwidth);
598		__entry->dirty_rate	= KBps(dirty_rate);
599		__entry->dirty_ratelimit = KBps(wb->dirty_ratelimit);
600		__entry->task_ratelimit	= KBps(task_ratelimit);
601		__entry->balanced_dirty_ratelimit =
602					KBps(wb->balanced_dirty_ratelimit);
603		__entry->cgroup_ino	= __trace_wb_assign_cgroup(wb);
604	),
605
606	TP_printk("bdi %s: "
607		  "write_bw=%lu awrite_bw=%lu dirty_rate=%lu "
608		  "dirty_ratelimit=%lu task_ratelimit=%lu "
609		  "balanced_dirty_ratelimit=%lu cgroup_ino=%lu",
610		  __entry->bdi,
611		  __entry->write_bw,		/* write bandwidth */
612		  __entry->avg_write_bw,	/* avg write bandwidth */
613		  __entry->dirty_rate,		/* bdi dirty rate */
614		  __entry->dirty_ratelimit,	/* base ratelimit */
615		  __entry->task_ratelimit, /* ratelimit with position control */
616		  __entry->balanced_dirty_ratelimit, /* the balanced ratelimit */
617		  (unsigned long)__entry->cgroup_ino
618	)
619);
620
621TRACE_EVENT(balance_dirty_pages,
622
623	TP_PROTO(struct bdi_writeback *wb,
624		 unsigned long thresh,
625		 unsigned long bg_thresh,
626		 unsigned long dirty,
627		 unsigned long bdi_thresh,
628		 unsigned long bdi_dirty,
629		 unsigned long dirty_ratelimit,
630		 unsigned long task_ratelimit,
631		 unsigned long dirtied,
632		 unsigned long period,
633		 long pause,
634		 unsigned long start_time),
635
636	TP_ARGS(wb, thresh, bg_thresh, dirty, bdi_thresh, bdi_dirty,
637		dirty_ratelimit, task_ratelimit,
638		dirtied, period, pause, start_time),
639
640	TP_STRUCT__entry(
641		__array(	 char,	bdi, 32)
642		__field(unsigned long,	limit)
643		__field(unsigned long,	setpoint)
644		__field(unsigned long,	dirty)
645		__field(unsigned long,	bdi_setpoint)
646		__field(unsigned long,	bdi_dirty)
647		__field(unsigned long,	dirty_ratelimit)
648		__field(unsigned long,	task_ratelimit)
649		__field(unsigned int,	dirtied)
650		__field(unsigned int,	dirtied_pause)
651		__field(unsigned long,	paused)
652		__field(	 long,	pause)
653		__field(unsigned long,	period)
654		__field(	 long,	think)
655		__field(ino_t,		cgroup_ino)
656	),
657
658	TP_fast_assign(
659		unsigned long freerun = (thresh + bg_thresh) / 2;
660		strscpy_pad(__entry->bdi, bdi_dev_name(wb->bdi), 32);
661
662		__entry->limit		= global_wb_domain.dirty_limit;
663		__entry->setpoint	= (global_wb_domain.dirty_limit +
664						freerun) / 2;
665		__entry->dirty		= dirty;
666		__entry->bdi_setpoint	= __entry->setpoint *
667						bdi_thresh / (thresh + 1);
668		__entry->bdi_dirty	= bdi_dirty;
669		__entry->dirty_ratelimit = KBps(dirty_ratelimit);
670		__entry->task_ratelimit	= KBps(task_ratelimit);
671		__entry->dirtied	= dirtied;
672		__entry->dirtied_pause	= current->nr_dirtied_pause;
673		__entry->think		= current->dirty_paused_when == 0 ? 0 :
674			 (long)(jiffies - current->dirty_paused_when) * 1000/HZ;
675		__entry->period		= period * 1000 / HZ;
676		__entry->pause		= pause * 1000 / HZ;
677		__entry->paused		= (jiffies - start_time) * 1000 / HZ;
678		__entry->cgroup_ino	= __trace_wb_assign_cgroup(wb);
679	),
680
681
682	TP_printk("bdi %s: "
683		  "limit=%lu setpoint=%lu dirty=%lu "
684		  "bdi_setpoint=%lu bdi_dirty=%lu "
685		  "dirty_ratelimit=%lu task_ratelimit=%lu "
686		  "dirtied=%u dirtied_pause=%u "
687		  "paused=%lu pause=%ld period=%lu think=%ld cgroup_ino=%lu",
688		  __entry->bdi,
689		  __entry->limit,
690		  __entry->setpoint,
691		  __entry->dirty,
692		  __entry->bdi_setpoint,
693		  __entry->bdi_dirty,
694		  __entry->dirty_ratelimit,
695		  __entry->task_ratelimit,
696		  __entry->dirtied,
697		  __entry->dirtied_pause,
698		  __entry->paused,	/* ms */
699		  __entry->pause,	/* ms */
700		  __entry->period,	/* ms */
701		  __entry->think,	/* ms */
702		  (unsigned long)__entry->cgroup_ino
703	  )
704);
705
706TRACE_EVENT(writeback_sb_inodes_requeue,
707
708	TP_PROTO(struct inode *inode),
709	TP_ARGS(inode),
710
711	TP_STRUCT__entry(
712		__array(char, name, 32)
713		__field(ino_t, ino)
714		__field(unsigned long, state)
715		__field(unsigned long, dirtied_when)
716		__field(ino_t, cgroup_ino)
717	),
718
719	TP_fast_assign(
720		strscpy_pad(__entry->name,
721			    bdi_dev_name(inode_to_bdi(inode)), 32);
722		__entry->ino		= inode->i_ino;
723		__entry->state		= inode->i_state;
724		__entry->dirtied_when	= inode->dirtied_when;
725		__entry->cgroup_ino	= __trace_wb_assign_cgroup(inode_to_wb(inode));
726	),
727
728	TP_printk("bdi %s: ino=%lu state=%s dirtied_when=%lu age=%lu cgroup_ino=%lu",
729		  __entry->name,
730		  (unsigned long)__entry->ino,
731		  show_inode_state(__entry->state),
732		  __entry->dirtied_when,
733		  (jiffies - __entry->dirtied_when) / HZ,
734		  (unsigned long)__entry->cgroup_ino
735	)
736);
737
738DECLARE_EVENT_CLASS(writeback_single_inode_template,
739
740	TP_PROTO(struct inode *inode,
741		 struct writeback_control *wbc,
742		 unsigned long nr_to_write
743	),
744
745	TP_ARGS(inode, wbc, nr_to_write),
746
747	TP_STRUCT__entry(
748		__array(char, name, 32)
749		__field(ino_t, ino)
750		__field(unsigned long, state)
751		__field(unsigned long, dirtied_when)
752		__field(unsigned long, writeback_index)
753		__field(long, nr_to_write)
754		__field(unsigned long, wrote)
755		__field(ino_t, cgroup_ino)
756	),
757
758	TP_fast_assign(
759		strscpy_pad(__entry->name,
760			    bdi_dev_name(inode_to_bdi(inode)), 32);
761		__entry->ino		= inode->i_ino;
762		__entry->state		= inode->i_state;
763		__entry->dirtied_when	= inode->dirtied_when;
764		__entry->writeback_index = inode->i_mapping->writeback_index;
765		__entry->nr_to_write	= nr_to_write;
766		__entry->wrote		= nr_to_write - wbc->nr_to_write;
767		__entry->cgroup_ino	= __trace_wbc_assign_cgroup(wbc);
768	),
769
770	TP_printk("bdi %s: ino=%lu state=%s dirtied_when=%lu age=%lu "
771		  "index=%lu to_write=%ld wrote=%lu cgroup_ino=%lu",
772		  __entry->name,
773		  (unsigned long)__entry->ino,
774		  show_inode_state(__entry->state),
775		  __entry->dirtied_when,
776		  (jiffies - __entry->dirtied_when) / HZ,
777		  __entry->writeback_index,
778		  __entry->nr_to_write,
779		  __entry->wrote,
780		  (unsigned long)__entry->cgroup_ino
781	)
782);
783
784DEFINE_EVENT(writeback_single_inode_template, writeback_single_inode_start,
785	TP_PROTO(struct inode *inode,
786		 struct writeback_control *wbc,
787		 unsigned long nr_to_write),
788	TP_ARGS(inode, wbc, nr_to_write)
789);
790
791DEFINE_EVENT(writeback_single_inode_template, writeback_single_inode,
792	TP_PROTO(struct inode *inode,
793		 struct writeback_control *wbc,
794		 unsigned long nr_to_write),
795	TP_ARGS(inode, wbc, nr_to_write)
796);
797
798DECLARE_EVENT_CLASS(writeback_inode_template,
799	TP_PROTO(struct inode *inode),
800
801	TP_ARGS(inode),
802
803	TP_STRUCT__entry(
804		__field(	dev_t,	dev			)
805		__field(	ino_t,	ino			)
806		__field(unsigned long,	state			)
807		__field(	__u16, mode			)
808		__field(unsigned long, dirtied_when		)
809	),
810
811	TP_fast_assign(
812		__entry->dev	= inode->i_sb->s_dev;
813		__entry->ino	= inode->i_ino;
814		__entry->state	= inode->i_state;
815		__entry->mode	= inode->i_mode;
816		__entry->dirtied_when = inode->dirtied_when;
817	),
818
819	TP_printk("dev %d,%d ino %lu dirtied %lu state %s mode 0%o",
820		  MAJOR(__entry->dev), MINOR(__entry->dev),
821		  (unsigned long)__entry->ino, __entry->dirtied_when,
822		  show_inode_state(__entry->state), __entry->mode)
823);
824
825DEFINE_EVENT(writeback_inode_template, writeback_lazytime,
826	TP_PROTO(struct inode *inode),
827
828	TP_ARGS(inode)
829);
830
831DEFINE_EVENT(writeback_inode_template, writeback_lazytime_iput,
832	TP_PROTO(struct inode *inode),
833
834	TP_ARGS(inode)
835);
836
837DEFINE_EVENT(writeback_inode_template, writeback_dirty_inode_enqueue,
838
839	TP_PROTO(struct inode *inode),
840
841	TP_ARGS(inode)
842);
843
844/*
845 * Inode writeback list tracking.
846 */
847
848DEFINE_EVENT(writeback_inode_template, sb_mark_inode_writeback,
849	TP_PROTO(struct inode *inode),
850	TP_ARGS(inode)
851);
852
853DEFINE_EVENT(writeback_inode_template, sb_clear_inode_writeback,
854	TP_PROTO(struct inode *inode),
855	TP_ARGS(inode)
856);
857
858#endif /* _TRACE_WRITEBACK_H */
859
860/* This part must be outside protection */
861#include <trace/define_trace.h>
862