1// SPDX-License-Identifier: GPL-2.0
2/*
3 * trace_events_synth - synthetic trace events
4 *
5 * Copyright (C) 2015, 2020 Tom Zanussi <tom.zanussi@linux.intel.com>
6 */
7
8#include <linux/module.h>
9#include <linux/kallsyms.h>
10#include <linux/security.h>
11#include <linux/mutex.h>
12#include <linux/slab.h>
13#include <linux/stacktrace.h>
14#include <linux/rculist.h>
15#include <linux/tracefs.h>
16
17/* for gfp flag names */
18#include <linux/trace_events.h>
19#include <trace/events/mmflags.h>
20#include "trace_probe.h"
21#include "trace_probe_kernel.h"
22
23#include "trace_synth.h"
24
25#undef ERRORS
26#define ERRORS	\
27	C(BAD_NAME,		"Illegal name"),		\
28	C(INVALID_CMD,		"Command must be of the form: <name> field[;field] ..."),\
29	C(INVALID_DYN_CMD,	"Command must be of the form: s or -:[synthetic/]<name> field[;field] ..."),\
30	C(EVENT_EXISTS,		"Event already exists"),	\
31	C(TOO_MANY_FIELDS,	"Too many fields"),		\
32	C(INCOMPLETE_TYPE,	"Incomplete type"),		\
33	C(INVALID_TYPE,		"Invalid type"),		\
34	C(INVALID_FIELD,        "Invalid field"),		\
35	C(INVALID_ARRAY_SPEC,	"Invalid array specification"),
36
37#undef C
38#define C(a, b)		SYNTH_ERR_##a
39
40enum { ERRORS };
41
42#undef C
43#define C(a, b)		b
44
45static const char *err_text[] = { ERRORS };
46
47static DEFINE_MUTEX(lastcmd_mutex);
48static char *last_cmd;
49
50static int errpos(const char *str)
51{
52	int ret = 0;
53
54	mutex_lock(&lastcmd_mutex);
55	if (!str || !last_cmd)
56		goto out;
57
58	ret = err_pos(last_cmd, str);
59 out:
60	mutex_unlock(&lastcmd_mutex);
61	return ret;
62}
63
64static void last_cmd_set(const char *str)
65{
66	if (!str)
67		return;
68
69	mutex_lock(&lastcmd_mutex);
70	kfree(last_cmd);
71	last_cmd = kstrdup(str, GFP_KERNEL);
72	mutex_unlock(&lastcmd_mutex);
73}
74
75static void synth_err(u8 err_type, u16 err_pos)
76{
77	mutex_lock(&lastcmd_mutex);
78	if (!last_cmd)
79		goto out;
80
81	tracing_log_err(NULL, "synthetic_events", last_cmd, err_text,
82			err_type, err_pos);
83 out:
84	mutex_unlock(&lastcmd_mutex);
85}
86
87static int create_synth_event(const char *raw_command);
88static int synth_event_show(struct seq_file *m, struct dyn_event *ev);
89static int synth_event_release(struct dyn_event *ev);
90static bool synth_event_is_busy(struct dyn_event *ev);
91static bool synth_event_match(const char *system, const char *event,
92			int argc, const char **argv, struct dyn_event *ev);
93
94static struct dyn_event_operations synth_event_ops = {
95	.create = create_synth_event,
96	.show = synth_event_show,
97	.is_busy = synth_event_is_busy,
98	.free = synth_event_release,
99	.match = synth_event_match,
100};
101
102static bool is_synth_event(struct dyn_event *ev)
103{
104	return ev->ops == &synth_event_ops;
105}
106
107static struct synth_event *to_synth_event(struct dyn_event *ev)
108{
109	return container_of(ev, struct synth_event, devent);
110}
111
112static bool synth_event_is_busy(struct dyn_event *ev)
113{
114	struct synth_event *event = to_synth_event(ev);
115
116	return event->ref != 0;
117}
118
119static bool synth_event_match(const char *system, const char *event,
120			int argc, const char **argv, struct dyn_event *ev)
121{
122	struct synth_event *sev = to_synth_event(ev);
123
124	return strcmp(sev->name, event) == 0 &&
125		(!system || strcmp(system, SYNTH_SYSTEM) == 0);
126}
127
128struct synth_trace_event {
129	struct trace_entry	ent;
130	union trace_synth_field	fields[];
131};
132
133static int synth_event_define_fields(struct trace_event_call *call)
134{
135	struct synth_trace_event trace;
136	int offset = offsetof(typeof(trace), fields);
137	struct synth_event *event = call->data;
138	unsigned int i, size, n_u64;
139	char *name, *type;
140	bool is_signed;
141	int ret = 0;
142
143	for (i = 0, n_u64 = 0; i < event->n_fields; i++) {
144		size = event->fields[i]->size;
145		is_signed = event->fields[i]->is_signed;
146		type = event->fields[i]->type;
147		name = event->fields[i]->name;
148		ret = trace_define_field(call, type, name, offset, size,
149					 is_signed, FILTER_OTHER);
150		if (ret)
151			break;
152
153		event->fields[i]->offset = n_u64;
154
155		if (event->fields[i]->is_string && !event->fields[i]->is_dynamic) {
156			offset += STR_VAR_LEN_MAX;
157			n_u64 += STR_VAR_LEN_MAX / sizeof(u64);
158		} else {
159			offset += sizeof(u64);
160			n_u64++;
161		}
162	}
163
164	event->n_u64 = n_u64;
165
166	return ret;
167}
168
169static bool synth_field_signed(char *type)
170{
171	if (str_has_prefix(type, "u"))
172		return false;
173	if (strcmp(type, "gfp_t") == 0)
174		return false;
175
176	return true;
177}
178
179static int synth_field_is_string(char *type)
180{
181	if (strstr(type, "char[") != NULL)
182		return true;
183
184	return false;
185}
186
187static int synth_field_is_stack(char *type)
188{
189	if (strstr(type, "long[") != NULL)
190		return true;
191
192	return false;
193}
194
195static int synth_field_string_size(char *type)
196{
197	char buf[4], *end, *start;
198	unsigned int len;
199	int size, err;
200
201	start = strstr(type, "char[");
202	if (start == NULL)
203		return -EINVAL;
204	start += sizeof("char[") - 1;
205
206	end = strchr(type, ']');
207	if (!end || end < start || type + strlen(type) > end + 1)
208		return -EINVAL;
209
210	len = end - start;
211	if (len > 3)
212		return -EINVAL;
213
214	if (len == 0)
215		return 0; /* variable-length string */
216
217	strncpy(buf, start, len);
218	buf[len] = '\0';
219
220	err = kstrtouint(buf, 0, &size);
221	if (err)
222		return err;
223
224	if (size > STR_VAR_LEN_MAX)
225		return -EINVAL;
226
227	return size;
228}
229
230static int synth_field_size(char *type)
231{
232	int size = 0;
233
234	if (strcmp(type, "s64") == 0)
235		size = sizeof(s64);
236	else if (strcmp(type, "u64") == 0)
237		size = sizeof(u64);
238	else if (strcmp(type, "s32") == 0)
239		size = sizeof(s32);
240	else if (strcmp(type, "u32") == 0)
241		size = sizeof(u32);
242	else if (strcmp(type, "s16") == 0)
243		size = sizeof(s16);
244	else if (strcmp(type, "u16") == 0)
245		size = sizeof(u16);
246	else if (strcmp(type, "s8") == 0)
247		size = sizeof(s8);
248	else if (strcmp(type, "u8") == 0)
249		size = sizeof(u8);
250	else if (strcmp(type, "char") == 0)
251		size = sizeof(char);
252	else if (strcmp(type, "unsigned char") == 0)
253		size = sizeof(unsigned char);
254	else if (strcmp(type, "int") == 0)
255		size = sizeof(int);
256	else if (strcmp(type, "unsigned int") == 0)
257		size = sizeof(unsigned int);
258	else if (strcmp(type, "long") == 0)
259		size = sizeof(long);
260	else if (strcmp(type, "unsigned long") == 0)
261		size = sizeof(unsigned long);
262	else if (strcmp(type, "bool") == 0)
263		size = sizeof(bool);
264	else if (strcmp(type, "pid_t") == 0)
265		size = sizeof(pid_t);
266	else if (strcmp(type, "gfp_t") == 0)
267		size = sizeof(gfp_t);
268	else if (synth_field_is_string(type))
269		size = synth_field_string_size(type);
270	else if (synth_field_is_stack(type))
271		size = 0;
272
273	return size;
274}
275
276static const char *synth_field_fmt(char *type)
277{
278	const char *fmt = "%llu";
279
280	if (strcmp(type, "s64") == 0)
281		fmt = "%lld";
282	else if (strcmp(type, "u64") == 0)
283		fmt = "%llu";
284	else if (strcmp(type, "s32") == 0)
285		fmt = "%d";
286	else if (strcmp(type, "u32") == 0)
287		fmt = "%u";
288	else if (strcmp(type, "s16") == 0)
289		fmt = "%d";
290	else if (strcmp(type, "u16") == 0)
291		fmt = "%u";
292	else if (strcmp(type, "s8") == 0)
293		fmt = "%d";
294	else if (strcmp(type, "u8") == 0)
295		fmt = "%u";
296	else if (strcmp(type, "char") == 0)
297		fmt = "%d";
298	else if (strcmp(type, "unsigned char") == 0)
299		fmt = "%u";
300	else if (strcmp(type, "int") == 0)
301		fmt = "%d";
302	else if (strcmp(type, "unsigned int") == 0)
303		fmt = "%u";
304	else if (strcmp(type, "long") == 0)
305		fmt = "%ld";
306	else if (strcmp(type, "unsigned long") == 0)
307		fmt = "%lu";
308	else if (strcmp(type, "bool") == 0)
309		fmt = "%d";
310	else if (strcmp(type, "pid_t") == 0)
311		fmt = "%d";
312	else if (strcmp(type, "gfp_t") == 0)
313		fmt = "%x";
314	else if (synth_field_is_string(type))
315		fmt = "%.*s";
316	else if (synth_field_is_stack(type))
317		fmt = "%s";
318
319	return fmt;
320}
321
322static void print_synth_event_num_val(struct trace_seq *s,
323				      char *print_fmt, char *name,
324				      int size, union trace_synth_field *val, char *space)
325{
326	switch (size) {
327	case 1:
328		trace_seq_printf(s, print_fmt, name, val->as_u8, space);
329		break;
330
331	case 2:
332		trace_seq_printf(s, print_fmt, name, val->as_u16, space);
333		break;
334
335	case 4:
336		trace_seq_printf(s, print_fmt, name, val->as_u32, space);
337		break;
338
339	default:
340		trace_seq_printf(s, print_fmt, name, val->as_u64, space);
341		break;
342	}
343}
344
345static enum print_line_t print_synth_event(struct trace_iterator *iter,
346					   int flags,
347					   struct trace_event *event)
348{
349	struct trace_array *tr = iter->tr;
350	struct trace_seq *s = &iter->seq;
351	struct synth_trace_event *entry;
352	struct synth_event *se;
353	unsigned int i, j, n_u64;
354	char print_fmt[32];
355	const char *fmt;
356
357	entry = (struct synth_trace_event *)iter->ent;
358	se = container_of(event, struct synth_event, call.event);
359
360	trace_seq_printf(s, "%s: ", se->name);
361
362	for (i = 0, n_u64 = 0; i < se->n_fields; i++) {
363		if (trace_seq_has_overflowed(s))
364			goto end;
365
366		fmt = synth_field_fmt(se->fields[i]->type);
367
368		/* parameter types */
369		if (tr && tr->trace_flags & TRACE_ITER_VERBOSE)
370			trace_seq_printf(s, "%s ", fmt);
371
372		snprintf(print_fmt, sizeof(print_fmt), "%%s=%s%%s", fmt);
373
374		/* parameter values */
375		if (se->fields[i]->is_string) {
376			if (se->fields[i]->is_dynamic) {
377				union trace_synth_field *data = &entry->fields[n_u64];
378
379				trace_seq_printf(s, print_fmt, se->fields[i]->name,
380						 STR_VAR_LEN_MAX,
381						 (char *)entry + data->as_dynamic.offset,
382						 i == se->n_fields - 1 ? "" : " ");
383				n_u64++;
384			} else {
385				trace_seq_printf(s, print_fmt, se->fields[i]->name,
386						 STR_VAR_LEN_MAX,
387						 (char *)&entry->fields[n_u64].as_u64,
388						 i == se->n_fields - 1 ? "" : " ");
389				n_u64 += STR_VAR_LEN_MAX / sizeof(u64);
390			}
391		} else if (se->fields[i]->is_stack) {
392			union trace_synth_field *data = &entry->fields[n_u64];
393			unsigned long *p = (void *)entry + data->as_dynamic.offset;
394
395			trace_seq_printf(s, "%s=STACK:\n", se->fields[i]->name);
396			for (j = 1; j < data->as_dynamic.len / sizeof(long); j++)
397				trace_seq_printf(s, "=> %pS\n", (void *)p[j]);
398			n_u64++;
399		} else {
400			struct trace_print_flags __flags[] = {
401			    __def_gfpflag_names, {-1, NULL} };
402			char *space = (i == se->n_fields - 1 ? "" : " ");
403
404			print_synth_event_num_val(s, print_fmt,
405						  se->fields[i]->name,
406						  se->fields[i]->size,
407						  &entry->fields[n_u64],
408						  space);
409
410			if (strcmp(se->fields[i]->type, "gfp_t") == 0) {
411				trace_seq_puts(s, " (");
412				trace_print_flags_seq(s, "|",
413						      entry->fields[n_u64].as_u64,
414						      __flags);
415				trace_seq_putc(s, ')');
416			}
417			n_u64++;
418		}
419	}
420end:
421	trace_seq_putc(s, '\n');
422
423	return trace_handle_return(s);
424}
425
426static struct trace_event_functions synth_event_funcs = {
427	.trace		= print_synth_event
428};
429
430static unsigned int trace_string(struct synth_trace_event *entry,
431				 struct synth_event *event,
432				 char *str_val,
433				 bool is_dynamic,
434				 unsigned int data_size,
435				 unsigned int *n_u64)
436{
437	unsigned int len = 0;
438	char *str_field;
439	int ret;
440
441	if (is_dynamic) {
442		union trace_synth_field *data = &entry->fields[*n_u64];
443
444		len = fetch_store_strlen((unsigned long)str_val);
445		data->as_dynamic.offset = struct_size(entry, fields, event->n_u64) + data_size;
446		data->as_dynamic.len = len;
447
448		ret = fetch_store_string((unsigned long)str_val, &entry->fields[*n_u64], entry);
449
450		(*n_u64)++;
451	} else {
452		str_field = (char *)&entry->fields[*n_u64].as_u64;
453
454#ifdef CONFIG_ARCH_HAS_NON_OVERLAPPING_ADDRESS_SPACE
455		if ((unsigned long)str_val < TASK_SIZE)
456			ret = strncpy_from_user_nofault(str_field, (const void __user *)str_val, STR_VAR_LEN_MAX);
457		else
458#endif
459			ret = strncpy_from_kernel_nofault(str_field, str_val, STR_VAR_LEN_MAX);
460
461		if (ret < 0)
462			strcpy(str_field, FAULT_STRING);
463
464		(*n_u64) += STR_VAR_LEN_MAX / sizeof(u64);
465	}
466
467	return len;
468}
469
470static unsigned int trace_stack(struct synth_trace_event *entry,
471				 struct synth_event *event,
472				 long *stack,
473				 unsigned int data_size,
474				 unsigned int *n_u64)
475{
476	union trace_synth_field *data = &entry->fields[*n_u64];
477	unsigned int len;
478	u32 data_offset;
479	void *data_loc;
480
481	data_offset = struct_size(entry, fields, event->n_u64);
482	data_offset += data_size;
483
484	for (len = 0; len < HIST_STACKTRACE_DEPTH; len++) {
485		if (!stack[len])
486			break;
487	}
488
489	len *= sizeof(long);
490
491	/* Find the dynamic section to copy the stack into. */
492	data_loc = (void *)entry + data_offset;
493	memcpy(data_loc, stack, len);
494
495	/* Fill in the field that holds the offset/len combo */
496
497	data->as_dynamic.offset = data_offset;
498	data->as_dynamic.len = len;
499
500	(*n_u64)++;
501
502	return len;
503}
504
505static notrace void trace_event_raw_event_synth(void *__data,
506						u64 *var_ref_vals,
507						unsigned int *var_ref_idx)
508{
509	unsigned int i, n_u64, val_idx, len, data_size = 0;
510	struct trace_event_file *trace_file = __data;
511	struct synth_trace_event *entry;
512	struct trace_event_buffer fbuffer;
513	struct trace_buffer *buffer;
514	struct synth_event *event;
515	int fields_size = 0;
516
517	event = trace_file->event_call->data;
518
519	if (trace_trigger_soft_disabled(trace_file))
520		return;
521
522	fields_size = event->n_u64 * sizeof(u64);
523
524	for (i = 0; i < event->n_dynamic_fields; i++) {
525		unsigned int field_pos = event->dynamic_fields[i]->field_pos;
526		char *str_val;
527
528		val_idx = var_ref_idx[field_pos];
529		str_val = (char *)(long)var_ref_vals[val_idx];
530
531		if (event->dynamic_fields[i]->is_stack) {
532			/* reserve one extra element for size */
533			len = *((unsigned long *)str_val) + 1;
534			len *= sizeof(unsigned long);
535		} else {
536			len = fetch_store_strlen((unsigned long)str_val);
537		}
538
539		fields_size += len;
540	}
541
542	/*
543	 * Avoid ring buffer recursion detection, as this event
544	 * is being performed within another event.
545	 */
546	buffer = trace_file->tr->array_buffer.buffer;
547	ring_buffer_nest_start(buffer);
548
549	entry = trace_event_buffer_reserve(&fbuffer, trace_file,
550					   sizeof(*entry) + fields_size);
551	if (!entry)
552		goto out;
553
554	for (i = 0, n_u64 = 0; i < event->n_fields; i++) {
555		val_idx = var_ref_idx[i];
556		if (event->fields[i]->is_string) {
557			char *str_val = (char *)(long)var_ref_vals[val_idx];
558
559			len = trace_string(entry, event, str_val,
560					   event->fields[i]->is_dynamic,
561					   data_size, &n_u64);
562			data_size += len; /* only dynamic string increments */
563		} else if (event->fields[i]->is_stack) {
564			long *stack = (long *)(long)var_ref_vals[val_idx];
565
566			len = trace_stack(entry, event, stack,
567					   data_size, &n_u64);
568			data_size += len;
569		} else {
570			struct synth_field *field = event->fields[i];
571			u64 val = var_ref_vals[val_idx];
572
573			switch (field->size) {
574			case 1:
575				entry->fields[n_u64].as_u8 = (u8)val;
576				break;
577
578			case 2:
579				entry->fields[n_u64].as_u16 = (u16)val;
580				break;
581
582			case 4:
583				entry->fields[n_u64].as_u32 = (u32)val;
584				break;
585
586			default:
587				entry->fields[n_u64].as_u64 = val;
588				break;
589			}
590			n_u64++;
591		}
592	}
593
594	trace_event_buffer_commit(&fbuffer);
595out:
596	ring_buffer_nest_end(buffer);
597}
598
599static void free_synth_event_print_fmt(struct trace_event_call *call)
600{
601	if (call) {
602		kfree(call->print_fmt);
603		call->print_fmt = NULL;
604	}
605}
606
607static int __set_synth_event_print_fmt(struct synth_event *event,
608				       char *buf, int len)
609{
610	const char *fmt;
611	int pos = 0;
612	int i;
613
614	/* When len=0, we just calculate the needed length */
615#define LEN_OR_ZERO (len ? len - pos : 0)
616
617	pos += snprintf(buf + pos, LEN_OR_ZERO, "\"");
618	for (i = 0; i < event->n_fields; i++) {
619		fmt = synth_field_fmt(event->fields[i]->type);
620		pos += snprintf(buf + pos, LEN_OR_ZERO, "%s=%s%s",
621				event->fields[i]->name, fmt,
622				i == event->n_fields - 1 ? "" : ", ");
623	}
624	pos += snprintf(buf + pos, LEN_OR_ZERO, "\"");
625
626	for (i = 0; i < event->n_fields; i++) {
627		if (event->fields[i]->is_string &&
628		    event->fields[i]->is_dynamic)
629			pos += snprintf(buf + pos, LEN_OR_ZERO,
630				", __get_str(%s)", event->fields[i]->name);
631		else if (event->fields[i]->is_stack)
632			pos += snprintf(buf + pos, LEN_OR_ZERO,
633				", __get_stacktrace(%s)", event->fields[i]->name);
634		else
635			pos += snprintf(buf + pos, LEN_OR_ZERO,
636					", REC->%s", event->fields[i]->name);
637	}
638
639#undef LEN_OR_ZERO
640
641	/* return the length of print_fmt */
642	return pos;
643}
644
645static int set_synth_event_print_fmt(struct trace_event_call *call)
646{
647	struct synth_event *event = call->data;
648	char *print_fmt;
649	int len;
650
651	/* First: called with 0 length to calculate the needed length */
652	len = __set_synth_event_print_fmt(event, NULL, 0);
653
654	print_fmt = kmalloc(len + 1, GFP_KERNEL);
655	if (!print_fmt)
656		return -ENOMEM;
657
658	/* Second: actually write the @print_fmt */
659	__set_synth_event_print_fmt(event, print_fmt, len + 1);
660	call->print_fmt = print_fmt;
661
662	return 0;
663}
664
665static void free_synth_field(struct synth_field *field)
666{
667	kfree(field->type);
668	kfree(field->name);
669	kfree(field);
670}
671
672static int check_field_version(const char *prefix, const char *field_type,
673			       const char *field_name)
674{
675	/*
676	 * For backward compatibility, the old synthetic event command
677	 * format did not require semicolons, and in order to not
678	 * break user space, that old format must still work. If a new
679	 * feature is added, then the format that uses the new feature
680	 * will be required to have semicolons, as nothing that uses
681	 * the old format would be using the new, yet to be created,
682	 * feature. When a new feature is added, this will detect it,
683	 * and return a number greater than 1, and require the format
684	 * to use semicolons.
685	 */
686	return 1;
687}
688
689static struct synth_field *parse_synth_field(int argc, char **argv,
690					     int *consumed, int *field_version)
691{
692	const char *prefix = NULL, *field_type = argv[0], *field_name, *array;
693	struct synth_field *field;
694	int len, ret = -ENOMEM;
695	struct seq_buf s;
696	ssize_t size;
697
698	if (!strcmp(field_type, "unsigned")) {
699		if (argc < 3) {
700			synth_err(SYNTH_ERR_INCOMPLETE_TYPE, errpos(field_type));
701			return ERR_PTR(-EINVAL);
702		}
703		prefix = "unsigned ";
704		field_type = argv[1];
705		field_name = argv[2];
706		*consumed += 3;
707	} else {
708		field_name = argv[1];
709		*consumed += 2;
710	}
711
712	if (!field_name) {
713		synth_err(SYNTH_ERR_INVALID_FIELD, errpos(field_type));
714		return ERR_PTR(-EINVAL);
715	}
716
717	*field_version = check_field_version(prefix, field_type, field_name);
718
719	field = kzalloc(sizeof(*field), GFP_KERNEL);
720	if (!field)
721		return ERR_PTR(-ENOMEM);
722
723	len = strlen(field_name);
724	array = strchr(field_name, '[');
725	if (array)
726		len -= strlen(array);
727
728	field->name = kmemdup_nul(field_name, len, GFP_KERNEL);
729	if (!field->name)
730		goto free;
731
732	if (!is_good_name(field->name)) {
733		synth_err(SYNTH_ERR_BAD_NAME, errpos(field_name));
734		ret = -EINVAL;
735		goto free;
736	}
737
738	len = strlen(field_type) + 1;
739
740	if (array)
741		len += strlen(array);
742
743	if (prefix)
744		len += strlen(prefix);
745
746	field->type = kzalloc(len, GFP_KERNEL);
747	if (!field->type)
748		goto free;
749
750	seq_buf_init(&s, field->type, len);
751	if (prefix)
752		seq_buf_puts(&s, prefix);
753	seq_buf_puts(&s, field_type);
754	if (array)
755		seq_buf_puts(&s, array);
756	if (WARN_ON_ONCE(!seq_buf_buffer_left(&s)))
757		goto free;
758
759	s.buffer[s.len] = '\0';
760
761	size = synth_field_size(field->type);
762	if (size < 0) {
763		if (array)
764			synth_err(SYNTH_ERR_INVALID_ARRAY_SPEC, errpos(field_name));
765		else
766			synth_err(SYNTH_ERR_INVALID_TYPE, errpos(field_type));
767		ret = -EINVAL;
768		goto free;
769	} else if (size == 0) {
770		if (synth_field_is_string(field->type) ||
771		    synth_field_is_stack(field->type)) {
772			char *type;
773
774			len = sizeof("__data_loc ") + strlen(field->type) + 1;
775			type = kzalloc(len, GFP_KERNEL);
776			if (!type)
777				goto free;
778
779			seq_buf_init(&s, type, len);
780			seq_buf_puts(&s, "__data_loc ");
781			seq_buf_puts(&s, field->type);
782
783			if (WARN_ON_ONCE(!seq_buf_buffer_left(&s)))
784				goto free;
785			s.buffer[s.len] = '\0';
786
787			kfree(field->type);
788			field->type = type;
789
790			field->is_dynamic = true;
791			size = sizeof(u64);
792		} else {
793			synth_err(SYNTH_ERR_INVALID_TYPE, errpos(field_type));
794			ret = -EINVAL;
795			goto free;
796		}
797	}
798	field->size = size;
799
800	if (synth_field_is_string(field->type))
801		field->is_string = true;
802	else if (synth_field_is_stack(field->type))
803		field->is_stack = true;
804
805	field->is_signed = synth_field_signed(field->type);
806 out:
807	return field;
808 free:
809	free_synth_field(field);
810	field = ERR_PTR(ret);
811	goto out;
812}
813
814static void free_synth_tracepoint(struct tracepoint *tp)
815{
816	if (!tp)
817		return;
818
819	kfree(tp->name);
820	kfree(tp);
821}
822
823static struct tracepoint *alloc_synth_tracepoint(char *name)
824{
825	struct tracepoint *tp;
826
827	tp = kzalloc(sizeof(*tp), GFP_KERNEL);
828	if (!tp)
829		return ERR_PTR(-ENOMEM);
830
831	tp->name = kstrdup(name, GFP_KERNEL);
832	if (!tp->name) {
833		kfree(tp);
834		return ERR_PTR(-ENOMEM);
835	}
836
837	return tp;
838}
839
840struct synth_event *find_synth_event(const char *name)
841{
842	struct dyn_event *pos;
843	struct synth_event *event;
844
845	for_each_dyn_event(pos) {
846		if (!is_synth_event(pos))
847			continue;
848		event = to_synth_event(pos);
849		if (strcmp(event->name, name) == 0)
850			return event;
851	}
852
853	return NULL;
854}
855
856static struct trace_event_fields synth_event_fields_array[] = {
857	{ .type = TRACE_FUNCTION_TYPE,
858	  .define_fields = synth_event_define_fields },
859	{}
860};
861
862static int register_synth_event(struct synth_event *event)
863{
864	struct trace_event_call *call = &event->call;
865	int ret = 0;
866
867	event->call.class = &event->class;
868	event->class.system = kstrdup(SYNTH_SYSTEM, GFP_KERNEL);
869	if (!event->class.system) {
870		ret = -ENOMEM;
871		goto out;
872	}
873
874	event->tp = alloc_synth_tracepoint(event->name);
875	if (IS_ERR(event->tp)) {
876		ret = PTR_ERR(event->tp);
877		event->tp = NULL;
878		goto out;
879	}
880
881	INIT_LIST_HEAD(&call->class->fields);
882	call->event.funcs = &synth_event_funcs;
883	call->class->fields_array = synth_event_fields_array;
884
885	ret = register_trace_event(&call->event);
886	if (!ret) {
887		ret = -ENODEV;
888		goto out;
889	}
890	call->flags = TRACE_EVENT_FL_TRACEPOINT;
891	call->class->reg = trace_event_reg;
892	call->class->probe = trace_event_raw_event_synth;
893	call->data = event;
894	call->tp = event->tp;
895
896	ret = trace_add_event_call(call);
897	if (ret) {
898		pr_warn("Failed to register synthetic event: %s\n",
899			trace_event_name(call));
900		goto err;
901	}
902
903	ret = set_synth_event_print_fmt(call);
904	/* unregister_trace_event() will be called inside */
905	if (ret < 0)
906		trace_remove_event_call(call);
907 out:
908	return ret;
909 err:
910	unregister_trace_event(&call->event);
911	goto out;
912}
913
914static int unregister_synth_event(struct synth_event *event)
915{
916	struct trace_event_call *call = &event->call;
917	int ret;
918
919	ret = trace_remove_event_call(call);
920
921	return ret;
922}
923
924static void free_synth_event(struct synth_event *event)
925{
926	unsigned int i;
927
928	if (!event)
929		return;
930
931	for (i = 0; i < event->n_fields; i++)
932		free_synth_field(event->fields[i]);
933
934	kfree(event->fields);
935	kfree(event->dynamic_fields);
936	kfree(event->name);
937	kfree(event->class.system);
938	free_synth_tracepoint(event->tp);
939	free_synth_event_print_fmt(&event->call);
940	kfree(event);
941}
942
943static struct synth_event *alloc_synth_event(const char *name, int n_fields,
944					     struct synth_field **fields)
945{
946	unsigned int i, j, n_dynamic_fields = 0;
947	struct synth_event *event;
948
949	event = kzalloc(sizeof(*event), GFP_KERNEL);
950	if (!event) {
951		event = ERR_PTR(-ENOMEM);
952		goto out;
953	}
954
955	event->name = kstrdup(name, GFP_KERNEL);
956	if (!event->name) {
957		kfree(event);
958		event = ERR_PTR(-ENOMEM);
959		goto out;
960	}
961
962	event->fields = kcalloc(n_fields, sizeof(*event->fields), GFP_KERNEL);
963	if (!event->fields) {
964		free_synth_event(event);
965		event = ERR_PTR(-ENOMEM);
966		goto out;
967	}
968
969	for (i = 0; i < n_fields; i++)
970		if (fields[i]->is_dynamic)
971			n_dynamic_fields++;
972
973	if (n_dynamic_fields) {
974		event->dynamic_fields = kcalloc(n_dynamic_fields,
975						sizeof(*event->dynamic_fields),
976						GFP_KERNEL);
977		if (!event->dynamic_fields) {
978			free_synth_event(event);
979			event = ERR_PTR(-ENOMEM);
980			goto out;
981		}
982	}
983
984	dyn_event_init(&event->devent, &synth_event_ops);
985
986	for (i = 0, j = 0; i < n_fields; i++) {
987		fields[i]->field_pos = i;
988		event->fields[i] = fields[i];
989
990		if (fields[i]->is_dynamic)
991			event->dynamic_fields[j++] = fields[i];
992	}
993	event->n_dynamic_fields = j;
994	event->n_fields = n_fields;
995 out:
996	return event;
997}
998
999static int synth_event_check_arg_fn(void *data)
1000{
1001	struct dynevent_arg_pair *arg_pair = data;
1002	int size;
1003
1004	size = synth_field_size((char *)arg_pair->lhs);
1005	if (size == 0) {
1006		if (strstr((char *)arg_pair->lhs, "["))
1007			return 0;
1008	}
1009
1010	return size ? 0 : -EINVAL;
1011}
1012
1013/**
1014 * synth_event_add_field - Add a new field to a synthetic event cmd
1015 * @cmd: A pointer to the dynevent_cmd struct representing the new event
1016 * @type: The type of the new field to add
1017 * @name: The name of the new field to add
1018 *
1019 * Add a new field to a synthetic event cmd object.  Field ordering is in
1020 * the same order the fields are added.
1021 *
1022 * See synth_field_size() for available types. If field_name contains
1023 * [n] the field is considered to be an array.
1024 *
1025 * Return: 0 if successful, error otherwise.
1026 */
1027int synth_event_add_field(struct dynevent_cmd *cmd, const char *type,
1028			  const char *name)
1029{
1030	struct dynevent_arg_pair arg_pair;
1031	int ret;
1032
1033	if (cmd->type != DYNEVENT_TYPE_SYNTH)
1034		return -EINVAL;
1035
1036	if (!type || !name)
1037		return -EINVAL;
1038
1039	dynevent_arg_pair_init(&arg_pair, 0, ';');
1040
1041	arg_pair.lhs = type;
1042	arg_pair.rhs = name;
1043
1044	ret = dynevent_arg_pair_add(cmd, &arg_pair, synth_event_check_arg_fn);
1045	if (ret)
1046		return ret;
1047
1048	if (++cmd->n_fields > SYNTH_FIELDS_MAX)
1049		ret = -EINVAL;
1050
1051	return ret;
1052}
1053EXPORT_SYMBOL_GPL(synth_event_add_field);
1054
1055/**
1056 * synth_event_add_field_str - Add a new field to a synthetic event cmd
1057 * @cmd: A pointer to the dynevent_cmd struct representing the new event
1058 * @type_name: The type and name of the new field to add, as a single string
1059 *
1060 * Add a new field to a synthetic event cmd object, as a single
1061 * string.  The @type_name string is expected to be of the form 'type
1062 * name', which will be appended by ';'.  No sanity checking is done -
1063 * what's passed in is assumed to already be well-formed.  Field
1064 * ordering is in the same order the fields are added.
1065 *
1066 * See synth_field_size() for available types. If field_name contains
1067 * [n] the field is considered to be an array.
1068 *
1069 * Return: 0 if successful, error otherwise.
1070 */
1071int synth_event_add_field_str(struct dynevent_cmd *cmd, const char *type_name)
1072{
1073	struct dynevent_arg arg;
1074	int ret;
1075
1076	if (cmd->type != DYNEVENT_TYPE_SYNTH)
1077		return -EINVAL;
1078
1079	if (!type_name)
1080		return -EINVAL;
1081
1082	dynevent_arg_init(&arg, ';');
1083
1084	arg.str = type_name;
1085
1086	ret = dynevent_arg_add(cmd, &arg, NULL);
1087	if (ret)
1088		return ret;
1089
1090	if (++cmd->n_fields > SYNTH_FIELDS_MAX)
1091		ret = -EINVAL;
1092
1093	return ret;
1094}
1095EXPORT_SYMBOL_GPL(synth_event_add_field_str);
1096
1097/**
1098 * synth_event_add_fields - Add multiple fields to a synthetic event cmd
1099 * @cmd: A pointer to the dynevent_cmd struct representing the new event
1100 * @fields: An array of type/name field descriptions
1101 * @n_fields: The number of field descriptions contained in the fields array
1102 *
1103 * Add a new set of fields to a synthetic event cmd object.  The event
1104 * fields that will be defined for the event should be passed in as an
1105 * array of struct synth_field_desc, and the number of elements in the
1106 * array passed in as n_fields.  Field ordering will retain the
1107 * ordering given in the fields array.
1108 *
1109 * See synth_field_size() for available types. If field_name contains
1110 * [n] the field is considered to be an array.
1111 *
1112 * Return: 0 if successful, error otherwise.
1113 */
1114int synth_event_add_fields(struct dynevent_cmd *cmd,
1115			   struct synth_field_desc *fields,
1116			   unsigned int n_fields)
1117{
1118	unsigned int i;
1119	int ret = 0;
1120
1121	for (i = 0; i < n_fields; i++) {
1122		if (fields[i].type == NULL || fields[i].name == NULL) {
1123			ret = -EINVAL;
1124			break;
1125		}
1126
1127		ret = synth_event_add_field(cmd, fields[i].type, fields[i].name);
1128		if (ret)
1129			break;
1130	}
1131
1132	return ret;
1133}
1134EXPORT_SYMBOL_GPL(synth_event_add_fields);
1135
1136/**
1137 * __synth_event_gen_cmd_start - Start a synthetic event command from arg list
1138 * @cmd: A pointer to the dynevent_cmd struct representing the new event
1139 * @name: The name of the synthetic event
1140 * @mod: The module creating the event, NULL if not created from a module
1141 * @...: Variable number of arg (pairs), one pair for each field
1142 *
1143 * NOTE: Users normally won't want to call this function directly, but
1144 * rather use the synth_event_gen_cmd_start() wrapper, which
1145 * automatically adds a NULL to the end of the arg list.  If this
1146 * function is used directly, make sure the last arg in the variable
1147 * arg list is NULL.
1148 *
1149 * Generate a synthetic event command to be executed by
1150 * synth_event_gen_cmd_end().  This function can be used to generate
1151 * the complete command or only the first part of it; in the latter
1152 * case, synth_event_add_field(), synth_event_add_field_str(), or
1153 * synth_event_add_fields() can be used to add more fields following
1154 * this.
1155 *
1156 * There should be an even number variable args, each pair consisting
1157 * of a type followed by a field name.
1158 *
1159 * See synth_field_size() for available types. If field_name contains
1160 * [n] the field is considered to be an array.
1161 *
1162 * Return: 0 if successful, error otherwise.
1163 */
1164int __synth_event_gen_cmd_start(struct dynevent_cmd *cmd, const char *name,
1165				struct module *mod, ...)
1166{
1167	struct dynevent_arg arg;
1168	va_list args;
1169	int ret;
1170
1171	cmd->event_name = name;
1172	cmd->private_data = mod;
1173
1174	if (cmd->type != DYNEVENT_TYPE_SYNTH)
1175		return -EINVAL;
1176
1177	dynevent_arg_init(&arg, 0);
1178	arg.str = name;
1179	ret = dynevent_arg_add(cmd, &arg, NULL);
1180	if (ret)
1181		return ret;
1182
1183	va_start(args, mod);
1184	for (;;) {
1185		const char *type, *name;
1186
1187		type = va_arg(args, const char *);
1188		if (!type)
1189			break;
1190		name = va_arg(args, const char *);
1191		if (!name)
1192			break;
1193
1194		if (++cmd->n_fields > SYNTH_FIELDS_MAX) {
1195			ret = -EINVAL;
1196			break;
1197		}
1198
1199		ret = synth_event_add_field(cmd, type, name);
1200		if (ret)
1201			break;
1202	}
1203	va_end(args);
1204
1205	return ret;
1206}
1207EXPORT_SYMBOL_GPL(__synth_event_gen_cmd_start);
1208
1209/**
1210 * synth_event_gen_cmd_array_start - Start synthetic event command from an array
1211 * @cmd: A pointer to the dynevent_cmd struct representing the new event
1212 * @name: The name of the synthetic event
1213 * @mod: The module creating the event, NULL if not created from a module
1214 * @fields: An array of type/name field descriptions
1215 * @n_fields: The number of field descriptions contained in the fields array
1216 *
1217 * Generate a synthetic event command to be executed by
1218 * synth_event_gen_cmd_end().  This function can be used to generate
1219 * the complete command or only the first part of it; in the latter
1220 * case, synth_event_add_field(), synth_event_add_field_str(), or
1221 * synth_event_add_fields() can be used to add more fields following
1222 * this.
1223 *
1224 * The event fields that will be defined for the event should be
1225 * passed in as an array of struct synth_field_desc, and the number of
1226 * elements in the array passed in as n_fields.  Field ordering will
1227 * retain the ordering given in the fields array.
1228 *
1229 * See synth_field_size() for available types. If field_name contains
1230 * [n] the field is considered to be an array.
1231 *
1232 * Return: 0 if successful, error otherwise.
1233 */
1234int synth_event_gen_cmd_array_start(struct dynevent_cmd *cmd, const char *name,
1235				    struct module *mod,
1236				    struct synth_field_desc *fields,
1237				    unsigned int n_fields)
1238{
1239	struct dynevent_arg arg;
1240	unsigned int i;
1241	int ret = 0;
1242
1243	cmd->event_name = name;
1244	cmd->private_data = mod;
1245
1246	if (cmd->type != DYNEVENT_TYPE_SYNTH)
1247		return -EINVAL;
1248
1249	if (n_fields > SYNTH_FIELDS_MAX)
1250		return -EINVAL;
1251
1252	dynevent_arg_init(&arg, 0);
1253	arg.str = name;
1254	ret = dynevent_arg_add(cmd, &arg, NULL);
1255	if (ret)
1256		return ret;
1257
1258	for (i = 0; i < n_fields; i++) {
1259		if (fields[i].type == NULL || fields[i].name == NULL)
1260			return -EINVAL;
1261
1262		ret = synth_event_add_field(cmd, fields[i].type, fields[i].name);
1263		if (ret)
1264			break;
1265	}
1266
1267	return ret;
1268}
1269EXPORT_SYMBOL_GPL(synth_event_gen_cmd_array_start);
1270
1271static int __create_synth_event(const char *name, const char *raw_fields)
1272{
1273	char **argv, *field_str, *tmp_fields, *saved_fields = NULL;
1274	struct synth_field *field, *fields[SYNTH_FIELDS_MAX];
1275	int consumed, cmd_version = 1, n_fields_this_loop;
1276	int i, argc, n_fields = 0, ret = 0;
1277	struct synth_event *event = NULL;
1278
1279	/*
1280	 * Argument syntax:
1281	 *  - Add synthetic event: <event_name> field[;field] ...
1282	 *  - Remove synthetic event: !<event_name> field[;field] ...
1283	 *      where 'field' = type field_name
1284	 */
1285
1286	if (name[0] == '\0') {
1287		synth_err(SYNTH_ERR_INVALID_CMD, 0);
1288		return -EINVAL;
1289	}
1290
1291	if (!is_good_name(name)) {
1292		synth_err(SYNTH_ERR_BAD_NAME, errpos(name));
1293		return -EINVAL;
1294	}
1295
1296	mutex_lock(&event_mutex);
1297
1298	event = find_synth_event(name);
1299	if (event) {
1300		synth_err(SYNTH_ERR_EVENT_EXISTS, errpos(name));
1301		ret = -EEXIST;
1302		goto err;
1303	}
1304
1305	tmp_fields = saved_fields = kstrdup(raw_fields, GFP_KERNEL);
1306	if (!tmp_fields) {
1307		ret = -ENOMEM;
1308		goto err;
1309	}
1310
1311	while ((field_str = strsep(&tmp_fields, ";")) != NULL) {
1312		argv = argv_split(GFP_KERNEL, field_str, &argc);
1313		if (!argv) {
1314			ret = -ENOMEM;
1315			goto err;
1316		}
1317
1318		if (!argc) {
1319			argv_free(argv);
1320			continue;
1321		}
1322
1323		n_fields_this_loop = 0;
1324		consumed = 0;
1325		while (argc > consumed) {
1326			int field_version;
1327
1328			field = parse_synth_field(argc - consumed,
1329						  argv + consumed, &consumed,
1330						  &field_version);
1331			if (IS_ERR(field)) {
1332				ret = PTR_ERR(field);
1333				goto err_free_arg;
1334			}
1335
1336			/*
1337			 * Track the highest version of any field we
1338			 * found in the command.
1339			 */
1340			if (field_version > cmd_version)
1341				cmd_version = field_version;
1342
1343			/*
1344			 * Now sort out what is and isn't valid for
1345			 * each supported version.
1346			 *
1347			 * If we see more than 1 field per loop, it
1348			 * means we have multiple fields between
1349			 * semicolons, and that's something we no
1350			 * longer support in a version 2 or greater
1351			 * command.
1352			 */
1353			if (cmd_version > 1 && n_fields_this_loop >= 1) {
1354				synth_err(SYNTH_ERR_INVALID_CMD, errpos(field_str));
1355				ret = -EINVAL;
1356				goto err_free_arg;
1357			}
1358
1359			if (n_fields == SYNTH_FIELDS_MAX) {
1360				synth_err(SYNTH_ERR_TOO_MANY_FIELDS, 0);
1361				ret = -EINVAL;
1362				goto err_free_arg;
1363			}
1364			fields[n_fields++] = field;
1365
1366			n_fields_this_loop++;
1367		}
1368		argv_free(argv);
1369
1370		if (consumed < argc) {
1371			synth_err(SYNTH_ERR_INVALID_CMD, 0);
1372			ret = -EINVAL;
1373			goto err;
1374		}
1375
1376	}
1377
1378	if (n_fields == 0) {
1379		synth_err(SYNTH_ERR_INVALID_CMD, 0);
1380		ret = -EINVAL;
1381		goto err;
1382	}
1383
1384	event = alloc_synth_event(name, n_fields, fields);
1385	if (IS_ERR(event)) {
1386		ret = PTR_ERR(event);
1387		event = NULL;
1388		goto err;
1389	}
1390	ret = register_synth_event(event);
1391	if (!ret)
1392		dyn_event_add(&event->devent, &event->call);
1393	else
1394		free_synth_event(event);
1395 out:
1396	mutex_unlock(&event_mutex);
1397
1398	kfree(saved_fields);
1399
1400	return ret;
1401 err_free_arg:
1402	argv_free(argv);
1403 err:
1404	for (i = 0; i < n_fields; i++)
1405		free_synth_field(fields[i]);
1406
1407	goto out;
1408}
1409
1410/**
1411 * synth_event_create - Create a new synthetic event
1412 * @name: The name of the new synthetic event
1413 * @fields: An array of type/name field descriptions
1414 * @n_fields: The number of field descriptions contained in the fields array
1415 * @mod: The module creating the event, NULL if not created from a module
1416 *
1417 * Create a new synthetic event with the given name under the
1418 * trace/events/synthetic/ directory.  The event fields that will be
1419 * defined for the event should be passed in as an array of struct
1420 * synth_field_desc, and the number elements in the array passed in as
1421 * n_fields. Field ordering will retain the ordering given in the
1422 * fields array.
1423 *
1424 * If the new synthetic event is being created from a module, the mod
1425 * param must be non-NULL.  This will ensure that the trace buffer
1426 * won't contain unreadable events.
1427 *
1428 * The new synth event should be deleted using synth_event_delete()
1429 * function.  The new synthetic event can be generated from modules or
1430 * other kernel code using trace_synth_event() and related functions.
1431 *
1432 * Return: 0 if successful, error otherwise.
1433 */
1434int synth_event_create(const char *name, struct synth_field_desc *fields,
1435		       unsigned int n_fields, struct module *mod)
1436{
1437	struct dynevent_cmd cmd;
1438	char *buf;
1439	int ret;
1440
1441	buf = kzalloc(MAX_DYNEVENT_CMD_LEN, GFP_KERNEL);
1442	if (!buf)
1443		return -ENOMEM;
1444
1445	synth_event_cmd_init(&cmd, buf, MAX_DYNEVENT_CMD_LEN);
1446
1447	ret = synth_event_gen_cmd_array_start(&cmd, name, mod,
1448					      fields, n_fields);
1449	if (ret)
1450		goto out;
1451
1452	ret = synth_event_gen_cmd_end(&cmd);
1453 out:
1454	kfree(buf);
1455
1456	return ret;
1457}
1458EXPORT_SYMBOL_GPL(synth_event_create);
1459
1460static int destroy_synth_event(struct synth_event *se)
1461{
1462	int ret;
1463
1464	if (se->ref)
1465		return -EBUSY;
1466
1467	if (trace_event_dyn_busy(&se->call))
1468		return -EBUSY;
1469
1470	ret = unregister_synth_event(se);
1471	if (!ret) {
1472		dyn_event_remove(&se->devent);
1473		free_synth_event(se);
1474	}
1475
1476	return ret;
1477}
1478
1479/**
1480 * synth_event_delete - Delete a synthetic event
1481 * @event_name: The name of the new synthetic event
1482 *
1483 * Delete a synthetic event that was created with synth_event_create().
1484 *
1485 * Return: 0 if successful, error otherwise.
1486 */
1487int synth_event_delete(const char *event_name)
1488{
1489	struct synth_event *se = NULL;
1490	struct module *mod = NULL;
1491	int ret = -ENOENT;
1492
1493	mutex_lock(&event_mutex);
1494	se = find_synth_event(event_name);
1495	if (se) {
1496		mod = se->mod;
1497		ret = destroy_synth_event(se);
1498	}
1499	mutex_unlock(&event_mutex);
1500
1501	if (mod) {
1502		/*
1503		 * It is safest to reset the ring buffer if the module
1504		 * being unloaded registered any events that were
1505		 * used. The only worry is if a new module gets
1506		 * loaded, and takes on the same id as the events of
1507		 * this module. When printing out the buffer, traced
1508		 * events left over from this module may be passed to
1509		 * the new module events and unexpected results may
1510		 * occur.
1511		 */
1512		tracing_reset_all_online_cpus();
1513	}
1514
1515	return ret;
1516}
1517EXPORT_SYMBOL_GPL(synth_event_delete);
1518
1519static int check_command(const char *raw_command)
1520{
1521	char **argv = NULL, *cmd, *saved_cmd, *name_and_field;
1522	int argc, ret = 0;
1523
1524	cmd = saved_cmd = kstrdup(raw_command, GFP_KERNEL);
1525	if (!cmd)
1526		return -ENOMEM;
1527
1528	name_and_field = strsep(&cmd, ";");
1529	if (!name_and_field) {
1530		ret = -EINVAL;
1531		goto free;
1532	}
1533
1534	if (name_and_field[0] == '!')
1535		goto free;
1536
1537	argv = argv_split(GFP_KERNEL, name_and_field, &argc);
1538	if (!argv) {
1539		ret = -ENOMEM;
1540		goto free;
1541	}
1542	argv_free(argv);
1543
1544	if (argc < 3)
1545		ret = -EINVAL;
1546free:
1547	kfree(saved_cmd);
1548
1549	return ret;
1550}
1551
1552static int create_or_delete_synth_event(const char *raw_command)
1553{
1554	char *name = NULL, *fields, *p;
1555	int ret = 0;
1556
1557	raw_command = skip_spaces(raw_command);
1558	if (raw_command[0] == '\0')
1559		return ret;
1560
1561	last_cmd_set(raw_command);
1562
1563	ret = check_command(raw_command);
1564	if (ret) {
1565		synth_err(SYNTH_ERR_INVALID_CMD, 0);
1566		return ret;
1567	}
1568
1569	p = strpbrk(raw_command, " \t");
1570	if (!p && raw_command[0] != '!') {
1571		synth_err(SYNTH_ERR_INVALID_CMD, 0);
1572		ret = -EINVAL;
1573		goto free;
1574	}
1575
1576	name = kmemdup_nul(raw_command, p ? p - raw_command : strlen(raw_command), GFP_KERNEL);
1577	if (!name)
1578		return -ENOMEM;
1579
1580	if (name[0] == '!') {
1581		ret = synth_event_delete(name + 1);
1582		goto free;
1583	}
1584
1585	fields = skip_spaces(p);
1586
1587	ret = __create_synth_event(name, fields);
1588free:
1589	kfree(name);
1590
1591	return ret;
1592}
1593
1594static int synth_event_run_command(struct dynevent_cmd *cmd)
1595{
1596	struct synth_event *se;
1597	int ret;
1598
1599	ret = create_or_delete_synth_event(cmd->seq.buffer);
1600	if (ret)
1601		return ret;
1602
1603	se = find_synth_event(cmd->event_name);
1604	if (WARN_ON(!se))
1605		return -ENOENT;
1606
1607	se->mod = cmd->private_data;
1608
1609	return ret;
1610}
1611
1612/**
1613 * synth_event_cmd_init - Initialize a synthetic event command object
1614 * @cmd: A pointer to the dynevent_cmd struct representing the new event
1615 * @buf: A pointer to the buffer used to build the command
1616 * @maxlen: The length of the buffer passed in @buf
1617 *
1618 * Initialize a synthetic event command object.  Use this before
1619 * calling any of the other dyenvent_cmd functions.
1620 */
1621void synth_event_cmd_init(struct dynevent_cmd *cmd, char *buf, int maxlen)
1622{
1623	dynevent_cmd_init(cmd, buf, maxlen, DYNEVENT_TYPE_SYNTH,
1624			  synth_event_run_command);
1625}
1626EXPORT_SYMBOL_GPL(synth_event_cmd_init);
1627
1628static inline int
1629__synth_event_trace_init(struct trace_event_file *file,
1630			 struct synth_event_trace_state *trace_state)
1631{
1632	int ret = 0;
1633
1634	memset(trace_state, '\0', sizeof(*trace_state));
1635
1636	/*
1637	 * Normal event tracing doesn't get called at all unless the
1638	 * ENABLED bit is set (which attaches the probe thus allowing
1639	 * this code to be called, etc).  Because this is called
1640	 * directly by the user, we don't have that but we still need
1641	 * to honor not logging when disabled.  For the iterated
1642	 * trace case, we save the enabled state upon start and just
1643	 * ignore the following data calls.
1644	 */
1645	if (!(file->flags & EVENT_FILE_FL_ENABLED) ||
1646	    trace_trigger_soft_disabled(file)) {
1647		trace_state->disabled = true;
1648		ret = -ENOENT;
1649		goto out;
1650	}
1651
1652	trace_state->event = file->event_call->data;
1653out:
1654	return ret;
1655}
1656
1657static inline int
1658__synth_event_trace_start(struct trace_event_file *file,
1659			  struct synth_event_trace_state *trace_state,
1660			  int dynamic_fields_size)
1661{
1662	int entry_size, fields_size = 0;
1663	int ret = 0;
1664
1665	fields_size = trace_state->event->n_u64 * sizeof(u64);
1666	fields_size += dynamic_fields_size;
1667
1668	/*
1669	 * Avoid ring buffer recursion detection, as this event
1670	 * is being performed within another event.
1671	 */
1672	trace_state->buffer = file->tr->array_buffer.buffer;
1673	ring_buffer_nest_start(trace_state->buffer);
1674
1675	entry_size = sizeof(*trace_state->entry) + fields_size;
1676	trace_state->entry = trace_event_buffer_reserve(&trace_state->fbuffer,
1677							file,
1678							entry_size);
1679	if (!trace_state->entry) {
1680		ring_buffer_nest_end(trace_state->buffer);
1681		ret = -EINVAL;
1682	}
1683
1684	return ret;
1685}
1686
1687static inline void
1688__synth_event_trace_end(struct synth_event_trace_state *trace_state)
1689{
1690	trace_event_buffer_commit(&trace_state->fbuffer);
1691
1692	ring_buffer_nest_end(trace_state->buffer);
1693}
1694
1695/**
1696 * synth_event_trace - Trace a synthetic event
1697 * @file: The trace_event_file representing the synthetic event
1698 * @n_vals: The number of values in vals
1699 * @...: Variable number of args containing the event values
1700 *
1701 * Trace a synthetic event using the values passed in the variable
1702 * argument list.
1703 *
1704 * The argument list should be a list 'n_vals' u64 values.  The number
1705 * of vals must match the number of field in the synthetic event, and
1706 * must be in the same order as the synthetic event fields.
1707 *
1708 * All vals should be cast to u64, and string vals are just pointers
1709 * to strings, cast to u64.  Strings will be copied into space
1710 * reserved in the event for the string, using these pointers.
1711 *
1712 * Return: 0 on success, err otherwise.
1713 */
1714int synth_event_trace(struct trace_event_file *file, unsigned int n_vals, ...)
1715{
1716	unsigned int i, n_u64, len, data_size = 0;
1717	struct synth_event_trace_state state;
1718	va_list args;
1719	int ret;
1720
1721	ret = __synth_event_trace_init(file, &state);
1722	if (ret) {
1723		if (ret == -ENOENT)
1724			ret = 0; /* just disabled, not really an error */
1725		return ret;
1726	}
1727
1728	if (state.event->n_dynamic_fields) {
1729		va_start(args, n_vals);
1730
1731		for (i = 0; i < state.event->n_fields; i++) {
1732			u64 val = va_arg(args, u64);
1733
1734			if (state.event->fields[i]->is_string &&
1735			    state.event->fields[i]->is_dynamic) {
1736				char *str_val = (char *)(long)val;
1737
1738				data_size += strlen(str_val) + 1;
1739			}
1740		}
1741
1742		va_end(args);
1743	}
1744
1745	ret = __synth_event_trace_start(file, &state, data_size);
1746	if (ret)
1747		return ret;
1748
1749	if (n_vals != state.event->n_fields) {
1750		ret = -EINVAL;
1751		goto out;
1752	}
1753
1754	data_size = 0;
1755
1756	va_start(args, n_vals);
1757	for (i = 0, n_u64 = 0; i < state.event->n_fields; i++) {
1758		u64 val;
1759
1760		val = va_arg(args, u64);
1761
1762		if (state.event->fields[i]->is_string) {
1763			char *str_val = (char *)(long)val;
1764
1765			len = trace_string(state.entry, state.event, str_val,
1766					   state.event->fields[i]->is_dynamic,
1767					   data_size, &n_u64);
1768			data_size += len; /* only dynamic string increments */
1769		} else {
1770			struct synth_field *field = state.event->fields[i];
1771
1772			switch (field->size) {
1773			case 1:
1774				state.entry->fields[n_u64].as_u8 = (u8)val;
1775				break;
1776
1777			case 2:
1778				state.entry->fields[n_u64].as_u16 = (u16)val;
1779				break;
1780
1781			case 4:
1782				state.entry->fields[n_u64].as_u32 = (u32)val;
1783				break;
1784
1785			default:
1786				state.entry->fields[n_u64].as_u64 = val;
1787				break;
1788			}
1789			n_u64++;
1790		}
1791	}
1792	va_end(args);
1793out:
1794	__synth_event_trace_end(&state);
1795
1796	return ret;
1797}
1798EXPORT_SYMBOL_GPL(synth_event_trace);
1799
1800/**
1801 * synth_event_trace_array - Trace a synthetic event from an array
1802 * @file: The trace_event_file representing the synthetic event
1803 * @vals: Array of values
1804 * @n_vals: The number of values in vals
1805 *
1806 * Trace a synthetic event using the values passed in as 'vals'.
1807 *
1808 * The 'vals' array is just an array of 'n_vals' u64.  The number of
1809 * vals must match the number of field in the synthetic event, and
1810 * must be in the same order as the synthetic event fields.
1811 *
1812 * All vals should be cast to u64, and string vals are just pointers
1813 * to strings, cast to u64.  Strings will be copied into space
1814 * reserved in the event for the string, using these pointers.
1815 *
1816 * Return: 0 on success, err otherwise.
1817 */
1818int synth_event_trace_array(struct trace_event_file *file, u64 *vals,
1819			    unsigned int n_vals)
1820{
1821	unsigned int i, n_u64, field_pos, len, data_size = 0;
1822	struct synth_event_trace_state state;
1823	char *str_val;
1824	int ret;
1825
1826	ret = __synth_event_trace_init(file, &state);
1827	if (ret) {
1828		if (ret == -ENOENT)
1829			ret = 0; /* just disabled, not really an error */
1830		return ret;
1831	}
1832
1833	if (state.event->n_dynamic_fields) {
1834		for (i = 0; i < state.event->n_dynamic_fields; i++) {
1835			field_pos = state.event->dynamic_fields[i]->field_pos;
1836			str_val = (char *)(long)vals[field_pos];
1837			len = strlen(str_val) + 1;
1838			data_size += len;
1839		}
1840	}
1841
1842	ret = __synth_event_trace_start(file, &state, data_size);
1843	if (ret)
1844		return ret;
1845
1846	if (n_vals != state.event->n_fields) {
1847		ret = -EINVAL;
1848		goto out;
1849	}
1850
1851	data_size = 0;
1852
1853	for (i = 0, n_u64 = 0; i < state.event->n_fields; i++) {
1854		if (state.event->fields[i]->is_string) {
1855			char *str_val = (char *)(long)vals[i];
1856
1857			len = trace_string(state.entry, state.event, str_val,
1858					   state.event->fields[i]->is_dynamic,
1859					   data_size, &n_u64);
1860			data_size += len; /* only dynamic string increments */
1861		} else {
1862			struct synth_field *field = state.event->fields[i];
1863			u64 val = vals[i];
1864
1865			switch (field->size) {
1866			case 1:
1867				state.entry->fields[n_u64].as_u8 = (u8)val;
1868				break;
1869
1870			case 2:
1871				state.entry->fields[n_u64].as_u16 = (u16)val;
1872				break;
1873
1874			case 4:
1875				state.entry->fields[n_u64].as_u32 = (u32)val;
1876				break;
1877
1878			default:
1879				state.entry->fields[n_u64].as_u64 = val;
1880				break;
1881			}
1882			n_u64++;
1883		}
1884	}
1885out:
1886	__synth_event_trace_end(&state);
1887
1888	return ret;
1889}
1890EXPORT_SYMBOL_GPL(synth_event_trace_array);
1891
1892/**
1893 * synth_event_trace_start - Start piecewise synthetic event trace
1894 * @file: The trace_event_file representing the synthetic event
1895 * @trace_state: A pointer to object tracking the piecewise trace state
1896 *
1897 * Start the trace of a synthetic event field-by-field rather than all
1898 * at once.
1899 *
1900 * This function 'opens' an event trace, which means space is reserved
1901 * for the event in the trace buffer, after which the event's
1902 * individual field values can be set through either
1903 * synth_event_add_next_val() or synth_event_add_val().
1904 *
1905 * A pointer to a trace_state object is passed in, which will keep
1906 * track of the current event trace state until the event trace is
1907 * closed (and the event finally traced) using
1908 * synth_event_trace_end().
1909 *
1910 * Note that synth_event_trace_end() must be called after all values
1911 * have been added for each event trace, regardless of whether adding
1912 * all field values succeeded or not.
1913 *
1914 * Note also that for a given event trace, all fields must be added
1915 * using either synth_event_add_next_val() or synth_event_add_val()
1916 * but not both together or interleaved.
1917 *
1918 * Return: 0 on success, err otherwise.
1919 */
1920int synth_event_trace_start(struct trace_event_file *file,
1921			    struct synth_event_trace_state *trace_state)
1922{
1923	int ret;
1924
1925	if (!trace_state)
1926		return -EINVAL;
1927
1928	ret = __synth_event_trace_init(file, trace_state);
1929	if (ret) {
1930		if (ret == -ENOENT)
1931			ret = 0; /* just disabled, not really an error */
1932		return ret;
1933	}
1934
1935	if (trace_state->event->n_dynamic_fields)
1936		return -ENOTSUPP;
1937
1938	ret = __synth_event_trace_start(file, trace_state, 0);
1939
1940	return ret;
1941}
1942EXPORT_SYMBOL_GPL(synth_event_trace_start);
1943
1944static int __synth_event_add_val(const char *field_name, u64 val,
1945				 struct synth_event_trace_state *trace_state)
1946{
1947	struct synth_field *field = NULL;
1948	struct synth_trace_event *entry;
1949	struct synth_event *event;
1950	int i, ret = 0;
1951
1952	if (!trace_state) {
1953		ret = -EINVAL;
1954		goto out;
1955	}
1956
1957	/* can't mix add_next_synth_val() with add_synth_val() */
1958	if (field_name) {
1959		if (trace_state->add_next) {
1960			ret = -EINVAL;
1961			goto out;
1962		}
1963		trace_state->add_name = true;
1964	} else {
1965		if (trace_state->add_name) {
1966			ret = -EINVAL;
1967			goto out;
1968		}
1969		trace_state->add_next = true;
1970	}
1971
1972	if (trace_state->disabled)
1973		goto out;
1974
1975	event = trace_state->event;
1976	if (trace_state->add_name) {
1977		for (i = 0; i < event->n_fields; i++) {
1978			field = event->fields[i];
1979			if (strcmp(field->name, field_name) == 0)
1980				break;
1981		}
1982		if (!field) {
1983			ret = -EINVAL;
1984			goto out;
1985		}
1986	} else {
1987		if (trace_state->cur_field >= event->n_fields) {
1988			ret = -EINVAL;
1989			goto out;
1990		}
1991		field = event->fields[trace_state->cur_field++];
1992	}
1993
1994	entry = trace_state->entry;
1995	if (field->is_string) {
1996		char *str_val = (char *)(long)val;
1997		char *str_field;
1998
1999		if (field->is_dynamic) { /* add_val can't do dynamic strings */
2000			ret = -EINVAL;
2001			goto out;
2002		}
2003
2004		if (!str_val) {
2005			ret = -EINVAL;
2006			goto out;
2007		}
2008
2009		str_field = (char *)&entry->fields[field->offset];
2010		strscpy(str_field, str_val, STR_VAR_LEN_MAX);
2011	} else {
2012		switch (field->size) {
2013		case 1:
2014			trace_state->entry->fields[field->offset].as_u8 = (u8)val;
2015			break;
2016
2017		case 2:
2018			trace_state->entry->fields[field->offset].as_u16 = (u16)val;
2019			break;
2020
2021		case 4:
2022			trace_state->entry->fields[field->offset].as_u32 = (u32)val;
2023			break;
2024
2025		default:
2026			trace_state->entry->fields[field->offset].as_u64 = val;
2027			break;
2028		}
2029	}
2030 out:
2031	return ret;
2032}
2033
2034/**
2035 * synth_event_add_next_val - Add the next field's value to an open synth trace
2036 * @val: The value to set the next field to
2037 * @trace_state: A pointer to object tracking the piecewise trace state
2038 *
2039 * Set the value of the next field in an event that's been opened by
2040 * synth_event_trace_start().
2041 *
2042 * The val param should be the value cast to u64.  If the value points
2043 * to a string, the val param should be a char * cast to u64.
2044 *
2045 * This function assumes all the fields in an event are to be set one
2046 * after another - successive calls to this function are made, one for
2047 * each field, in the order of the fields in the event, until all
2048 * fields have been set.  If you'd rather set each field individually
2049 * without regard to ordering, synth_event_add_val() can be used
2050 * instead.
2051 *
2052 * Note however that synth_event_add_next_val() and
2053 * synth_event_add_val() can't be intermixed for a given event trace -
2054 * one or the other but not both can be used at the same time.
2055 *
2056 * Note also that synth_event_trace_end() must be called after all
2057 * values have been added for each event trace, regardless of whether
2058 * adding all field values succeeded or not.
2059 *
2060 * Return: 0 on success, err otherwise.
2061 */
2062int synth_event_add_next_val(u64 val,
2063			     struct synth_event_trace_state *trace_state)
2064{
2065	return __synth_event_add_val(NULL, val, trace_state);
2066}
2067EXPORT_SYMBOL_GPL(synth_event_add_next_val);
2068
2069/**
2070 * synth_event_add_val - Add a named field's value to an open synth trace
2071 * @field_name: The name of the synthetic event field value to set
2072 * @val: The value to set the named field to
2073 * @trace_state: A pointer to object tracking the piecewise trace state
2074 *
2075 * Set the value of the named field in an event that's been opened by
2076 * synth_event_trace_start().
2077 *
2078 * The val param should be the value cast to u64.  If the value points
2079 * to a string, the val param should be a char * cast to u64.
2080 *
2081 * This function looks up the field name, and if found, sets the field
2082 * to the specified value.  This lookup makes this function more
2083 * expensive than synth_event_add_next_val(), so use that or the
2084 * none-piecewise synth_event_trace() instead if efficiency is more
2085 * important.
2086 *
2087 * Note however that synth_event_add_next_val() and
2088 * synth_event_add_val() can't be intermixed for a given event trace -
2089 * one or the other but not both can be used at the same time.
2090 *
2091 * Note also that synth_event_trace_end() must be called after all
2092 * values have been added for each event trace, regardless of whether
2093 * adding all field values succeeded or not.
2094 *
2095 * Return: 0 on success, err otherwise.
2096 */
2097int synth_event_add_val(const char *field_name, u64 val,
2098			struct synth_event_trace_state *trace_state)
2099{
2100	return __synth_event_add_val(field_name, val, trace_state);
2101}
2102EXPORT_SYMBOL_GPL(synth_event_add_val);
2103
2104/**
2105 * synth_event_trace_end - End piecewise synthetic event trace
2106 * @trace_state: A pointer to object tracking the piecewise trace state
2107 *
2108 * End the trace of a synthetic event opened by
2109 * synth_event_trace__start().
2110 *
2111 * This function 'closes' an event trace, which basically means that
2112 * it commits the reserved event and cleans up other loose ends.
2113 *
2114 * A pointer to a trace_state object is passed in, which will keep
2115 * track of the current event trace state opened with
2116 * synth_event_trace_start().
2117 *
2118 * Note that this function must be called after all values have been
2119 * added for each event trace, regardless of whether adding all field
2120 * values succeeded or not.
2121 *
2122 * Return: 0 on success, err otherwise.
2123 */
2124int synth_event_trace_end(struct synth_event_trace_state *trace_state)
2125{
2126	if (!trace_state)
2127		return -EINVAL;
2128
2129	__synth_event_trace_end(trace_state);
2130
2131	return 0;
2132}
2133EXPORT_SYMBOL_GPL(synth_event_trace_end);
2134
2135static int create_synth_event(const char *raw_command)
2136{
2137	char *fields, *p;
2138	const char *name;
2139	int len, ret = 0;
2140
2141	raw_command = skip_spaces(raw_command);
2142	if (raw_command[0] == '\0')
2143		return ret;
2144
2145	last_cmd_set(raw_command);
2146
2147	name = raw_command;
2148
2149	/* Don't try to process if not our system */
2150	if (name[0] != 's' || name[1] != ':')
2151		return -ECANCELED;
2152	name += 2;
2153
2154	p = strpbrk(raw_command, " \t");
2155	if (!p) {
2156		synth_err(SYNTH_ERR_INVALID_CMD, 0);
2157		return -EINVAL;
2158	}
2159
2160	fields = skip_spaces(p);
2161
2162	/* This interface accepts group name prefix */
2163	if (strchr(name, '/')) {
2164		len = str_has_prefix(name, SYNTH_SYSTEM "/");
2165		if (len == 0) {
2166			synth_err(SYNTH_ERR_INVALID_DYN_CMD, 0);
2167			return -EINVAL;
2168		}
2169		name += len;
2170	}
2171
2172	len = name - raw_command;
2173
2174	ret = check_command(raw_command + len);
2175	if (ret) {
2176		synth_err(SYNTH_ERR_INVALID_CMD, 0);
2177		return ret;
2178	}
2179
2180	name = kmemdup_nul(raw_command + len, p - raw_command - len, GFP_KERNEL);
2181	if (!name)
2182		return -ENOMEM;
2183
2184	ret = __create_synth_event(name, fields);
2185
2186	kfree(name);
2187
2188	return ret;
2189}
2190
2191static int synth_event_release(struct dyn_event *ev)
2192{
2193	struct synth_event *event = to_synth_event(ev);
2194	int ret;
2195
2196	if (event->ref)
2197		return -EBUSY;
2198
2199	if (trace_event_dyn_busy(&event->call))
2200		return -EBUSY;
2201
2202	ret = unregister_synth_event(event);
2203	if (ret)
2204		return ret;
2205
2206	dyn_event_remove(ev);
2207	free_synth_event(event);
2208	return 0;
2209}
2210
2211static int __synth_event_show(struct seq_file *m, struct synth_event *event)
2212{
2213	struct synth_field *field;
2214	unsigned int i;
2215	char *type, *t;
2216
2217	seq_printf(m, "%s\t", event->name);
2218
2219	for (i = 0; i < event->n_fields; i++) {
2220		field = event->fields[i];
2221
2222		type = field->type;
2223		t = strstr(type, "__data_loc");
2224		if (t) { /* __data_loc belongs in format but not event desc */
2225			t += sizeof("__data_loc");
2226			type = t;
2227		}
2228
2229		/* parameter values */
2230		seq_printf(m, "%s %s%s", type, field->name,
2231			   i == event->n_fields - 1 ? "" : "; ");
2232	}
2233
2234	seq_putc(m, '\n');
2235
2236	return 0;
2237}
2238
2239static int synth_event_show(struct seq_file *m, struct dyn_event *ev)
2240{
2241	struct synth_event *event = to_synth_event(ev);
2242
2243	seq_printf(m, "s:%s/", event->class.system);
2244
2245	return __synth_event_show(m, event);
2246}
2247
2248static int synth_events_seq_show(struct seq_file *m, void *v)
2249{
2250	struct dyn_event *ev = v;
2251
2252	if (!is_synth_event(ev))
2253		return 0;
2254
2255	return __synth_event_show(m, to_synth_event(ev));
2256}
2257
2258static const struct seq_operations synth_events_seq_op = {
2259	.start	= dyn_event_seq_start,
2260	.next	= dyn_event_seq_next,
2261	.stop	= dyn_event_seq_stop,
2262	.show	= synth_events_seq_show,
2263};
2264
2265static int synth_events_open(struct inode *inode, struct file *file)
2266{
2267	int ret;
2268
2269	ret = security_locked_down(LOCKDOWN_TRACEFS);
2270	if (ret)
2271		return ret;
2272
2273	if ((file->f_mode & FMODE_WRITE) && (file->f_flags & O_TRUNC)) {
2274		ret = dyn_events_release_all(&synth_event_ops);
2275		if (ret < 0)
2276			return ret;
2277	}
2278
2279	return seq_open(file, &synth_events_seq_op);
2280}
2281
2282static ssize_t synth_events_write(struct file *file,
2283				  const char __user *buffer,
2284				  size_t count, loff_t *ppos)
2285{
2286	return trace_parse_run_command(file, buffer, count, ppos,
2287				       create_or_delete_synth_event);
2288}
2289
2290static const struct file_operations synth_events_fops = {
2291	.open           = synth_events_open,
2292	.write		= synth_events_write,
2293	.read           = seq_read,
2294	.llseek         = seq_lseek,
2295	.release        = seq_release,
2296};
2297
2298/*
2299 * Register dynevent at core_initcall. This allows kernel to setup kprobe
2300 * events in postcore_initcall without tracefs.
2301 */
2302static __init int trace_events_synth_init_early(void)
2303{
2304	int err = 0;
2305
2306	err = dyn_event_register(&synth_event_ops);
2307	if (err)
2308		pr_warn("Could not register synth_event_ops\n");
2309
2310	return err;
2311}
2312core_initcall(trace_events_synth_init_early);
2313
2314static __init int trace_events_synth_init(void)
2315{
2316	struct dentry *entry = NULL;
2317	int err = 0;
2318	err = tracing_init_dentry();
2319	if (err)
2320		goto err;
2321
2322	entry = tracefs_create_file("synthetic_events", TRACE_MODE_WRITE,
2323				    NULL, NULL, &synth_events_fops);
2324	if (!entry) {
2325		err = -ENODEV;
2326		goto err;
2327	}
2328
2329	return err;
2330 err:
2331	pr_warn("Could not create tracefs 'synthetic_events' entry\n");
2332
2333	return err;
2334}
2335
2336fs_initcall(trace_events_synth_init);
2337