1#include <linux/types.h>
2#include "event.h"
3#include "debug.h"
4#include "machine.h"
5#include "sort.h"
6#include "string.h"
7#include "strlist.h"
8#include "thread.h"
9#include "thread_map.h"
10
11static const char *perf_event__names[] = {
12	[0]					= "TOTAL",
13	[PERF_RECORD_MMAP]			= "MMAP",
14	[PERF_RECORD_MMAP2]			= "MMAP2",
15	[PERF_RECORD_LOST]			= "LOST",
16	[PERF_RECORD_COMM]			= "COMM",
17	[PERF_RECORD_EXIT]			= "EXIT",
18	[PERF_RECORD_THROTTLE]			= "THROTTLE",
19	[PERF_RECORD_UNTHROTTLE]		= "UNTHROTTLE",
20	[PERF_RECORD_FORK]			= "FORK",
21	[PERF_RECORD_READ]			= "READ",
22	[PERF_RECORD_SAMPLE]			= "SAMPLE",
23	[PERF_RECORD_HEADER_ATTR]		= "ATTR",
24	[PERF_RECORD_HEADER_EVENT_TYPE]		= "EVENT_TYPE",
25	[PERF_RECORD_HEADER_TRACING_DATA]	= "TRACING_DATA",
26	[PERF_RECORD_HEADER_BUILD_ID]		= "BUILD_ID",
27	[PERF_RECORD_FINISHED_ROUND]		= "FINISHED_ROUND",
28};
29
30const char *perf_event__name(unsigned int id)
31{
32	if (id >= ARRAY_SIZE(perf_event__names))
33		return "INVALID";
34	if (!perf_event__names[id])
35		return "UNKNOWN";
36	return perf_event__names[id];
37}
38
39static struct perf_sample synth_sample = {
40	.pid	   = -1,
41	.tid	   = -1,
42	.time	   = -1,
43	.stream_id = -1,
44	.cpu	   = -1,
45	.period	   = 1,
46};
47
48static pid_t perf_event__get_comm_tgid(pid_t pid, char *comm, size_t len)
49{
50	char filename[PATH_MAX];
51	char bf[BUFSIZ];
52	FILE *fp;
53	size_t size = 0;
54	pid_t tgid = -1;
55
56	snprintf(filename, sizeof(filename), "/proc/%d/status", pid);
57
58	fp = fopen(filename, "r");
59	if (fp == NULL) {
60		pr_debug("couldn't open %s\n", filename);
61		return 0;
62	}
63
64	while (!comm[0] || (tgid < 0)) {
65		if (fgets(bf, sizeof(bf), fp) == NULL) {
66			pr_warning("couldn't get COMM and pgid, malformed %s\n",
67				   filename);
68			break;
69		}
70
71		if (memcmp(bf, "Name:", 5) == 0) {
72			char *name = bf + 5;
73			while (*name && isspace(*name))
74				++name;
75			size = strlen(name) - 1;
76			if (size >= len)
77				size = len - 1;
78			memcpy(comm, name, size);
79			comm[size] = '\0';
80
81		} else if (memcmp(bf, "Tgid:", 5) == 0) {
82			char *tgids = bf + 5;
83			while (*tgids && isspace(*tgids))
84				++tgids;
85			tgid = atoi(tgids);
86		}
87	}
88
89	fclose(fp);
90
91	return tgid;
92}
93
94static pid_t perf_event__synthesize_comm(struct perf_tool *tool,
95					 union perf_event *event, pid_t pid,
96					 int full,
97					 perf_event__handler_t process,
98					 struct machine *machine)
99{
100	char filename[PATH_MAX];
101	size_t size;
102	DIR *tasks;
103	struct dirent dirent, *next;
104	pid_t tgid;
105
106	memset(&event->comm, 0, sizeof(event->comm));
107
108	tgid = perf_event__get_comm_tgid(pid, event->comm.comm,
109					 sizeof(event->comm.comm));
110	if (tgid < 0)
111		goto out;
112
113	event->comm.pid = tgid;
114	event->comm.header.type = PERF_RECORD_COMM;
115
116	size = strlen(event->comm.comm) + 1;
117	size = PERF_ALIGN(size, sizeof(u64));
118	memset(event->comm.comm + size, 0, machine->id_hdr_size);
119	event->comm.header.size = (sizeof(event->comm) -
120				(sizeof(event->comm.comm) - size) +
121				machine->id_hdr_size);
122	if (!full) {
123		event->comm.tid = pid;
124
125		if (process(tool, event, &synth_sample, machine) != 0)
126			return -1;
127
128		goto out;
129	}
130
131	snprintf(filename, sizeof(filename), "/proc/%d/task", pid);
132
133	tasks = opendir(filename);
134	if (tasks == NULL) {
135		pr_debug("couldn't open %s\n", filename);
136		return 0;
137	}
138
139	while (!readdir_r(tasks, &dirent, &next) && next) {
140		char *end;
141		pid = strtol(dirent.d_name, &end, 10);
142		if (*end)
143			continue;
144
145		/* already have tgid; jut want to update the comm */
146		(void) perf_event__get_comm_tgid(pid, event->comm.comm,
147					 sizeof(event->comm.comm));
148
149		size = strlen(event->comm.comm) + 1;
150		size = PERF_ALIGN(size, sizeof(u64));
151		memset(event->comm.comm + size, 0, machine->id_hdr_size);
152		event->comm.header.size = (sizeof(event->comm) -
153					  (sizeof(event->comm.comm) - size) +
154					  machine->id_hdr_size);
155
156		event->comm.tid = pid;
157
158		if (process(tool, event, &synth_sample, machine) != 0) {
159			tgid = -1;
160			break;
161		}
162	}
163
164	closedir(tasks);
165out:
166	return tgid;
167}
168
169static int perf_event__synthesize_mmap_events(struct perf_tool *tool,
170					      union perf_event *event,
171					      pid_t pid, pid_t tgid,
172					      perf_event__handler_t process,
173					      struct machine *machine)
174{
175	char filename[PATH_MAX];
176	FILE *fp;
177	int rc = 0;
178
179	snprintf(filename, sizeof(filename), "/proc/%d/maps", pid);
180
181	fp = fopen(filename, "r");
182	if (fp == NULL) {
183		/*
184		 * We raced with a task exiting - just return:
185		 */
186		pr_debug("couldn't open %s\n", filename);
187		return -1;
188	}
189
190	event->header.type = PERF_RECORD_MMAP;
191	/*
192	 * Just like the kernel, see __perf_event_mmap in kernel/perf_event.c
193	 */
194	event->header.misc = PERF_RECORD_MISC_USER;
195
196	while (1) {
197		char bf[BUFSIZ];
198		char prot[5];
199		char execname[PATH_MAX];
200		char anonstr[] = "//anon";
201		size_t size;
202		ssize_t n;
203
204		if (fgets(bf, sizeof(bf), fp) == NULL)
205			break;
206
207		/* ensure null termination since stack will be reused. */
208		strcpy(execname, "");
209
210		/* 00400000-0040c000 r-xp 00000000 fd:01 41038  /bin/cat */
211		n = sscanf(bf, "%"PRIx64"-%"PRIx64" %s %"PRIx64" %*x:%*x %*u %s\n",
212		       &event->mmap.start, &event->mmap.len, prot,
213		       &event->mmap.pgoff,
214		       execname);
215
216		if (n != 5)
217			continue;
218
219		if (prot[2] != 'x')
220			continue;
221
222		if (!strcmp(execname, ""))
223			strcpy(execname, anonstr);
224
225		size = strlen(execname) + 1;
226		memcpy(event->mmap.filename, execname, size);
227		size = PERF_ALIGN(size, sizeof(u64));
228		event->mmap.len -= event->mmap.start;
229		event->mmap.header.size = (sizeof(event->mmap) -
230					(sizeof(event->mmap.filename) - size));
231		memset(event->mmap.filename + size, 0, machine->id_hdr_size);
232		event->mmap.header.size += machine->id_hdr_size;
233		event->mmap.pid = tgid;
234		event->mmap.tid = pid;
235
236		if (process(tool, event, &synth_sample, machine) != 0) {
237			rc = -1;
238			break;
239		}
240	}
241
242	fclose(fp);
243	return rc;
244}
245
246int perf_event__synthesize_modules(struct perf_tool *tool,
247				   perf_event__handler_t process,
248				   struct machine *machine)
249{
250	int rc = 0;
251	struct rb_node *nd;
252	struct map_groups *kmaps = &machine->kmaps;
253	union perf_event *event = zalloc((sizeof(event->mmap) +
254					  machine->id_hdr_size));
255	if (event == NULL) {
256		pr_debug("Not enough memory synthesizing mmap event "
257			 "for kernel modules\n");
258		return -1;
259	}
260
261	event->header.type = PERF_RECORD_MMAP;
262
263	/*
264	 * kernel uses 0 for user space maps, see kernel/perf_event.c
265	 * __perf_event_mmap
266	 */
267	if (machine__is_host(machine))
268		event->header.misc = PERF_RECORD_MISC_KERNEL;
269	else
270		event->header.misc = PERF_RECORD_MISC_GUEST_KERNEL;
271
272	for (nd = rb_first(&kmaps->maps[MAP__FUNCTION]);
273	     nd; nd = rb_next(nd)) {
274		size_t size;
275		struct map *pos = rb_entry(nd, struct map, rb_node);
276
277		if (pos->dso->kernel)
278			continue;
279
280		size = PERF_ALIGN(pos->dso->long_name_len + 1, sizeof(u64));
281		event->mmap.header.type = PERF_RECORD_MMAP;
282		event->mmap.header.size = (sizeof(event->mmap) -
283				        (sizeof(event->mmap.filename) - size));
284		memset(event->mmap.filename + size, 0, machine->id_hdr_size);
285		event->mmap.header.size += machine->id_hdr_size;
286		event->mmap.start = pos->start;
287		event->mmap.len   = pos->end - pos->start;
288		event->mmap.pid   = machine->pid;
289
290		memcpy(event->mmap.filename, pos->dso->long_name,
291		       pos->dso->long_name_len + 1);
292		if (process(tool, event, &synth_sample, machine) != 0) {
293			rc = -1;
294			break;
295		}
296	}
297
298	free(event);
299	return rc;
300}
301
302static int __event__synthesize_thread(union perf_event *comm_event,
303				      union perf_event *mmap_event,
304				      pid_t pid, int full,
305					  perf_event__handler_t process,
306				      struct perf_tool *tool,
307				      struct machine *machine)
308{
309	pid_t tgid = perf_event__synthesize_comm(tool, comm_event, pid, full,
310						 process, machine);
311	if (tgid == -1)
312		return -1;
313	return perf_event__synthesize_mmap_events(tool, mmap_event, pid, tgid,
314						  process, machine);
315}
316
317int perf_event__synthesize_thread_map(struct perf_tool *tool,
318				      struct thread_map *threads,
319				      perf_event__handler_t process,
320				      struct machine *machine)
321{
322	union perf_event *comm_event, *mmap_event;
323	int err = -1, thread, j;
324
325	comm_event = malloc(sizeof(comm_event->comm) + machine->id_hdr_size);
326	if (comm_event == NULL)
327		goto out;
328
329	mmap_event = malloc(sizeof(mmap_event->mmap) + machine->id_hdr_size);
330	if (mmap_event == NULL)
331		goto out_free_comm;
332
333	err = 0;
334	for (thread = 0; thread < threads->nr; ++thread) {
335		if (__event__synthesize_thread(comm_event, mmap_event,
336					       threads->map[thread], 0,
337					       process, tool, machine)) {
338			err = -1;
339			break;
340		}
341
342		/*
343		 * comm.pid is set to thread group id by
344		 * perf_event__synthesize_comm
345		 */
346		if ((int) comm_event->comm.pid != threads->map[thread]) {
347			bool need_leader = true;
348
349			/* is thread group leader in thread_map? */
350			for (j = 0; j < threads->nr; ++j) {
351				if ((int) comm_event->comm.pid == threads->map[j]) {
352					need_leader = false;
353					break;
354				}
355			}
356
357			/* if not, generate events for it */
358			if (need_leader &&
359			    __event__synthesize_thread(comm_event,
360						      mmap_event,
361						      comm_event->comm.pid, 0,
362						      process, tool, machine)) {
363				err = -1;
364				break;
365			}
366		}
367	}
368	free(mmap_event);
369out_free_comm:
370	free(comm_event);
371out:
372	return err;
373}
374
375int perf_event__synthesize_threads(struct perf_tool *tool,
376				   perf_event__handler_t process,
377				   struct machine *machine)
378{
379	DIR *proc;
380	struct dirent dirent, *next;
381	union perf_event *comm_event, *mmap_event;
382	int err = -1;
383
384	comm_event = malloc(sizeof(comm_event->comm) + machine->id_hdr_size);
385	if (comm_event == NULL)
386		goto out;
387
388	mmap_event = malloc(sizeof(mmap_event->mmap) + machine->id_hdr_size);
389	if (mmap_event == NULL)
390		goto out_free_comm;
391
392	proc = opendir("/proc");
393	if (proc == NULL)
394		goto out_free_mmap;
395
396	while (!readdir_r(proc, &dirent, &next) && next) {
397		char *end;
398		pid_t pid = strtol(dirent.d_name, &end, 10);
399
400		if (*end) /* only interested in proper numerical dirents */
401			continue;
402		/*
403 		 * We may race with exiting thread, so don't stop just because
404 		 * one thread couldn't be synthesized.
405 		 */
406		__event__synthesize_thread(comm_event, mmap_event, pid, 1,
407					   process, tool, machine);
408	}
409
410	err = 0;
411	closedir(proc);
412out_free_mmap:
413	free(mmap_event);
414out_free_comm:
415	free(comm_event);
416out:
417	return err;
418}
419
420struct process_symbol_args {
421	const char *name;
422	u64	   start;
423};
424
425static int find_symbol_cb(void *arg, const char *name, char type,
426			  u64 start)
427{
428	struct process_symbol_args *args = arg;
429
430	/*
431	 * Must be a function or at least an alias, as in PARISC64, where "_text" is
432	 * an 'A' to the same address as "_stext".
433	 */
434	if (!(symbol_type__is_a(type, MAP__FUNCTION) ||
435	      type == 'A') || strcmp(name, args->name))
436		return 0;
437
438	args->start = start;
439	return 1;
440}
441
442int perf_event__synthesize_kernel_mmap(struct perf_tool *tool,
443				       perf_event__handler_t process,
444				       struct machine *machine,
445				       const char *symbol_name)
446{
447	size_t size;
448	const char *filename, *mmap_name;
449	char path[PATH_MAX];
450	char name_buff[PATH_MAX];
451	struct map *map;
452	int err;
453	/*
454	 * We should get this from /sys/kernel/sections/.text, but till that is
455	 * available use this, and after it is use this as a fallback for older
456	 * kernels.
457	 */
458	struct process_symbol_args args = { .name = symbol_name, };
459	union perf_event *event = zalloc((sizeof(event->mmap) +
460					  machine->id_hdr_size));
461	if (event == NULL) {
462		pr_debug("Not enough memory synthesizing mmap event "
463			 "for kernel modules\n");
464		return -1;
465	}
466
467	mmap_name = machine__mmap_name(machine, name_buff, sizeof(name_buff));
468	if (machine__is_host(machine)) {
469		/*
470		 * kernel uses PERF_RECORD_MISC_USER for user space maps,
471		 * see kernel/perf_event.c __perf_event_mmap
472		 */
473		event->header.misc = PERF_RECORD_MISC_KERNEL;
474		filename = "/proc/kallsyms";
475	} else {
476		event->header.misc = PERF_RECORD_MISC_GUEST_KERNEL;
477		if (machine__is_default_guest(machine))
478			filename = (char *) symbol_conf.default_guest_kallsyms;
479		else {
480			sprintf(path, "%s/proc/kallsyms", machine->root_dir);
481			filename = path;
482		}
483	}
484
485	if (kallsyms__parse(filename, &args, find_symbol_cb) <= 0) {
486		free(event);
487		return -ENOENT;
488	}
489
490	map = machine->vmlinux_maps[MAP__FUNCTION];
491	size = snprintf(event->mmap.filename, sizeof(event->mmap.filename),
492			"%s%s", mmap_name, symbol_name) + 1;
493	size = PERF_ALIGN(size, sizeof(u64));
494	event->mmap.header.type = PERF_RECORD_MMAP;
495	event->mmap.header.size = (sizeof(event->mmap) -
496			(sizeof(event->mmap.filename) - size) + machine->id_hdr_size);
497	event->mmap.pgoff = args.start;
498	event->mmap.start = map->start;
499	event->mmap.len   = map->end - event->mmap.start;
500	event->mmap.pid   = machine->pid;
501
502	err = process(tool, event, &synth_sample, machine);
503	free(event);
504
505	return err;
506}
507
508size_t perf_event__fprintf_comm(union perf_event *event, FILE *fp)
509{
510	return fprintf(fp, ": %s:%d\n", event->comm.comm, event->comm.tid);
511}
512
513int perf_event__process_comm(struct perf_tool *tool __maybe_unused,
514			     union perf_event *event,
515			     struct perf_sample *sample __maybe_unused,
516			     struct machine *machine)
517{
518	return machine__process_comm_event(machine, event);
519}
520
521int perf_event__process_lost(struct perf_tool *tool __maybe_unused,
522			     union perf_event *event,
523			     struct perf_sample *sample __maybe_unused,
524			     struct machine *machine)
525{
526	return machine__process_lost_event(machine, event);
527}
528
529size_t perf_event__fprintf_mmap(union perf_event *event, FILE *fp)
530{
531	return fprintf(fp, " %d/%d: [%#" PRIx64 "(%#" PRIx64 ") @ %#" PRIx64 "]: %s\n",
532		       event->mmap.pid, event->mmap.tid, event->mmap.start,
533		       event->mmap.len, event->mmap.pgoff, event->mmap.filename);
534}
535
536size_t perf_event__fprintf_mmap2(union perf_event *event, FILE *fp)
537{
538	return fprintf(fp, " %d/%d: [%#" PRIx64 "(%#" PRIx64 ") @ %#" PRIx64
539			   " %02x:%02x %"PRIu64" %"PRIu64"]: %s\n",
540		       event->mmap2.pid, event->mmap2.tid, event->mmap2.start,
541		       event->mmap2.len, event->mmap2.pgoff, event->mmap2.maj,
542		       event->mmap2.min, event->mmap2.ino,
543		       event->mmap2.ino_generation,
544		       event->mmap2.filename);
545}
546
547int perf_event__process_mmap(struct perf_tool *tool __maybe_unused,
548			     union perf_event *event,
549			     struct perf_sample *sample __maybe_unused,
550			     struct machine *machine)
551{
552	return machine__process_mmap_event(machine, event);
553}
554
555int perf_event__process_mmap2(struct perf_tool *tool __maybe_unused,
556			     union perf_event *event,
557			     struct perf_sample *sample __maybe_unused,
558			     struct machine *machine)
559{
560	return machine__process_mmap2_event(machine, event);
561}
562
563size_t perf_event__fprintf_task(union perf_event *event, FILE *fp)
564{
565	return fprintf(fp, "(%d:%d):(%d:%d)\n",
566		       event->fork.pid, event->fork.tid,
567		       event->fork.ppid, event->fork.ptid);
568}
569
570int perf_event__process_fork(struct perf_tool *tool __maybe_unused,
571			     union perf_event *event,
572			     struct perf_sample *sample __maybe_unused,
573			     struct machine *machine)
574{
575	return machine__process_fork_event(machine, event);
576}
577
578int perf_event__process_exit(struct perf_tool *tool __maybe_unused,
579			     union perf_event *event,
580			     struct perf_sample *sample __maybe_unused,
581			     struct machine *machine)
582{
583	return machine__process_exit_event(machine, event);
584}
585
586size_t perf_event__fprintf(union perf_event *event, FILE *fp)
587{
588	size_t ret = fprintf(fp, "PERF_RECORD_%s",
589			     perf_event__name(event->header.type));
590
591	switch (event->header.type) {
592	case PERF_RECORD_COMM:
593		ret += perf_event__fprintf_comm(event, fp);
594		break;
595	case PERF_RECORD_FORK:
596	case PERF_RECORD_EXIT:
597		ret += perf_event__fprintf_task(event, fp);
598		break;
599	case PERF_RECORD_MMAP:
600		ret += perf_event__fprintf_mmap(event, fp);
601		break;
602	case PERF_RECORD_MMAP2:
603		ret += perf_event__fprintf_mmap2(event, fp);
604		break;
605	default:
606		ret += fprintf(fp, "\n");
607	}
608
609	return ret;
610}
611
612int perf_event__process(struct perf_tool *tool __maybe_unused,
613			union perf_event *event,
614			struct perf_sample *sample __maybe_unused,
615			struct machine *machine)
616{
617	return machine__process_event(machine, event);
618}
619
620void thread__find_addr_map(struct thread *self,
621			   struct machine *machine, u8 cpumode,
622			   enum map_type type, u64 addr,
623			   struct addr_location *al)
624{
625	struct map_groups *mg = &self->mg;
626	bool load_map = false;
627
628	al->thread = self;
629	al->addr = addr;
630	al->cpumode = cpumode;
631	al->filtered = false;
632
633	if (machine == NULL) {
634		al->map = NULL;
635		return;
636	}
637
638	if (cpumode == PERF_RECORD_MISC_KERNEL && perf_host) {
639		al->level = 'k';
640		mg = &machine->kmaps;
641		load_map = true;
642	} else if (cpumode == PERF_RECORD_MISC_USER && perf_host) {
643		al->level = '.';
644	} else if (cpumode == PERF_RECORD_MISC_GUEST_KERNEL && perf_guest) {
645		al->level = 'g';
646		mg = &machine->kmaps;
647		load_map = true;
648	} else {
649		/*
650		 * 'u' means guest os user space.
651		 * TODO: We don't support guest user space. Might support late.
652		 */
653		if (cpumode == PERF_RECORD_MISC_GUEST_USER && perf_guest)
654			al->level = 'u';
655		else
656			al->level = 'H';
657		al->map = NULL;
658
659		if ((cpumode == PERF_RECORD_MISC_GUEST_USER ||
660			cpumode == PERF_RECORD_MISC_GUEST_KERNEL) &&
661			!perf_guest)
662			al->filtered = true;
663		if ((cpumode == PERF_RECORD_MISC_USER ||
664			cpumode == PERF_RECORD_MISC_KERNEL) &&
665			!perf_host)
666			al->filtered = true;
667
668		return;
669	}
670try_again:
671	al->map = map_groups__find(mg, type, al->addr);
672	if (al->map == NULL) {
673		/*
674		 * If this is outside of all known maps, and is a negative
675		 * address, try to look it up in the kernel dso, as it might be
676		 * a vsyscall or vdso (which executes in user-mode).
677		 *
678		 * XXX This is nasty, we should have a symbol list in the
679		 * "[vdso]" dso, but for now lets use the old trick of looking
680		 * in the whole kernel symbol list.
681		 */
682		if ((long long)al->addr < 0 &&
683		    cpumode == PERF_RECORD_MISC_USER &&
684		    machine && mg != &machine->kmaps) {
685			mg = &machine->kmaps;
686			goto try_again;
687		}
688	} else {
689		/*
690		 * Kernel maps might be changed when loading symbols so loading
691		 * must be done prior to using kernel maps.
692		 */
693		if (load_map)
694			map__load(al->map, machine->symbol_filter);
695		al->addr = al->map->map_ip(al->map, al->addr);
696	}
697}
698
699void thread__find_addr_location(struct thread *thread, struct machine *machine,
700				u8 cpumode, enum map_type type, u64 addr,
701				struct addr_location *al)
702{
703	thread__find_addr_map(thread, machine, cpumode, type, addr, al);
704	if (al->map != NULL)
705		al->sym = map__find_symbol(al->map, al->addr,
706					   machine->symbol_filter);
707	else
708		al->sym = NULL;
709}
710
711int perf_event__preprocess_sample(const union perf_event *event,
712				  struct machine *machine,
713				  struct addr_location *al,
714				  struct perf_sample *sample)
715{
716	u8 cpumode = event->header.misc & PERF_RECORD_MISC_CPUMODE_MASK;
717	struct thread *thread = machine__findnew_thread(machine, sample->pid,
718							sample->pid);
719
720	if (thread == NULL)
721		return -1;
722
723	if (symbol_conf.comm_list &&
724	    !strlist__has_entry(symbol_conf.comm_list, thread->comm))
725		goto out_filtered;
726
727	dump_printf(" ... thread: %s:%d\n", thread->comm, thread->tid);
728	/*
729	 * Have we already created the kernel maps for this machine?
730	 *
731	 * This should have happened earlier, when we processed the kernel MMAP
732	 * events, but for older perf.data files there was no such thing, so do
733	 * it now.
734	 */
735	if (cpumode == PERF_RECORD_MISC_KERNEL &&
736	    machine->vmlinux_maps[MAP__FUNCTION] == NULL)
737		machine__create_kernel_maps(machine);
738
739	thread__find_addr_map(thread, machine, cpumode, MAP__FUNCTION,
740			      sample->ip, al);
741	dump_printf(" ...... dso: %s\n",
742		    al->map ? al->map->dso->long_name :
743			al->level == 'H' ? "[hypervisor]" : "<not found>");
744	al->sym = NULL;
745	al->cpu = sample->cpu;
746
747	if (al->map) {
748		struct dso *dso = al->map->dso;
749
750		if (symbol_conf.dso_list &&
751		    (!dso || !(strlist__has_entry(symbol_conf.dso_list,
752						  dso->short_name) ||
753			       (dso->short_name != dso->long_name &&
754				strlist__has_entry(symbol_conf.dso_list,
755						   dso->long_name)))))
756			goto out_filtered;
757
758		al->sym = map__find_symbol(al->map, al->addr,
759					   machine->symbol_filter);
760	}
761
762	if (symbol_conf.sym_list &&
763		(!al->sym || !strlist__has_entry(symbol_conf.sym_list,
764						al->sym->name)))
765		goto out_filtered;
766
767	return 0;
768
769out_filtered:
770	al->filtered = true;
771	return 0;
772}
773