mirror of
				git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
				synced 2025-09-04 20:19:47 +08:00 
			
		
		
		
	 907b223651
			
		
	
	
		907b223651
		
	
	
	
	
		
			
			Currently bpftool contains a mix of GPL-only and GPL or BSD2 licensed files. Make sure all files are dual licensed under GPLv2 and BSD-2-Clause. Signed-off-by: Jakub Kicinski <jakub.kicinski@netronome.com> Acked-by: Roman Gushchin <guro@fb.com> Acked-by: YueHaibing <yuehaibing@huawei.com> Acked-by: Yonghong Song <yhs@fb.com> Acked-by: Stanislav Fomichev <sdf@google.com> Acked-by: Sean Young <sean@mess.org> Acked-by: Jiri Benc <jbenc@redhat.com> Acked-by: David Calavera <david.calavera@gmail.com> Acked-by: Andrey Ignatov <rdna@fb.com> Acked-by: Joe Stringer <joe@wand.net.nz> Acked-by: David Ahern <dsahern@gmail.com> Acked-by: Alexei Starovoitov <alexei.starovoitov@gmail.com> Acked-by: Petar Penkov <ppenkov@stanford.edu> Acked-by: Sandipan Das <sandipan@linux.ibm.com> Acked-by: Prashant Bhole <bhole_prashant_q7@lab.ntt.co.jp> Acked-by: Stephen Hemminger <stephen@networkplumber.org> Acked-by: John Fastabend <john.fastabend@gmail.com> Acked-by: Taeung Song <treeze.taeung@gmail.com> Acked-by: Jiri Olsa <jolsa@kernel.org> Acked-by: Daniel Borkmann <daniel@iogearbox.net> CC: okash.khawaja@gmail.com Signed-off-by: Daniel Borkmann <daniel@iogearbox.net>
		
			
				
	
	
		
			312 lines
		
	
	
		
			6.6 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			312 lines
		
	
	
		
			6.6 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: (GPL-2.0-only OR BSD-2-Clause)
 | |
| /* Copyright (C) 2018 Netronome Systems, Inc. */
 | |
| /* This program is free software; you can redistribute it and/or
 | |
|  * modify it under the terms of version 2 of the GNU General Public
 | |
|  * License as published by the Free Software Foundation.
 | |
|  */
 | |
| #include <errno.h>
 | |
| #include <fcntl.h>
 | |
| #include <libbpf.h>
 | |
| #include <poll.h>
 | |
| #include <signal.h>
 | |
| #include <stdbool.h>
 | |
| #include <stdio.h>
 | |
| #include <stdlib.h>
 | |
| #include <string.h>
 | |
| #include <time.h>
 | |
| #include <unistd.h>
 | |
| #include <linux/bpf.h>
 | |
| #include <linux/perf_event.h>
 | |
| #include <sys/ioctl.h>
 | |
| #include <sys/mman.h>
 | |
| #include <sys/syscall.h>
 | |
| 
 | |
| #include <bpf.h>
 | |
| #include <perf-sys.h>
 | |
| 
 | |
| #include "main.h"
 | |
| 
 | |
| #define MMAP_PAGE_CNT	16
 | |
| 
 | |
| static bool stop;
 | |
| 
 | |
| struct event_ring_info {
 | |
| 	int fd;
 | |
| 	int key;
 | |
| 	unsigned int cpu;
 | |
| 	void *mem;
 | |
| };
 | |
| 
 | |
| struct perf_event_sample {
 | |
| 	struct perf_event_header header;
 | |
| 	u64 time;
 | |
| 	__u32 size;
 | |
| 	unsigned char data[];
 | |
| };
 | |
| 
 | |
| static void int_exit(int signo)
 | |
| {
 | |
| 	fprintf(stderr, "Stopping...\n");
 | |
| 	stop = true;
 | |
| }
 | |
| 
 | |
| static enum bpf_perf_event_ret
 | |
| print_bpf_output(struct perf_event_header *event, void *private_data)
 | |
| {
 | |
| 	struct perf_event_sample *e = container_of(event, struct perf_event_sample,
 | |
| 						   header);
 | |
| 	struct event_ring_info *ring = private_data;
 | |
| 	struct {
 | |
| 		struct perf_event_header header;
 | |
| 		__u64 id;
 | |
| 		__u64 lost;
 | |
| 	} *lost = (typeof(lost))event;
 | |
| 
 | |
| 	if (json_output) {
 | |
| 		jsonw_start_object(json_wtr);
 | |
| 		jsonw_name(json_wtr, "type");
 | |
| 		jsonw_uint(json_wtr, e->header.type);
 | |
| 		jsonw_name(json_wtr, "cpu");
 | |
| 		jsonw_uint(json_wtr, ring->cpu);
 | |
| 		jsonw_name(json_wtr, "index");
 | |
| 		jsonw_uint(json_wtr, ring->key);
 | |
| 		if (e->header.type == PERF_RECORD_SAMPLE) {
 | |
| 			jsonw_name(json_wtr, "timestamp");
 | |
| 			jsonw_uint(json_wtr, e->time);
 | |
| 			jsonw_name(json_wtr, "data");
 | |
| 			print_data_json(e->data, e->size);
 | |
| 		} else if (e->header.type == PERF_RECORD_LOST) {
 | |
| 			jsonw_name(json_wtr, "lost");
 | |
| 			jsonw_start_object(json_wtr);
 | |
| 			jsonw_name(json_wtr, "id");
 | |
| 			jsonw_uint(json_wtr, lost->id);
 | |
| 			jsonw_name(json_wtr, "count");
 | |
| 			jsonw_uint(json_wtr, lost->lost);
 | |
| 			jsonw_end_object(json_wtr);
 | |
| 		}
 | |
| 		jsonw_end_object(json_wtr);
 | |
| 	} else {
 | |
| 		if (e->header.type == PERF_RECORD_SAMPLE) {
 | |
| 			printf("== @%lld.%09lld CPU: %d index: %d =====\n",
 | |
| 			       e->time / 1000000000ULL, e->time % 1000000000ULL,
 | |
| 			       ring->cpu, ring->key);
 | |
| 			fprint_hex(stdout, e->data, e->size, " ");
 | |
| 			printf("\n");
 | |
| 		} else if (e->header.type == PERF_RECORD_LOST) {
 | |
| 			printf("lost %lld events\n", lost->lost);
 | |
| 		} else {
 | |
| 			printf("unknown event type=%d size=%d\n",
 | |
| 			       e->header.type, e->header.size);
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return LIBBPF_PERF_EVENT_CONT;
 | |
| }
 | |
| 
 | |
| static void
 | |
| perf_event_read(struct event_ring_info *ring, void **buf, size_t *buf_len)
 | |
| {
 | |
| 	enum bpf_perf_event_ret ret;
 | |
| 
 | |
| 	ret = bpf_perf_event_read_simple(ring->mem,
 | |
| 					 MMAP_PAGE_CNT * get_page_size(),
 | |
| 					 get_page_size(), buf, buf_len,
 | |
| 					 print_bpf_output, ring);
 | |
| 	if (ret != LIBBPF_PERF_EVENT_CONT) {
 | |
| 		fprintf(stderr, "perf read loop failed with %d\n", ret);
 | |
| 		stop = true;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static int perf_mmap_size(void)
 | |
| {
 | |
| 	return get_page_size() * (MMAP_PAGE_CNT + 1);
 | |
| }
 | |
| 
 | |
| static void *perf_event_mmap(int fd)
 | |
| {
 | |
| 	int mmap_size = perf_mmap_size();
 | |
| 	void *base;
 | |
| 
 | |
| 	base = mmap(NULL, mmap_size, PROT_READ | PROT_WRITE, MAP_SHARED, fd, 0);
 | |
| 	if (base == MAP_FAILED) {
 | |
| 		p_err("event mmap failed: %s\n", strerror(errno));
 | |
| 		return NULL;
 | |
| 	}
 | |
| 
 | |
| 	return base;
 | |
| }
 | |
| 
 | |
| static void perf_event_unmap(void *mem)
 | |
| {
 | |
| 	if (munmap(mem, perf_mmap_size()))
 | |
| 		fprintf(stderr, "Can't unmap ring memory!\n");
 | |
| }
 | |
| 
 | |
| static int bpf_perf_event_open(int map_fd, int key, int cpu)
 | |
| {
 | |
| 	struct perf_event_attr attr = {
 | |
| 		.sample_type = PERF_SAMPLE_RAW | PERF_SAMPLE_TIME,
 | |
| 		.type = PERF_TYPE_SOFTWARE,
 | |
| 		.config = PERF_COUNT_SW_BPF_OUTPUT,
 | |
| 	};
 | |
| 	int pmu_fd;
 | |
| 
 | |
| 	pmu_fd = sys_perf_event_open(&attr, -1, cpu, -1, 0);
 | |
| 	if (pmu_fd < 0) {
 | |
| 		p_err("failed to open perf event %d for CPU %d", key, cpu);
 | |
| 		return -1;
 | |
| 	}
 | |
| 
 | |
| 	if (bpf_map_update_elem(map_fd, &key, &pmu_fd, BPF_ANY)) {
 | |
| 		p_err("failed to update map for event %d for CPU %d", key, cpu);
 | |
| 		goto err_close;
 | |
| 	}
 | |
| 	if (ioctl(pmu_fd, PERF_EVENT_IOC_ENABLE, 0)) {
 | |
| 		p_err("failed to enable event %d for CPU %d", key, cpu);
 | |
| 		goto err_close;
 | |
| 	}
 | |
| 
 | |
| 	return pmu_fd;
 | |
| 
 | |
| err_close:
 | |
| 	close(pmu_fd);
 | |
| 	return -1;
 | |
| }
 | |
| 
 | |
| int do_event_pipe(int argc, char **argv)
 | |
| {
 | |
| 	int i, nfds, map_fd, index = -1, cpu = -1;
 | |
| 	struct bpf_map_info map_info = {};
 | |
| 	struct event_ring_info *rings;
 | |
| 	size_t tmp_buf_sz = 0;
 | |
| 	void *tmp_buf = NULL;
 | |
| 	struct pollfd *pfds;
 | |
| 	__u32 map_info_len;
 | |
| 	bool do_all = true;
 | |
| 
 | |
| 	map_info_len = sizeof(map_info);
 | |
| 	map_fd = map_parse_fd_and_info(&argc, &argv, &map_info, &map_info_len);
 | |
| 	if (map_fd < 0)
 | |
| 		return -1;
 | |
| 
 | |
| 	if (map_info.type != BPF_MAP_TYPE_PERF_EVENT_ARRAY) {
 | |
| 		p_err("map is not a perf event array");
 | |
| 		goto err_close_map;
 | |
| 	}
 | |
| 
 | |
| 	while (argc) {
 | |
| 		if (argc < 2) {
 | |
| 			BAD_ARG();
 | |
| 			goto err_close_map;
 | |
| 		}
 | |
| 
 | |
| 		if (is_prefix(*argv, "cpu")) {
 | |
| 			char *endptr;
 | |
| 
 | |
| 			NEXT_ARG();
 | |
| 			cpu = strtoul(*argv, &endptr, 0);
 | |
| 			if (*endptr) {
 | |
| 				p_err("can't parse %s as CPU ID", **argv);
 | |
| 				goto err_close_map;
 | |
| 			}
 | |
| 
 | |
| 			NEXT_ARG();
 | |
| 		} else if (is_prefix(*argv, "index")) {
 | |
| 			char *endptr;
 | |
| 
 | |
| 			NEXT_ARG();
 | |
| 			index = strtoul(*argv, &endptr, 0);
 | |
| 			if (*endptr) {
 | |
| 				p_err("can't parse %s as index", **argv);
 | |
| 				goto err_close_map;
 | |
| 			}
 | |
| 
 | |
| 			NEXT_ARG();
 | |
| 		} else {
 | |
| 			BAD_ARG();
 | |
| 			goto err_close_map;
 | |
| 		}
 | |
| 
 | |
| 		do_all = false;
 | |
| 	}
 | |
| 
 | |
| 	if (!do_all) {
 | |
| 		if (index == -1 || cpu == -1) {
 | |
| 			p_err("cpu and index must be specified together");
 | |
| 			goto err_close_map;
 | |
| 		}
 | |
| 
 | |
| 		nfds = 1;
 | |
| 	} else {
 | |
| 		nfds = min(get_possible_cpus(), map_info.max_entries);
 | |
| 		cpu = 0;
 | |
| 		index = 0;
 | |
| 	}
 | |
| 
 | |
| 	rings = calloc(nfds, sizeof(rings[0]));
 | |
| 	if (!rings)
 | |
| 		goto err_close_map;
 | |
| 
 | |
| 	pfds = calloc(nfds, sizeof(pfds[0]));
 | |
| 	if (!pfds)
 | |
| 		goto err_free_rings;
 | |
| 
 | |
| 	for (i = 0; i < nfds; i++) {
 | |
| 		rings[i].cpu = cpu + i;
 | |
| 		rings[i].key = index + i;
 | |
| 
 | |
| 		rings[i].fd = bpf_perf_event_open(map_fd, rings[i].key,
 | |
| 						  rings[i].cpu);
 | |
| 		if (rings[i].fd < 0)
 | |
| 			goto err_close_fds_prev;
 | |
| 
 | |
| 		rings[i].mem = perf_event_mmap(rings[i].fd);
 | |
| 		if (!rings[i].mem)
 | |
| 			goto err_close_fds_current;
 | |
| 
 | |
| 		pfds[i].fd = rings[i].fd;
 | |
| 		pfds[i].events = POLLIN;
 | |
| 	}
 | |
| 
 | |
| 	signal(SIGINT, int_exit);
 | |
| 	signal(SIGHUP, int_exit);
 | |
| 	signal(SIGTERM, int_exit);
 | |
| 
 | |
| 	if (json_output)
 | |
| 		jsonw_start_array(json_wtr);
 | |
| 
 | |
| 	while (!stop) {
 | |
| 		poll(pfds, nfds, 200);
 | |
| 		for (i = 0; i < nfds; i++)
 | |
| 			perf_event_read(&rings[i], &tmp_buf, &tmp_buf_sz);
 | |
| 	}
 | |
| 	free(tmp_buf);
 | |
| 
 | |
| 	if (json_output)
 | |
| 		jsonw_end_array(json_wtr);
 | |
| 
 | |
| 	for (i = 0; i < nfds; i++) {
 | |
| 		perf_event_unmap(rings[i].mem);
 | |
| 		close(rings[i].fd);
 | |
| 	}
 | |
| 	free(pfds);
 | |
| 	free(rings);
 | |
| 	close(map_fd);
 | |
| 
 | |
| 	return 0;
 | |
| 
 | |
| err_close_fds_prev:
 | |
| 	while (i--) {
 | |
| 		perf_event_unmap(rings[i].mem);
 | |
| err_close_fds_current:
 | |
| 		close(rings[i].fd);
 | |
| 	}
 | |
| 	free(pfds);
 | |
| err_free_rings:
 | |
| 	free(rings);
 | |
| err_close_map:
 | |
| 	close(map_fd);
 | |
| 	return -1;
 | |
| }
 |