mirror of
				git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
				synced 2025-09-04 20:19:47 +08:00 
			
		
		
		
	 b25acdafd3
			
		
	
	
		b25acdafd3
		
	
	
	
	
		
			
			We have switched to memcg-based memory accouting and thus the rlimit is not needed any more. LIBBPF_STRICT_AUTO_RLIMIT_MEMLOCK was introduced in libbpf for backward compatibility, so we can use it instead now. This patch also removes the useless header sys/resource.h from many files in samples/bpf. Signed-off-by: Yafang Shao <laoar.shao@gmail.com> Signed-off-by: Andrii Nakryiko <andrii@kernel.org> Link: https://lore.kernel.org/bpf/20220409125958.92629-2-laoar.shao@gmail.com
		
			
				
	
	
		
			540 lines
		
	
	
		
			12 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			540 lines
		
	
	
		
			12 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0-only
 | |
| /*
 | |
|  * Copyright (c) 2016 Facebook
 | |
|  */
 | |
| #define _GNU_SOURCE
 | |
| #include <linux/types.h>
 | |
| #include <stdio.h>
 | |
| #include <unistd.h>
 | |
| #include <linux/bpf.h>
 | |
| #include <errno.h>
 | |
| #include <string.h>
 | |
| #include <assert.h>
 | |
| #include <sched.h>
 | |
| #include <sys/wait.h>
 | |
| #include <sys/stat.h>
 | |
| #include <fcntl.h>
 | |
| #include <stdlib.h>
 | |
| #include <time.h>
 | |
| 
 | |
| #include <bpf/bpf.h>
 | |
| #include "bpf_util.h"
 | |
| 
 | |
| #define min(a, b) ((a) < (b) ? (a) : (b))
 | |
| #ifndef offsetof
 | |
| # define offsetof(TYPE, MEMBER)	((size_t)&((TYPE *)0)->MEMBER)
 | |
| #endif
 | |
| #define container_of(ptr, type, member) ({			\
 | |
| 	const typeof( ((type *)0)->member ) *__mptr = (ptr);	\
 | |
| 	(type *)( (char *)__mptr - offsetof(type,member) );})
 | |
| 
 | |
| static int nr_cpus;
 | |
| static unsigned long long *dist_keys;
 | |
| static unsigned int dist_key_counts;
 | |
| 
 | |
| struct list_head {
 | |
| 	struct list_head *next, *prev;
 | |
| };
 | |
| 
 | |
| static inline void INIT_LIST_HEAD(struct list_head *list)
 | |
| {
 | |
| 	list->next = list;
 | |
| 	list->prev = list;
 | |
| }
 | |
| 
 | |
| static inline int list_empty(const struct list_head *head)
 | |
| {
 | |
| 	return head->next == head;
 | |
| }
 | |
| 
 | |
| static inline void __list_add(struct list_head *new,
 | |
| 			      struct list_head *prev,
 | |
| 			      struct list_head *next)
 | |
| {
 | |
| 	next->prev = new;
 | |
| 	new->next = next;
 | |
| 	new->prev = prev;
 | |
| 	prev->next = new;
 | |
| }
 | |
| 
 | |
| static inline void list_add(struct list_head *new, struct list_head *head)
 | |
| {
 | |
| 	__list_add(new, head, head->next);
 | |
| }
 | |
| 
 | |
| static inline void __list_del(struct list_head *prev, struct list_head *next)
 | |
| {
 | |
| 	next->prev = prev;
 | |
| 	prev->next = next;
 | |
| }
 | |
| 
 | |
| static inline void __list_del_entry(struct list_head *entry)
 | |
| {
 | |
| 	__list_del(entry->prev, entry->next);
 | |
| }
 | |
| 
 | |
| static inline void list_move(struct list_head *list, struct list_head *head)
 | |
| {
 | |
| 	__list_del_entry(list);
 | |
| 	list_add(list, head);
 | |
| }
 | |
| 
 | |
| #define list_entry(ptr, type, member) \
 | |
| 	container_of(ptr, type, member)
 | |
| 
 | |
| #define list_last_entry(ptr, type, member) \
 | |
| 	list_entry((ptr)->prev, type, member)
 | |
| 
 | |
| struct pfect_lru_node {
 | |
| 	struct list_head list;
 | |
| 	unsigned long long key;
 | |
| };
 | |
| 
 | |
| struct pfect_lru {
 | |
| 	struct list_head list;
 | |
| 	struct pfect_lru_node *free_nodes;
 | |
| 	unsigned int cur_size;
 | |
| 	unsigned int lru_size;
 | |
| 	unsigned int nr_unique;
 | |
| 	unsigned int nr_misses;
 | |
| 	unsigned int total;
 | |
| 	int map_fd;
 | |
| };
 | |
| 
 | |
| static void pfect_lru_init(struct pfect_lru *lru, unsigned int lru_size,
 | |
| 			   unsigned int nr_possible_elems)
 | |
| {
 | |
| 	lru->map_fd = bpf_map_create(BPF_MAP_TYPE_HASH, NULL,
 | |
| 				     sizeof(unsigned long long),
 | |
| 				     sizeof(struct pfect_lru_node *),
 | |
| 				     nr_possible_elems, NULL);
 | |
| 	assert(lru->map_fd != -1);
 | |
| 
 | |
| 	lru->free_nodes = malloc(lru_size * sizeof(struct pfect_lru_node));
 | |
| 	assert(lru->free_nodes);
 | |
| 
 | |
| 	INIT_LIST_HEAD(&lru->list);
 | |
| 	lru->cur_size = 0;
 | |
| 	lru->lru_size = lru_size;
 | |
| 	lru->nr_unique = lru->nr_misses = lru->total = 0;
 | |
| }
 | |
| 
 | |
| static void pfect_lru_destroy(struct pfect_lru *lru)
 | |
| {
 | |
| 	close(lru->map_fd);
 | |
| 	free(lru->free_nodes);
 | |
| }
 | |
| 
 | |
| static int pfect_lru_lookup_or_insert(struct pfect_lru *lru,
 | |
| 				      unsigned long long key)
 | |
| {
 | |
| 	struct pfect_lru_node *node = NULL;
 | |
| 	int seen = 0;
 | |
| 
 | |
| 	lru->total++;
 | |
| 	if (!bpf_map_lookup_elem(lru->map_fd, &key, &node)) {
 | |
| 		if (node) {
 | |
| 			list_move(&node->list, &lru->list);
 | |
| 			return 1;
 | |
| 		}
 | |
| 		seen = 1;
 | |
| 	}
 | |
| 
 | |
| 	if (lru->cur_size < lru->lru_size) {
 | |
| 		node =  &lru->free_nodes[lru->cur_size++];
 | |
| 		INIT_LIST_HEAD(&node->list);
 | |
| 	} else {
 | |
| 		struct pfect_lru_node *null_node = NULL;
 | |
| 
 | |
| 		node = list_last_entry(&lru->list,
 | |
| 				       struct pfect_lru_node,
 | |
| 				       list);
 | |
| 		bpf_map_update_elem(lru->map_fd, &node->key, &null_node, BPF_EXIST);
 | |
| 	}
 | |
| 
 | |
| 	node->key = key;
 | |
| 	list_move(&node->list, &lru->list);
 | |
| 
 | |
| 	lru->nr_misses++;
 | |
| 	if (seen) {
 | |
| 		assert(!bpf_map_update_elem(lru->map_fd, &key, &node, BPF_EXIST));
 | |
| 	} else {
 | |
| 		lru->nr_unique++;
 | |
| 		assert(!bpf_map_update_elem(lru->map_fd, &key, &node, BPF_NOEXIST));
 | |
| 	}
 | |
| 
 | |
| 	return seen;
 | |
| }
 | |
| 
 | |
| static unsigned int read_keys(const char *dist_file,
 | |
| 			      unsigned long long **keys)
 | |
| {
 | |
| 	struct stat fst;
 | |
| 	unsigned long long *retkeys;
 | |
| 	unsigned int counts = 0;
 | |
| 	int dist_fd;
 | |
| 	char *b, *l;
 | |
| 	int i;
 | |
| 
 | |
| 	dist_fd = open(dist_file, 0);
 | |
| 	assert(dist_fd != -1);
 | |
| 
 | |
| 	assert(fstat(dist_fd, &fst) == 0);
 | |
| 	b = malloc(fst.st_size);
 | |
| 	assert(b);
 | |
| 
 | |
| 	assert(read(dist_fd, b, fst.st_size) == fst.st_size);
 | |
| 	close(dist_fd);
 | |
| 	for (i = 0; i < fst.st_size; i++) {
 | |
| 		if (b[i] == '\n')
 | |
| 			counts++;
 | |
| 	}
 | |
| 	counts++; /* in case the last line has no \n */
 | |
| 
 | |
| 	retkeys = malloc(counts * sizeof(unsigned long long));
 | |
| 	assert(retkeys);
 | |
| 
 | |
| 	counts = 0;
 | |
| 	for (l = strtok(b, "\n"); l; l = strtok(NULL, "\n"))
 | |
| 		retkeys[counts++] = strtoull(l, NULL, 10);
 | |
| 	free(b);
 | |
| 
 | |
| 	*keys = retkeys;
 | |
| 
 | |
| 	return counts;
 | |
| }
 | |
| 
 | |
| static int create_map(int map_type, int map_flags, unsigned int size)
 | |
| {
 | |
| 	LIBBPF_OPTS(bpf_map_create_opts, opts,
 | |
| 		.map_flags = map_flags,
 | |
| 	);
 | |
| 	int map_fd;
 | |
| 
 | |
| 	map_fd = bpf_map_create(map_type, NULL, sizeof(unsigned long long),
 | |
| 				sizeof(unsigned long long), size, &opts);
 | |
| 
 | |
| 	if (map_fd == -1)
 | |
| 		perror("bpf_create_map");
 | |
| 
 | |
| 	return map_fd;
 | |
| }
 | |
| 
 | |
| static int sched_next_online(int pid, int next_to_try)
 | |
| {
 | |
| 	cpu_set_t cpuset;
 | |
| 
 | |
| 	if (next_to_try == nr_cpus)
 | |
| 		return -1;
 | |
| 
 | |
| 	while (next_to_try < nr_cpus) {
 | |
| 		CPU_ZERO(&cpuset);
 | |
| 		CPU_SET(next_to_try++, &cpuset);
 | |
| 		if (!sched_setaffinity(pid, sizeof(cpuset), &cpuset))
 | |
| 			break;
 | |
| 	}
 | |
| 
 | |
| 	return next_to_try;
 | |
| }
 | |
| 
 | |
| static void run_parallel(unsigned int tasks, void (*fn)(int i, void *data),
 | |
| 			 void *data)
 | |
| {
 | |
| 	int next_sched_cpu = 0;
 | |
| 	pid_t pid[tasks];
 | |
| 	int i;
 | |
| 
 | |
| 	for (i = 0; i < tasks; i++) {
 | |
| 		pid[i] = fork();
 | |
| 		if (pid[i] == 0) {
 | |
| 			next_sched_cpu = sched_next_online(0, next_sched_cpu);
 | |
| 			fn(i, data);
 | |
| 			exit(0);
 | |
| 		} else if (pid[i] == -1) {
 | |
| 			printf("couldn't spawn #%d process\n", i);
 | |
| 			exit(1);
 | |
| 		}
 | |
| 		/* It is mostly redundant and just allow the parent
 | |
| 		 * process to update next_shced_cpu for the next child
 | |
| 		 * process
 | |
| 		 */
 | |
| 		next_sched_cpu = sched_next_online(pid[i], next_sched_cpu);
 | |
| 	}
 | |
| 	for (i = 0; i < tasks; i++) {
 | |
| 		int status;
 | |
| 
 | |
| 		assert(waitpid(pid[i], &status, 0) == pid[i]);
 | |
| 		assert(status == 0);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static void do_test_lru_dist(int task, void *data)
 | |
| {
 | |
| 	unsigned int nr_misses = 0;
 | |
| 	struct pfect_lru pfect_lru;
 | |
| 	unsigned long long key, value = 1234;
 | |
| 	unsigned int i;
 | |
| 
 | |
| 	unsigned int lru_map_fd = ((unsigned int *)data)[0];
 | |
| 	unsigned int lru_size = ((unsigned int *)data)[1];
 | |
| 	unsigned long long key_offset = task * dist_key_counts;
 | |
| 
 | |
| 	pfect_lru_init(&pfect_lru, lru_size, dist_key_counts);
 | |
| 
 | |
| 	for (i = 0; i < dist_key_counts; i++) {
 | |
| 		key = dist_keys[i] + key_offset;
 | |
| 
 | |
| 		pfect_lru_lookup_or_insert(&pfect_lru, key);
 | |
| 
 | |
| 		if (!bpf_map_lookup_elem(lru_map_fd, &key, &value))
 | |
| 			continue;
 | |
| 
 | |
| 		if (bpf_map_update_elem(lru_map_fd, &key, &value, BPF_NOEXIST)) {
 | |
| 			printf("bpf_map_update_elem(lru_map_fd, %llu): errno:%d\n",
 | |
| 			       key, errno);
 | |
| 			assert(0);
 | |
| 		}
 | |
| 
 | |
| 		nr_misses++;
 | |
| 	}
 | |
| 
 | |
| 	printf("    task:%d BPF LRU: nr_unique:%u(/%u) nr_misses:%u(/%u)\n",
 | |
| 	       task, pfect_lru.nr_unique, dist_key_counts, nr_misses,
 | |
| 	       dist_key_counts);
 | |
| 	printf("    task:%d Perfect LRU: nr_unique:%u(/%u) nr_misses:%u(/%u)\n",
 | |
| 	       task, pfect_lru.nr_unique, pfect_lru.total,
 | |
| 	       pfect_lru.nr_misses, pfect_lru.total);
 | |
| 
 | |
| 	pfect_lru_destroy(&pfect_lru);
 | |
| 	close(lru_map_fd);
 | |
| }
 | |
| 
 | |
| static void test_parallel_lru_dist(int map_type, int map_flags,
 | |
| 				   int nr_tasks, unsigned int lru_size)
 | |
| {
 | |
| 	int child_data[2];
 | |
| 	int lru_map_fd;
 | |
| 
 | |
| 	printf("%s (map_type:%d map_flags:0x%X):\n", __func__, map_type,
 | |
| 	       map_flags);
 | |
| 
 | |
| 	if (map_flags & BPF_F_NO_COMMON_LRU)
 | |
| 		lru_map_fd = create_map(map_type, map_flags,
 | |
| 					nr_cpus * lru_size);
 | |
| 	else
 | |
| 		lru_map_fd = create_map(map_type, map_flags,
 | |
| 					nr_tasks * lru_size);
 | |
| 	assert(lru_map_fd != -1);
 | |
| 
 | |
| 	child_data[0] = lru_map_fd;
 | |
| 	child_data[1] = lru_size;
 | |
| 
 | |
| 	run_parallel(nr_tasks, do_test_lru_dist, child_data);
 | |
| 
 | |
| 	close(lru_map_fd);
 | |
| }
 | |
| 
 | |
| static void test_lru_loss0(int map_type, int map_flags)
 | |
| {
 | |
| 	unsigned long long key, value[nr_cpus];
 | |
| 	unsigned int old_unused_losses = 0;
 | |
| 	unsigned int new_unused_losses = 0;
 | |
| 	unsigned int used_losses = 0;
 | |
| 	int map_fd;
 | |
| 
 | |
| 	printf("%s (map_type:%d map_flags:0x%X): ", __func__, map_type,
 | |
| 	       map_flags);
 | |
| 
 | |
| 	assert(sched_next_online(0, 0) != -1);
 | |
| 
 | |
| 	if (map_flags & BPF_F_NO_COMMON_LRU)
 | |
| 		map_fd = create_map(map_type, map_flags, 900 * nr_cpus);
 | |
| 	else
 | |
| 		map_fd = create_map(map_type, map_flags, 900);
 | |
| 
 | |
| 	assert(map_fd != -1);
 | |
| 
 | |
| 	value[0] = 1234;
 | |
| 
 | |
| 	for (key = 1; key <= 1000; key++) {
 | |
| 		int start_key, end_key;
 | |
| 
 | |
| 		assert(bpf_map_update_elem(map_fd, &key, value, BPF_NOEXIST) == 0);
 | |
| 
 | |
| 		start_key = 101;
 | |
| 		end_key = min(key, 900);
 | |
| 
 | |
| 		while (start_key <= end_key) {
 | |
| 			bpf_map_lookup_elem(map_fd, &start_key, value);
 | |
| 			start_key++;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	for (key = 1; key <= 1000; key++) {
 | |
| 		if (bpf_map_lookup_elem(map_fd, &key, value)) {
 | |
| 			if (key <= 100)
 | |
| 				old_unused_losses++;
 | |
| 			else if (key <= 900)
 | |
| 				used_losses++;
 | |
| 			else
 | |
| 				new_unused_losses++;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	close(map_fd);
 | |
| 
 | |
| 	printf("older-elem-losses:%d(/100) active-elem-losses:%d(/800) "
 | |
| 	       "newer-elem-losses:%d(/100)\n",
 | |
| 	       old_unused_losses, used_losses, new_unused_losses);
 | |
| }
 | |
| 
 | |
| static void test_lru_loss1(int map_type, int map_flags)
 | |
| {
 | |
| 	unsigned long long key, value[nr_cpus];
 | |
| 	int map_fd;
 | |
| 	unsigned int nr_losses = 0;
 | |
| 
 | |
| 	printf("%s (map_type:%d map_flags:0x%X): ", __func__, map_type,
 | |
| 	       map_flags);
 | |
| 
 | |
| 	assert(sched_next_online(0, 0) != -1);
 | |
| 
 | |
| 	if (map_flags & BPF_F_NO_COMMON_LRU)
 | |
| 		map_fd = create_map(map_type, map_flags, 1000 * nr_cpus);
 | |
| 	else
 | |
| 		map_fd = create_map(map_type, map_flags, 1000);
 | |
| 
 | |
| 	assert(map_fd != -1);
 | |
| 
 | |
| 	value[0] = 1234;
 | |
| 
 | |
| 	for (key = 1; key <= 1000; key++)
 | |
| 		assert(!bpf_map_update_elem(map_fd, &key, value, BPF_NOEXIST));
 | |
| 
 | |
| 	for (key = 1; key <= 1000; key++) {
 | |
| 		if (bpf_map_lookup_elem(map_fd, &key, value))
 | |
| 			nr_losses++;
 | |
| 	}
 | |
| 
 | |
| 	close(map_fd);
 | |
| 
 | |
| 	printf("nr_losses:%d(/1000)\n", nr_losses);
 | |
| }
 | |
| 
 | |
| static void do_test_parallel_lru_loss(int task, void *data)
 | |
| {
 | |
| 	const unsigned int nr_stable_elems = 1000;
 | |
| 	const unsigned int nr_repeats = 100000;
 | |
| 
 | |
| 	int map_fd = *(int *)data;
 | |
| 	unsigned long long stable_base;
 | |
| 	unsigned long long key, value[nr_cpus];
 | |
| 	unsigned long long next_ins_key;
 | |
| 	unsigned int nr_losses = 0;
 | |
| 	unsigned int i;
 | |
| 
 | |
| 	stable_base = task * nr_repeats * 2 + 1;
 | |
| 	next_ins_key = stable_base;
 | |
| 	value[0] = 1234;
 | |
| 	for (i = 0; i < nr_stable_elems; i++) {
 | |
| 		assert(bpf_map_update_elem(map_fd, &next_ins_key, value,
 | |
| 				       BPF_NOEXIST) == 0);
 | |
| 		next_ins_key++;
 | |
| 	}
 | |
| 
 | |
| 	for (i = 0; i < nr_repeats; i++) {
 | |
| 		int rn;
 | |
| 
 | |
| 		rn = rand();
 | |
| 
 | |
| 		if (rn % 10) {
 | |
| 			key = rn % nr_stable_elems + stable_base;
 | |
| 			bpf_map_lookup_elem(map_fd, &key, value);
 | |
| 		} else {
 | |
| 			bpf_map_update_elem(map_fd, &next_ins_key, value,
 | |
| 					BPF_NOEXIST);
 | |
| 			next_ins_key++;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	key = stable_base;
 | |
| 	for (i = 0; i < nr_stable_elems; i++) {
 | |
| 		if (bpf_map_lookup_elem(map_fd, &key, value))
 | |
| 			nr_losses++;
 | |
| 		key++;
 | |
| 	}
 | |
| 
 | |
| 	printf("    task:%d nr_losses:%u\n", task, nr_losses);
 | |
| }
 | |
| 
 | |
| static void test_parallel_lru_loss(int map_type, int map_flags, int nr_tasks)
 | |
| {
 | |
| 	int map_fd;
 | |
| 
 | |
| 	printf("%s (map_type:%d map_flags:0x%X):\n", __func__, map_type,
 | |
| 	       map_flags);
 | |
| 
 | |
| 	/* Give 20% more than the active working set */
 | |
| 	if (map_flags & BPF_F_NO_COMMON_LRU)
 | |
| 		map_fd = create_map(map_type, map_flags,
 | |
| 				    nr_cpus * (1000 + 200));
 | |
| 	else
 | |
| 		map_fd = create_map(map_type, map_flags,
 | |
| 				    nr_tasks * (1000 + 200));
 | |
| 
 | |
| 	assert(map_fd != -1);
 | |
| 
 | |
| 	run_parallel(nr_tasks, do_test_parallel_lru_loss, &map_fd);
 | |
| 
 | |
| 	close(map_fd);
 | |
| }
 | |
| 
 | |
| int main(int argc, char **argv)
 | |
| {
 | |
| 	int map_flags[] = {0, BPF_F_NO_COMMON_LRU};
 | |
| 	const char *dist_file;
 | |
| 	int nr_tasks = 1;
 | |
| 	int lru_size;
 | |
| 	int f;
 | |
| 
 | |
| 	if (argc < 4) {
 | |
| 		printf("Usage: %s <dist-file> <lru-size> <nr-tasks>\n",
 | |
| 		       argv[0]);
 | |
| 		return -1;
 | |
| 	}
 | |
| 
 | |
| 	dist_file = argv[1];
 | |
| 	lru_size = atoi(argv[2]);
 | |
| 	nr_tasks = atoi(argv[3]);
 | |
| 
 | |
| 	setbuf(stdout, NULL);
 | |
| 
 | |
| 	srand(time(NULL));
 | |
| 
 | |
| 	nr_cpus = bpf_num_possible_cpus();
 | |
| 	assert(nr_cpus != -1);
 | |
| 	printf("nr_cpus:%d\n\n", nr_cpus);
 | |
| 
 | |
| 	nr_tasks = min(nr_tasks, nr_cpus);
 | |
| 
 | |
| 	dist_key_counts = read_keys(dist_file, &dist_keys);
 | |
| 	if (!dist_key_counts) {
 | |
| 		printf("%s has no key\n", dist_file);
 | |
| 		return -1;
 | |
| 	}
 | |
| 
 | |
| 	for (f = 0; f < sizeof(map_flags) / sizeof(*map_flags); f++) {
 | |
| 		test_lru_loss0(BPF_MAP_TYPE_LRU_HASH, map_flags[f]);
 | |
| 		test_lru_loss1(BPF_MAP_TYPE_LRU_HASH, map_flags[f]);
 | |
| 		test_parallel_lru_loss(BPF_MAP_TYPE_LRU_HASH, map_flags[f],
 | |
| 				       nr_tasks);
 | |
| 		test_parallel_lru_dist(BPF_MAP_TYPE_LRU_HASH, map_flags[f],
 | |
| 				       nr_tasks, lru_size);
 | |
| 		printf("\n");
 | |
| 	}
 | |
| 
 | |
| 	free(dist_keys);
 | |
| 
 | |
| 	return 0;
 | |
| }
 |