mirror of
git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-09-04 20:19:47 +08:00
Core
----
- Allow live renaming when an interface is up
- Add retpoline wrappers for tc, improving considerably the
performances of complex queue discipline configurations.
- Add inet drop monitor support.
- A few GRO performance improvements.
- Add infrastructure for atomic dev stats, addressing long standing
data races.
- De-duplicate common code between OVS and conntrack offloading
infrastructure.
- A bunch of UBSAN_BOUNDS/FORTIFY_SOURCE improvements.
- Netfilter: introduce packet parser for tunneled packets
- Replace IPVS timer-based estimators with kthreads to scale up
the workload with the number of available CPUs.
- Add the helper support for connection-tracking OVS offload.
BPF
---
- Support for user defined BPF objects: the use case is to allocate
own objects, build own object hierarchies and use the building
blocks to build own data structures flexibly, for example, linked
lists in BPF.
- Make cgroup local storage available to non-cgroup attached BPF
programs.
- Avoid unnecessary deadlock detection and failures wrt BPF task
storage helpers.
- A relevant bunch of BPF verifier fixes and improvements.
- Veristat tool improvements to support custom filtering, sorting,
and replay of results.
- Add LLVM disassembler as default library for dumping JITed code.
- Lots of new BPF documentation for various BPF maps.
- Add bpf_rcu_read_{,un}lock() support for sleepable programs.
- Add RCU grace period chaining to BPF to wait for the completion
of access from both sleepable and non-sleepable BPF programs.
- Add support storing struct task_struct objects as kptrs in maps.
- Improve helper UAPI by explicitly defining BPF_FUNC_xxx integer
values.
- Add libbpf *_opts API-variants for bpf_*_get_fd_by_id() functions.
Protocols
---------
- TCP: implement Protective Load Balancing across switch links.
- TCP: allow dynamically disabling TCP-MD5 static key, reverting
back to fast[er]-path.
- UDP: Introduce optional per-netns hash lookup table.
- IPv6: simplify and cleanup sockets disposal.
- Netlink: support different type policies for each generic
netlink operation.
- MPTCP: add MSG_FASTOPEN and FastOpen listener side support.
- MPTCP: add netlink notification support for listener sockets
events.
- SCTP: add VRF support, allowing sctp sockets binding to VRF
devices.
- Add bridging MAC Authentication Bypass (MAB) support.
- Extensions for Ethernet VPN bridging implementation to better
support multicast scenarios.
- More work for Wi-Fi 7 support, comprising conversion of all
the existing drivers to internal TX queue usage.
- IPSec: introduce a new offload type (packet offload) allowing
complete header processing and crypto offloading.
- IPSec: extended ack support for more descriptive XFRM error
reporting.
- RXRPC: increase SACK table size and move processing into a
per-local endpoint kernel thread, reducing considerably the
required locking.
- IEEE 802154: synchronous send frame and extended filtering
support, initial support for scanning available 15.4 networks.
- Tun: bump the link speed from 10Mbps to 10Gbps.
- Tun/VirtioNet: implement UDP segmentation offload support.
Driver API
----------
- PHY/SFP: improve power level switching between standard
level 1 and the higher power levels.
- New API for netdev <-> devlink_port linkage.
- PTP: convert existing drivers to new frequency adjustment
implementation.
- DSA: add support for rx offloading.
- Autoload DSA tagging driver when dynamically changing protocol.
- Add new PCP and APPTRUST attributes to Data Center Bridging.
- Add configuration support for 800Gbps link speed.
- Add devlink port function attribute to enable/disable RoCE and
migratable.
- Extend devlink-rate to support strict prioriry and weighted fair
queuing.
- Add devlink support to directly reading from region memory.
- New device tree helper to fetch MAC address from nvmem.
- New big TCP helper to simplify temporary header stripping.
New hardware / drivers
----------------------
- Ethernet:
- Marvel Octeon CNF95N and CN10KB Ethernet Switches.
- Marvel Prestera AC5X Ethernet Switch.
- WangXun 10 Gigabit NIC.
- Motorcomm yt8521 Gigabit Ethernet.
- Microchip ksz9563 Gigabit Ethernet Switch.
- Microsoft Azure Network Adapter.
- Linux Automation 10Base-T1L adapter.
- PHY:
- Aquantia AQR112 and AQR412.
- Motorcomm YT8531S.
- PTP:
- Orolia ART-CARD.
- WiFi:
- MediaTek Wi-Fi 7 (802.11be) devices.
- RealTek rtw8821cu, rtw8822bu, rtw8822cu and rtw8723du USB
devices.
- Bluetooth:
- Broadcom BCM4377/4378/4387 Bluetooth chipsets.
- Realtek RTL8852BE and RTL8723DS.
- Cypress.CYW4373A0 WiFi + Bluetooth combo device.
Drivers
-------
- CAN:
- gs_usb: bus error reporting support.
- kvaser_usb: listen only and bus error reporting support.
- Ethernet NICs:
- Intel (100G):
- extend action skbedit to RX queue mapping.
- implement devlink-rate support.
- support direct read from memory.
- nVidia/Mellanox (mlx5):
- SW steering improvements, increasing rules update rate.
- Support for enhanced events compression.
- extend H/W offload packet manipulation capabilities.
- implement IPSec packet offload mode.
- nVidia/Mellanox (mlx4):
- better big TCP support.
- Netronome Ethernet NICs (nfp):
- IPsec offload support.
- add support for multicast filter.
- Broadcom:
- RSS and PTP support improvements.
- AMD/SolarFlare:
- netlink extened ack improvements.
- add basic flower matches to offload, and related stats.
- Virtual NICs:
- ibmvnic: introduce affinity hint support.
- small / embedded:
- FreeScale fec: add initial XDP support.
- Marvel mv643xx_eth: support MII/GMII/RGMII modes for Kirkwood.
- TI am65-cpsw: add suspend/resume support.
- Mediatek MT7986: add RX wireless wthernet dispatch support.
- Realtek 8169: enable GRO software interrupt coalescing per
default.
- Ethernet high-speed switches:
- Microchip (sparx5):
- add support for Sparx5 TC/flower H/W offload via VCAP.
- Mellanox mlxsw:
- add 802.1X and MAC Authentication Bypass offload support.
- add ip6gre support.
- Embedded Ethernet switches:
- Mediatek (mtk_eth_soc):
- improve PCS implementation, add DSA untag support.
- enable flow offload support.
- Renesas:
- add rswitch R-Car Gen4 gPTP support.
- Microchip (lan966x):
- add full XDP support.
- add TC H/W offload via VCAP.
- enable PTP on bridge interfaces.
- Microchip (ksz8):
- add MTU support for KSZ8 series.
- Qualcomm 802.11ax WiFi (ath11k):
- support configuring channel dwell time during scan.
- MediaTek WiFi (mt76):
- enable Wireless Ethernet Dispatch (WED) offload support.
- add ack signal support.
- enable coredump support.
- remain_on_channel support.
- Intel WiFi (iwlwifi):
- enable Wi-Fi 7 Extremely High Throughput (EHT) PHY capabilities.
- 320 MHz channels support.
- RealTek WiFi (rtw89):
- new dynamic header firmware format support.
- wake-over-WLAN support.
Signed-off-by: Paolo Abeni <pabeni@redhat.com>
-----BEGIN PGP SIGNATURE-----
iQJGBAABCAAwFiEEg1AjqC77wbdLX2LbKSR5jcyPE6QFAmOYXUcSHHBhYmVuaUBy
ZWRoYXQuY29tAAoJECkkeY3MjxOk8zQP/R7BZtbJMTPiWkRnSoKHnAyupDVwrz5U
ktukLkwPsCyJuEbAjgxrxf4EEEQ9uq2FFlxNSYuKiiQMqIpFxV6KED7LCUygn4Tc
kxtkp0Q+5XiqisWlQmtfExf2OjuuPqcjV9tWCDBI6GebKUbfNwY/eI44RcMu4BSv
DzIlW5GkX/kZAPqnnuqaLsN3FudDTJHGEAD7NbA++7wJ076RWYSLXlFv0Z+SCSPS
H8/PEG0/ZK/65rIWMAFRClJ9BNIDwGVgp0GrsIvs1gqbRUOlA1hl1rDM21TqtNFf
5QPQT7sIfTcCE/nerxKJD5JE3JyP+XRlRn96PaRw3rt4MgI6I/EOj/HOKQ5tMCNc
oPiqb7N70+hkLZyr42qX+vN9eDPjp2koEQm7EO2Zs+/534/zWDs24Zfk/Aa1ps0I
Fa82oGjAgkBhGe/FZ6i5cYoLcyxqRqZV1Ws9XQMl72qRC7/BwvNbIW6beLpCRyeM
yYIU+0e9dEm+wHQEdh2niJuVtR63hy8tvmPx56lyh+6u0+pondkwbfSiC5aD3kAC
ikKsN5DyEsdXyiBAlytCEBxnaOjQy4RAz+3YXSiS0eBNacXp03UUrNGx4Pzpu/D0
QLFJhBnMFFCgy5to8/DvKnrTPgZdSURwqbIUcZdvU21f1HLR8tUTpaQnYffc/Whm
V8gnt1EL+0cc
=CbJC
-----END PGP SIGNATURE-----
Merge tag 'net-next-6.2' of git://git.kernel.org/pub/scm/linux/kernel/git/netdev/net-next
Pull networking updates from Paolo Abeni:
"Core:
- Allow live renaming when an interface is up
- Add retpoline wrappers for tc, improving considerably the
performances of complex queue discipline configurations
- Add inet drop monitor support
- A few GRO performance improvements
- Add infrastructure for atomic dev stats, addressing long standing
data races
- De-duplicate common code between OVS and conntrack offloading
infrastructure
- A bunch of UBSAN_BOUNDS/FORTIFY_SOURCE improvements
- Netfilter: introduce packet parser for tunneled packets
- Replace IPVS timer-based estimators with kthreads to scale up the
workload with the number of available CPUs
- Add the helper support for connection-tracking OVS offload
BPF:
- Support for user defined BPF objects: the use case is to allocate
own objects, build own object hierarchies and use the building
blocks to build own data structures flexibly, for example, linked
lists in BPF
- Make cgroup local storage available to non-cgroup attached BPF
programs
- Avoid unnecessary deadlock detection and failures wrt BPF task
storage helpers
- A relevant bunch of BPF verifier fixes and improvements
- Veristat tool improvements to support custom filtering, sorting,
and replay of results
- Add LLVM disassembler as default library for dumping JITed code
- Lots of new BPF documentation for various BPF maps
- Add bpf_rcu_read_{,un}lock() support for sleepable programs
- Add RCU grace period chaining to BPF to wait for the completion of
access from both sleepable and non-sleepable BPF programs
- Add support storing struct task_struct objects as kptrs in maps
- Improve helper UAPI by explicitly defining BPF_FUNC_xxx integer
values
- Add libbpf *_opts API-variants for bpf_*_get_fd_by_id() functions
Protocols:
- TCP: implement Protective Load Balancing across switch links
- TCP: allow dynamically disabling TCP-MD5 static key, reverting back
to fast[er]-path
- UDP: Introduce optional per-netns hash lookup table
- IPv6: simplify and cleanup sockets disposal
- Netlink: support different type policies for each generic netlink
operation
- MPTCP: add MSG_FASTOPEN and FastOpen listener side support
- MPTCP: add netlink notification support for listener sockets events
- SCTP: add VRF support, allowing sctp sockets binding to VRF devices
- Add bridging MAC Authentication Bypass (MAB) support
- Extensions for Ethernet VPN bridging implementation to better
support multicast scenarios
- More work for Wi-Fi 7 support, comprising conversion of all the
existing drivers to internal TX queue usage
- IPSec: introduce a new offload type (packet offload) allowing
complete header processing and crypto offloading
- IPSec: extended ack support for more descriptive XFRM error
reporting
- RXRPC: increase SACK table size and move processing into a
per-local endpoint kernel thread, reducing considerably the
required locking
- IEEE 802154: synchronous send frame and extended filtering support,
initial support for scanning available 15.4 networks
- Tun: bump the link speed from 10Mbps to 10Gbps
- Tun/VirtioNet: implement UDP segmentation offload support
Driver API:
- PHY/SFP: improve power level switching between standard level 1 and
the higher power levels
- New API for netdev <-> devlink_port linkage
- PTP: convert existing drivers to new frequency adjustment
implementation
- DSA: add support for rx offloading
- Autoload DSA tagging driver when dynamically changing protocol
- Add new PCP and APPTRUST attributes to Data Center Bridging
- Add configuration support for 800Gbps link speed
- Add devlink port function attribute to enable/disable RoCE and
migratable
- Extend devlink-rate to support strict prioriry and weighted fair
queuing
- Add devlink support to directly reading from region memory
- New device tree helper to fetch MAC address from nvmem
- New big TCP helper to simplify temporary header stripping
New hardware / drivers:
- Ethernet:
- Marvel Octeon CNF95N and CN10KB Ethernet Switches
- Marvel Prestera AC5X Ethernet Switch
- WangXun 10 Gigabit NIC
- Motorcomm yt8521 Gigabit Ethernet
- Microchip ksz9563 Gigabit Ethernet Switch
- Microsoft Azure Network Adapter
- Linux Automation 10Base-T1L adapter
- PHY:
- Aquantia AQR112 and AQR412
- Motorcomm YT8531S
- PTP:
- Orolia ART-CARD
- WiFi:
- MediaTek Wi-Fi 7 (802.11be) devices
- RealTek rtw8821cu, rtw8822bu, rtw8822cu and rtw8723du USB
devices
- Bluetooth:
- Broadcom BCM4377/4378/4387 Bluetooth chipsets
- Realtek RTL8852BE and RTL8723DS
- Cypress.CYW4373A0 WiFi + Bluetooth combo device
Drivers:
- CAN:
- gs_usb: bus error reporting support
- kvaser_usb: listen only and bus error reporting support
- Ethernet NICs:
- Intel (100G):
- extend action skbedit to RX queue mapping
- implement devlink-rate support
- support direct read from memory
- nVidia/Mellanox (mlx5):
- SW steering improvements, increasing rules update rate
- Support for enhanced events compression
- extend H/W offload packet manipulation capabilities
- implement IPSec packet offload mode
- nVidia/Mellanox (mlx4):
- better big TCP support
- Netronome Ethernet NICs (nfp):
- IPsec offload support
- add support for multicast filter
- Broadcom:
- RSS and PTP support improvements
- AMD/SolarFlare:
- netlink extened ack improvements
- add basic flower matches to offload, and related stats
- Virtual NICs:
- ibmvnic: introduce affinity hint support
- small / embedded:
- FreeScale fec: add initial XDP support
- Marvel mv643xx_eth: support MII/GMII/RGMII modes for Kirkwood
- TI am65-cpsw: add suspend/resume support
- Mediatek MT7986: add RX wireless wthernet dispatch support
- Realtek 8169: enable GRO software interrupt coalescing per
default
- Ethernet high-speed switches:
- Microchip (sparx5):
- add support for Sparx5 TC/flower H/W offload via VCAP
- Mellanox mlxsw:
- add 802.1X and MAC Authentication Bypass offload support
- add ip6gre support
- Embedded Ethernet switches:
- Mediatek (mtk_eth_soc):
- improve PCS implementation, add DSA untag support
- enable flow offload support
- Renesas:
- add rswitch R-Car Gen4 gPTP support
- Microchip (lan966x):
- add full XDP support
- add TC H/W offload via VCAP
- enable PTP on bridge interfaces
- Microchip (ksz8):
- add MTU support for KSZ8 series
- Qualcomm 802.11ax WiFi (ath11k):
- support configuring channel dwell time during scan
- MediaTek WiFi (mt76):
- enable Wireless Ethernet Dispatch (WED) offload support
- add ack signal support
- enable coredump support
- remain_on_channel support
- Intel WiFi (iwlwifi):
- enable Wi-Fi 7 Extremely High Throughput (EHT) PHY capabilities
- 320 MHz channels support
- RealTek WiFi (rtw89):
- new dynamic header firmware format support
- wake-over-WLAN support"
* tag 'net-next-6.2' of git://git.kernel.org/pub/scm/linux/kernel/git/netdev/net-next: (2002 commits)
ipvs: fix type warning in do_div() on 32 bit
net: lan966x: Remove a useless test in lan966x_ptp_add_trap()
net: ipa: add IPA v4.7 support
dt-bindings: net: qcom,ipa: Add SM6350 compatible
bnxt: Use generic HBH removal helper in tx path
IPv6/GRO: generic helper to remove temporary HBH/jumbo header in driver
selftests: forwarding: Add bridge MDB test
selftests: forwarding: Rename bridge_mdb test
bridge: mcast: Support replacement of MDB port group entries
bridge: mcast: Allow user space to specify MDB entry routing protocol
bridge: mcast: Allow user space to add (*, G) with a source list and filter mode
bridge: mcast: Add support for (*, G) with a source list and filter mode
bridge: mcast: Avoid arming group timer when (S, G) corresponds to a source
bridge: mcast: Add a flag for user installed source entries
bridge: mcast: Expose __br_multicast_del_group_src()
bridge: mcast: Expose br_multicast_new_group_src()
bridge: mcast: Add a centralized error path
bridge: mcast: Place netlink policy before validation functions
bridge: mcast: Split (*, G) and (S, G) addition into different functions
bridge: mcast: Do not derive entry type from its filter mode
...
814 lines
20 KiB
C
814 lines
20 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* Resizable, Scalable, Concurrent Hash Table
|
|
*
|
|
* Copyright (c) 2014-2015 Thomas Graf <tgraf@suug.ch>
|
|
* Copyright (c) 2008-2014 Patrick McHardy <kaber@trash.net>
|
|
*/
|
|
|
|
/**************************************************************************
|
|
* Self Test
|
|
**************************************************************************/
|
|
|
|
#include <linux/init.h>
|
|
#include <linux/jhash.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/kthread.h>
|
|
#include <linux/module.h>
|
|
#include <linux/rcupdate.h>
|
|
#include <linux/rhashtable.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/random.h>
|
|
#include <linux/vmalloc.h>
|
|
#include <linux/wait.h>
|
|
|
|
#define MAX_ENTRIES 1000000
|
|
#define TEST_INSERT_FAIL INT_MAX
|
|
|
|
static int parm_entries = 50000;
|
|
module_param(parm_entries, int, 0);
|
|
MODULE_PARM_DESC(parm_entries, "Number of entries to add (default: 50000)");
|
|
|
|
static int runs = 4;
|
|
module_param(runs, int, 0);
|
|
MODULE_PARM_DESC(runs, "Number of test runs per variant (default: 4)");
|
|
|
|
static int max_size = 0;
|
|
module_param(max_size, int, 0);
|
|
MODULE_PARM_DESC(max_size, "Maximum table size (default: calculated)");
|
|
|
|
static bool shrinking = false;
|
|
module_param(shrinking, bool, 0);
|
|
MODULE_PARM_DESC(shrinking, "Enable automatic shrinking (default: off)");
|
|
|
|
static int size = 8;
|
|
module_param(size, int, 0);
|
|
MODULE_PARM_DESC(size, "Initial size hint of table (default: 8)");
|
|
|
|
static int tcount = 10;
|
|
module_param(tcount, int, 0);
|
|
MODULE_PARM_DESC(tcount, "Number of threads to spawn (default: 10)");
|
|
|
|
static bool enomem_retry = false;
|
|
module_param(enomem_retry, bool, 0);
|
|
MODULE_PARM_DESC(enomem_retry, "Retry insert even if -ENOMEM was returned (default: off)");
|
|
|
|
struct test_obj_val {
|
|
int id;
|
|
int tid;
|
|
};
|
|
|
|
struct test_obj {
|
|
struct test_obj_val value;
|
|
struct rhash_head node;
|
|
};
|
|
|
|
struct test_obj_rhl {
|
|
struct test_obj_val value;
|
|
struct rhlist_head list_node;
|
|
};
|
|
|
|
struct thread_data {
|
|
unsigned int entries;
|
|
int id;
|
|
struct task_struct *task;
|
|
struct test_obj *objs;
|
|
};
|
|
|
|
static u32 my_hashfn(const void *data, u32 len, u32 seed)
|
|
{
|
|
const struct test_obj_rhl *obj = data;
|
|
|
|
return (obj->value.id % 10);
|
|
}
|
|
|
|
static int my_cmpfn(struct rhashtable_compare_arg *arg, const void *obj)
|
|
{
|
|
const struct test_obj_rhl *test_obj = obj;
|
|
const struct test_obj_val *val = arg->key;
|
|
|
|
return test_obj->value.id - val->id;
|
|
}
|
|
|
|
static struct rhashtable_params test_rht_params = {
|
|
.head_offset = offsetof(struct test_obj, node),
|
|
.key_offset = offsetof(struct test_obj, value),
|
|
.key_len = sizeof(struct test_obj_val),
|
|
.hashfn = jhash,
|
|
};
|
|
|
|
static struct rhashtable_params test_rht_params_dup = {
|
|
.head_offset = offsetof(struct test_obj_rhl, list_node),
|
|
.key_offset = offsetof(struct test_obj_rhl, value),
|
|
.key_len = sizeof(struct test_obj_val),
|
|
.hashfn = jhash,
|
|
.obj_hashfn = my_hashfn,
|
|
.obj_cmpfn = my_cmpfn,
|
|
.nelem_hint = 128,
|
|
.automatic_shrinking = false,
|
|
};
|
|
|
|
static atomic_t startup_count;
|
|
static DECLARE_WAIT_QUEUE_HEAD(startup_wait);
|
|
|
|
static int insert_retry(struct rhashtable *ht, struct test_obj *obj,
|
|
const struct rhashtable_params params)
|
|
{
|
|
int err, retries = -1, enomem_retries = 0;
|
|
|
|
do {
|
|
retries++;
|
|
cond_resched();
|
|
err = rhashtable_insert_fast(ht, &obj->node, params);
|
|
if (err == -ENOMEM && enomem_retry) {
|
|
enomem_retries++;
|
|
err = -EBUSY;
|
|
}
|
|
} while (err == -EBUSY);
|
|
|
|
if (enomem_retries)
|
|
pr_info(" %u insertions retried after -ENOMEM\n",
|
|
enomem_retries);
|
|
|
|
return err ? : retries;
|
|
}
|
|
|
|
static int __init test_rht_lookup(struct rhashtable *ht, struct test_obj *array,
|
|
unsigned int entries)
|
|
{
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < entries; i++) {
|
|
struct test_obj *obj;
|
|
bool expected = !(i % 2);
|
|
struct test_obj_val key = {
|
|
.id = i,
|
|
};
|
|
|
|
if (array[i / 2].value.id == TEST_INSERT_FAIL)
|
|
expected = false;
|
|
|
|
obj = rhashtable_lookup_fast(ht, &key, test_rht_params);
|
|
|
|
if (expected && !obj) {
|
|
pr_warn("Test failed: Could not find key %u\n", key.id);
|
|
return -ENOENT;
|
|
} else if (!expected && obj) {
|
|
pr_warn("Test failed: Unexpected entry found for key %u\n",
|
|
key.id);
|
|
return -EEXIST;
|
|
} else if (expected && obj) {
|
|
if (obj->value.id != i) {
|
|
pr_warn("Test failed: Lookup value mismatch %u!=%u\n",
|
|
obj->value.id, i);
|
|
return -EINVAL;
|
|
}
|
|
}
|
|
|
|
cond_resched_rcu();
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void test_bucket_stats(struct rhashtable *ht, unsigned int entries)
|
|
{
|
|
unsigned int total = 0, chain_len = 0;
|
|
struct rhashtable_iter hti;
|
|
struct rhash_head *pos;
|
|
|
|
rhashtable_walk_enter(ht, &hti);
|
|
rhashtable_walk_start(&hti);
|
|
|
|
while ((pos = rhashtable_walk_next(&hti))) {
|
|
if (PTR_ERR(pos) == -EAGAIN) {
|
|
pr_info("Info: encountered resize\n");
|
|
chain_len++;
|
|
continue;
|
|
} else if (IS_ERR(pos)) {
|
|
pr_warn("Test failed: rhashtable_walk_next() error: %ld\n",
|
|
PTR_ERR(pos));
|
|
break;
|
|
}
|
|
|
|
total++;
|
|
}
|
|
|
|
rhashtable_walk_stop(&hti);
|
|
rhashtable_walk_exit(&hti);
|
|
|
|
pr_info(" Traversal complete: counted=%u, nelems=%u, entries=%d, table-jumps=%u\n",
|
|
total, atomic_read(&ht->nelems), entries, chain_len);
|
|
|
|
if (total != atomic_read(&ht->nelems) || total != entries)
|
|
pr_warn("Test failed: Total count mismatch ^^^");
|
|
}
|
|
|
|
static s64 __init test_rhashtable(struct rhashtable *ht, struct test_obj *array,
|
|
unsigned int entries)
|
|
{
|
|
struct test_obj *obj;
|
|
int err;
|
|
unsigned int i, insert_retries = 0;
|
|
s64 start, end;
|
|
|
|
/*
|
|
* Insertion Test:
|
|
* Insert entries into table with all keys even numbers
|
|
*/
|
|
pr_info(" Adding %d keys\n", entries);
|
|
start = ktime_get_ns();
|
|
for (i = 0; i < entries; i++) {
|
|
struct test_obj *obj = &array[i];
|
|
|
|
obj->value.id = i * 2;
|
|
err = insert_retry(ht, obj, test_rht_params);
|
|
if (err > 0)
|
|
insert_retries += err;
|
|
else if (err)
|
|
return err;
|
|
}
|
|
|
|
if (insert_retries)
|
|
pr_info(" %u insertions retried due to memory pressure\n",
|
|
insert_retries);
|
|
|
|
test_bucket_stats(ht, entries);
|
|
rcu_read_lock();
|
|
test_rht_lookup(ht, array, entries);
|
|
rcu_read_unlock();
|
|
|
|
test_bucket_stats(ht, entries);
|
|
|
|
pr_info(" Deleting %d keys\n", entries);
|
|
for (i = 0; i < entries; i++) {
|
|
struct test_obj_val key = {
|
|
.id = i * 2,
|
|
};
|
|
|
|
if (array[i].value.id != TEST_INSERT_FAIL) {
|
|
obj = rhashtable_lookup_fast(ht, &key, test_rht_params);
|
|
BUG_ON(!obj);
|
|
|
|
rhashtable_remove_fast(ht, &obj->node, test_rht_params);
|
|
}
|
|
|
|
cond_resched();
|
|
}
|
|
|
|
end = ktime_get_ns();
|
|
pr_info(" Duration of test: %lld ns\n", end - start);
|
|
|
|
return end - start;
|
|
}
|
|
|
|
static struct rhashtable ht;
|
|
static struct rhltable rhlt;
|
|
|
|
static int __init test_rhltable(unsigned int entries)
|
|
{
|
|
struct test_obj_rhl *rhl_test_objects;
|
|
unsigned long *obj_in_table;
|
|
unsigned int i, j, k;
|
|
int ret, err;
|
|
|
|
if (entries == 0)
|
|
entries = 1;
|
|
|
|
rhl_test_objects = vzalloc(array_size(entries,
|
|
sizeof(*rhl_test_objects)));
|
|
if (!rhl_test_objects)
|
|
return -ENOMEM;
|
|
|
|
ret = -ENOMEM;
|
|
obj_in_table = vzalloc(array_size(sizeof(unsigned long),
|
|
BITS_TO_LONGS(entries)));
|
|
if (!obj_in_table)
|
|
goto out_free;
|
|
|
|
err = rhltable_init(&rhlt, &test_rht_params);
|
|
if (WARN_ON(err))
|
|
goto out_free;
|
|
|
|
k = get_random_u32();
|
|
ret = 0;
|
|
for (i = 0; i < entries; i++) {
|
|
rhl_test_objects[i].value.id = k;
|
|
err = rhltable_insert(&rhlt, &rhl_test_objects[i].list_node,
|
|
test_rht_params);
|
|
if (WARN(err, "error %d on element %d\n", err, i))
|
|
break;
|
|
if (err == 0)
|
|
set_bit(i, obj_in_table);
|
|
}
|
|
|
|
if (err)
|
|
ret = err;
|
|
|
|
pr_info("test %d add/delete pairs into rhlist\n", entries);
|
|
for (i = 0; i < entries; i++) {
|
|
struct rhlist_head *h, *pos;
|
|
struct test_obj_rhl *obj;
|
|
struct test_obj_val key = {
|
|
.id = k,
|
|
};
|
|
bool found;
|
|
|
|
rcu_read_lock();
|
|
h = rhltable_lookup(&rhlt, &key, test_rht_params);
|
|
if (WARN(!h, "key not found during iteration %d of %d", i, entries)) {
|
|
rcu_read_unlock();
|
|
break;
|
|
}
|
|
|
|
if (i) {
|
|
j = i - 1;
|
|
rhl_for_each_entry_rcu(obj, pos, h, list_node) {
|
|
if (WARN(pos == &rhl_test_objects[j].list_node, "old element found, should be gone"))
|
|
break;
|
|
}
|
|
}
|
|
|
|
cond_resched_rcu();
|
|
|
|
found = false;
|
|
|
|
rhl_for_each_entry_rcu(obj, pos, h, list_node) {
|
|
if (pos == &rhl_test_objects[i].list_node) {
|
|
found = true;
|
|
break;
|
|
}
|
|
}
|
|
|
|
rcu_read_unlock();
|
|
|
|
if (WARN(!found, "element %d not found", i))
|
|
break;
|
|
|
|
err = rhltable_remove(&rhlt, &rhl_test_objects[i].list_node, test_rht_params);
|
|
WARN(err, "rhltable_remove: err %d for iteration %d\n", err, i);
|
|
if (err == 0)
|
|
clear_bit(i, obj_in_table);
|
|
}
|
|
|
|
if (ret == 0 && err)
|
|
ret = err;
|
|
|
|
for (i = 0; i < entries; i++) {
|
|
WARN(test_bit(i, obj_in_table), "elem %d allegedly still present", i);
|
|
|
|
err = rhltable_insert(&rhlt, &rhl_test_objects[i].list_node,
|
|
test_rht_params);
|
|
if (WARN(err, "error %d on element %d\n", err, i))
|
|
break;
|
|
if (err == 0)
|
|
set_bit(i, obj_in_table);
|
|
}
|
|
|
|
pr_info("test %d random rhlist add/delete operations\n", entries);
|
|
for (j = 0; j < entries; j++) {
|
|
u32 i = get_random_u32_below(entries);
|
|
u32 prand = get_random_u32_below(4);
|
|
|
|
cond_resched();
|
|
|
|
err = rhltable_remove(&rhlt, &rhl_test_objects[i].list_node, test_rht_params);
|
|
if (test_bit(i, obj_in_table)) {
|
|
clear_bit(i, obj_in_table);
|
|
if (WARN(err, "cannot remove element at slot %d", i))
|
|
continue;
|
|
} else {
|
|
if (WARN(err != -ENOENT, "removed non-existent element %d, error %d not %d",
|
|
i, err, -ENOENT))
|
|
continue;
|
|
}
|
|
|
|
if (prand & 1) {
|
|
err = rhltable_insert(&rhlt, &rhl_test_objects[i].list_node, test_rht_params);
|
|
if (err == 0) {
|
|
if (WARN(test_and_set_bit(i, obj_in_table), "succeeded to insert same object %d", i))
|
|
continue;
|
|
} else {
|
|
if (WARN(!test_bit(i, obj_in_table), "failed to insert object %d", i))
|
|
continue;
|
|
}
|
|
}
|
|
|
|
if (prand & 2) {
|
|
i = get_random_u32_below(entries);
|
|
if (test_bit(i, obj_in_table)) {
|
|
err = rhltable_remove(&rhlt, &rhl_test_objects[i].list_node, test_rht_params);
|
|
WARN(err, "cannot remove element at slot %d", i);
|
|
if (err == 0)
|
|
clear_bit(i, obj_in_table);
|
|
} else {
|
|
err = rhltable_insert(&rhlt, &rhl_test_objects[i].list_node, test_rht_params);
|
|
WARN(err, "failed to insert object %d", i);
|
|
if (err == 0)
|
|
set_bit(i, obj_in_table);
|
|
}
|
|
}
|
|
}
|
|
|
|
for (i = 0; i < entries; i++) {
|
|
cond_resched();
|
|
err = rhltable_remove(&rhlt, &rhl_test_objects[i].list_node, test_rht_params);
|
|
if (test_bit(i, obj_in_table)) {
|
|
if (WARN(err, "cannot remove element at slot %d", i))
|
|
continue;
|
|
} else {
|
|
if (WARN(err != -ENOENT, "removed non-existent element, error %d not %d",
|
|
err, -ENOENT))
|
|
continue;
|
|
}
|
|
}
|
|
|
|
rhltable_destroy(&rhlt);
|
|
out_free:
|
|
vfree(rhl_test_objects);
|
|
vfree(obj_in_table);
|
|
return ret;
|
|
}
|
|
|
|
static int __init test_rhashtable_max(struct test_obj *array,
|
|
unsigned int entries)
|
|
{
|
|
unsigned int i;
|
|
int err;
|
|
|
|
test_rht_params.max_size = roundup_pow_of_two(entries / 8);
|
|
err = rhashtable_init(&ht, &test_rht_params);
|
|
if (err)
|
|
return err;
|
|
|
|
for (i = 0; i < ht.max_elems; i++) {
|
|
struct test_obj *obj = &array[i];
|
|
|
|
obj->value.id = i * 2;
|
|
err = insert_retry(&ht, obj, test_rht_params);
|
|
if (err < 0)
|
|
return err;
|
|
}
|
|
|
|
err = insert_retry(&ht, &array[ht.max_elems], test_rht_params);
|
|
if (err == -E2BIG) {
|
|
err = 0;
|
|
} else {
|
|
pr_info("insert element %u should have failed with %d, got %d\n",
|
|
ht.max_elems, -E2BIG, err);
|
|
if (err == 0)
|
|
err = -1;
|
|
}
|
|
|
|
rhashtable_destroy(&ht);
|
|
|
|
return err;
|
|
}
|
|
|
|
static unsigned int __init print_ht(struct rhltable *rhlt)
|
|
{
|
|
struct rhashtable *ht;
|
|
const struct bucket_table *tbl;
|
|
char buff[512] = "";
|
|
int offset = 0;
|
|
unsigned int i, cnt = 0;
|
|
|
|
ht = &rhlt->ht;
|
|
/* Take the mutex to avoid RCU warning */
|
|
mutex_lock(&ht->mutex);
|
|
tbl = rht_dereference(ht->tbl, ht);
|
|
for (i = 0; i < tbl->size; i++) {
|
|
struct rhash_head *pos, *next;
|
|
struct test_obj_rhl *p;
|
|
|
|
pos = rht_ptr_exclusive(tbl->buckets + i);
|
|
next = !rht_is_a_nulls(pos) ? rht_dereference(pos->next, ht) : NULL;
|
|
|
|
if (!rht_is_a_nulls(pos)) {
|
|
offset += sprintf(buff + offset, "\nbucket[%d] -> ", i);
|
|
}
|
|
|
|
while (!rht_is_a_nulls(pos)) {
|
|
struct rhlist_head *list = container_of(pos, struct rhlist_head, rhead);
|
|
offset += sprintf(buff + offset, "[[");
|
|
do {
|
|
pos = &list->rhead;
|
|
list = rht_dereference(list->next, ht);
|
|
p = rht_obj(ht, pos);
|
|
|
|
offset += sprintf(buff + offset, " val %d (tid=%d)%s", p->value.id, p->value.tid,
|
|
list? ", " : " ");
|
|
cnt++;
|
|
} while (list);
|
|
|
|
pos = next,
|
|
next = !rht_is_a_nulls(pos) ?
|
|
rht_dereference(pos->next, ht) : NULL;
|
|
|
|
offset += sprintf(buff + offset, "]]%s", !rht_is_a_nulls(pos) ? " -> " : "");
|
|
}
|
|
}
|
|
printk(KERN_ERR "\n---- ht: ----%s\n-------------\n", buff);
|
|
mutex_unlock(&ht->mutex);
|
|
|
|
return cnt;
|
|
}
|
|
|
|
static int __init test_insert_dup(struct test_obj_rhl *rhl_test_objects,
|
|
int cnt, bool slow)
|
|
{
|
|
struct rhltable *rhlt;
|
|
unsigned int i, ret;
|
|
const char *key;
|
|
int err = 0;
|
|
|
|
rhlt = kmalloc(sizeof(*rhlt), GFP_KERNEL);
|
|
if (WARN_ON(!rhlt))
|
|
return -EINVAL;
|
|
|
|
err = rhltable_init(rhlt, &test_rht_params_dup);
|
|
if (WARN_ON(err)) {
|
|
kfree(rhlt);
|
|
return err;
|
|
}
|
|
|
|
for (i = 0; i < cnt; i++) {
|
|
rhl_test_objects[i].value.tid = i;
|
|
key = rht_obj(&rhlt->ht, &rhl_test_objects[i].list_node.rhead);
|
|
key += test_rht_params_dup.key_offset;
|
|
|
|
if (slow) {
|
|
err = PTR_ERR(rhashtable_insert_slow(&rhlt->ht, key,
|
|
&rhl_test_objects[i].list_node.rhead));
|
|
if (err == -EAGAIN)
|
|
err = 0;
|
|
} else
|
|
err = rhltable_insert(rhlt,
|
|
&rhl_test_objects[i].list_node,
|
|
test_rht_params_dup);
|
|
if (WARN(err, "error %d on element %d/%d (%s)\n", err, i, cnt, slow? "slow" : "fast"))
|
|
goto skip_print;
|
|
}
|
|
|
|
ret = print_ht(rhlt);
|
|
WARN(ret != cnt, "missing rhltable elements (%d != %d, %s)\n", ret, cnt, slow? "slow" : "fast");
|
|
|
|
skip_print:
|
|
rhltable_destroy(rhlt);
|
|
kfree(rhlt);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int __init test_insert_duplicates_run(void)
|
|
{
|
|
struct test_obj_rhl rhl_test_objects[3] = {};
|
|
|
|
pr_info("test inserting duplicates\n");
|
|
|
|
/* two different values that map to same bucket */
|
|
rhl_test_objects[0].value.id = 1;
|
|
rhl_test_objects[1].value.id = 21;
|
|
|
|
/* and another duplicate with same as [0] value
|
|
* which will be second on the bucket list */
|
|
rhl_test_objects[2].value.id = rhl_test_objects[0].value.id;
|
|
|
|
test_insert_dup(rhl_test_objects, 2, false);
|
|
test_insert_dup(rhl_test_objects, 3, false);
|
|
test_insert_dup(rhl_test_objects, 2, true);
|
|
test_insert_dup(rhl_test_objects, 3, true);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int thread_lookup_test(struct thread_data *tdata)
|
|
{
|
|
unsigned int entries = tdata->entries;
|
|
int i, err = 0;
|
|
|
|
for (i = 0; i < entries; i++) {
|
|
struct test_obj *obj;
|
|
struct test_obj_val key = {
|
|
.id = i,
|
|
.tid = tdata->id,
|
|
};
|
|
|
|
obj = rhashtable_lookup_fast(&ht, &key, test_rht_params);
|
|
if (obj && (tdata->objs[i].value.id == TEST_INSERT_FAIL)) {
|
|
pr_err(" found unexpected object %d-%d\n", key.tid, key.id);
|
|
err++;
|
|
} else if (!obj && (tdata->objs[i].value.id != TEST_INSERT_FAIL)) {
|
|
pr_err(" object %d-%d not found!\n", key.tid, key.id);
|
|
err++;
|
|
} else if (obj && memcmp(&obj->value, &key, sizeof(key))) {
|
|
pr_err(" wrong object returned (got %d-%d, expected %d-%d)\n",
|
|
obj->value.tid, obj->value.id, key.tid, key.id);
|
|
err++;
|
|
}
|
|
|
|
cond_resched();
|
|
}
|
|
return err;
|
|
}
|
|
|
|
static int threadfunc(void *data)
|
|
{
|
|
int i, step, err = 0, insert_retries = 0;
|
|
struct thread_data *tdata = data;
|
|
|
|
if (atomic_dec_and_test(&startup_count))
|
|
wake_up(&startup_wait);
|
|
if (wait_event_interruptible(startup_wait, atomic_read(&startup_count) == -1)) {
|
|
pr_err(" thread[%d]: interrupted\n", tdata->id);
|
|
goto out;
|
|
}
|
|
|
|
for (i = 0; i < tdata->entries; i++) {
|
|
tdata->objs[i].value.id = i;
|
|
tdata->objs[i].value.tid = tdata->id;
|
|
err = insert_retry(&ht, &tdata->objs[i], test_rht_params);
|
|
if (err > 0) {
|
|
insert_retries += err;
|
|
} else if (err) {
|
|
pr_err(" thread[%d]: rhashtable_insert_fast failed\n",
|
|
tdata->id);
|
|
goto out;
|
|
}
|
|
}
|
|
if (insert_retries)
|
|
pr_info(" thread[%d]: %u insertions retried due to memory pressure\n",
|
|
tdata->id, insert_retries);
|
|
|
|
err = thread_lookup_test(tdata);
|
|
if (err) {
|
|
pr_err(" thread[%d]: rhashtable_lookup_test failed\n",
|
|
tdata->id);
|
|
goto out;
|
|
}
|
|
|
|
for (step = 10; step > 0; step--) {
|
|
for (i = 0; i < tdata->entries; i += step) {
|
|
if (tdata->objs[i].value.id == TEST_INSERT_FAIL)
|
|
continue;
|
|
err = rhashtable_remove_fast(&ht, &tdata->objs[i].node,
|
|
test_rht_params);
|
|
if (err) {
|
|
pr_err(" thread[%d]: rhashtable_remove_fast failed\n",
|
|
tdata->id);
|
|
goto out;
|
|
}
|
|
tdata->objs[i].value.id = TEST_INSERT_FAIL;
|
|
|
|
cond_resched();
|
|
}
|
|
err = thread_lookup_test(tdata);
|
|
if (err) {
|
|
pr_err(" thread[%d]: rhashtable_lookup_test (2) failed\n",
|
|
tdata->id);
|
|
goto out;
|
|
}
|
|
}
|
|
out:
|
|
while (!kthread_should_stop()) {
|
|
set_current_state(TASK_INTERRUPTIBLE);
|
|
schedule();
|
|
}
|
|
return err;
|
|
}
|
|
|
|
static int __init test_rht_init(void)
|
|
{
|
|
unsigned int entries;
|
|
int i, err, started_threads = 0, failed_threads = 0;
|
|
u64 total_time = 0;
|
|
struct thread_data *tdata;
|
|
struct test_obj *objs;
|
|
|
|
if (parm_entries < 0)
|
|
parm_entries = 1;
|
|
|
|
entries = min(parm_entries, MAX_ENTRIES);
|
|
|
|
test_rht_params.automatic_shrinking = shrinking;
|
|
test_rht_params.max_size = max_size ? : roundup_pow_of_two(entries);
|
|
test_rht_params.nelem_hint = size;
|
|
|
|
objs = vzalloc(array_size(sizeof(struct test_obj),
|
|
test_rht_params.max_size + 1));
|
|
if (!objs)
|
|
return -ENOMEM;
|
|
|
|
pr_info("Running rhashtable test nelem=%d, max_size=%d, shrinking=%d\n",
|
|
size, max_size, shrinking);
|
|
|
|
for (i = 0; i < runs; i++) {
|
|
s64 time;
|
|
|
|
pr_info("Test %02d:\n", i);
|
|
memset(objs, 0, test_rht_params.max_size * sizeof(struct test_obj));
|
|
|
|
err = rhashtable_init(&ht, &test_rht_params);
|
|
if (err < 0) {
|
|
pr_warn("Test failed: Unable to initialize hashtable: %d\n",
|
|
err);
|
|
continue;
|
|
}
|
|
|
|
time = test_rhashtable(&ht, objs, entries);
|
|
rhashtable_destroy(&ht);
|
|
if (time < 0) {
|
|
vfree(objs);
|
|
pr_warn("Test failed: return code %lld\n", time);
|
|
return -EINVAL;
|
|
}
|
|
|
|
total_time += time;
|
|
}
|
|
|
|
pr_info("test if its possible to exceed max_size %d: %s\n",
|
|
test_rht_params.max_size, test_rhashtable_max(objs, entries) == 0 ?
|
|
"no, ok" : "YES, failed");
|
|
vfree(objs);
|
|
|
|
do_div(total_time, runs);
|
|
pr_info("Average test time: %llu\n", total_time);
|
|
|
|
test_insert_duplicates_run();
|
|
|
|
if (!tcount)
|
|
return 0;
|
|
|
|
pr_info("Testing concurrent rhashtable access from %d threads\n",
|
|
tcount);
|
|
atomic_set(&startup_count, tcount);
|
|
tdata = vzalloc(array_size(tcount, sizeof(struct thread_data)));
|
|
if (!tdata)
|
|
return -ENOMEM;
|
|
objs = vzalloc(array3_size(sizeof(struct test_obj), tcount, entries));
|
|
if (!objs) {
|
|
vfree(tdata);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
test_rht_params.max_size = max_size ? :
|
|
roundup_pow_of_two(tcount * entries);
|
|
err = rhashtable_init(&ht, &test_rht_params);
|
|
if (err < 0) {
|
|
pr_warn("Test failed: Unable to initialize hashtable: %d\n",
|
|
err);
|
|
vfree(tdata);
|
|
vfree(objs);
|
|
return -EINVAL;
|
|
}
|
|
for (i = 0; i < tcount; i++) {
|
|
tdata[i].id = i;
|
|
tdata[i].entries = entries;
|
|
tdata[i].objs = objs + i * entries;
|
|
tdata[i].task = kthread_run(threadfunc, &tdata[i],
|
|
"rhashtable_thrad[%d]", i);
|
|
if (IS_ERR(tdata[i].task)) {
|
|
pr_err(" kthread_run failed for thread %d\n", i);
|
|
atomic_dec(&startup_count);
|
|
} else {
|
|
started_threads++;
|
|
}
|
|
}
|
|
if (wait_event_interruptible(startup_wait, atomic_read(&startup_count) == 0))
|
|
pr_err(" wait_event interruptible failed\n");
|
|
/* count is 0 now, set it to -1 and wake up all threads together */
|
|
atomic_dec(&startup_count);
|
|
wake_up_all(&startup_wait);
|
|
for (i = 0; i < tcount; i++) {
|
|
if (IS_ERR(tdata[i].task))
|
|
continue;
|
|
if ((err = kthread_stop(tdata[i].task))) {
|
|
pr_warn("Test failed: thread %d returned: %d\n",
|
|
i, err);
|
|
failed_threads++;
|
|
}
|
|
}
|
|
rhashtable_destroy(&ht);
|
|
vfree(tdata);
|
|
vfree(objs);
|
|
|
|
/*
|
|
* rhltable_remove is very expensive, default values can cause test
|
|
* to run for 2 minutes or more, use a smaller number instead.
|
|
*/
|
|
err = test_rhltable(entries / 16);
|
|
pr_info("Started %d threads, %d failed, rhltable test returns %d\n",
|
|
started_threads, failed_threads, err);
|
|
return 0;
|
|
}
|
|
|
|
static void __exit test_rht_exit(void)
|
|
{
|
|
}
|
|
|
|
module_init(test_rht_init);
|
|
module_exit(test_rht_exit);
|
|
|
|
MODULE_LICENSE("GPL v2");
|