| // SPDX-License-Identifier: GPL-2.0 |
| /* |
| * Copyright (C) 2015-2019 Jason A. Donenfeld <Jason@zx2c4.com>. All Rights Reserved. |
| */ |
| |
| #include "peerlookup.h" |
| #include "peer.h" |
| #include "noise.h" |
| |
| static struct hlist_head *pubkey_bucket(struct pubkey_hashtable *table, |
| const u8 pubkey[NOISE_PUBLIC_KEY_LEN]) |
| { |
| /* siphash gives us a secure 64bit number based on a random key. Since |
| * the bits are uniformly distributed, we can then mask off to get the |
| * bits we need. |
| */ |
| const u64 hash = siphash(pubkey, NOISE_PUBLIC_KEY_LEN, &table->key); |
| |
| return &table->hashtable[hash & (HASH_SIZE(table->hashtable) - 1)]; |
| } |
| |
| struct pubkey_hashtable *wg_pubkey_hashtable_alloc(void) |
| { |
| struct pubkey_hashtable *table = kvmalloc(sizeof(*table), GFP_KERNEL); |
| |
| if (!table) |
| return NULL; |
| |
| get_random_bytes(&table->key, sizeof(table->key)); |
| hash_init(table->hashtable); |
| mutex_init(&table->lock); |
| return table; |
| } |
| |
| void wg_pubkey_hashtable_add(struct pubkey_hashtable *table, |
| struct wg_peer *peer) |
| { |
| mutex_lock(&table->lock); |
| hlist_add_head_rcu(&peer->pubkey_hash, |
| pubkey_bucket(table, peer->handshake.remote_static)); |
| mutex_unlock(&table->lock); |
| } |
| |
| void wg_pubkey_hashtable_remove(struct pubkey_hashtable *table, |
| struct wg_peer *peer) |
| { |
| mutex_lock(&table->lock); |
| hlist_del_init_rcu(&peer->pubkey_hash); |
| mutex_unlock(&table->lock); |
| } |
| |
| /* Returns a strong reference to a peer */ |
| struct wg_peer * |
| wg_pubkey_hashtable_lookup(struct pubkey_hashtable *table, |
| const u8 pubkey[NOISE_PUBLIC_KEY_LEN]) |
| { |
| struct wg_peer *iter_peer, *peer = NULL; |
| |
| rcu_read_lock_bh(); |
| hlist_for_each_entry_rcu_bh(iter_peer, pubkey_bucket(table, pubkey), |
| pubkey_hash) { |
| if (!memcmp(pubkey, iter_peer->handshake.remote_static, |
| NOISE_PUBLIC_KEY_LEN)) { |
| peer = iter_peer; |
| break; |
| } |
| } |
| peer = wg_peer_get_maybe_zero(peer); |
| rcu_read_unlock_bh(); |
| return peer; |
| } |
| |
| static struct hlist_head *index_bucket(struct index_hashtable *table, |
| const __le32 index) |
| { |
| /* Since the indices are random and thus all bits are uniformly |
| * distributed, we can find its bucket simply by masking. |
| */ |
| return &table->hashtable[(__force u32)index & |
| (HASH_SIZE(table->hashtable) - 1)]; |
| } |
| |
| struct index_hashtable *wg_index_hashtable_alloc(void) |
| { |
| struct index_hashtable *table = kvmalloc(sizeof(*table), GFP_KERNEL); |
| |
| if (!table) |
| return NULL; |
| |
| hash_init(table->hashtable); |
| spin_lock_init(&table->lock); |
| return table; |
| } |
| |
| /* At the moment, we limit ourselves to 2^20 total peers, which generally might |
| * amount to 2^20*3 items in this hashtable. The algorithm below works by |
| * picking a random number and testing it. We can see that these limits mean we |
| * usually succeed pretty quickly: |
| * |
| * >>> def calculation(tries, size): |
| * ... return (size / 2**32)**(tries - 1) * (1 - (size / 2**32)) |
| * ... |
| * >>> calculation(1, 2**20 * 3) |
| * 0.999267578125 |
| * >>> calculation(2, 2**20 * 3) |
| * 0.0007318854331970215 |
| * >>> calculation(3, 2**20 * 3) |
| * 5.360489012673497e-07 |
| * >>> calculation(4, 2**20 * 3) |
| * 3.9261394135792216e-10 |
| * |
| * At the moment, we don't do any masking, so this algorithm isn't exactly |
| * constant time in either the random guessing or in the hash list lookup. We |
| * could require a minimum of 3 tries, which would successfully mask the |
| * guessing. this would not, however, help with the growing hash lengths, which |
| * is another thing to consider moving forward. |
| */ |
| |
| __le32 wg_index_hashtable_insert(struct index_hashtable *table, |
| struct index_hashtable_entry *entry) |
| { |
| struct index_hashtable_entry *existing_entry; |
| |
| spin_lock_bh(&table->lock); |
| hlist_del_init_rcu(&entry->index_hash); |
| spin_unlock_bh(&table->lock); |
| |
| rcu_read_lock_bh(); |
| |
| search_unused_slot: |
| /* First we try to find an unused slot, randomly, while unlocked. */ |
| entry->index = (__force __le32)get_random_u32(); |
| hlist_for_each_entry_rcu_bh(existing_entry, |
| index_bucket(table, entry->index), |
| index_hash) { |
| if (existing_entry->index == entry->index) |
| /* If it's already in use, we continue searching. */ |
| goto search_unused_slot; |
| } |
| |
| /* Once we've found an unused slot, we lock it, and then double-check |
| * that nobody else stole it from us. |
| */ |
| spin_lock_bh(&table->lock); |
| hlist_for_each_entry_rcu_bh(existing_entry, |
| index_bucket(table, entry->index), |
| index_hash) { |
| if (existing_entry->index == entry->index) { |
| spin_unlock_bh(&table->lock); |
| /* If it was stolen, we start over. */ |
| goto search_unused_slot; |
| } |
| } |
| /* Otherwise, we know we have it exclusively (since we're locked), |
| * so we insert. |
| */ |
| hlist_add_head_rcu(&entry->index_hash, |
| index_bucket(table, entry->index)); |
| spin_unlock_bh(&table->lock); |
| |
| rcu_read_unlock_bh(); |
| |
| return entry->index; |
| } |
| |
| bool wg_index_hashtable_replace(struct index_hashtable *table, |
| struct index_hashtable_entry *old, |
| struct index_hashtable_entry *new) |
| { |
| bool ret; |
| |
| spin_lock_bh(&table->lock); |
| ret = !hlist_unhashed(&old->index_hash); |
| if (unlikely(!ret)) |
| goto out; |
| |
| new->index = old->index; |
| hlist_replace_rcu(&old->index_hash, &new->index_hash); |
| |
| /* Calling init here NULLs out index_hash, and in fact after this |
| * function returns, it's theoretically possible for this to get |
| * reinserted elsewhere. That means the RCU lookup below might either |
| * terminate early or jump between buckets, in which case the packet |
| * simply gets dropped, which isn't terrible. |
| */ |
| INIT_HLIST_NODE(&old->index_hash); |
| out: |
| spin_unlock_bh(&table->lock); |
| return ret; |
| } |
| |
| void wg_index_hashtable_remove(struct index_hashtable *table, |
| struct index_hashtable_entry *entry) |
| { |
| spin_lock_bh(&table->lock); |
| hlist_del_init_rcu(&entry->index_hash); |
| spin_unlock_bh(&table->lock); |
| } |
| |
| /* Returns a strong reference to a entry->peer */ |
| struct index_hashtable_entry * |
| wg_index_hashtable_lookup(struct index_hashtable *table, |
| const enum index_hashtable_type type_mask, |
| const __le32 index, struct wg_peer **peer) |
| { |
| struct index_hashtable_entry *iter_entry, *entry = NULL; |
| |
| rcu_read_lock_bh(); |
| hlist_for_each_entry_rcu_bh(iter_entry, index_bucket(table, index), |
| index_hash) { |
| if (iter_entry->index == index) { |
| if (likely(iter_entry->type & type_mask)) |
| entry = iter_entry; |
| break; |
| } |
| } |
| if (likely(entry)) { |
| entry->peer = wg_peer_get_maybe_zero(entry->peer); |
| if (likely(entry->peer)) |
| *peer = entry->peer; |
| else |
| entry = NULL; |
| } |
| rcu_read_unlock_bh(); |
| return entry; |
| } |