mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/herbert/cryptodev-2.6.git
synced 2026-04-18 03:23:53 -04:00
This was done entirely with mindless brute force, using
git grep -l '\<k[vmz]*alloc_objs*(.*, GFP_KERNEL)' |
xargs sed -i 's/\(alloc_objs*(.*\), GFP_KERNEL)/\1)/'
to convert the new alloc_obj() users that had a simple GFP_KERNEL
argument to just drop that argument.
Note that due to the extreme simplicity of the scripting, any slightly
more complex cases spread over multiple lines would not be triggered:
they definitely exist, but this covers the vast bulk of the cases, and
the resulting diff is also then easier to check automatically.
For the same reason the 'flex' versions will be done as a separate
conversion.
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
200 lines
4.4 KiB
C
200 lines
4.4 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Implementation of the hash table type.
|
|
*
|
|
* Author : Stephen Smalley, <stephen.smalley.work@gmail.com>
|
|
*/
|
|
|
|
#include <linux/kernel.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/errno.h>
|
|
#include "hashtab.h"
|
|
#include "security.h"
|
|
|
|
static struct kmem_cache *hashtab_node_cachep __ro_after_init;
|
|
|
|
/*
|
|
* Here we simply round the number of elements up to the nearest power of two.
|
|
* I tried also other options like rounding down or rounding to the closest
|
|
* power of two (up or down based on which is closer), but I was unable to
|
|
* find any significant difference in lookup/insert performance that would
|
|
* justify switching to a different (less intuitive) formula. It could be that
|
|
* a different formula is actually more optimal, but any future changes here
|
|
* should be supported with performance/memory usage data.
|
|
*
|
|
* The total memory used by the htable arrays (only) with Fedora policy loaded
|
|
* is approximately 163 KB at the time of writing.
|
|
*/
|
|
static u32 hashtab_compute_size(u32 nel)
|
|
{
|
|
return nel == 0 ? 0 : roundup_pow_of_two(nel);
|
|
}
|
|
|
|
int hashtab_init(struct hashtab *h, u32 nel_hint)
|
|
{
|
|
u32 size = hashtab_compute_size(nel_hint);
|
|
|
|
/* should already be zeroed, but better be safe */
|
|
h->nel = 0;
|
|
h->size = 0;
|
|
h->htable = NULL;
|
|
|
|
if (size) {
|
|
h->htable = kzalloc_objs(*h->htable, size,
|
|
GFP_KERNEL | __GFP_NOWARN);
|
|
if (!h->htable)
|
|
return -ENOMEM;
|
|
h->size = size;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
int __hashtab_insert(struct hashtab *h, struct hashtab_node **dst, void *key,
|
|
void *datum)
|
|
{
|
|
struct hashtab_node *newnode;
|
|
|
|
newnode = kmem_cache_zalloc(hashtab_node_cachep, GFP_KERNEL);
|
|
if (!newnode)
|
|
return -ENOMEM;
|
|
newnode->key = key;
|
|
newnode->datum = datum;
|
|
newnode->next = *dst;
|
|
*dst = newnode;
|
|
|
|
h->nel++;
|
|
return 0;
|
|
}
|
|
|
|
void hashtab_destroy(struct hashtab *h)
|
|
{
|
|
u32 i;
|
|
struct hashtab_node *cur, *temp;
|
|
|
|
for (i = 0; i < h->size; i++) {
|
|
cur = h->htable[i];
|
|
while (cur) {
|
|
temp = cur;
|
|
cur = cur->next;
|
|
kmem_cache_free(hashtab_node_cachep, temp);
|
|
}
|
|
h->htable[i] = NULL;
|
|
}
|
|
|
|
kfree(h->htable);
|
|
h->htable = NULL;
|
|
}
|
|
|
|
int hashtab_map(struct hashtab *h, int (*apply)(void *k, void *d, void *args),
|
|
void *args)
|
|
{
|
|
u32 i;
|
|
int ret;
|
|
struct hashtab_node *cur;
|
|
|
|
for (i = 0; i < h->size; i++) {
|
|
cur = h->htable[i];
|
|
while (cur) {
|
|
ret = apply(cur->key, cur->datum, args);
|
|
if (ret)
|
|
return ret;
|
|
cur = cur->next;
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
#ifdef CONFIG_SECURITY_SELINUX_DEBUG
|
|
void hashtab_stat(struct hashtab *h, struct hashtab_info *info)
|
|
{
|
|
u32 i, chain_len, slots_used, max_chain_len;
|
|
u64 chain2_len_sum;
|
|
struct hashtab_node *cur;
|
|
|
|
slots_used = 0;
|
|
max_chain_len = 0;
|
|
chain2_len_sum = 0;
|
|
for (i = 0; i < h->size; i++) {
|
|
cur = h->htable[i];
|
|
if (cur) {
|
|
slots_used++;
|
|
chain_len = 0;
|
|
while (cur) {
|
|
chain_len++;
|
|
cur = cur->next;
|
|
}
|
|
|
|
if (chain_len > max_chain_len)
|
|
max_chain_len = chain_len;
|
|
|
|
chain2_len_sum += (u64)chain_len * chain_len;
|
|
}
|
|
}
|
|
|
|
info->slots_used = slots_used;
|
|
info->max_chain_len = max_chain_len;
|
|
info->chain2_len_sum = chain2_len_sum;
|
|
}
|
|
#endif /* CONFIG_SECURITY_SELINUX_DEBUG */
|
|
|
|
int hashtab_duplicate(struct hashtab *new, const struct hashtab *orig,
|
|
int (*copy)(struct hashtab_node *new,
|
|
const struct hashtab_node *orig, void *args),
|
|
int (*destroy)(void *k, void *d, void *args), void *args)
|
|
{
|
|
const struct hashtab_node *orig_cur;
|
|
struct hashtab_node *cur, *tmp, *tail;
|
|
u32 i;
|
|
int rc;
|
|
|
|
memset(new, 0, sizeof(*new));
|
|
|
|
new->htable = kzalloc_objs(*new->htable, orig->size);
|
|
if (!new->htable)
|
|
return -ENOMEM;
|
|
|
|
new->size = orig->size;
|
|
|
|
for (i = 0; i < orig->size; i++) {
|
|
tail = NULL;
|
|
for (orig_cur = orig->htable[i]; orig_cur;
|
|
orig_cur = orig_cur->next) {
|
|
tmp = kmem_cache_zalloc(hashtab_node_cachep,
|
|
GFP_KERNEL);
|
|
if (!tmp)
|
|
goto error;
|
|
rc = copy(tmp, orig_cur, args);
|
|
if (rc) {
|
|
kmem_cache_free(hashtab_node_cachep, tmp);
|
|
goto error;
|
|
}
|
|
tmp->next = NULL;
|
|
if (!tail)
|
|
new->htable[i] = tmp;
|
|
else
|
|
tail->next = tmp;
|
|
tail = tmp;
|
|
new->nel++;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
|
|
error:
|
|
for (i = 0; i < new->size; i++) {
|
|
for (cur = new->htable[i]; cur; cur = tmp) {
|
|
tmp = cur->next;
|
|
destroy(cur->key, cur->datum, args);
|
|
kmem_cache_free(hashtab_node_cachep, cur);
|
|
}
|
|
}
|
|
kfree(new->htable);
|
|
memset(new, 0, sizeof(*new));
|
|
return -ENOMEM;
|
|
}
|
|
|
|
void __init hashtab_cache_init(void)
|
|
{
|
|
hashtab_node_cachep = KMEM_CACHE(hashtab_node, SLAB_PANIC);
|
|
}
|