[PATCH 1/3] random: unify early init crng load accounting

From: Jason A. Donenfeld
Date: Sat Feb 12 2022 - 18:10:42 EST


crng_fast_load() and crng_slow_load() have different semantics:

- crng_fast_load() xors and accounts with crng_init_cnt.
- crng_slow_load() hashes and doesn't account.

However add_hwgenerator_randomness() can afford to hash (it's called
from a kthread), and it should account. Additionally, ones that can
afford to hash don't need to take a trylock but can take a normal lock.
So, we combine these into one function, crng_pre_init_inject(), which
allows us to control these in a uniform way. This will make it simpler
later to simplify this all down when the time comes for that.

Cc: Dominik Brodowski <linux@xxxxxxxxxxxxxxxxxxxx>
Cc: Theodore Ts'o <tytso@xxxxxxx>
Signed-off-by: Jason A. Donenfeld <Jason@xxxxxxxxx>
---
drivers/char/random.c | 107 ++++++++++++++++++++++--------------------
1 file changed, 55 insertions(+), 52 deletions(-)

diff --git a/drivers/char/random.c b/drivers/char/random.c
index 8088348190e6..a128bb947bd4 100644
--- a/drivers/char/random.c
+++ b/drivers/char/random.c
@@ -423,72 +423,74 @@ static void crng_make_state(u32 chacha_state[CHACHA_STATE_WORDS],
}

/*
- * This function is for crng_init < 2 only.
+ * This function is for crng_init < 2 only. It loads entropy directly
+ * into the crng's key, without going through the input pool. It is,
+ * generally speaking, not very safe, but we use this only at early
+ * boot time when it's better to have something there rather than
+ * nothing.
*
- * crng_fast_load() can be called by code in the interrupt service
- * path. So we can't afford to dilly-dally. Returns the number of
- * bytes processed from cp.
+ * There are two paths, a slow one and a fast one. The slow one
+ * hashes the input along with the current key. The fast one simply
+ * xors it in, and should only be used from interrupt context.
+ *
+ * If account is set, then the crng_init_cnt counter is incremented.
+ * This shouldn't be set by functions like add_device_randomness(),
+ * where we can't trust the buffer passed to it is guaranteed to be
+ * unpredictable (so it might not have any entropy at all).
+ *
+ * Returns the number of bytes processed from cp, which is bounded by
+ * CRNG_INIT_CNT_THRESH if account is true.
*/
-static size_t crng_fast_load(const void *cp, size_t len)
+static size_t crng_pre_init_inject(const void *cp, size_t len,
+ bool fast, bool account)
{
static int crng_init_cnt = 0;
unsigned long flags;
- u8 *src = (u8 *)cp;
- size_t ret = 0;
+ const u8 *src = cp;
+
+ if (fast) {
+ if (!spin_trylock_irqsave(&base_crng.lock, flags))
+ return 0;
+ } else
+ spin_lock_irqsave(&base_crng.lock, flags);

- if (!spin_trylock_irqsave(&base_crng.lock, flags))
- return 0;
if (crng_init != 0) {
spin_unlock_irqrestore(&base_crng.lock, flags);
return 0;
}
- while (len > 0 && crng_init_cnt < CRNG_INIT_CNT_THRESH) {
- base_crng.key[crng_init_cnt % sizeof(base_crng.key)] ^= *src;
- src++; crng_init_cnt++; len--; ret++;
- }
- if (crng_init_cnt >= CRNG_INIT_CNT_THRESH) {
- ++base_crng.generation;
- crng_init = 1;
- }
- spin_unlock_irqrestore(&base_crng.lock, flags);
- if (crng_init == 1)
- pr_notice("fast init done\n");
- return ret;
-}

-/*
- * This function is for crng_init < 2 only.
- *
- * crng_slow_load() is called by add_device_randomness, which has two
- * attributes. (1) We can't trust the buffer passed to it is
- * guaranteed to be unpredictable (so it might not have any entropy at
- * all), and (2) it doesn't have the performance constraints of
- * crng_fast_load().
- *
- * So, we simply hash the contents in with the current key. Finally,
- * we do *not* advance crng_init_cnt since buffer we may get may be
- * something like a fixed DMI table (for example), which might very
- * well be unique to the machine, but is otherwise unvarying.
- */
-static void crng_slow_load(const void *cp, size_t len)
-{
- unsigned long flags;
- struct blake2s_state hash;
+ if (account)
+ len = min_t(size_t, len, CRNG_INIT_CNT_THRESH - crng_init_cnt);

- blake2s_init(&hash, sizeof(base_crng.key));
+ if (fast) {
+ size_t i;

- if (!spin_trylock_irqsave(&base_crng.lock, flags))
- return;
- if (crng_init != 0) {
- spin_unlock_irqrestore(&base_crng.lock, flags);
- return;
+ for (i = 0; i < len; ++i)
+ base_crng.key[(crng_init_cnt + i) %
+ sizeof(base_crng.key)] ^= *src++;
+ } else {
+ struct blake2s_state hash;
+
+ blake2s_init(&hash, sizeof(base_crng.key));
+ blake2s_update(&hash, base_crng.key, sizeof(base_crng.key));
+ blake2s_update(&hash, cp, len);
+ blake2s_final(&hash, base_crng.key);
}

- blake2s_update(&hash, base_crng.key, sizeof(base_crng.key));
- blake2s_update(&hash, cp, len);
- blake2s_final(&hash, base_crng.key);
+ if (account) {
+ crng_init_cnt += len;
+ if (crng_init_cnt >= CRNG_INIT_CNT_THRESH) {
+ ++base_crng.generation;
+ crng_init = 1;
+ }
+ }

spin_unlock_irqrestore(&base_crng.lock, flags);
+
+ if (crng_init == 1)
+ pr_notice("fast init done\n");
+
+ return len;
}

static void _get_random_bytes(void *buf, size_t nbytes)
@@ -1006,7 +1008,7 @@ void add_device_randomness(const void *buf, size_t size)
unsigned long flags;

if (!crng_ready() && size)
- crng_slow_load(buf, size);
+ crng_pre_init_inject(buf, size, false, false);

spin_lock_irqsave(&input_pool.lock, flags);
_mix_pool_bytes(buf, size);
@@ -1123,7 +1125,7 @@ void add_hwgenerator_randomness(const void *buffer, size_t count,
size_t entropy)
{
if (unlikely(crng_init == 0)) {
- size_t ret = crng_fast_load(buffer, count);
+ size_t ret = crng_pre_init_inject(buffer, count, false, true);
mix_pool_bytes(buffer, ret);
count -= ret;
buffer += ret;
@@ -1279,7 +1281,8 @@ void add_interrupt_randomness(int irq)

if (unlikely(crng_init == 0)) {
if (new_count >= 64 &&
- crng_fast_load(fast_pool->pool, sizeof(fast_pool->pool)) > 0) {
+ crng_pre_init_inject(fast_pool->pool, sizeof(fast_pool->pool),
+ true, true) > 0) {
atomic_set(&fast_pool->count, 0);
fast_pool->last = now;

--
2.35.0