From 61598093bbdd283a7edc367d900f223070ead8d2 Mon Sep 17 00:00:00 2001
From: hc <hc@nodka.com>
Date: Fri, 10 May 2024 07:43:03 +0000
Subject: [PATCH] add ax88772C AX88772C_eeprom_tools

---
 kernel/kernel/scs.c |  210 ++++++++++++++-------------------------------------
 1 files changed, 59 insertions(+), 151 deletions(-)

diff --git a/kernel/kernel/scs.c b/kernel/kernel/scs.c
index ad74d13..579841b 100644
--- a/kernel/kernel/scs.c
+++ b/kernel/kernel/scs.c
@@ -8,46 +8,23 @@
 #include <linux/cpuhotplug.h>
 #include <linux/kasan.h>
 #include <linux/mm.h>
-#include <linux/mmzone.h>
 #include <linux/scs.h>
-#include <linux/slab.h>
 #include <linux/vmalloc.h>
 #include <linux/vmstat.h>
-#include <asm/scs.h>
 
-static inline void *__scs_base(struct task_struct *tsk)
+static void __scs_account(void *s, int account)
 {
-	/*
-	 * To minimize risk the of exposure, architectures may clear a
-	 * task's thread_info::shadow_call_stack while that task is
-	 * running, and only save/restore the active shadow call stack
-	 * pointer when the usual register may be clobbered (e.g. across
-	 * context switches).
-	 *
-	 * The shadow call stack is aligned to SCS_SIZE, and grows
-	 * upwards, so we can mask out the low bits to extract the base
-	 * when the task is not running.
-	 */
-	return (void *)((unsigned long)task_scs(tsk) & ~(SCS_SIZE - 1));
-}
+	struct page *scs_page = vmalloc_to_page(s);
 
-static inline unsigned long *scs_magic(void *s)
-{
-	return (unsigned long *)(s + SCS_SIZE) - 1;
+	mod_node_page_state(page_pgdat(scs_page), NR_KERNEL_SCS_KB,
+			    account * (SCS_SIZE / SZ_1K));
 }
-
-static inline void scs_set_magic(void *s)
-{
-	*scs_magic(s) = SCS_END_MAGIC;
-}
-
-#ifdef CONFIG_SHADOW_CALL_STACK_VMAP
 
 /* Matches NR_CACHED_STACKS for VMAP_STACK */
 #define NR_CACHED_SCS 2
 static DEFINE_PER_CPU(void *, scs_cache[NR_CACHED_SCS]);
 
-static void *scs_alloc(int node)
+static void *__scs_alloc(int node)
 {
 	int i;
 	void *s;
@@ -55,44 +32,54 @@
 	for (i = 0; i < NR_CACHED_SCS; i++) {
 		s = this_cpu_xchg(scs_cache[i], NULL);
 		if (s) {
+			kasan_unpoison_vmalloc(s, SCS_SIZE);
 			memset(s, 0, SCS_SIZE);
-			goto out;
+			return s;
 		}
 	}
 
+	return __vmalloc_node_range(SCS_SIZE, 1, VMALLOC_START, VMALLOC_END,
+				    GFP_SCS, PAGE_KERNEL, 0, node,
+				    __builtin_return_address(0));
+}
+
+void *scs_alloc(int node)
+{
+	void *s;
+
+	s = __scs_alloc(node);
+	if (!s)
+		return NULL;
+
+	*__scs_magic(s) = SCS_END_MAGIC;
+
 	/*
-	 * We allocate a full page for the shadow stack, which should be
-	 * more than we need. Check the assumption nevertheless.
+	 * Poison the allocation to catch unintentional accesses to
+	 * the shadow stack when KASAN is enabled.
 	 */
-	BUILD_BUG_ON(SCS_SIZE > PAGE_SIZE);
-
-	s = __vmalloc_node_range(PAGE_SIZE, SCS_SIZE,
-				 VMALLOC_START, VMALLOC_END,
-				 GFP_SCS, PAGE_KERNEL, 0,
-				 node, __builtin_return_address(0));
-
-out:
-	if (s)
-		scs_set_magic(s);
-	/* TODO: poison for KASAN, unpoison in scs_free */
-
+	kasan_poison_vmalloc(s, SCS_SIZE);
+	__scs_account(s, 1);
 	return s;
 }
 
-static void scs_free(void *s)
+void scs_free(void *s)
 {
 	int i;
+
+	__scs_account(s, -1);
+
+	/*
+	 * We cannot sleep as this can be called in interrupt context,
+	 * so use this_cpu_cmpxchg to update the cache, and vfree_atomic
+	 * to free the stack.
+	 */
 
 	for (i = 0; i < NR_CACHED_SCS; i++)
 		if (this_cpu_cmpxchg(scs_cache[i], 0, s) == NULL)
 			return;
 
+	kasan_unpoison_vmalloc(s, SCS_SIZE);
 	vfree_atomic(s);
-}
-
-static struct page *__scs_page(struct task_struct *tsk)
-{
-	return vmalloc_to_page(__scs_base(tsk));
 }
 
 static int scs_cleanup(unsigned int cpu)
@@ -110,137 +97,58 @@
 
 void __init scs_init(void)
 {
-	WARN_ON(cpuhp_setup_state(CPUHP_BP_PREPARE_DYN, "scs:scs_cache", NULL,
-			scs_cleanup) < 0);
-}
-
-#else /* !CONFIG_SHADOW_CALL_STACK_VMAP */
-
-static struct kmem_cache *scs_cache;
-
-static inline void *scs_alloc(int node)
-{
-	void *s;
-
-	s = kmem_cache_alloc_node(scs_cache, GFP_SCS, node);
-	if (s) {
-		scs_set_magic(s);
-		/*
-		 * Poison the allocation to catch unintentional accesses to
-		 * the shadow stack when KASAN is enabled.
-		 */
-		kasan_poison_object_data(scs_cache, s);
-	}
-
-	return s;
-}
-
-static inline void scs_free(void *s)
-{
-	kasan_unpoison_object_data(scs_cache, s);
-	kmem_cache_free(scs_cache, s);
-}
-
-static struct page *__scs_page(struct task_struct *tsk)
-{
-	return virt_to_page(__scs_base(tsk));
-}
-
-void __init scs_init(void)
-{
-	scs_cache = kmem_cache_create("scs_cache", SCS_SIZE, SCS_SIZE,
-				0, NULL);
-	WARN_ON(!scs_cache);
-}
-
-#endif /* CONFIG_SHADOW_CALL_STACK_VMAP */
-
-void scs_task_reset(struct task_struct *tsk)
-{
-	/*
-	 * Reset the shadow stack to the base address in case the task
-	 * is reused.
-	 */
-	task_set_scs(tsk, __scs_base(tsk));
-}
-
-static void scs_account(struct task_struct *tsk, int account)
-{
-	mod_zone_page_state(page_zone(__scs_page(tsk)), NR_KERNEL_SCS_BYTES,
-		account * SCS_SIZE);
+	cpuhp_setup_state(CPUHP_BP_PREPARE_DYN, "scs:scs_cache", NULL,
+			  scs_cleanup);
 }
 
 int scs_prepare(struct task_struct *tsk, int node)
 {
-	void *s;
+	void *s = scs_alloc(node);
 
-	s = scs_alloc(node);
 	if (!s)
 		return -ENOMEM;
 
-	task_set_scs(tsk, s);
-	scs_account(tsk, 1);
-
+	task_scs(tsk) = task_scs_sp(tsk) = s;
 	return 0;
-}
-
-#ifdef CONFIG_DEBUG_STACK_USAGE
-static inline unsigned long scs_used(struct task_struct *tsk)
-{
-	unsigned long *p = __scs_base(tsk);
-	unsigned long *end = scs_magic(p);
-	unsigned long s = (unsigned long)p;
-
-	while (p < end && READ_ONCE_NOCHECK(*p))
-		p++;
-
-	return (unsigned long)p - s;
 }
 
 static void scs_check_usage(struct task_struct *tsk)
 {
-	static DEFINE_SPINLOCK(lock);
 	static unsigned long highest;
-	unsigned long used = scs_used(tsk);
 
-	if (used <= highest)
+	unsigned long *p, prev, curr = highest, used = 0;
+
+	if (!IS_ENABLED(CONFIG_DEBUG_STACK_USAGE))
 		return;
 
-	spin_lock(&lock);
-
-	if (used > highest) {
-		pr_info("%s (%d): highest shadow stack usage: %lu bytes\n",
-			tsk->comm, task_pid_nr(tsk), used);
-		highest = used;
+	for (p = task_scs(tsk); p < __scs_magic(tsk); ++p) {
+		if (!READ_ONCE_NOCHECK(*p))
+			break;
+		used += sizeof(*p);
 	}
 
-	spin_unlock(&lock);
-}
-#else
-static inline void scs_check_usage(struct task_struct *tsk)
-{
-}
-#endif
+	while (used > curr) {
+		prev = cmpxchg_relaxed(&highest, curr, used);
 
-bool scs_corrupted(struct task_struct *tsk)
-{
-	unsigned long *magic = scs_magic(__scs_base(tsk));
+		if (prev == curr) {
+			pr_info("%s (%d): highest shadow stack usage: %lu bytes\n",
+				tsk->comm, task_pid_nr(tsk), used);
+			break;
+		}
 
-	return READ_ONCE_NOCHECK(*magic) != SCS_END_MAGIC;
+		curr = prev;
+	}
 }
 
 void scs_release(struct task_struct *tsk)
 {
-	void *s;
+	void *s = task_scs(tsk);
 
-	s = __scs_base(tsk);
 	if (!s)
 		return;
 
-	WARN_ON(scs_corrupted(tsk));
+	WARN(task_scs_end_corrupted(tsk),
+	     "corrupted shadow stack detected when freeing task\n");
 	scs_check_usage(tsk);
-
-	scs_account(tsk, -1);
-	task_set_scs(tsk, NULL);
 	scs_free(s);
 }

--
Gitblit v1.6.2