.. | .. |
---|
7 | 7 | * |
---|
8 | 8 | * Some code borrowed from https://github.com/xairy/kasan-prototype by |
---|
9 | 9 | * Andrey Konovalov <andreyknvl@gmail.com> |
---|
10 | | - * |
---|
11 | | - * This program is free software; you can redistribute it and/or modify |
---|
12 | | - * it under the terms of the GNU General Public License version 2 as |
---|
13 | | - * published by the Free Software Foundation. |
---|
14 | | - * |
---|
15 | 10 | */ |
---|
16 | | - |
---|
17 | | -#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt |
---|
18 | | -#define DISABLE_BRANCH_PROFILING |
---|
19 | 11 | |
---|
20 | 12 | #include <linux/export.h> |
---|
21 | 13 | #include <linux/interrupt.h> |
---|
22 | 14 | #include <linux/init.h> |
---|
23 | 15 | #include <linux/kasan.h> |
---|
24 | 16 | #include <linux/kernel.h> |
---|
| 17 | +#include <linux/kfence.h> |
---|
25 | 18 | #include <linux/kmemleak.h> |
---|
26 | 19 | #include <linux/linkage.h> |
---|
27 | 20 | #include <linux/memblock.h> |
---|
.. | .. |
---|
52 | 45 | s8 shadow_value = *(s8 *)kasan_mem_to_shadow((void *)addr); |
---|
53 | 46 | |
---|
54 | 47 | if (unlikely(shadow_value)) { |
---|
55 | | - s8 last_accessible_byte = addr & KASAN_SHADOW_MASK; |
---|
| 48 | + s8 last_accessible_byte = addr & KASAN_GRANULE_MASK; |
---|
56 | 49 | return unlikely(last_accessible_byte >= shadow_value); |
---|
57 | 50 | } |
---|
58 | 51 | |
---|
.. | .. |
---|
68 | 61 | * Access crosses 8(shadow size)-byte boundary. Such access maps |
---|
69 | 62 | * into 2 shadow bytes, so we need to check them both. |
---|
70 | 63 | */ |
---|
71 | | - if (unlikely(((addr + size - 1) & KASAN_SHADOW_MASK) < size - 1)) |
---|
| 64 | + if (unlikely(((addr + size - 1) & KASAN_GRANULE_MASK) < size - 1)) |
---|
72 | 65 | return *shadow_addr || memory_is_poisoned_1(addr + size - 1); |
---|
73 | 66 | |
---|
74 | 67 | return memory_is_poisoned_1(addr + size - 1); |
---|
.. | .. |
---|
79 | 72 | u16 *shadow_addr = (u16 *)kasan_mem_to_shadow((void *)addr); |
---|
80 | 73 | |
---|
81 | 74 | /* Unaligned 16-bytes access maps into 3 shadow bytes. */ |
---|
82 | | - if (unlikely(!IS_ALIGNED(addr, KASAN_SHADOW_SCALE_SIZE))) |
---|
| 75 | + if (unlikely(!IS_ALIGNED(addr, KASAN_GRANULE_SIZE))) |
---|
83 | 76 | return *shadow_addr || memory_is_poisoned_1(addr + 15); |
---|
84 | 77 | |
---|
85 | 78 | return *shadow_addr; |
---|
.. | .. |
---|
140 | 133 | s8 *last_shadow = (s8 *)kasan_mem_to_shadow((void *)last_byte); |
---|
141 | 134 | |
---|
142 | 135 | if (unlikely(ret != (unsigned long)last_shadow || |
---|
143 | | - ((long)(last_byte & KASAN_SHADOW_MASK) >= *last_shadow))) |
---|
| 136 | + ((long)(last_byte & KASAN_GRANULE_MASK) >= *last_shadow))) |
---|
144 | 137 | return true; |
---|
145 | 138 | } |
---|
146 | 139 | return false; |
---|
.. | .. |
---|
166 | 159 | return memory_is_poisoned_n(addr, size); |
---|
167 | 160 | } |
---|
168 | 161 | |
---|
169 | | -static __always_inline void check_memory_region_inline(unsigned long addr, |
---|
| 162 | +static __always_inline bool check_region_inline(unsigned long addr, |
---|
170 | 163 | size_t size, bool write, |
---|
171 | 164 | unsigned long ret_ip) |
---|
172 | 165 | { |
---|
173 | 166 | if (unlikely(size == 0)) |
---|
174 | | - return; |
---|
| 167 | + return true; |
---|
| 168 | + |
---|
| 169 | + if (unlikely(addr + size < addr)) |
---|
| 170 | + return !kasan_report(addr, size, write, ret_ip); |
---|
175 | 171 | |
---|
176 | 172 | if (unlikely((void *)addr < |
---|
177 | 173 | kasan_shadow_to_mem((void *)KASAN_SHADOW_START))) { |
---|
178 | | - kasan_report(addr, size, write, ret_ip); |
---|
179 | | - return; |
---|
| 174 | + return !kasan_report(addr, size, write, ret_ip); |
---|
180 | 175 | } |
---|
181 | 176 | |
---|
182 | 177 | if (likely(!memory_is_poisoned(addr, size))) |
---|
183 | | - return; |
---|
| 178 | + return true; |
---|
184 | 179 | |
---|
185 | | - kasan_report(addr, size, write, ret_ip); |
---|
| 180 | + return !kasan_report(addr, size, write, ret_ip); |
---|
186 | 181 | } |
---|
187 | 182 | |
---|
188 | | -void check_memory_region(unsigned long addr, size_t size, bool write, |
---|
189 | | - unsigned long ret_ip) |
---|
| 183 | +bool kasan_check_range(unsigned long addr, size_t size, bool write, |
---|
| 184 | + unsigned long ret_ip) |
---|
190 | 185 | { |
---|
191 | | - check_memory_region_inline(addr, size, write, ret_ip); |
---|
| 186 | + return check_region_inline(addr, size, write, ret_ip); |
---|
| 187 | +} |
---|
| 188 | + |
---|
| 189 | +bool kasan_byte_accessible(const void *addr) |
---|
| 190 | +{ |
---|
| 191 | + s8 shadow_byte = READ_ONCE(*(s8 *)kasan_mem_to_shadow(addr)); |
---|
| 192 | + |
---|
| 193 | + return shadow_byte >= 0 && shadow_byte < KASAN_GRANULE_SIZE; |
---|
192 | 194 | } |
---|
193 | 195 | |
---|
194 | 196 | void kasan_cache_shrink(struct kmem_cache *cache) |
---|
195 | 197 | { |
---|
196 | | - quarantine_remove_cache(cache); |
---|
| 198 | + kasan_quarantine_remove_cache(cache); |
---|
197 | 199 | } |
---|
198 | 200 | |
---|
199 | 201 | void kasan_cache_shutdown(struct kmem_cache *cache) |
---|
200 | 202 | { |
---|
201 | 203 | if (!__kmem_cache_empty(cache)) |
---|
202 | | - quarantine_remove_cache(cache); |
---|
| 204 | + kasan_quarantine_remove_cache(cache); |
---|
203 | 205 | } |
---|
204 | 206 | |
---|
205 | 207 | static void register_global(struct kasan_global *global) |
---|
206 | 208 | { |
---|
207 | | - size_t aligned_size = round_up(global->size, KASAN_SHADOW_SCALE_SIZE); |
---|
| 209 | + size_t aligned_size = round_up(global->size, KASAN_GRANULE_SIZE); |
---|
208 | 210 | |
---|
209 | | - kasan_unpoison_shadow(global->beg, global->size); |
---|
| 211 | + kasan_unpoison(global->beg, global->size, false); |
---|
210 | 212 | |
---|
211 | | - kasan_poison_shadow(global->beg + aligned_size, |
---|
212 | | - global->size_with_redzone - aligned_size, |
---|
213 | | - KASAN_GLOBAL_REDZONE); |
---|
| 213 | + kasan_poison(global->beg + aligned_size, |
---|
| 214 | + global->size_with_redzone - aligned_size, |
---|
| 215 | + KASAN_GLOBAL_REDZONE, false); |
---|
214 | 216 | } |
---|
215 | 217 | |
---|
216 | 218 | void __asan_register_globals(struct kasan_global *globals, size_t size) |
---|
.. | .. |
---|
230 | 232 | #define DEFINE_ASAN_LOAD_STORE(size) \ |
---|
231 | 233 | void __asan_load##size(unsigned long addr) \ |
---|
232 | 234 | { \ |
---|
233 | | - check_memory_region_inline(addr, size, false, _RET_IP_);\ |
---|
| 235 | + check_region_inline(addr, size, false, _RET_IP_); \ |
---|
234 | 236 | } \ |
---|
235 | 237 | EXPORT_SYMBOL(__asan_load##size); \ |
---|
236 | 238 | __alias(__asan_load##size) \ |
---|
.. | .. |
---|
238 | 240 | EXPORT_SYMBOL(__asan_load##size##_noabort); \ |
---|
239 | 241 | void __asan_store##size(unsigned long addr) \ |
---|
240 | 242 | { \ |
---|
241 | | - check_memory_region_inline(addr, size, true, _RET_IP_); \ |
---|
| 243 | + check_region_inline(addr, size, true, _RET_IP_); \ |
---|
242 | 244 | } \ |
---|
243 | 245 | EXPORT_SYMBOL(__asan_store##size); \ |
---|
244 | 246 | __alias(__asan_store##size) \ |
---|
.. | .. |
---|
253 | 255 | |
---|
254 | 256 | void __asan_loadN(unsigned long addr, size_t size) |
---|
255 | 257 | { |
---|
256 | | - check_memory_region(addr, size, false, _RET_IP_); |
---|
| 258 | + kasan_check_range(addr, size, false, _RET_IP_); |
---|
257 | 259 | } |
---|
258 | 260 | EXPORT_SYMBOL(__asan_loadN); |
---|
259 | 261 | |
---|
.. | .. |
---|
263 | 265 | |
---|
264 | 266 | void __asan_storeN(unsigned long addr, size_t size) |
---|
265 | 267 | { |
---|
266 | | - check_memory_region(addr, size, true, _RET_IP_); |
---|
| 268 | + kasan_check_range(addr, size, true, _RET_IP_); |
---|
267 | 269 | } |
---|
268 | 270 | EXPORT_SYMBOL(__asan_storeN); |
---|
269 | 271 | |
---|
.. | .. |
---|
278 | 280 | /* Emitted by compiler to poison alloca()ed objects. */ |
---|
279 | 281 | void __asan_alloca_poison(unsigned long addr, size_t size) |
---|
280 | 282 | { |
---|
281 | | - size_t rounded_up_size = round_up(size, KASAN_SHADOW_SCALE_SIZE); |
---|
| 283 | + size_t rounded_up_size = round_up(size, KASAN_GRANULE_SIZE); |
---|
282 | 284 | size_t padding_size = round_up(size, KASAN_ALLOCA_REDZONE_SIZE) - |
---|
283 | 285 | rounded_up_size; |
---|
284 | | - size_t rounded_down_size = round_down(size, KASAN_SHADOW_SCALE_SIZE); |
---|
| 286 | + size_t rounded_down_size = round_down(size, KASAN_GRANULE_SIZE); |
---|
285 | 287 | |
---|
286 | 288 | const void *left_redzone = (const void *)(addr - |
---|
287 | 289 | KASAN_ALLOCA_REDZONE_SIZE); |
---|
.. | .. |
---|
289 | 291 | |
---|
290 | 292 | WARN_ON(!IS_ALIGNED(addr, KASAN_ALLOCA_REDZONE_SIZE)); |
---|
291 | 293 | |
---|
292 | | - kasan_unpoison_shadow((const void *)(addr + rounded_down_size), |
---|
293 | | - size - rounded_down_size); |
---|
294 | | - kasan_poison_shadow(left_redzone, KASAN_ALLOCA_REDZONE_SIZE, |
---|
295 | | - KASAN_ALLOCA_LEFT); |
---|
296 | | - kasan_poison_shadow(right_redzone, |
---|
297 | | - padding_size + KASAN_ALLOCA_REDZONE_SIZE, |
---|
298 | | - KASAN_ALLOCA_RIGHT); |
---|
| 294 | + kasan_unpoison((const void *)(addr + rounded_down_size), |
---|
| 295 | + size - rounded_down_size, false); |
---|
| 296 | + kasan_poison(left_redzone, KASAN_ALLOCA_REDZONE_SIZE, |
---|
| 297 | + KASAN_ALLOCA_LEFT, false); |
---|
| 298 | + kasan_poison(right_redzone, padding_size + KASAN_ALLOCA_REDZONE_SIZE, |
---|
| 299 | + KASAN_ALLOCA_RIGHT, false); |
---|
299 | 300 | } |
---|
300 | 301 | EXPORT_SYMBOL(__asan_alloca_poison); |
---|
301 | 302 | |
---|
.. | .. |
---|
305 | 306 | if (unlikely(!stack_top || stack_top > stack_bottom)) |
---|
306 | 307 | return; |
---|
307 | 308 | |
---|
308 | | - kasan_unpoison_shadow(stack_top, stack_bottom - stack_top); |
---|
| 309 | + kasan_unpoison(stack_top, stack_bottom - stack_top, false); |
---|
309 | 310 | } |
---|
310 | 311 | EXPORT_SYMBOL(__asan_allocas_unpoison); |
---|
311 | 312 | |
---|
.. | .. |
---|
323 | 324 | DEFINE_ASAN_SET_SHADOW(f3); |
---|
324 | 325 | DEFINE_ASAN_SET_SHADOW(f5); |
---|
325 | 326 | DEFINE_ASAN_SET_SHADOW(f8); |
---|
| 327 | + |
---|
| 328 | +void kasan_record_aux_stack(void *addr) |
---|
| 329 | +{ |
---|
| 330 | + struct page *page = kasan_addr_to_page(addr); |
---|
| 331 | + struct kmem_cache *cache; |
---|
| 332 | + struct kasan_alloc_meta *alloc_meta; |
---|
| 333 | + void *object; |
---|
| 334 | + |
---|
| 335 | + if (is_kfence_address(addr) || !(page && PageSlab(page))) |
---|
| 336 | + return; |
---|
| 337 | + |
---|
| 338 | + cache = page->slab_cache; |
---|
| 339 | + object = nearest_obj(cache, page, addr); |
---|
| 340 | + alloc_meta = kasan_get_alloc_meta(cache, object); |
---|
| 341 | + if (!alloc_meta) |
---|
| 342 | + return; |
---|
| 343 | + |
---|
| 344 | + alloc_meta->aux_stack[1] = alloc_meta->aux_stack[0]; |
---|
| 345 | + alloc_meta->aux_stack[0] = kasan_save_stack(GFP_NOWAIT); |
---|
| 346 | +} |
---|
| 347 | + |
---|
| 348 | +void kasan_set_free_info(struct kmem_cache *cache, |
---|
| 349 | + void *object, u8 tag) |
---|
| 350 | +{ |
---|
| 351 | + struct kasan_free_meta *free_meta; |
---|
| 352 | + |
---|
| 353 | + free_meta = kasan_get_free_meta(cache, object); |
---|
| 354 | + if (!free_meta) |
---|
| 355 | + return; |
---|
| 356 | + |
---|
| 357 | + kasan_set_track(&free_meta->free_track, GFP_NOWAIT); |
---|
| 358 | + /* The object was freed and has free track set. */ |
---|
| 359 | + *(u8 *)kasan_mem_to_shadow(object) = KASAN_KMALLOC_FREETRACK; |
---|
| 360 | +} |
---|
| 361 | + |
---|
| 362 | +struct kasan_track *kasan_get_free_track(struct kmem_cache *cache, |
---|
| 363 | + void *object, u8 tag) |
---|
| 364 | +{ |
---|
| 365 | + if (*(u8 *)kasan_mem_to_shadow(object) != KASAN_KMALLOC_FREETRACK) |
---|
| 366 | + return NULL; |
---|
| 367 | + /* Free meta must be present with KASAN_KMALLOC_FREETRACK. */ |
---|
| 368 | + return &kasan_get_free_meta(cache, object)->free_track; |
---|
| 369 | +} |
---|