1 | //===-- asan_poisoning.cpp ------------------------------------------------===// |
2 | // |
3 | // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. |
4 | // See https://llvm.org/LICENSE.txt for license information. |
5 | // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception |
6 | // |
7 | //===----------------------------------------------------------------------===// |
8 | // |
9 | // This file is a part of AddressSanitizer, an address sanity checker. |
10 | // |
11 | // Shadow memory poisoning by ASan RTL and by user application. |
12 | //===----------------------------------------------------------------------===// |
13 | |
14 | #include "asan_poisoning.h" |
15 | |
16 | #include "asan_report.h" |
17 | #include "asan_stack.h" |
18 | #include "sanitizer_common/sanitizer_atomic.h" |
19 | #include "sanitizer_common/sanitizer_common.h" |
20 | #include "sanitizer_common/sanitizer_flags.h" |
21 | #include "sanitizer_common/sanitizer_interface_internal.h" |
22 | #include "sanitizer_common/sanitizer_libc.h" |
23 | #include "sanitizer_common/sanitizer_ring_buffer.h" |
24 | #include "sanitizer_common/sanitizer_stackdepot.h" |
25 | |
26 | namespace __asan { |
27 | |
28 | using PoisonRecordRingBuffer = RingBuffer<PoisonRecord>; |
29 | |
30 | static atomic_uint8_t can_poison_memory; |
31 | |
32 | static Mutex poison_records_mutex; |
33 | static PoisonRecordRingBuffer *poison_records |
34 | SANITIZER_GUARDED_BY(poison_records_mutex) = nullptr; |
35 | |
36 | void AddPoisonRecord(const PoisonRecord &new_record) { |
37 | if (flags()->poison_history_size <= 0) |
38 | return; |
39 | |
40 | GenericScopedLock<Mutex> l(&poison_records_mutex); |
41 | |
42 | if (poison_records == nullptr) |
43 | poison_records = PoisonRecordRingBuffer::New(Size: flags()->poison_history_size); |
44 | |
45 | poison_records->push(t: new_record); |
46 | } |
47 | |
48 | bool FindPoisonRecord(uptr addr, PoisonRecord &match) { |
49 | if (flags()->poison_history_size <= 0) |
50 | return false; |
51 | |
52 | GenericScopedLock<Mutex> l(&poison_records_mutex); |
53 | |
54 | if (poison_records) { |
55 | for (unsigned int i = 0; i < poison_records->size(); i++) { |
56 | PoisonRecord record = (*poison_records)[i]; |
57 | if (record.begin <= addr && addr < record.end) { |
58 | internal_memcpy(dest: &match, src: &record, n: sizeof(record)); |
59 | return true; |
60 | } |
61 | } |
62 | } |
63 | |
64 | return false; |
65 | } |
66 | |
67 | void SANITIZER_ACQUIRE(poison_records_mutex) AcquirePoisonRecords() { |
68 | poison_records_mutex.Lock(); |
69 | } |
70 | |
71 | void SANITIZER_RELEASE(poison_records_mutex) ReleasePoisonRecords() { |
72 | poison_records_mutex.Unlock(); |
73 | } |
74 | |
75 | void SetCanPoisonMemory(bool value) { |
76 | atomic_store(a: &can_poison_memory, v: value, mo: memory_order_release); |
77 | } |
78 | |
79 | bool CanPoisonMemory() { |
80 | return atomic_load(a: &can_poison_memory, mo: memory_order_acquire); |
81 | } |
82 | |
83 | void PoisonShadow(uptr addr, uptr size, u8 value) { |
84 | if (value && !CanPoisonMemory()) return; |
85 | CHECK(AddrIsAlignedByGranularity(addr)); |
86 | CHECK(AddrIsInMem(addr)); |
87 | CHECK(AddrIsAlignedByGranularity(addr + size)); |
88 | CHECK(AddrIsInMem(addr + size - ASAN_SHADOW_GRANULARITY)); |
89 | CHECK(REAL(memset)); |
90 | FastPoisonShadow(aligned_beg: addr, aligned_size: size, value); |
91 | } |
92 | |
93 | void PoisonShadowPartialRightRedzone(uptr addr, |
94 | uptr size, |
95 | uptr redzone_size, |
96 | u8 value) { |
97 | if (!CanPoisonMemory()) return; |
98 | CHECK(AddrIsAlignedByGranularity(addr)); |
99 | CHECK(AddrIsInMem(addr)); |
100 | FastPoisonShadowPartialRightRedzone(aligned_addr: addr, size, redzone_size, value); |
101 | } |
102 | |
103 | struct ShadowSegmentEndpoint { |
104 | u8 *chunk; |
105 | s8 offset; // in [0, ASAN_SHADOW_GRANULARITY) |
106 | s8 value; // = *chunk; |
107 | |
108 | explicit ShadowSegmentEndpoint(uptr address) { |
109 | chunk = (u8*)MemToShadow(p: address); |
110 | offset = address & (ASAN_SHADOW_GRANULARITY - 1); |
111 | value = *chunk; |
112 | } |
113 | }; |
114 | |
115 | void AsanPoisonOrUnpoisonIntraObjectRedzone(uptr ptr, uptr size, bool poison) { |
116 | uptr end = ptr + size; |
117 | if (Verbosity()) { |
118 | Printf(format: "__asan_%spoison_intra_object_redzone [%p,%p) %zd\n" , |
119 | poison ? "" : "un" , (void *)ptr, (void *)end, size); |
120 | if (Verbosity() >= 2) |
121 | PRINT_CURRENT_STACK(); |
122 | } |
123 | CHECK(size); |
124 | CHECK_LE(size, 4096); |
125 | CHECK(IsAligned(end, ASAN_SHADOW_GRANULARITY)); |
126 | if (!IsAligned(a: ptr, ASAN_SHADOW_GRANULARITY)) { |
127 | *(u8 *)MemToShadow(p: ptr) = |
128 | poison ? static_cast<u8>(ptr % ASAN_SHADOW_GRANULARITY) : 0; |
129 | ptr |= ASAN_SHADOW_GRANULARITY - 1; |
130 | ptr++; |
131 | } |
132 | for (; ptr < end; ptr += ASAN_SHADOW_GRANULARITY) |
133 | *(u8*)MemToShadow(p: ptr) = poison ? kAsanIntraObjectRedzone : 0; |
134 | } |
135 | |
136 | } // namespace __asan |
137 | |
138 | // ---------------------- Interface ---------------- {{{1 |
139 | using namespace __asan; |
140 | |
141 | // Current implementation of __asan_(un)poison_memory_region doesn't check |
142 | // that user program (un)poisons the memory it owns. It poisons memory |
143 | // conservatively, and unpoisons progressively to make sure asan shadow |
144 | // mapping invariant is preserved (see detailed mapping description here: |
145 | // https://github.com/google/sanitizers/wiki/AddressSanitizerAlgorithm). |
146 | // |
147 | // * if user asks to poison region [left, right), the program poisons |
148 | // at least [left, AlignDown(right)). |
149 | // * if user asks to unpoison region [left, right), the program unpoisons |
150 | // at most [AlignDown(left), right). |
151 | void __asan_poison_memory_region(void const volatile *addr, uptr size) { |
152 | if (!flags()->allow_user_poisoning || size == 0) return; |
153 | uptr beg_addr = (uptr)addr; |
154 | uptr end_addr = beg_addr + size; |
155 | VPrintf(3, "Trying to poison memory region [%p, %p)\n" , (void *)beg_addr, |
156 | (void *)end_addr); |
157 | |
158 | if (flags()->poison_history_size > 0) { |
159 | GET_STACK_TRACE(/*max_size=*/16, /*fast=*/false); |
160 | u32 current_tid = GetCurrentTidOrInvalid(); |
161 | |
162 | u32 stack_id = StackDepotPut(stack); |
163 | |
164 | PoisonRecord record; |
165 | record.stack_id = stack_id; |
166 | record.thread_id = current_tid; |
167 | record.begin = beg_addr; |
168 | record.end = end_addr; |
169 | AddPoisonRecord(new_record: record); |
170 | } |
171 | |
172 | ShadowSegmentEndpoint beg(beg_addr); |
173 | ShadowSegmentEndpoint end(end_addr); |
174 | if (beg.chunk == end.chunk) { |
175 | CHECK_LT(beg.offset, end.offset); |
176 | s8 value = beg.value; |
177 | CHECK_EQ(value, end.value); |
178 | // We can only poison memory if the byte in end.offset is unaddressable. |
179 | // No need to re-poison memory if it is poisoned already. |
180 | if (value > 0 && value <= end.offset) { |
181 | if (beg.offset > 0) { |
182 | *beg.chunk = Min(a: value, b: beg.offset); |
183 | } else { |
184 | *beg.chunk = kAsanUserPoisonedMemoryMagic; |
185 | } |
186 | } |
187 | return; |
188 | } |
189 | CHECK_LT(beg.chunk, end.chunk); |
190 | if (beg.offset > 0) { |
191 | // Mark bytes from beg.offset as unaddressable. |
192 | if (beg.value == 0) { |
193 | *beg.chunk = beg.offset; |
194 | } else { |
195 | *beg.chunk = Min(a: beg.value, b: beg.offset); |
196 | } |
197 | beg.chunk++; |
198 | } |
199 | REAL(memset)(beg.chunk, kAsanUserPoisonedMemoryMagic, end.chunk - beg.chunk); |
200 | // Poison if byte in end.offset is unaddressable. |
201 | if (end.value > 0 && end.value <= end.offset) { |
202 | *end.chunk = kAsanUserPoisonedMemoryMagic; |
203 | } |
204 | } |
205 | |
206 | void __asan_unpoison_memory_region(void const volatile *addr, uptr size) { |
207 | if (!flags()->allow_user_poisoning || size == 0) return; |
208 | uptr beg_addr = (uptr)addr; |
209 | uptr end_addr = beg_addr + size; |
210 | VPrintf(3, "Trying to unpoison memory region [%p, %p)\n" , (void *)beg_addr, |
211 | (void *)end_addr); |
212 | |
213 | // Note: we don't need to update the poison tracking here. Since the shadow |
214 | // memory will be unpoisoned, the poison tracking ring buffer entries will be |
215 | // ignored. |
216 | |
217 | ShadowSegmentEndpoint beg(beg_addr); |
218 | ShadowSegmentEndpoint end(end_addr); |
219 | if (beg.chunk == end.chunk) { |
220 | CHECK_LT(beg.offset, end.offset); |
221 | s8 value = beg.value; |
222 | CHECK_EQ(value, end.value); |
223 | // We unpoison memory bytes up to enbytes up to end.offset if it is not |
224 | // unpoisoned already. |
225 | if (value != 0) { |
226 | *beg.chunk = Max(a: value, b: end.offset); |
227 | } |
228 | return; |
229 | } |
230 | CHECK_LT(beg.chunk, end.chunk); |
231 | REAL(memset)(beg.chunk, 0, end.chunk - beg.chunk); |
232 | if (end.offset > 0 && end.value != 0) { |
233 | *end.chunk = Max(a: end.value, b: end.offset); |
234 | } |
235 | } |
236 | |
237 | int __asan_address_is_poisoned(void const volatile *addr) { |
238 | return __asan::AddressIsPoisoned(a: (uptr)addr); |
239 | } |
240 | |
241 | uptr __asan_region_is_poisoned(uptr beg, uptr size) { |
242 | if (!size) |
243 | return 0; |
244 | uptr end = beg + size; |
245 | if (!AddrIsInMem(a: beg)) |
246 | return beg; |
247 | if (!AddrIsInMem(a: end)) |
248 | return end; |
249 | CHECK_LT(beg, end); |
250 | uptr aligned_b = RoundUpTo(size: beg, ASAN_SHADOW_GRANULARITY); |
251 | uptr aligned_e = RoundDownTo(x: end, ASAN_SHADOW_GRANULARITY); |
252 | uptr shadow_beg = MemToShadow(p: aligned_b); |
253 | uptr shadow_end = MemToShadow(p: aligned_e); |
254 | // First check the first and the last application bytes, |
255 | // then check the ASAN_SHADOW_GRANULARITY-aligned region by calling |
256 | // mem_is_zero on the corresponding shadow. |
257 | if (!__asan::AddressIsPoisoned(a: beg) && !__asan::AddressIsPoisoned(a: end - 1) && |
258 | (shadow_end <= shadow_beg || |
259 | __sanitizer::mem_is_zero(mem: (const char *)shadow_beg, |
260 | size: shadow_end - shadow_beg))) |
261 | return 0; |
262 | // The fast check failed, so we have a poisoned byte somewhere. |
263 | // Find it slowly. |
264 | for (; beg < end; beg++) |
265 | if (__asan::AddressIsPoisoned(a: beg)) |
266 | return beg; |
267 | UNREACHABLE("mem_is_zero returned false, but poisoned byte was not found" ); |
268 | return 0; |
269 | } |
270 | |
271 | #define CHECK_SMALL_REGION(p, size, isWrite) \ |
272 | do { \ |
273 | uptr __p = reinterpret_cast<uptr>(p); \ |
274 | uptr __size = size; \ |
275 | if (UNLIKELY(__asan::AddressIsPoisoned(__p) || \ |
276 | __asan::AddressIsPoisoned(__p + __size - 1))) { \ |
277 | GET_CURRENT_PC_BP_SP; \ |
278 | uptr __bad = __asan_region_is_poisoned(__p, __size); \ |
279 | __asan_report_error(pc, bp, sp, __bad, isWrite, __size, 0);\ |
280 | } \ |
281 | } while (false) |
282 | |
283 | |
284 | extern "C" SANITIZER_INTERFACE_ATTRIBUTE |
285 | u16 __sanitizer_unaligned_load16(const uu16 *p) { |
286 | CHECK_SMALL_REGION(p, sizeof(*p), false); |
287 | return *p; |
288 | } |
289 | |
290 | extern "C" SANITIZER_INTERFACE_ATTRIBUTE |
291 | u32 __sanitizer_unaligned_load32(const uu32 *p) { |
292 | CHECK_SMALL_REGION(p, sizeof(*p), false); |
293 | return *p; |
294 | } |
295 | |
296 | extern "C" SANITIZER_INTERFACE_ATTRIBUTE |
297 | u64 __sanitizer_unaligned_load64(const uu64 *p) { |
298 | CHECK_SMALL_REGION(p, sizeof(*p), false); |
299 | return *p; |
300 | } |
301 | |
302 | extern "C" SANITIZER_INTERFACE_ATTRIBUTE |
303 | void __sanitizer_unaligned_store16(uu16 *p, u16 x) { |
304 | CHECK_SMALL_REGION(p, sizeof(*p), true); |
305 | *p = x; |
306 | } |
307 | |
308 | extern "C" SANITIZER_INTERFACE_ATTRIBUTE |
309 | void __sanitizer_unaligned_store32(uu32 *p, u32 x) { |
310 | CHECK_SMALL_REGION(p, sizeof(*p), true); |
311 | *p = x; |
312 | } |
313 | |
314 | extern "C" SANITIZER_INTERFACE_ATTRIBUTE |
315 | void __sanitizer_unaligned_store64(uu64 *p, u64 x) { |
316 | CHECK_SMALL_REGION(p, sizeof(*p), true); |
317 | *p = x; |
318 | } |
319 | |
320 | extern "C" SANITIZER_INTERFACE_ATTRIBUTE |
321 | void __asan_poison_cxx_array_cookie(uptr p) { |
322 | if (SANITIZER_WORDSIZE != 64) return; |
323 | if (!flags()->poison_array_cookie) return; |
324 | uptr s = MEM_TO_SHADOW(p); |
325 | *reinterpret_cast<u8*>(s) = kAsanArrayCookieMagic; |
326 | } |
327 | |
328 | extern "C" SANITIZER_INTERFACE_ATTRIBUTE |
329 | uptr __asan_load_cxx_array_cookie(uptr *p) { |
330 | if (SANITIZER_WORDSIZE != 64) return *p; |
331 | if (!flags()->poison_array_cookie) return *p; |
332 | uptr s = MEM_TO_SHADOW(reinterpret_cast<uptr>(p)); |
333 | u8 sval = *reinterpret_cast<u8*>(s); |
334 | if (sval == kAsanArrayCookieMagic) return *p; |
335 | // If sval is not kAsanArrayCookieMagic it can only be freed memory, |
336 | // which means that we are going to get double-free. So, return 0 to avoid |
337 | // infinite loop of destructors. We don't want to report a double-free here |
338 | // though, so print a warning just in case. |
339 | // CHECK_EQ(sval, kAsanHeapFreeMagic); |
340 | if (sval == kAsanHeapFreeMagic) { |
341 | Report(format: "AddressSanitizer: loaded array cookie from free-d memory; " |
342 | "expect a double-free report\n" ); |
343 | return 0; |
344 | } |
345 | // The cookie may remain unpoisoned if e.g. it comes from a custom |
346 | // operator new defined inside a class. |
347 | return *p; |
348 | } |
349 | |
350 | // This is a simplified version of __asan_(un)poison_memory_region, which |
351 | // assumes that left border of region to be poisoned is properly aligned. |
352 | static void PoisonAlignedStackMemory(uptr addr, uptr size, bool do_poison) { |
353 | if (size == 0) return; |
354 | uptr aligned_size = size & ~(ASAN_SHADOW_GRANULARITY - 1); |
355 | PoisonShadow(addr, size: aligned_size, |
356 | value: do_poison ? kAsanStackUseAfterScopeMagic : 0); |
357 | if (size == aligned_size) |
358 | return; |
359 | s8 end_offset = (s8)(size - aligned_size); |
360 | s8* shadow_end = (s8*)MemToShadow(p: addr + aligned_size); |
361 | s8 end_value = *shadow_end; |
362 | if (do_poison) { |
363 | // If possible, mark all the bytes mapping to last shadow byte as |
364 | // unaddressable. |
365 | if (end_value > 0 && end_value <= end_offset) |
366 | *shadow_end = (s8)kAsanStackUseAfterScopeMagic; |
367 | } else { |
368 | // If necessary, mark few first bytes mapping to last shadow byte |
369 | // as addressable |
370 | if (end_value != 0) |
371 | *shadow_end = Max(a: end_value, b: end_offset); |
372 | } |
373 | } |
374 | |
375 | void __asan_set_shadow_00(uptr addr, uptr size) { |
376 | REAL(memset)((void *)addr, 0, size); |
377 | } |
378 | |
379 | void __asan_set_shadow_01(uptr addr, uptr size) { |
380 | REAL(memset)((void *)addr, 0x01, size); |
381 | } |
382 | |
383 | void __asan_set_shadow_02(uptr addr, uptr size) { |
384 | REAL(memset)((void *)addr, 0x02, size); |
385 | } |
386 | |
387 | void __asan_set_shadow_03(uptr addr, uptr size) { |
388 | REAL(memset)((void *)addr, 0x03, size); |
389 | } |
390 | |
391 | void __asan_set_shadow_04(uptr addr, uptr size) { |
392 | REAL(memset)((void *)addr, 0x04, size); |
393 | } |
394 | |
395 | void __asan_set_shadow_05(uptr addr, uptr size) { |
396 | REAL(memset)((void *)addr, 0x05, size); |
397 | } |
398 | |
399 | void __asan_set_shadow_06(uptr addr, uptr size) { |
400 | REAL(memset)((void *)addr, 0x06, size); |
401 | } |
402 | |
403 | void __asan_set_shadow_07(uptr addr, uptr size) { |
404 | REAL(memset)((void *)addr, 0x07, size); |
405 | } |
406 | |
407 | void __asan_set_shadow_f1(uptr addr, uptr size) { |
408 | REAL(memset)((void *)addr, 0xf1, size); |
409 | } |
410 | |
411 | void __asan_set_shadow_f2(uptr addr, uptr size) { |
412 | REAL(memset)((void *)addr, 0xf2, size); |
413 | } |
414 | |
415 | void __asan_set_shadow_f3(uptr addr, uptr size) { |
416 | REAL(memset)((void *)addr, 0xf3, size); |
417 | } |
418 | |
419 | void __asan_set_shadow_f5(uptr addr, uptr size) { |
420 | REAL(memset)((void *)addr, 0xf5, size); |
421 | } |
422 | |
423 | void __asan_set_shadow_f8(uptr addr, uptr size) { |
424 | REAL(memset)((void *)addr, 0xf8, size); |
425 | } |
426 | |
427 | void __asan_poison_stack_memory(uptr addr, uptr size) { |
428 | VReport(1, "poisoning: %p %zx\n" , (void *)addr, size); |
429 | PoisonAlignedStackMemory(addr, size, do_poison: true); |
430 | } |
431 | |
432 | void __asan_unpoison_stack_memory(uptr addr, uptr size) { |
433 | VReport(1, "unpoisoning: %p %zx\n" , (void *)addr, size); |
434 | PoisonAlignedStackMemory(addr, size, do_poison: false); |
435 | } |
436 | |
437 | static void FixUnalignedStorage(uptr storage_beg, uptr storage_end, |
438 | uptr &old_beg, uptr &old_end, uptr &new_beg, |
439 | uptr &new_end) { |
440 | constexpr uptr granularity = ASAN_SHADOW_GRANULARITY; |
441 | if (UNLIKELY(!AddrIsAlignedByGranularity(storage_end))) { |
442 | uptr end_down = RoundDownTo(x: storage_end, boundary: granularity); |
443 | // Ignore the last unaligned granule if the storage is followed by |
444 | // unpoisoned byte, because we can't poison the prefix anyway. Don't call |
445 | // AddressIsPoisoned at all if container changes does not affect the last |
446 | // granule at all. |
447 | if ((((old_end != new_end) && Max(a: old_end, b: new_end) > end_down) || |
448 | ((old_beg != new_beg) && Max(a: old_beg, b: new_beg) > end_down)) && |
449 | !AddressIsPoisoned(a: storage_end)) { |
450 | old_beg = Min(a: end_down, b: old_beg); |
451 | old_end = Min(a: end_down, b: old_end); |
452 | new_beg = Min(a: end_down, b: new_beg); |
453 | new_end = Min(a: end_down, b: new_end); |
454 | } |
455 | } |
456 | |
457 | // Handle misaligned begin and cut it off. |
458 | if (UNLIKELY(!AddrIsAlignedByGranularity(storage_beg))) { |
459 | uptr beg_up = RoundUpTo(size: storage_beg, boundary: granularity); |
460 | // The first unaligned granule needs special handling only if we had bytes |
461 | // there before and will have none after. |
462 | if ((new_beg == new_end || new_beg >= beg_up) && old_beg != old_end && |
463 | old_beg < beg_up) { |
464 | // Keep granule prefix outside of the storage unpoisoned. |
465 | uptr beg_down = RoundDownTo(x: storage_beg, boundary: granularity); |
466 | *(u8 *)MemToShadow(p: beg_down) = storage_beg - beg_down; |
467 | old_beg = Max(a: beg_up, b: old_beg); |
468 | old_end = Max(a: beg_up, b: old_end); |
469 | new_beg = Max(a: beg_up, b: new_beg); |
470 | new_end = Max(a: beg_up, b: new_end); |
471 | } |
472 | } |
473 | } |
474 | |
475 | void __sanitizer_annotate_contiguous_container(const void *beg_p, |
476 | const void *end_p, |
477 | const void *old_mid_p, |
478 | const void *new_mid_p) { |
479 | if (!flags()->detect_container_overflow) |
480 | return; |
481 | VPrintf(3, "contiguous_container: %p %p %p %p\n" , beg_p, end_p, old_mid_p, |
482 | new_mid_p); |
483 | uptr storage_beg = reinterpret_cast<uptr>(beg_p); |
484 | uptr storage_end = reinterpret_cast<uptr>(end_p); |
485 | uptr old_end = reinterpret_cast<uptr>(old_mid_p); |
486 | uptr new_end = reinterpret_cast<uptr>(new_mid_p); |
487 | uptr old_beg = storage_beg; |
488 | uptr new_beg = storage_beg; |
489 | uptr granularity = ASAN_SHADOW_GRANULARITY; |
490 | if (!(storage_beg <= old_end && storage_beg <= new_end && |
491 | old_end <= storage_end && new_end <= storage_end)) { |
492 | GET_STACK_TRACE_FATAL_HERE; |
493 | ReportBadParamsToAnnotateContiguousContainer(beg: storage_beg, end: storage_end, |
494 | old_mid: old_end, new_mid: new_end, stack: &stack); |
495 | } |
496 | CHECK_LE(storage_end - storage_beg, |
497 | FIRST_32_SECOND_64(1UL << 30, 1ULL << 40)); // Sanity check. |
498 | |
499 | if (old_end == new_end) |
500 | return; // Nothing to do here. |
501 | |
502 | FixUnalignedStorage(storage_beg, storage_end, old_beg, old_end, new_beg, |
503 | new_end); |
504 | |
505 | uptr a = RoundDownTo(x: Min(a: old_end, b: new_end), boundary: granularity); |
506 | uptr c = RoundUpTo(size: Max(a: old_end, b: new_end), boundary: granularity); |
507 | uptr d1 = RoundDownTo(x: old_end, boundary: granularity); |
508 | // uptr d2 = RoundUpTo(old_mid, granularity); |
509 | // Currently we should be in this state: |
510 | // [a, d1) is good, [d2, c) is bad, [d1, d2) is partially good. |
511 | // Make a quick sanity check that we are indeed in this state. |
512 | // |
513 | // FIXME: Two of these three checks are disabled until we fix |
514 | // https://github.com/google/sanitizers/issues/258. |
515 | // if (d1 != d2) |
516 | // DCHECK_EQ(*(u8*)MemToShadow(d1), old_mid - d1); |
517 | // |
518 | // NOTE: curly brackets for the "if" below to silence a MSVC warning. |
519 | if (a + granularity <= d1) { |
520 | DCHECK_EQ(*(u8 *)MemToShadow(a), 0); |
521 | } |
522 | // if (d2 + granularity <= c && c <= end) |
523 | // DCHECK_EQ(*(u8 *)MemToShadow(c - granularity), |
524 | // kAsanContiguousContainerOOBMagic); |
525 | |
526 | uptr b1 = RoundDownTo(x: new_end, boundary: granularity); |
527 | uptr b2 = RoundUpTo(size: new_end, boundary: granularity); |
528 | // New state: |
529 | // [a, b1) is good, [b2, c) is bad, [b1, b2) is partially good. |
530 | if (b1 > a) |
531 | PoisonShadow(addr: a, size: b1 - a, value: 0); |
532 | else if (c > b2) |
533 | PoisonShadow(addr: b2, size: c - b2, value: kAsanContiguousContainerOOBMagic); |
534 | if (b1 != b2) { |
535 | CHECK_EQ(b2 - b1, granularity); |
536 | *(u8 *)MemToShadow(p: b1) = static_cast<u8>(new_end - b1); |
537 | } |
538 | } |
539 | |
540 | // Annotates a double ended contiguous memory area like std::deque's chunk. |
541 | // It allows detecting buggy accesses to allocated but not used begining |
542 | // or end items of such a container. |
543 | void __sanitizer_annotate_double_ended_contiguous_container( |
544 | const void *storage_beg_p, const void *storage_end_p, |
545 | const void *old_container_beg_p, const void *old_container_end_p, |
546 | const void *new_container_beg_p, const void *new_container_end_p) { |
547 | if (!flags()->detect_container_overflow) |
548 | return; |
549 | |
550 | VPrintf(3, "contiguous_container: %p %p %p %p %p %p\n" , storage_beg_p, |
551 | storage_end_p, old_container_beg_p, old_container_end_p, |
552 | new_container_beg_p, new_container_end_p); |
553 | |
554 | uptr storage_beg = reinterpret_cast<uptr>(storage_beg_p); |
555 | uptr storage_end = reinterpret_cast<uptr>(storage_end_p); |
556 | uptr old_beg = reinterpret_cast<uptr>(old_container_beg_p); |
557 | uptr old_end = reinterpret_cast<uptr>(old_container_end_p); |
558 | uptr new_beg = reinterpret_cast<uptr>(new_container_beg_p); |
559 | uptr new_end = reinterpret_cast<uptr>(new_container_end_p); |
560 | |
561 | constexpr uptr granularity = ASAN_SHADOW_GRANULARITY; |
562 | |
563 | if (!(old_beg <= old_end && new_beg <= new_end) || |
564 | !(storage_beg <= new_beg && new_end <= storage_end) || |
565 | !(storage_beg <= old_beg && old_end <= storage_end)) { |
566 | GET_STACK_TRACE_FATAL_HERE; |
567 | ReportBadParamsToAnnotateDoubleEndedContiguousContainer( |
568 | storage_beg, storage_end, old_container_beg: old_beg, old_container_end: old_end, new_container_beg: new_beg, new_container_end: new_end, stack: &stack); |
569 | } |
570 | CHECK_LE(storage_end - storage_beg, |
571 | FIRST_32_SECOND_64(1UL << 30, 1ULL << 40)); // Sanity check. |
572 | |
573 | if ((old_beg == old_end && new_beg == new_end) || |
574 | (old_beg == new_beg && old_end == new_end)) |
575 | return; // Nothing to do here. |
576 | |
577 | FixUnalignedStorage(storage_beg, storage_end, old_beg, old_end, new_beg, |
578 | new_end); |
579 | |
580 | // Handle non-intersecting new/old containers separately have simpler |
581 | // intersecting case. |
582 | if (old_beg == old_end || new_beg == new_end || new_end <= old_beg || |
583 | old_end <= new_beg) { |
584 | if (old_beg != old_end) { |
585 | // Poisoning the old container. |
586 | uptr a = RoundDownTo(x: old_beg, boundary: granularity); |
587 | uptr b = RoundUpTo(size: old_end, boundary: granularity); |
588 | PoisonShadow(addr: a, size: b - a, value: kAsanContiguousContainerOOBMagic); |
589 | } |
590 | |
591 | if (new_beg != new_end) { |
592 | // Unpoisoning the new container. |
593 | uptr a = RoundDownTo(x: new_beg, boundary: granularity); |
594 | uptr b = RoundDownTo(x: new_end, boundary: granularity); |
595 | PoisonShadow(addr: a, size: b - a, value: 0); |
596 | if (!AddrIsAlignedByGranularity(a: new_end)) |
597 | *(u8 *)MemToShadow(p: b) = static_cast<u8>(new_end - b); |
598 | } |
599 | |
600 | return; |
601 | } |
602 | |
603 | // Intersection of old and new containers is not empty. |
604 | CHECK_LT(new_beg, old_end); |
605 | CHECK_GT(new_end, old_beg); |
606 | |
607 | if (new_beg < old_beg) { |
608 | // Round down because we can't poison prefixes. |
609 | uptr a = RoundDownTo(x: new_beg, boundary: granularity); |
610 | // Round down and ignore the [c, old_beg) as its state defined by unchanged |
611 | // [old_beg, old_end). |
612 | uptr c = RoundDownTo(x: old_beg, boundary: granularity); |
613 | PoisonShadow(addr: a, size: c - a, value: 0); |
614 | } else if (new_beg > old_beg) { |
615 | // Round down and poison [a, old_beg) because it was unpoisoned only as a |
616 | // prefix. |
617 | uptr a = RoundDownTo(x: old_beg, boundary: granularity); |
618 | // Round down and ignore the [c, new_beg) as its state defined by unchanged |
619 | // [new_beg, old_end). |
620 | uptr c = RoundDownTo(x: new_beg, boundary: granularity); |
621 | |
622 | PoisonShadow(addr: a, size: c - a, value: kAsanContiguousContainerOOBMagic); |
623 | } |
624 | |
625 | if (new_end > old_end) { |
626 | // Round down to poison the prefix. |
627 | uptr a = RoundDownTo(x: old_end, boundary: granularity); |
628 | // Round down and handle remainder below. |
629 | uptr c = RoundDownTo(x: new_end, boundary: granularity); |
630 | PoisonShadow(addr: a, size: c - a, value: 0); |
631 | if (!AddrIsAlignedByGranularity(a: new_end)) |
632 | *(u8 *)MemToShadow(p: c) = static_cast<u8>(new_end - c); |
633 | } else if (new_end < old_end) { |
634 | // Round up and handle remained below. |
635 | uptr a2 = RoundUpTo(size: new_end, boundary: granularity); |
636 | // Round up to poison entire granule as we had nothing in [old_end, c2). |
637 | uptr c2 = RoundUpTo(size: old_end, boundary: granularity); |
638 | PoisonShadow(addr: a2, size: c2 - a2, value: kAsanContiguousContainerOOBMagic); |
639 | |
640 | if (!AddrIsAlignedByGranularity(a: new_end)) { |
641 | uptr a = RoundDownTo(x: new_end, boundary: granularity); |
642 | *(u8 *)MemToShadow(p: a) = static_cast<u8>(new_end - a); |
643 | } |
644 | } |
645 | } |
646 | |
647 | // Marks the specified number of bytes in a granule as accessible or |
648 | // poisones the whole granule with kAsanContiguousContainerOOBMagic value. |
649 | static void SetContainerGranule(uptr ptr, u8 n) { |
650 | constexpr uptr granularity = ASAN_SHADOW_GRANULARITY; |
651 | u8 s = (n == granularity) ? 0 : (n ? n : kAsanContiguousContainerOOBMagic); |
652 | *(u8 *)MemToShadow(p: ptr) = s; |
653 | } |
654 | |
655 | // Performs a byte-by-byte copy of ASan annotations (shadow memory values). |
656 | // Result may be different due to ASan limitations, but result cannot lead |
657 | // to false positives (more memory than requested may get unpoisoned). |
658 | static void SlowCopyContainerAnnotations(uptr src_beg, uptr src_end, |
659 | uptr dst_beg, uptr dst_end) { |
660 | constexpr uptr granularity = ASAN_SHADOW_GRANULARITY; |
661 | uptr dst_end_down = RoundDownTo(x: dst_end, boundary: granularity); |
662 | uptr src_ptr = src_beg; |
663 | uptr dst_ptr = dst_beg; |
664 | |
665 | while (dst_ptr < dst_end) { |
666 | uptr granule_beg = RoundDownTo(x: dst_ptr, boundary: granularity); |
667 | uptr granule_end = granule_beg + granularity; |
668 | uptr unpoisoned_bytes = 0; |
669 | |
670 | uptr end = Min(a: granule_end, b: dst_end); |
671 | for (; dst_ptr != end; ++dst_ptr, ++src_ptr) |
672 | if (!AddressIsPoisoned(a: src_ptr)) |
673 | unpoisoned_bytes = dst_ptr - granule_beg + 1; |
674 | |
675 | if (dst_ptr == dst_end && dst_end != dst_end_down && |
676 | !AddressIsPoisoned(a: dst_end)) |
677 | continue; |
678 | |
679 | if (unpoisoned_bytes != 0 || granule_beg >= dst_beg) |
680 | SetContainerGranule(ptr: granule_beg, n: unpoisoned_bytes); |
681 | else if (!AddressIsPoisoned(a: dst_beg)) |
682 | SetContainerGranule(ptr: granule_beg, n: dst_beg - granule_beg); |
683 | } |
684 | } |
685 | |
686 | // Performs a byte-by-byte copy of ASan annotations (shadow memory values), |
687 | // going through bytes in reversed order, but not reversing annotations. |
688 | // Result may be different due to ASan limitations, but result cannot lead |
689 | // to false positives (more memory than requested may get unpoisoned). |
690 | static void SlowReversedCopyContainerAnnotations(uptr src_beg, uptr src_end, |
691 | uptr dst_beg, uptr dst_end) { |
692 | constexpr uptr granularity = ASAN_SHADOW_GRANULARITY; |
693 | uptr dst_end_down = RoundDownTo(x: dst_end, boundary: granularity); |
694 | uptr src_ptr = src_end; |
695 | uptr dst_ptr = dst_end; |
696 | |
697 | while (dst_ptr > dst_beg) { |
698 | uptr granule_beg = RoundDownTo(x: dst_ptr - 1, boundary: granularity); |
699 | uptr unpoisoned_bytes = 0; |
700 | |
701 | uptr end = Max(a: granule_beg, b: dst_beg); |
702 | for (; dst_ptr != end; --dst_ptr, --src_ptr) |
703 | if (unpoisoned_bytes == 0 && !AddressIsPoisoned(a: src_ptr - 1)) |
704 | unpoisoned_bytes = dst_ptr - granule_beg; |
705 | |
706 | if (dst_ptr >= dst_end_down && !AddressIsPoisoned(a: dst_end)) |
707 | continue; |
708 | |
709 | if (granule_beg == dst_ptr || unpoisoned_bytes != 0) |
710 | SetContainerGranule(ptr: granule_beg, n: unpoisoned_bytes); |
711 | else if (!AddressIsPoisoned(a: dst_beg)) |
712 | SetContainerGranule(ptr: granule_beg, n: dst_beg - granule_beg); |
713 | } |
714 | } |
715 | |
716 | // A helper function for __sanitizer_copy_contiguous_container_annotations, |
717 | // has assumption about begin and end of the container. |
718 | // Should not be used stand alone. |
719 | static void CopyContainerFirstGranuleAnnotation(uptr src_beg, uptr dst_beg) { |
720 | constexpr uptr granularity = ASAN_SHADOW_GRANULARITY; |
721 | // First granule |
722 | uptr src_beg_down = RoundDownTo(x: src_beg, boundary: granularity); |
723 | uptr dst_beg_down = RoundDownTo(x: dst_beg, boundary: granularity); |
724 | if (dst_beg_down == dst_beg) |
725 | return; |
726 | if (!AddressIsPoisoned(a: src_beg)) |
727 | *(u8 *)MemToShadow(p: dst_beg_down) = *(u8 *)MemToShadow(p: src_beg_down); |
728 | else if (!AddressIsPoisoned(a: dst_beg)) |
729 | SetContainerGranule(ptr: dst_beg_down, n: dst_beg - dst_beg_down); |
730 | } |
731 | |
732 | // A helper function for __sanitizer_copy_contiguous_container_annotations, |
733 | // has assumption about begin and end of the container. |
734 | // Should not be used stand alone. |
735 | static void CopyContainerLastGranuleAnnotation(uptr src_end, uptr dst_end) { |
736 | constexpr uptr granularity = ASAN_SHADOW_GRANULARITY; |
737 | // Last granule |
738 | uptr src_end_down = RoundDownTo(x: src_end, boundary: granularity); |
739 | uptr dst_end_down = RoundDownTo(x: dst_end, boundary: granularity); |
740 | if (dst_end_down == dst_end || !AddressIsPoisoned(a: dst_end)) |
741 | return; |
742 | if (AddressIsPoisoned(a: src_end)) |
743 | *(u8 *)MemToShadow(p: dst_end_down) = *(u8 *)MemToShadow(p: src_end_down); |
744 | else |
745 | SetContainerGranule(ptr: dst_end_down, n: src_end - src_end_down); |
746 | } |
747 | |
748 | // This function copies ASan memory annotations (poisoned/unpoisoned states) |
749 | // from one buffer to another. |
750 | // It's main purpose is to help with relocating trivially relocatable objects, |
751 | // which memory may be poisoned, without calling copy constructor. |
752 | // However, it does not move memory content itself, only annotations. |
753 | // If the buffers aren't aligned (the distance between buffers isn't |
754 | // granule-aligned) |
755 | // // src_beg % granularity != dst_beg % granularity |
756 | // the function handles this by going byte by byte, slowing down performance. |
757 | // The old buffer annotations are not removed. If necessary, |
758 | // user can unpoison old buffer with __asan_unpoison_memory_region. |
759 | void __sanitizer_copy_contiguous_container_annotations(const void *src_beg_p, |
760 | const void *src_end_p, |
761 | const void *dst_beg_p, |
762 | const void *dst_end_p) { |
763 | if (!flags()->detect_container_overflow) |
764 | return; |
765 | |
766 | VPrintf(3, "contiguous_container_src: %p %p\n" , src_beg_p, src_end_p); |
767 | VPrintf(3, "contiguous_container_dst: %p %p\n" , dst_beg_p, dst_end_p); |
768 | |
769 | uptr src_beg = reinterpret_cast<uptr>(src_beg_p); |
770 | uptr src_end = reinterpret_cast<uptr>(src_end_p); |
771 | uptr dst_beg = reinterpret_cast<uptr>(dst_beg_p); |
772 | uptr dst_end = reinterpret_cast<uptr>(dst_end_p); |
773 | |
774 | constexpr uptr granularity = ASAN_SHADOW_GRANULARITY; |
775 | |
776 | if (src_beg > src_end || (dst_end - dst_beg) != (src_end - src_beg)) { |
777 | GET_STACK_TRACE_FATAL_HERE; |
778 | ReportBadParamsToCopyContiguousContainerAnnotations( |
779 | old_storage_beg: src_beg, old_storage_end: src_end, new_storage_beg: dst_beg, new_storage_end: dst_end, stack: &stack); |
780 | } |
781 | |
782 | if (src_beg == src_end || src_beg == dst_beg) |
783 | return; |
784 | // Due to support for overlapping buffers, we may have to copy elements |
785 | // in reversed order, when destination buffer starts in the middle of |
786 | // the source buffer (or shares first granule with it). |
787 | // |
788 | // When buffers are not granule-aligned (or distance between them, |
789 | // to be specific), annotatios have to be copied byte by byte. |
790 | // |
791 | // The only remaining edge cases involve edge granules, |
792 | // when the container starts or ends within a granule. |
793 | uptr src_beg_up = RoundUpTo(size: src_beg, boundary: granularity); |
794 | uptr src_end_up = RoundUpTo(size: src_end, boundary: granularity); |
795 | bool copy_in_reversed_order = src_beg < dst_beg && dst_beg <= src_end_up; |
796 | if (src_beg % granularity != dst_beg % granularity || |
797 | RoundDownTo(x: dst_end - 1, boundary: granularity) <= dst_beg) { |
798 | if (copy_in_reversed_order) |
799 | SlowReversedCopyContainerAnnotations(src_beg, src_end, dst_beg, dst_end); |
800 | else |
801 | SlowCopyContainerAnnotations(src_beg, src_end, dst_beg, dst_end); |
802 | return; |
803 | } |
804 | |
805 | // As buffers are granule-aligned, we can just copy annotations of granules |
806 | // from the middle. |
807 | uptr dst_beg_up = RoundUpTo(size: dst_beg, boundary: granularity); |
808 | uptr dst_end_down = RoundDownTo(x: dst_end, boundary: granularity); |
809 | if (copy_in_reversed_order) |
810 | CopyContainerLastGranuleAnnotation(src_end, dst_end); |
811 | else |
812 | CopyContainerFirstGranuleAnnotation(src_beg, dst_beg); |
813 | |
814 | if (dst_beg_up < dst_end_down) { |
815 | internal_memmove(dest: (u8 *)MemToShadow(p: dst_beg_up), |
816 | src: (u8 *)MemToShadow(p: src_beg_up), |
817 | n: (dst_end_down - dst_beg_up) / granularity); |
818 | } |
819 | |
820 | if (copy_in_reversed_order) |
821 | CopyContainerFirstGranuleAnnotation(src_beg, dst_beg); |
822 | else |
823 | CopyContainerLastGranuleAnnotation(src_end, dst_end); |
824 | } |
825 | |
826 | static const void *FindBadAddress(uptr begin, uptr end, bool poisoned) { |
827 | CHECK_LE(begin, end); |
828 | constexpr uptr kMaxRangeToCheck = 32; |
829 | if (end - begin > kMaxRangeToCheck * 2) { |
830 | if (auto *bad = FindBadAddress(begin, end: begin + kMaxRangeToCheck, poisoned)) |
831 | return bad; |
832 | if (auto *bad = FindBadAddress(begin: end - kMaxRangeToCheck, end, poisoned)) |
833 | return bad; |
834 | } |
835 | |
836 | for (uptr i = begin; i < end; ++i) |
837 | if (AddressIsPoisoned(a: i) != poisoned) |
838 | return reinterpret_cast<const void *>(i); |
839 | return nullptr; |
840 | } |
841 | |
842 | const void *__sanitizer_contiguous_container_find_bad_address( |
843 | const void *beg_p, const void *mid_p, const void *end_p) { |
844 | if (!flags()->detect_container_overflow) |
845 | return nullptr; |
846 | uptr granularity = ASAN_SHADOW_GRANULARITY; |
847 | uptr beg = reinterpret_cast<uptr>(beg_p); |
848 | uptr end = reinterpret_cast<uptr>(end_p); |
849 | uptr mid = reinterpret_cast<uptr>(mid_p); |
850 | CHECK_LE(beg, mid); |
851 | CHECK_LE(mid, end); |
852 | // If the byte after the storage is unpoisoned, everything in the granule |
853 | // before must stay unpoisoned. |
854 | uptr annotations_end = |
855 | (!AddrIsAlignedByGranularity(a: end) && !AddressIsPoisoned(a: end)) |
856 | ? RoundDownTo(x: end, boundary: granularity) |
857 | : end; |
858 | beg = Min(a: beg, b: annotations_end); |
859 | mid = Min(a: mid, b: annotations_end); |
860 | if (auto *bad = FindBadAddress(begin: beg, end: mid, poisoned: false)) |
861 | return bad; |
862 | if (auto *bad = FindBadAddress(begin: mid, end: annotations_end, poisoned: true)) |
863 | return bad; |
864 | return FindBadAddress(begin: annotations_end, end, poisoned: false); |
865 | } |
866 | |
867 | int __sanitizer_verify_contiguous_container(const void *beg_p, |
868 | const void *mid_p, |
869 | const void *end_p) { |
870 | return __sanitizer_contiguous_container_find_bad_address(beg_p, mid_p, |
871 | end_p) == nullptr; |
872 | } |
873 | |
874 | const void *__sanitizer_double_ended_contiguous_container_find_bad_address( |
875 | const void *storage_beg_p, const void *container_beg_p, |
876 | const void *container_end_p, const void *storage_end_p) { |
877 | if (!flags()->detect_container_overflow) |
878 | return nullptr; |
879 | uptr granularity = ASAN_SHADOW_GRANULARITY; |
880 | uptr storage_beg = reinterpret_cast<uptr>(storage_beg_p); |
881 | uptr storage_end = reinterpret_cast<uptr>(storage_end_p); |
882 | uptr beg = reinterpret_cast<uptr>(container_beg_p); |
883 | uptr end = reinterpret_cast<uptr>(container_end_p); |
884 | |
885 | // The prefix of the firs granule of the container is unpoisoned. |
886 | if (beg != end) |
887 | beg = Max(a: storage_beg, b: RoundDownTo(x: beg, boundary: granularity)); |
888 | |
889 | // If the byte after the storage is unpoisoned, the prefix of the last granule |
890 | // is unpoisoned. |
891 | uptr annotations_end = (!AddrIsAlignedByGranularity(a: storage_end) && |
892 | !AddressIsPoisoned(a: storage_end)) |
893 | ? RoundDownTo(x: storage_end, boundary: granularity) |
894 | : storage_end; |
895 | storage_beg = Min(a: storage_beg, b: annotations_end); |
896 | beg = Min(a: beg, b: annotations_end); |
897 | end = Min(a: end, b: annotations_end); |
898 | |
899 | if (auto *bad = FindBadAddress(begin: storage_beg, end: beg, poisoned: true)) |
900 | return bad; |
901 | if (auto *bad = FindBadAddress(begin: beg, end, poisoned: false)) |
902 | return bad; |
903 | if (auto *bad = FindBadAddress(begin: end, end: annotations_end, poisoned: true)) |
904 | return bad; |
905 | return FindBadAddress(begin: annotations_end, end: storage_end, poisoned: false); |
906 | } |
907 | |
908 | int __sanitizer_verify_double_ended_contiguous_container( |
909 | const void *storage_beg_p, const void *container_beg_p, |
910 | const void *container_end_p, const void *storage_end_p) { |
911 | return __sanitizer_double_ended_contiguous_container_find_bad_address( |
912 | storage_beg_p, container_beg_p, container_end_p, storage_end_p) == |
913 | nullptr; |
914 | } |
915 | |
916 | extern "C" SANITIZER_INTERFACE_ATTRIBUTE |
917 | void __asan_poison_intra_object_redzone(uptr ptr, uptr size) { |
918 | AsanPoisonOrUnpoisonIntraObjectRedzone(ptr, size, poison: true); |
919 | } |
920 | |
921 | extern "C" SANITIZER_INTERFACE_ATTRIBUTE |
922 | void __asan_unpoison_intra_object_redzone(uptr ptr, uptr size) { |
923 | AsanPoisonOrUnpoisonIntraObjectRedzone(ptr, size, poison: false); |
924 | } |
925 | |
926 | // --- Implementation of LSan-specific functions --- {{{1 |
927 | namespace __lsan { |
928 | bool WordIsPoisoned(uptr addr) { |
929 | return (__asan_region_is_poisoned(beg: addr, size: sizeof(uptr)) != 0); |
930 | } |
931 | } |
932 | |