This PR adds a `__sanitizer_copy_contiguous_container_annotations`
function, which copies annotations from one memory area to another. New
area is annotated in the same way as the old region at the beginning
(within limitations of ASan).
Overlapping case: The function supports overlapping containers, however
no assumptions should be made outside of no false positives in new
buffer area. (It doesn't modify old container annotations where it's not
necessary, false negatives may happen in edge granules of the new
container area.) I don't expect this function to be used with
overlapping buffers, but it's designed to work with them and not result
in incorrect ASan errors (false positives).
If buffers have granularity-aligned distance between them (`old_beg %
granularity == new_beg % granularity`), copying algorithm works faster.
If the distance is not granularity-aligned, annotations are copied byte
after byte.
```cpp
void __sanitizer_copy_contiguous_container_annotations(
const void *old_storage_beg_p, const void *old_storage_end_p,
const void *new_storage_beg_p, const void *new_storage_end_p) {
```
This function aims to help with short string annotations and similar
container annotations. Right now we change trait types of
`std::basic_string` when compiling with ASan and this function purpose
is reverting that change as soon as possible.
87f3407856/libcxx/include/string (L738-L751)
The goal is to not change `__trivially_relocatable` when compiling with
ASan. If this function is accepted and upstreamed, the next step is
creating a function like `__memcpy_with_asan` moving memory with ASan.
And then using this function instead of `__builtin__memcpy` while moving
trivially relocatable objects.
11a6799740/libcxx/include/__memory/uninitialized_algorithms.h (L644-L646)
---
I'm thinking if there is a good way to address fact that in a container
the new buffer is usually bigger than the previous one. We may add two
more arguments to the functions to address it (the beginning and the end
of the whole buffer.
Another potential change is removing `new_storage_end_p` as it's
redundant, because we require the same size.
Potential future work is creating a function `__asan_unsafe_memmove`,
which will be basically memmove, but with turned off instrumentation
(therefore it will allow copy data from poisoned area).
---------
Co-authored-by: Vitaly Buka <vitalybuka@google.com>
865 lines
32 KiB
C++
865 lines
32 KiB
C++
//===-- asan_poisoning.cpp ------------------------------------------------===//
|
|
//
|
|
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
// See https://llvm.org/LICENSE.txt for license information.
|
|
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
//
|
|
// This file is a part of AddressSanitizer, an address sanity checker.
|
|
//
|
|
// Shadow memory poisoning by ASan RTL and by user application.
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
#include "asan_poisoning.h"
|
|
|
|
#include "asan_report.h"
|
|
#include "asan_stack.h"
|
|
#include "sanitizer_common/sanitizer_atomic.h"
|
|
#include "sanitizer_common/sanitizer_common.h"
|
|
#include "sanitizer_common/sanitizer_flags.h"
|
|
#include "sanitizer_common/sanitizer_interface_internal.h"
|
|
#include "sanitizer_common/sanitizer_libc.h"
|
|
|
|
namespace __asan {
|
|
|
|
static atomic_uint8_t can_poison_memory;
|
|
|
|
void SetCanPoisonMemory(bool value) {
|
|
atomic_store(&can_poison_memory, value, memory_order_release);
|
|
}
|
|
|
|
bool CanPoisonMemory() {
|
|
return atomic_load(&can_poison_memory, memory_order_acquire);
|
|
}
|
|
|
|
void PoisonShadow(uptr addr, uptr size, u8 value) {
|
|
if (value && !CanPoisonMemory()) return;
|
|
CHECK(AddrIsAlignedByGranularity(addr));
|
|
CHECK(AddrIsInMem(addr));
|
|
CHECK(AddrIsAlignedByGranularity(addr + size));
|
|
CHECK(AddrIsInMem(addr + size - ASAN_SHADOW_GRANULARITY));
|
|
CHECK(REAL(memset));
|
|
FastPoisonShadow(addr, size, value);
|
|
}
|
|
|
|
void PoisonShadowPartialRightRedzone(uptr addr,
|
|
uptr size,
|
|
uptr redzone_size,
|
|
u8 value) {
|
|
if (!CanPoisonMemory()) return;
|
|
CHECK(AddrIsAlignedByGranularity(addr));
|
|
CHECK(AddrIsInMem(addr));
|
|
FastPoisonShadowPartialRightRedzone(addr, size, redzone_size, value);
|
|
}
|
|
|
|
struct ShadowSegmentEndpoint {
|
|
u8 *chunk;
|
|
s8 offset; // in [0, ASAN_SHADOW_GRANULARITY)
|
|
s8 value; // = *chunk;
|
|
|
|
explicit ShadowSegmentEndpoint(uptr address) {
|
|
chunk = (u8*)MemToShadow(address);
|
|
offset = address & (ASAN_SHADOW_GRANULARITY - 1);
|
|
value = *chunk;
|
|
}
|
|
};
|
|
|
|
void AsanPoisonOrUnpoisonIntraObjectRedzone(uptr ptr, uptr size, bool poison) {
|
|
uptr end = ptr + size;
|
|
if (Verbosity()) {
|
|
Printf("__asan_%spoison_intra_object_redzone [%p,%p) %zd\n",
|
|
poison ? "" : "un", (void *)ptr, (void *)end, size);
|
|
if (Verbosity() >= 2)
|
|
PRINT_CURRENT_STACK();
|
|
}
|
|
CHECK(size);
|
|
CHECK_LE(size, 4096);
|
|
CHECK(IsAligned(end, ASAN_SHADOW_GRANULARITY));
|
|
if (!IsAligned(ptr, ASAN_SHADOW_GRANULARITY)) {
|
|
*(u8 *)MemToShadow(ptr) =
|
|
poison ? static_cast<u8>(ptr % ASAN_SHADOW_GRANULARITY) : 0;
|
|
ptr |= ASAN_SHADOW_GRANULARITY - 1;
|
|
ptr++;
|
|
}
|
|
for (; ptr < end; ptr += ASAN_SHADOW_GRANULARITY)
|
|
*(u8*)MemToShadow(ptr) = poison ? kAsanIntraObjectRedzone : 0;
|
|
}
|
|
|
|
} // namespace __asan
|
|
|
|
// ---------------------- Interface ---------------- {{{1
|
|
using namespace __asan;
|
|
|
|
// Current implementation of __asan_(un)poison_memory_region doesn't check
|
|
// that user program (un)poisons the memory it owns. It poisons memory
|
|
// conservatively, and unpoisons progressively to make sure asan shadow
|
|
// mapping invariant is preserved (see detailed mapping description here:
|
|
// https://github.com/google/sanitizers/wiki/AddressSanitizerAlgorithm).
|
|
//
|
|
// * if user asks to poison region [left, right), the program poisons
|
|
// at least [left, AlignDown(right)).
|
|
// * if user asks to unpoison region [left, right), the program unpoisons
|
|
// at most [AlignDown(left), right).
|
|
void __asan_poison_memory_region(void const volatile *addr, uptr size) {
|
|
if (!flags()->allow_user_poisoning || size == 0) return;
|
|
uptr beg_addr = (uptr)addr;
|
|
uptr end_addr = beg_addr + size;
|
|
VPrintf(3, "Trying to poison memory region [%p, %p)\n", (void *)beg_addr,
|
|
(void *)end_addr);
|
|
ShadowSegmentEndpoint beg(beg_addr);
|
|
ShadowSegmentEndpoint end(end_addr);
|
|
if (beg.chunk == end.chunk) {
|
|
CHECK_LT(beg.offset, end.offset);
|
|
s8 value = beg.value;
|
|
CHECK_EQ(value, end.value);
|
|
// We can only poison memory if the byte in end.offset is unaddressable.
|
|
// No need to re-poison memory if it is poisoned already.
|
|
if (value > 0 && value <= end.offset) {
|
|
if (beg.offset > 0) {
|
|
*beg.chunk = Min(value, beg.offset);
|
|
} else {
|
|
*beg.chunk = kAsanUserPoisonedMemoryMagic;
|
|
}
|
|
}
|
|
return;
|
|
}
|
|
CHECK_LT(beg.chunk, end.chunk);
|
|
if (beg.offset > 0) {
|
|
// Mark bytes from beg.offset as unaddressable.
|
|
if (beg.value == 0) {
|
|
*beg.chunk = beg.offset;
|
|
} else {
|
|
*beg.chunk = Min(beg.value, beg.offset);
|
|
}
|
|
beg.chunk++;
|
|
}
|
|
REAL(memset)(beg.chunk, kAsanUserPoisonedMemoryMagic, end.chunk - beg.chunk);
|
|
// Poison if byte in end.offset is unaddressable.
|
|
if (end.value > 0 && end.value <= end.offset) {
|
|
*end.chunk = kAsanUserPoisonedMemoryMagic;
|
|
}
|
|
}
|
|
|
|
void __asan_unpoison_memory_region(void const volatile *addr, uptr size) {
|
|
if (!flags()->allow_user_poisoning || size == 0) return;
|
|
uptr beg_addr = (uptr)addr;
|
|
uptr end_addr = beg_addr + size;
|
|
VPrintf(3, "Trying to unpoison memory region [%p, %p)\n", (void *)beg_addr,
|
|
(void *)end_addr);
|
|
ShadowSegmentEndpoint beg(beg_addr);
|
|
ShadowSegmentEndpoint end(end_addr);
|
|
if (beg.chunk == end.chunk) {
|
|
CHECK_LT(beg.offset, end.offset);
|
|
s8 value = beg.value;
|
|
CHECK_EQ(value, end.value);
|
|
// We unpoison memory bytes up to enbytes up to end.offset if it is not
|
|
// unpoisoned already.
|
|
if (value != 0) {
|
|
*beg.chunk = Max(value, end.offset);
|
|
}
|
|
return;
|
|
}
|
|
CHECK_LT(beg.chunk, end.chunk);
|
|
REAL(memset)(beg.chunk, 0, end.chunk - beg.chunk);
|
|
if (end.offset > 0 && end.value != 0) {
|
|
*end.chunk = Max(end.value, end.offset);
|
|
}
|
|
}
|
|
|
|
int __asan_address_is_poisoned(void const volatile *addr) {
|
|
return __asan::AddressIsPoisoned((uptr)addr);
|
|
}
|
|
|
|
uptr __asan_region_is_poisoned(uptr beg, uptr size) {
|
|
if (!size)
|
|
return 0;
|
|
uptr end = beg + size;
|
|
if (!AddrIsInMem(beg))
|
|
return beg;
|
|
if (!AddrIsInMem(end))
|
|
return end;
|
|
CHECK_LT(beg, end);
|
|
uptr aligned_b = RoundUpTo(beg, ASAN_SHADOW_GRANULARITY);
|
|
uptr aligned_e = RoundDownTo(end, ASAN_SHADOW_GRANULARITY);
|
|
uptr shadow_beg = MemToShadow(aligned_b);
|
|
uptr shadow_end = MemToShadow(aligned_e);
|
|
// First check the first and the last application bytes,
|
|
// then check the ASAN_SHADOW_GRANULARITY-aligned region by calling
|
|
// mem_is_zero on the corresponding shadow.
|
|
if (!__asan::AddressIsPoisoned(beg) && !__asan::AddressIsPoisoned(end - 1) &&
|
|
(shadow_end <= shadow_beg ||
|
|
__sanitizer::mem_is_zero((const char *)shadow_beg,
|
|
shadow_end - shadow_beg)))
|
|
return 0;
|
|
// The fast check failed, so we have a poisoned byte somewhere.
|
|
// Find it slowly.
|
|
for (; beg < end; beg++)
|
|
if (__asan::AddressIsPoisoned(beg))
|
|
return beg;
|
|
UNREACHABLE("mem_is_zero returned false, but poisoned byte was not found");
|
|
return 0;
|
|
}
|
|
|
|
#define CHECK_SMALL_REGION(p, size, isWrite) \
|
|
do { \
|
|
uptr __p = reinterpret_cast<uptr>(p); \
|
|
uptr __size = size; \
|
|
if (UNLIKELY(__asan::AddressIsPoisoned(__p) || \
|
|
__asan::AddressIsPoisoned(__p + __size - 1))) { \
|
|
GET_CURRENT_PC_BP_SP; \
|
|
uptr __bad = __asan_region_is_poisoned(__p, __size); \
|
|
__asan_report_error(pc, bp, sp, __bad, isWrite, __size, 0);\
|
|
} \
|
|
} while (false)
|
|
|
|
|
|
extern "C" SANITIZER_INTERFACE_ATTRIBUTE
|
|
u16 __sanitizer_unaligned_load16(const uu16 *p) {
|
|
CHECK_SMALL_REGION(p, sizeof(*p), false);
|
|
return *p;
|
|
}
|
|
|
|
extern "C" SANITIZER_INTERFACE_ATTRIBUTE
|
|
u32 __sanitizer_unaligned_load32(const uu32 *p) {
|
|
CHECK_SMALL_REGION(p, sizeof(*p), false);
|
|
return *p;
|
|
}
|
|
|
|
extern "C" SANITIZER_INTERFACE_ATTRIBUTE
|
|
u64 __sanitizer_unaligned_load64(const uu64 *p) {
|
|
CHECK_SMALL_REGION(p, sizeof(*p), false);
|
|
return *p;
|
|
}
|
|
|
|
extern "C" SANITIZER_INTERFACE_ATTRIBUTE
|
|
void __sanitizer_unaligned_store16(uu16 *p, u16 x) {
|
|
CHECK_SMALL_REGION(p, sizeof(*p), true);
|
|
*p = x;
|
|
}
|
|
|
|
extern "C" SANITIZER_INTERFACE_ATTRIBUTE
|
|
void __sanitizer_unaligned_store32(uu32 *p, u32 x) {
|
|
CHECK_SMALL_REGION(p, sizeof(*p), true);
|
|
*p = x;
|
|
}
|
|
|
|
extern "C" SANITIZER_INTERFACE_ATTRIBUTE
|
|
void __sanitizer_unaligned_store64(uu64 *p, u64 x) {
|
|
CHECK_SMALL_REGION(p, sizeof(*p), true);
|
|
*p = x;
|
|
}
|
|
|
|
extern "C" SANITIZER_INTERFACE_ATTRIBUTE
|
|
void __asan_poison_cxx_array_cookie(uptr p) {
|
|
if (SANITIZER_WORDSIZE != 64) return;
|
|
if (!flags()->poison_array_cookie) return;
|
|
uptr s = MEM_TO_SHADOW(p);
|
|
*reinterpret_cast<u8*>(s) = kAsanArrayCookieMagic;
|
|
}
|
|
|
|
extern "C" SANITIZER_INTERFACE_ATTRIBUTE
|
|
uptr __asan_load_cxx_array_cookie(uptr *p) {
|
|
if (SANITIZER_WORDSIZE != 64) return *p;
|
|
if (!flags()->poison_array_cookie) return *p;
|
|
uptr s = MEM_TO_SHADOW(reinterpret_cast<uptr>(p));
|
|
u8 sval = *reinterpret_cast<u8*>(s);
|
|
if (sval == kAsanArrayCookieMagic) return *p;
|
|
// If sval is not kAsanArrayCookieMagic it can only be freed memory,
|
|
// which means that we are going to get double-free. So, return 0 to avoid
|
|
// infinite loop of destructors. We don't want to report a double-free here
|
|
// though, so print a warning just in case.
|
|
// CHECK_EQ(sval, kAsanHeapFreeMagic);
|
|
if (sval == kAsanHeapFreeMagic) {
|
|
Report("AddressSanitizer: loaded array cookie from free-d memory; "
|
|
"expect a double-free report\n");
|
|
return 0;
|
|
}
|
|
// The cookie may remain unpoisoned if e.g. it comes from a custom
|
|
// operator new defined inside a class.
|
|
return *p;
|
|
}
|
|
|
|
// This is a simplified version of __asan_(un)poison_memory_region, which
|
|
// assumes that left border of region to be poisoned is properly aligned.
|
|
static void PoisonAlignedStackMemory(uptr addr, uptr size, bool do_poison) {
|
|
if (size == 0) return;
|
|
uptr aligned_size = size & ~(ASAN_SHADOW_GRANULARITY - 1);
|
|
PoisonShadow(addr, aligned_size,
|
|
do_poison ? kAsanStackUseAfterScopeMagic : 0);
|
|
if (size == aligned_size)
|
|
return;
|
|
s8 end_offset = (s8)(size - aligned_size);
|
|
s8* shadow_end = (s8*)MemToShadow(addr + aligned_size);
|
|
s8 end_value = *shadow_end;
|
|
if (do_poison) {
|
|
// If possible, mark all the bytes mapping to last shadow byte as
|
|
// unaddressable.
|
|
if (end_value > 0 && end_value <= end_offset)
|
|
*shadow_end = (s8)kAsanStackUseAfterScopeMagic;
|
|
} else {
|
|
// If necessary, mark few first bytes mapping to last shadow byte
|
|
// as addressable
|
|
if (end_value != 0)
|
|
*shadow_end = Max(end_value, end_offset);
|
|
}
|
|
}
|
|
|
|
void __asan_set_shadow_00(uptr addr, uptr size) {
|
|
REAL(memset)((void *)addr, 0, size);
|
|
}
|
|
|
|
void __asan_set_shadow_01(uptr addr, uptr size) {
|
|
REAL(memset)((void *)addr, 0x01, size);
|
|
}
|
|
|
|
void __asan_set_shadow_02(uptr addr, uptr size) {
|
|
REAL(memset)((void *)addr, 0x02, size);
|
|
}
|
|
|
|
void __asan_set_shadow_03(uptr addr, uptr size) {
|
|
REAL(memset)((void *)addr, 0x03, size);
|
|
}
|
|
|
|
void __asan_set_shadow_04(uptr addr, uptr size) {
|
|
REAL(memset)((void *)addr, 0x04, size);
|
|
}
|
|
|
|
void __asan_set_shadow_05(uptr addr, uptr size) {
|
|
REAL(memset)((void *)addr, 0x05, size);
|
|
}
|
|
|
|
void __asan_set_shadow_06(uptr addr, uptr size) {
|
|
REAL(memset)((void *)addr, 0x06, size);
|
|
}
|
|
|
|
void __asan_set_shadow_07(uptr addr, uptr size) {
|
|
REAL(memset)((void *)addr, 0x07, size);
|
|
}
|
|
|
|
void __asan_set_shadow_f1(uptr addr, uptr size) {
|
|
REAL(memset)((void *)addr, 0xf1, size);
|
|
}
|
|
|
|
void __asan_set_shadow_f2(uptr addr, uptr size) {
|
|
REAL(memset)((void *)addr, 0xf2, size);
|
|
}
|
|
|
|
void __asan_set_shadow_f3(uptr addr, uptr size) {
|
|
REAL(memset)((void *)addr, 0xf3, size);
|
|
}
|
|
|
|
void __asan_set_shadow_f5(uptr addr, uptr size) {
|
|
REAL(memset)((void *)addr, 0xf5, size);
|
|
}
|
|
|
|
void __asan_set_shadow_f8(uptr addr, uptr size) {
|
|
REAL(memset)((void *)addr, 0xf8, size);
|
|
}
|
|
|
|
void __asan_poison_stack_memory(uptr addr, uptr size) {
|
|
VReport(1, "poisoning: %p %zx\n", (void *)addr, size);
|
|
PoisonAlignedStackMemory(addr, size, true);
|
|
}
|
|
|
|
void __asan_unpoison_stack_memory(uptr addr, uptr size) {
|
|
VReport(1, "unpoisoning: %p %zx\n", (void *)addr, size);
|
|
PoisonAlignedStackMemory(addr, size, false);
|
|
}
|
|
|
|
static void FixUnalignedStorage(uptr storage_beg, uptr storage_end,
|
|
uptr &old_beg, uptr &old_end, uptr &new_beg,
|
|
uptr &new_end) {
|
|
constexpr uptr granularity = ASAN_SHADOW_GRANULARITY;
|
|
if (UNLIKELY(!AddrIsAlignedByGranularity(storage_end))) {
|
|
uptr end_down = RoundDownTo(storage_end, granularity);
|
|
// Ignore the last unaligned granule if the storage is followed by
|
|
// unpoisoned byte, because we can't poison the prefix anyway. Don't call
|
|
// AddressIsPoisoned at all if container changes does not affect the last
|
|
// granule at all.
|
|
if ((((old_end != new_end) && Max(old_end, new_end) > end_down) ||
|
|
((old_beg != new_beg) && Max(old_beg, new_beg) > end_down)) &&
|
|
!AddressIsPoisoned(storage_end)) {
|
|
old_beg = Min(end_down, old_beg);
|
|
old_end = Min(end_down, old_end);
|
|
new_beg = Min(end_down, new_beg);
|
|
new_end = Min(end_down, new_end);
|
|
}
|
|
}
|
|
|
|
// Handle misaligned begin and cut it off.
|
|
if (UNLIKELY(!AddrIsAlignedByGranularity(storage_beg))) {
|
|
uptr beg_up = RoundUpTo(storage_beg, granularity);
|
|
// The first unaligned granule needs special handling only if we had bytes
|
|
// there before and will have none after.
|
|
if ((new_beg == new_end || new_beg >= beg_up) && old_beg != old_end &&
|
|
old_beg < beg_up) {
|
|
// Keep granule prefix outside of the storage unpoisoned.
|
|
uptr beg_down = RoundDownTo(storage_beg, granularity);
|
|
*(u8 *)MemToShadow(beg_down) = storage_beg - beg_down;
|
|
old_beg = Max(beg_up, old_beg);
|
|
old_end = Max(beg_up, old_end);
|
|
new_beg = Max(beg_up, new_beg);
|
|
new_end = Max(beg_up, new_end);
|
|
}
|
|
}
|
|
}
|
|
|
|
void __sanitizer_annotate_contiguous_container(const void *beg_p,
|
|
const void *end_p,
|
|
const void *old_mid_p,
|
|
const void *new_mid_p) {
|
|
if (!flags()->detect_container_overflow)
|
|
return;
|
|
VPrintf(3, "contiguous_container: %p %p %p %p\n", beg_p, end_p, old_mid_p,
|
|
new_mid_p);
|
|
uptr storage_beg = reinterpret_cast<uptr>(beg_p);
|
|
uptr storage_end = reinterpret_cast<uptr>(end_p);
|
|
uptr old_end = reinterpret_cast<uptr>(old_mid_p);
|
|
uptr new_end = reinterpret_cast<uptr>(new_mid_p);
|
|
uptr old_beg = storage_beg;
|
|
uptr new_beg = storage_beg;
|
|
uptr granularity = ASAN_SHADOW_GRANULARITY;
|
|
if (!(storage_beg <= old_end && storage_beg <= new_end &&
|
|
old_end <= storage_end && new_end <= storage_end)) {
|
|
GET_STACK_TRACE_FATAL_HERE;
|
|
ReportBadParamsToAnnotateContiguousContainer(storage_beg, storage_end,
|
|
old_end, new_end, &stack);
|
|
}
|
|
CHECK_LE(storage_end - storage_beg,
|
|
FIRST_32_SECOND_64(1UL << 30, 1ULL << 40)); // Sanity check.
|
|
|
|
if (old_end == new_end)
|
|
return; // Nothing to do here.
|
|
|
|
FixUnalignedStorage(storage_beg, storage_end, old_beg, old_end, new_beg,
|
|
new_end);
|
|
|
|
uptr a = RoundDownTo(Min(old_end, new_end), granularity);
|
|
uptr c = RoundUpTo(Max(old_end, new_end), granularity);
|
|
uptr d1 = RoundDownTo(old_end, granularity);
|
|
// uptr d2 = RoundUpTo(old_mid, granularity);
|
|
// Currently we should be in this state:
|
|
// [a, d1) is good, [d2, c) is bad, [d1, d2) is partially good.
|
|
// Make a quick sanity check that we are indeed in this state.
|
|
//
|
|
// FIXME: Two of these three checks are disabled until we fix
|
|
// https://github.com/google/sanitizers/issues/258.
|
|
// if (d1 != d2)
|
|
// DCHECK_EQ(*(u8*)MemToShadow(d1), old_mid - d1);
|
|
//
|
|
// NOTE: curly brackets for the "if" below to silence a MSVC warning.
|
|
if (a + granularity <= d1) {
|
|
DCHECK_EQ(*(u8 *)MemToShadow(a), 0);
|
|
}
|
|
// if (d2 + granularity <= c && c <= end)
|
|
// DCHECK_EQ(*(u8 *)MemToShadow(c - granularity),
|
|
// kAsanContiguousContainerOOBMagic);
|
|
|
|
uptr b1 = RoundDownTo(new_end, granularity);
|
|
uptr b2 = RoundUpTo(new_end, granularity);
|
|
// New state:
|
|
// [a, b1) is good, [b2, c) is bad, [b1, b2) is partially good.
|
|
if (b1 > a)
|
|
PoisonShadow(a, b1 - a, 0);
|
|
else if (c > b2)
|
|
PoisonShadow(b2, c - b2, kAsanContiguousContainerOOBMagic);
|
|
if (b1 != b2) {
|
|
CHECK_EQ(b2 - b1, granularity);
|
|
*(u8 *)MemToShadow(b1) = static_cast<u8>(new_end - b1);
|
|
}
|
|
}
|
|
|
|
// Annotates a double ended contiguous memory area like std::deque's chunk.
|
|
// It allows detecting buggy accesses to allocated but not used begining
|
|
// or end items of such a container.
|
|
void __sanitizer_annotate_double_ended_contiguous_container(
|
|
const void *storage_beg_p, const void *storage_end_p,
|
|
const void *old_container_beg_p, const void *old_container_end_p,
|
|
const void *new_container_beg_p, const void *new_container_end_p) {
|
|
if (!flags()->detect_container_overflow)
|
|
return;
|
|
|
|
VPrintf(3, "contiguous_container: %p %p %p %p %p %p\n", storage_beg_p,
|
|
storage_end_p, old_container_beg_p, old_container_end_p,
|
|
new_container_beg_p, new_container_end_p);
|
|
|
|
uptr storage_beg = reinterpret_cast<uptr>(storage_beg_p);
|
|
uptr storage_end = reinterpret_cast<uptr>(storage_end_p);
|
|
uptr old_beg = reinterpret_cast<uptr>(old_container_beg_p);
|
|
uptr old_end = reinterpret_cast<uptr>(old_container_end_p);
|
|
uptr new_beg = reinterpret_cast<uptr>(new_container_beg_p);
|
|
uptr new_end = reinterpret_cast<uptr>(new_container_end_p);
|
|
|
|
constexpr uptr granularity = ASAN_SHADOW_GRANULARITY;
|
|
|
|
if (!(old_beg <= old_end && new_beg <= new_end) ||
|
|
!(storage_beg <= new_beg && new_end <= storage_end) ||
|
|
!(storage_beg <= old_beg && old_end <= storage_end)) {
|
|
GET_STACK_TRACE_FATAL_HERE;
|
|
ReportBadParamsToAnnotateDoubleEndedContiguousContainer(
|
|
storage_beg, storage_end, old_beg, old_end, new_beg, new_end, &stack);
|
|
}
|
|
CHECK_LE(storage_end - storage_beg,
|
|
FIRST_32_SECOND_64(1UL << 30, 1ULL << 40)); // Sanity check.
|
|
|
|
if ((old_beg == old_end && new_beg == new_end) ||
|
|
(old_beg == new_beg && old_end == new_end))
|
|
return; // Nothing to do here.
|
|
|
|
FixUnalignedStorage(storage_beg, storage_end, old_beg, old_end, new_beg,
|
|
new_end);
|
|
|
|
// Handle non-intersecting new/old containers separately have simpler
|
|
// intersecting case.
|
|
if (old_beg == old_end || new_beg == new_end || new_end <= old_beg ||
|
|
old_end <= new_beg) {
|
|
if (old_beg != old_end) {
|
|
// Poisoning the old container.
|
|
uptr a = RoundDownTo(old_beg, granularity);
|
|
uptr b = RoundUpTo(old_end, granularity);
|
|
PoisonShadow(a, b - a, kAsanContiguousContainerOOBMagic);
|
|
}
|
|
|
|
if (new_beg != new_end) {
|
|
// Unpoisoning the new container.
|
|
uptr a = RoundDownTo(new_beg, granularity);
|
|
uptr b = RoundDownTo(new_end, granularity);
|
|
PoisonShadow(a, b - a, 0);
|
|
if (!AddrIsAlignedByGranularity(new_end))
|
|
*(u8 *)MemToShadow(b) = static_cast<u8>(new_end - b);
|
|
}
|
|
|
|
return;
|
|
}
|
|
|
|
// Intersection of old and new containers is not empty.
|
|
CHECK_LT(new_beg, old_end);
|
|
CHECK_GT(new_end, old_beg);
|
|
|
|
if (new_beg < old_beg) {
|
|
// Round down because we can't poison prefixes.
|
|
uptr a = RoundDownTo(new_beg, granularity);
|
|
// Round down and ignore the [c, old_beg) as its state defined by unchanged
|
|
// [old_beg, old_end).
|
|
uptr c = RoundDownTo(old_beg, granularity);
|
|
PoisonShadow(a, c - a, 0);
|
|
} else if (new_beg > old_beg) {
|
|
// Round down and poison [a, old_beg) because it was unpoisoned only as a
|
|
// prefix.
|
|
uptr a = RoundDownTo(old_beg, granularity);
|
|
// Round down and ignore the [c, new_beg) as its state defined by unchanged
|
|
// [new_beg, old_end).
|
|
uptr c = RoundDownTo(new_beg, granularity);
|
|
|
|
PoisonShadow(a, c - a, kAsanContiguousContainerOOBMagic);
|
|
}
|
|
|
|
if (new_end > old_end) {
|
|
// Round down to poison the prefix.
|
|
uptr a = RoundDownTo(old_end, granularity);
|
|
// Round down and handle remainder below.
|
|
uptr c = RoundDownTo(new_end, granularity);
|
|
PoisonShadow(a, c - a, 0);
|
|
if (!AddrIsAlignedByGranularity(new_end))
|
|
*(u8 *)MemToShadow(c) = static_cast<u8>(new_end - c);
|
|
} else if (new_end < old_end) {
|
|
// Round up and handle remained below.
|
|
uptr a2 = RoundUpTo(new_end, granularity);
|
|
// Round up to poison entire granule as we had nothing in [old_end, c2).
|
|
uptr c2 = RoundUpTo(old_end, granularity);
|
|
PoisonShadow(a2, c2 - a2, kAsanContiguousContainerOOBMagic);
|
|
|
|
if (!AddrIsAlignedByGranularity(new_end)) {
|
|
uptr a = RoundDownTo(new_end, granularity);
|
|
*(u8 *)MemToShadow(a) = static_cast<u8>(new_end - a);
|
|
}
|
|
}
|
|
}
|
|
|
|
// Marks the specified number of bytes in a granule as accessible or
|
|
// poisones the whole granule with kAsanContiguousContainerOOBMagic value.
|
|
static void SetContainerGranule(uptr ptr, u8 n) {
|
|
constexpr uptr granularity = ASAN_SHADOW_GRANULARITY;
|
|
u8 s = (n == granularity) ? 0 : (n ? n : kAsanContiguousContainerOOBMagic);
|
|
*(u8 *)MemToShadow(ptr) = s;
|
|
}
|
|
|
|
// Performs a byte-by-byte copy of ASan annotations (shadow memory values).
|
|
// Result may be different due to ASan limitations, but result cannot lead
|
|
// to false positives (more memory than requested may get unpoisoned).
|
|
static void SlowCopyContainerAnnotations(uptr src_beg, uptr src_end,
|
|
uptr dst_beg, uptr dst_end) {
|
|
constexpr uptr granularity = ASAN_SHADOW_GRANULARITY;
|
|
uptr dst_end_down = RoundDownTo(dst_end, granularity);
|
|
uptr src_ptr = src_beg;
|
|
uptr dst_ptr = dst_beg;
|
|
|
|
while (dst_ptr < dst_end) {
|
|
uptr granule_beg = RoundDownTo(dst_ptr, granularity);
|
|
uptr granule_end = granule_beg + granularity;
|
|
uptr unpoisoned_bytes = 0;
|
|
|
|
uptr end = Min(granule_end, dst_end);
|
|
for (; dst_ptr != end; ++dst_ptr, ++src_ptr)
|
|
if (!AddressIsPoisoned(src_ptr))
|
|
unpoisoned_bytes = dst_ptr - granule_beg + 1;
|
|
|
|
if (dst_ptr == dst_end && dst_end != dst_end_down &&
|
|
!AddressIsPoisoned(dst_end))
|
|
continue;
|
|
|
|
if (unpoisoned_bytes != 0 || granule_beg >= dst_beg)
|
|
SetContainerGranule(granule_beg, unpoisoned_bytes);
|
|
else if (!AddressIsPoisoned(dst_beg))
|
|
SetContainerGranule(granule_beg, dst_beg - granule_beg);
|
|
}
|
|
}
|
|
|
|
// Performs a byte-by-byte copy of ASan annotations (shadow memory values),
|
|
// going through bytes in reversed order, but not reversing annotations.
|
|
// Result may be different due to ASan limitations, but result cannot lead
|
|
// to false positives (more memory than requested may get unpoisoned).
|
|
static void SlowReversedCopyContainerAnnotations(uptr src_beg, uptr src_end,
|
|
uptr dst_beg, uptr dst_end) {
|
|
constexpr uptr granularity = ASAN_SHADOW_GRANULARITY;
|
|
uptr dst_end_down = RoundDownTo(dst_end, granularity);
|
|
uptr src_ptr = src_end;
|
|
uptr dst_ptr = dst_end;
|
|
|
|
while (dst_ptr > dst_beg) {
|
|
uptr granule_beg = RoundDownTo(dst_ptr - 1, granularity);
|
|
uptr unpoisoned_bytes = 0;
|
|
|
|
uptr end = Max(granule_beg, dst_beg);
|
|
for (; dst_ptr != end; --dst_ptr, --src_ptr)
|
|
if (unpoisoned_bytes == 0 && !AddressIsPoisoned(src_ptr - 1))
|
|
unpoisoned_bytes = dst_ptr - granule_beg;
|
|
|
|
if (dst_ptr >= dst_end_down && !AddressIsPoisoned(dst_end))
|
|
continue;
|
|
|
|
if (granule_beg == dst_ptr || unpoisoned_bytes != 0)
|
|
SetContainerGranule(granule_beg, unpoisoned_bytes);
|
|
else if (!AddressIsPoisoned(dst_beg))
|
|
SetContainerGranule(granule_beg, dst_beg - granule_beg);
|
|
}
|
|
}
|
|
|
|
// A helper function for __sanitizer_copy_contiguous_container_annotations,
|
|
// has assumption about begin and end of the container.
|
|
// Should not be used stand alone.
|
|
static void CopyContainerFirstGranuleAnnotation(uptr src_beg, uptr dst_beg) {
|
|
constexpr uptr granularity = ASAN_SHADOW_GRANULARITY;
|
|
// First granule
|
|
uptr src_beg_down = RoundDownTo(src_beg, granularity);
|
|
uptr dst_beg_down = RoundDownTo(dst_beg, granularity);
|
|
if (dst_beg_down == dst_beg)
|
|
return;
|
|
if (!AddressIsPoisoned(src_beg))
|
|
*(u8 *)MemToShadow(dst_beg_down) = *(u8 *)MemToShadow(src_beg_down);
|
|
else if (!AddressIsPoisoned(dst_beg))
|
|
SetContainerGranule(dst_beg_down, dst_beg - dst_beg_down);
|
|
}
|
|
|
|
// A helper function for __sanitizer_copy_contiguous_container_annotations,
|
|
// has assumption about begin and end of the container.
|
|
// Should not be used stand alone.
|
|
static void CopyContainerLastGranuleAnnotation(uptr src_end, uptr dst_end) {
|
|
constexpr uptr granularity = ASAN_SHADOW_GRANULARITY;
|
|
// Last granule
|
|
uptr src_end_down = RoundDownTo(src_end, granularity);
|
|
uptr dst_end_down = RoundDownTo(dst_end, granularity);
|
|
if (dst_end_down == dst_end || !AddressIsPoisoned(dst_end))
|
|
return;
|
|
if (AddressIsPoisoned(src_end))
|
|
*(u8 *)MemToShadow(dst_end_down) = *(u8 *)MemToShadow(src_end_down);
|
|
else
|
|
SetContainerGranule(dst_end_down, src_end - src_end_down);
|
|
}
|
|
|
|
// This function copies ASan memory annotations (poisoned/unpoisoned states)
|
|
// from one buffer to another.
|
|
// It's main purpose is to help with relocating trivially relocatable objects,
|
|
// which memory may be poisoned, without calling copy constructor.
|
|
// However, it does not move memory content itself, only annotations.
|
|
// If the buffers aren't aligned (the distance between buffers isn't
|
|
// granule-aligned)
|
|
// // src_beg % granularity != dst_beg % granularity
|
|
// the function handles this by going byte by byte, slowing down performance.
|
|
// The old buffer annotations are not removed. If necessary,
|
|
// user can unpoison old buffer with __asan_unpoison_memory_region.
|
|
void __sanitizer_copy_contiguous_container_annotations(const void *src_beg_p,
|
|
const void *src_end_p,
|
|
const void *dst_beg_p,
|
|
const void *dst_end_p) {
|
|
if (!flags()->detect_container_overflow)
|
|
return;
|
|
|
|
VPrintf(3, "contiguous_container_src: %p %p\n", src_beg_p, src_end_p);
|
|
VPrintf(3, "contiguous_container_dst: %p %p\n", dst_beg_p, dst_end_p);
|
|
|
|
uptr src_beg = reinterpret_cast<uptr>(src_beg_p);
|
|
uptr src_end = reinterpret_cast<uptr>(src_end_p);
|
|
uptr dst_beg = reinterpret_cast<uptr>(dst_beg_p);
|
|
uptr dst_end = reinterpret_cast<uptr>(dst_end_p);
|
|
|
|
constexpr uptr granularity = ASAN_SHADOW_GRANULARITY;
|
|
|
|
if (src_beg > src_end || (dst_end - dst_beg) != (src_end - src_beg)) {
|
|
GET_STACK_TRACE_FATAL_HERE;
|
|
ReportBadParamsToCopyContiguousContainerAnnotations(
|
|
src_beg, src_end, dst_beg, dst_end, &stack);
|
|
}
|
|
|
|
if (src_beg == src_end || src_beg == dst_beg)
|
|
return;
|
|
// Due to support for overlapping buffers, we may have to copy elements
|
|
// in reversed order, when destination buffer starts in the middle of
|
|
// the source buffer (or shares first granule with it).
|
|
//
|
|
// When buffers are not granule-aligned (or distance between them,
|
|
// to be specific), annotatios have to be copied byte by byte.
|
|
//
|
|
// The only remaining edge cases involve edge granules,
|
|
// when the container starts or ends within a granule.
|
|
uptr src_beg_up = RoundUpTo(src_beg, granularity);
|
|
uptr src_end_up = RoundUpTo(src_end, granularity);
|
|
bool copy_in_reversed_order = src_beg < dst_beg && dst_beg <= src_end_up;
|
|
if (src_beg % granularity != dst_beg % granularity ||
|
|
RoundDownTo(dst_end - 1, granularity) <= dst_beg) {
|
|
if (copy_in_reversed_order)
|
|
SlowReversedCopyContainerAnnotations(src_beg, src_end, dst_beg, dst_end);
|
|
else
|
|
SlowCopyContainerAnnotations(src_beg, src_end, dst_beg, dst_end);
|
|
return;
|
|
}
|
|
|
|
// As buffers are granule-aligned, we can just copy annotations of granules
|
|
// from the middle.
|
|
uptr dst_beg_up = RoundUpTo(dst_beg, granularity);
|
|
uptr dst_end_down = RoundDownTo(dst_end, granularity);
|
|
if (copy_in_reversed_order)
|
|
CopyContainerLastGranuleAnnotation(src_end, dst_end);
|
|
else
|
|
CopyContainerFirstGranuleAnnotation(src_beg, dst_beg);
|
|
|
|
if (dst_beg_up < dst_end_down) {
|
|
internal_memmove((u8 *)MemToShadow(dst_beg_up),
|
|
(u8 *)MemToShadow(src_beg_up),
|
|
(dst_end_down - dst_beg_up) / granularity);
|
|
}
|
|
|
|
if (copy_in_reversed_order)
|
|
CopyContainerFirstGranuleAnnotation(src_beg, dst_beg);
|
|
else
|
|
CopyContainerLastGranuleAnnotation(src_end, dst_end);
|
|
}
|
|
|
|
static const void *FindBadAddress(uptr begin, uptr end, bool poisoned) {
|
|
CHECK_LE(begin, end);
|
|
constexpr uptr kMaxRangeToCheck = 32;
|
|
if (end - begin > kMaxRangeToCheck * 2) {
|
|
if (auto *bad = FindBadAddress(begin, begin + kMaxRangeToCheck, poisoned))
|
|
return bad;
|
|
if (auto *bad = FindBadAddress(end - kMaxRangeToCheck, end, poisoned))
|
|
return bad;
|
|
}
|
|
|
|
for (uptr i = begin; i < end; ++i)
|
|
if (AddressIsPoisoned(i) != poisoned)
|
|
return reinterpret_cast<const void *>(i);
|
|
return nullptr;
|
|
}
|
|
|
|
const void *__sanitizer_contiguous_container_find_bad_address(
|
|
const void *beg_p, const void *mid_p, const void *end_p) {
|
|
if (!flags()->detect_container_overflow)
|
|
return nullptr;
|
|
uptr granularity = ASAN_SHADOW_GRANULARITY;
|
|
uptr beg = reinterpret_cast<uptr>(beg_p);
|
|
uptr end = reinterpret_cast<uptr>(end_p);
|
|
uptr mid = reinterpret_cast<uptr>(mid_p);
|
|
CHECK_LE(beg, mid);
|
|
CHECK_LE(mid, end);
|
|
// If the byte after the storage is unpoisoned, everything in the granule
|
|
// before must stay unpoisoned.
|
|
uptr annotations_end =
|
|
(!AddrIsAlignedByGranularity(end) && !AddressIsPoisoned(end))
|
|
? RoundDownTo(end, granularity)
|
|
: end;
|
|
beg = Min(beg, annotations_end);
|
|
mid = Min(mid, annotations_end);
|
|
if (auto *bad = FindBadAddress(beg, mid, false))
|
|
return bad;
|
|
if (auto *bad = FindBadAddress(mid, annotations_end, true))
|
|
return bad;
|
|
return FindBadAddress(annotations_end, end, false);
|
|
}
|
|
|
|
int __sanitizer_verify_contiguous_container(const void *beg_p,
|
|
const void *mid_p,
|
|
const void *end_p) {
|
|
return __sanitizer_contiguous_container_find_bad_address(beg_p, mid_p,
|
|
end_p) == nullptr;
|
|
}
|
|
|
|
const void *__sanitizer_double_ended_contiguous_container_find_bad_address(
|
|
const void *storage_beg_p, const void *container_beg_p,
|
|
const void *container_end_p, const void *storage_end_p) {
|
|
if (!flags()->detect_container_overflow)
|
|
return nullptr;
|
|
uptr granularity = ASAN_SHADOW_GRANULARITY;
|
|
uptr storage_beg = reinterpret_cast<uptr>(storage_beg_p);
|
|
uptr storage_end = reinterpret_cast<uptr>(storage_end_p);
|
|
uptr beg = reinterpret_cast<uptr>(container_beg_p);
|
|
uptr end = reinterpret_cast<uptr>(container_end_p);
|
|
|
|
// The prefix of the firs granule of the container is unpoisoned.
|
|
if (beg != end)
|
|
beg = Max(storage_beg, RoundDownTo(beg, granularity));
|
|
|
|
// If the byte after the storage is unpoisoned, the prefix of the last granule
|
|
// is unpoisoned.
|
|
uptr annotations_end = (!AddrIsAlignedByGranularity(storage_end) &&
|
|
!AddressIsPoisoned(storage_end))
|
|
? RoundDownTo(storage_end, granularity)
|
|
: storage_end;
|
|
storage_beg = Min(storage_beg, annotations_end);
|
|
beg = Min(beg, annotations_end);
|
|
end = Min(end, annotations_end);
|
|
|
|
if (auto *bad = FindBadAddress(storage_beg, beg, true))
|
|
return bad;
|
|
if (auto *bad = FindBadAddress(beg, end, false))
|
|
return bad;
|
|
if (auto *bad = FindBadAddress(end, annotations_end, true))
|
|
return bad;
|
|
return FindBadAddress(annotations_end, storage_end, false);
|
|
}
|
|
|
|
int __sanitizer_verify_double_ended_contiguous_container(
|
|
const void *storage_beg_p, const void *container_beg_p,
|
|
const void *container_end_p, const void *storage_end_p) {
|
|
return __sanitizer_double_ended_contiguous_container_find_bad_address(
|
|
storage_beg_p, container_beg_p, container_end_p, storage_end_p) ==
|
|
nullptr;
|
|
}
|
|
|
|
extern "C" SANITIZER_INTERFACE_ATTRIBUTE
|
|
void __asan_poison_intra_object_redzone(uptr ptr, uptr size) {
|
|
AsanPoisonOrUnpoisonIntraObjectRedzone(ptr, size, true);
|
|
}
|
|
|
|
extern "C" SANITIZER_INTERFACE_ATTRIBUTE
|
|
void __asan_unpoison_intra_object_redzone(uptr ptr, uptr size) {
|
|
AsanPoisonOrUnpoisonIntraObjectRedzone(ptr, size, false);
|
|
}
|
|
|
|
// --- Implementation of LSan-specific functions --- {{{1
|
|
namespace __lsan {
|
|
bool WordIsPoisoned(uptr addr) {
|
|
return (__asan_region_is_poisoned(addr, sizeof(uptr)) != 0);
|
|
}
|
|
}
|