// Protocol Buffers - Google's data interchange format // Copyright 2023 Google LLC. All rights reserved. // // Use of this source code is governed by a BSD-style // license that can be found in the LICENSE file or at // https://developers.google.com/open-source/licenses/bsd #ifndef UPB_MEM_INTERNAL_ARENA_H_ #define UPB_MEM_INTERNAL_ARENA_H_ #include #include #include #include "upb/port/sanitizers.h" // Must be last. #include "upb/port/def.inc" // This is QUITE an ugly hack, which specifies the number of pointers needed // to equal (or exceed) the storage required for one upb_Arena. // // We need this because the decoder inlines a upb_Arena for performance but // the full struct is not visible outside of arena.c. Yes, I know, it's awful. #ifndef NDEBUG #define UPB_ARENA_BASE_SIZE_HACK 10 #else #define UPB_ARENA_BASE_SIZE_HACK 9 #endif #define UPB_ARENA_SIZE_HACK \ (sizeof(void*) * (UPB_ARENA_BASE_SIZE_HACK + (UPB_XSAN_STRUCT_SIZE * 2))) + \ (sizeof(uint32_t) * 2) // LINT.IfChange(upb_Arena) struct upb_Arena { char* UPB_ONLYBITS(ptr); const UPB_NODEREF char* UPB_ONLYBITS(end); UPB_XSAN_MEMBER }; // LINT.ThenChange(//depot/google3/third_party/upb/bits/typescript/arena.ts:upb_Arena) #ifdef __cplusplus extern "C" { #endif void UPB_PRIVATE(_upb_Arena_SwapIn)(struct upb_Arena* des, const struct upb_Arena* src); void UPB_PRIVATE(_upb_Arena_SwapOut)(struct upb_Arena* des, const struct upb_Arena* src); UPB_INLINE size_t UPB_PRIVATE(_upb_ArenaHas)(const struct upb_Arena* a) { return (size_t)(a->UPB_ONLYBITS(end) - a->UPB_ONLYBITS(ptr)); } UPB_INLINE size_t UPB_PRIVATE(_upb_Arena_AllocSpan)(size_t size) { return UPB_ALIGN_MALLOC(size) + UPB_PRIVATE(kUpb_Asan_GuardSize); } UPB_INLINE bool UPB_PRIVATE(_upb_Arena_WasLastAllocFromCurrentBlock)( const struct upb_Arena* a, void* ptr, size_t size) { return UPB_PRIVATE(upb_Xsan_PtrEq)( (char*)ptr + UPB_PRIVATE(_upb_Arena_AllocSpan)(size), a->UPB_ONLYBITS(ptr)); } UPB_INLINE bool UPB_PRIVATE(_upb_Arena_IsAligned)(const void* ptr) { return (uintptr_t)ptr % UPB_MALLOC_ALIGN == 0; } UPB_API_INLINE void* upb_Arena_Malloc(struct upb_Arena* a, size_t size) { UPB_PRIVATE(upb_Xsan_AccessReadWrite)(UPB_XSAN(a)); size_t span = UPB_PRIVATE(_upb_Arena_AllocSpan)(size); if (UPB_UNLIKELY(UPB_PRIVATE(_upb_ArenaHas)(a) < span)) { void* UPB_PRIVATE(_upb_Arena_SlowMalloc)(struct upb_Arena * a, size_t size); return UPB_PRIVATE(_upb_Arena_SlowMalloc)(a, span); } // We have enough space to do a fast malloc. void* ret = a->UPB_ONLYBITS(ptr); a->UPB_ONLYBITS(ptr) += span; UPB_ASSERT(UPB_PRIVATE(_upb_Arena_IsAligned)(ret)); UPB_ASSERT(UPB_PRIVATE(_upb_Arena_IsAligned)(a->UPB_ONLYBITS(ptr))); return UPB_PRIVATE(upb_Xsan_NewUnpoisonedRegion)(UPB_XSAN(a), ret, size); } UPB_API_INLINE void upb_Arena_ShrinkLast(struct upb_Arena* a, void* ptr, size_t oldsize, size_t size) { UPB_ASSERT(ptr); UPB_ASSERT(size <= oldsize); UPB_PRIVATE(upb_Xsan_AccessReadWrite)(UPB_XSAN(a)); UPB_PRIVATE(upb_Xsan_ResizeUnpoisonedRegion)(ptr, oldsize, size); if (UPB_PRIVATE(_upb_Arena_WasLastAllocFromCurrentBlock)(a, ptr, oldsize)) { // We can reclaim some memory. a->UPB_ONLYBITS(ptr) -= UPB_ALIGN_MALLOC(oldsize) - UPB_ALIGN_MALLOC(size); } else { // We can't reclaim any memory, but we need to verify that `ptr` really // does represent the most recent allocation. #ifndef NDEBUG bool _upb_Arena_WasLastAlloc(struct upb_Arena * a, void* ptr, size_t oldsize); UPB_ASSERT(_upb_Arena_WasLastAlloc(a, ptr, oldsize)); #endif } } UPB_API_INLINE bool upb_Arena_TryExtend(struct upb_Arena* a, void* ptr, size_t oldsize, size_t size) { UPB_ASSERT(ptr); UPB_ASSERT(size > oldsize); UPB_PRIVATE(upb_Xsan_AccessReadWrite)(UPB_XSAN(a)); size_t extend = UPB_ALIGN_MALLOC(size) - UPB_ALIGN_MALLOC(oldsize); if (UPB_PRIVATE(_upb_Arena_WasLastAllocFromCurrentBlock)(a, ptr, oldsize) && UPB_PRIVATE(_upb_ArenaHas)(a) >= extend) { a->UPB_ONLYBITS(ptr) += extend; UPB_PRIVATE(upb_Xsan_ResizeUnpoisonedRegion)(ptr, oldsize, size); return true; } return false; } UPB_API_INLINE void* upb_Arena_Realloc(struct upb_Arena* a, void* ptr, size_t oldsize, size_t size) { UPB_PRIVATE(upb_Xsan_AccessReadWrite)(UPB_XSAN(a)); void* ret; if (ptr && (size <= oldsize || upb_Arena_TryExtend(a, ptr, oldsize, size))) { // We can extend or shrink in place. if (size <= oldsize && UPB_PRIVATE(_upb_Arena_WasLastAllocFromCurrentBlock)(a, ptr, oldsize)) { upb_Arena_ShrinkLast(a, ptr, oldsize, size); } ret = ptr; } else { // We need to copy into a new allocation. ret = upb_Arena_Malloc(a, size); if (ret && oldsize > 0) { memcpy(ret, ptr, UPB_MIN(oldsize, size)); } } if (ret) { // We want to invalidate pointers to the old region if hwasan is enabled, so // we poison and unpoison even if ptr == ret. However, if reallocation fails // we do not want to poison the old memory, or attempt to poison null. UPB_PRIVATE(upb_Xsan_PoisonRegion)(ptr, oldsize); return UPB_PRIVATE(upb_Xsan_NewUnpoisonedRegion)(UPB_XSAN(a), ret, size); } return ret; } #ifdef __cplusplus } /* extern "C" */ #endif #include "upb/port/undef.inc" #endif /* UPB_MEM_INTERNAL_ARENA_H_ */