mirror of
git://source.winehq.org/git/wine.git
synced 2024-11-01 08:13:18 +00:00
2168 lines
74 KiB
C
2168 lines
74 KiB
C
/*
|
|
* Win32 heap functions
|
|
*
|
|
* Copyright 1996 Alexandre Julliard
|
|
* Copyright 1998 Ulrich Weigand
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with this library; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301, USA
|
|
*/
|
|
|
|
#include <assert.h>
|
|
#include <stdlib.h>
|
|
#include <stdarg.h>
|
|
#include <stdio.h>
|
|
#include <string.h>
|
|
|
|
#define RUNNING_ON_VALGRIND 0 /* FIXME */
|
|
|
|
#include "ntstatus.h"
|
|
#define WIN32_NO_STATUS
|
|
#define NONAMELESSUNION
|
|
#include "windef.h"
|
|
#include "winnt.h"
|
|
#include "winternl.h"
|
|
#include "ntdll_misc.h"
|
|
#include "wine/list.h"
|
|
#include "wine/debug.h"
|
|
|
|
WINE_DEFAULT_DEBUG_CHANNEL(heap);
|
|
|
|
/* undocumented RtlWalkHeap structure */
|
|
|
|
struct rtl_heap_entry
|
|
{
|
|
LPVOID lpData;
|
|
SIZE_T cbData; /* differs from PROCESS_HEAP_ENTRY */
|
|
BYTE cbOverhead;
|
|
BYTE iRegionIndex;
|
|
WORD wFlags; /* value differs from PROCESS_HEAP_ENTRY */
|
|
union {
|
|
struct {
|
|
HANDLE hMem;
|
|
DWORD dwReserved[3];
|
|
} Block;
|
|
struct {
|
|
DWORD dwCommittedSize;
|
|
DWORD dwUnCommittedSize;
|
|
LPVOID lpFirstBlock;
|
|
LPVOID lpLastBlock;
|
|
} Region;
|
|
};
|
|
};
|
|
|
|
/* rtl_heap_entry flags, names made up */
|
|
|
|
#define RTL_HEAP_ENTRY_BUSY 0x0001
|
|
#define RTL_HEAP_ENTRY_REGION 0x0002
|
|
#define RTL_HEAP_ENTRY_BLOCK 0x0010
|
|
#define RTL_HEAP_ENTRY_UNCOMMITTED 0x1000
|
|
#define RTL_HEAP_ENTRY_COMMITTED 0x4000
|
|
#define RTL_HEAP_ENTRY_LFH 0x8000
|
|
|
|
|
|
/* header for heap blocks */
|
|
|
|
#define REGION_ALIGN 0x10000
|
|
#define BLOCK_ALIGN (2 * sizeof(void *))
|
|
|
|
struct block
|
|
{
|
|
WORD block_size; /* block size in multiple of BLOCK_ALIGN */
|
|
BYTE block_flags;
|
|
BYTE tail_size; /* unused size (used block) / high size bits (free block) */
|
|
WORD base_offset; /* offset to the region base in multiple of REGION_ALIGN */
|
|
WORD block_type;
|
|
};
|
|
|
|
C_ASSERT( sizeof(struct block) == 8 );
|
|
|
|
/* block specific flags */
|
|
|
|
#define BLOCK_FLAG_FREE 0x01
|
|
#define BLOCK_FLAG_PREV_FREE 0x02
|
|
#define BLOCK_FLAG_FREE_LINK 0x03
|
|
#define BLOCK_FLAG_LARGE 0x04
|
|
#define BLOCK_FLAG_USER_INFO 0x10 /* user flags up to 0xf0 */
|
|
#define BLOCK_FLAG_USER_MASK 0xf0
|
|
|
|
#define BLOCK_USER_FLAGS( heap_flags ) (((heap_flags) >> 4) & BLOCK_FLAG_USER_MASK)
|
|
#define HEAP_USER_FLAGS( block_flags ) (((block_flags) & BLOCK_FLAG_USER_MASK) << 4)
|
|
|
|
/* entry to link free blocks in free lists */
|
|
|
|
struct DECLSPEC_ALIGN(BLOCK_ALIGN) entry
|
|
{
|
|
struct block block;
|
|
struct list entry;
|
|
};
|
|
|
|
C_ASSERT( sizeof(struct entry) == 2 * BLOCK_ALIGN );
|
|
|
|
typedef struct
|
|
{
|
|
SIZE_T __pad[sizeof(SIZE_T) / sizeof(DWORD)];
|
|
struct list entry; /* entry in heap large blocks list */
|
|
SIZE_T data_size; /* size of user data */
|
|
SIZE_T block_size; /* total size of virtual memory block */
|
|
void *user_value;
|
|
struct block block;
|
|
} ARENA_LARGE;
|
|
|
|
/* block must be last and aligned */
|
|
C_ASSERT( sizeof(ARENA_LARGE) == offsetof(ARENA_LARGE, block) + sizeof(struct block) );
|
|
C_ASSERT( sizeof(ARENA_LARGE) == 4 * BLOCK_ALIGN );
|
|
|
|
#define BLOCK_TYPE_USED 0x5355
|
|
#define BLOCK_TYPE_DEAD 0xdead
|
|
#define BLOCK_TYPE_FREE 0x5246
|
|
#define BLOCK_TYPE_LARGE 0x614c
|
|
|
|
#define BLOCK_FILL_USED 0x55
|
|
#define BLOCK_FILL_TAIL 0xab
|
|
#define BLOCK_FILL_FREE 0xfeeefeee
|
|
|
|
#define ROUND_ADDR(addr, mask) ((void *)((UINT_PTR)(addr) & ~(UINT_PTR)(mask)))
|
|
#define ROUND_SIZE(size, mask) ((((SIZE_T)(size) + (mask)) & ~(SIZE_T)(mask)))
|
|
#define FIELD_MAX(type, field) (((SIZE_T)1 << (sizeof(((type *)0)->field) * 8)) - 1)
|
|
|
|
#define HEAP_MIN_BLOCK_SIZE ROUND_SIZE(sizeof(struct entry) + BLOCK_ALIGN, BLOCK_ALIGN - 1)
|
|
|
|
C_ASSERT( sizeof(struct block) <= HEAP_MIN_BLOCK_SIZE );
|
|
C_ASSERT( sizeof(struct entry) <= HEAP_MIN_BLOCK_SIZE );
|
|
|
|
/* used block size is coded into block_size */
|
|
#define HEAP_MAX_USED_BLOCK_SIZE (FIELD_MAX( struct block, block_size ) * BLOCK_ALIGN)
|
|
/* free block size is coded into block_size + tail_size */
|
|
#define HEAP_MAX_FREE_BLOCK_SIZE (HEAP_MAX_USED_BLOCK_SIZE + (FIELD_MAX( struct block, tail_size ) << 16) * BLOCK_ALIGN)
|
|
/* subheap distance is coded into base_offset */
|
|
#define HEAP_MAX_BLOCK_REGION_SIZE (FIELD_MAX( struct block, base_offset ) * REGION_ALIGN)
|
|
|
|
C_ASSERT( HEAP_MAX_USED_BLOCK_SIZE == 512 * 1024 * (sizeof(void *) / 4) - BLOCK_ALIGN );
|
|
C_ASSERT( HEAP_MAX_FREE_BLOCK_SIZE == 128 * 1024 * 1024 * (sizeof(void *) / 4) - BLOCK_ALIGN );
|
|
C_ASSERT( HEAP_MAX_BLOCK_REGION_SIZE >= HEAP_MAX_FREE_BLOCK_SIZE );
|
|
|
|
/* minimum size to start allocating large blocks */
|
|
#define HEAP_MIN_LARGE_BLOCK_SIZE (HEAP_MAX_USED_BLOCK_SIZE - 0x1000)
|
|
|
|
/* There will be a free list bucket for every arena size up to and including this value */
|
|
#define HEAP_MAX_SMALL_FREE_LIST 0x100
|
|
C_ASSERT( HEAP_MAX_SMALL_FREE_LIST % BLOCK_ALIGN == 0 );
|
|
#define HEAP_NB_SMALL_FREE_LISTS (((HEAP_MAX_SMALL_FREE_LIST - HEAP_MIN_BLOCK_SIZE) / BLOCK_ALIGN) + 1)
|
|
|
|
/* Max size of the blocks on the free lists above HEAP_MAX_SMALL_FREE_LIST */
|
|
static const SIZE_T free_list_sizes[] =
|
|
{
|
|
0x200, 0x400, 0x1000, ~(SIZE_T)0
|
|
};
|
|
#define HEAP_NB_FREE_LISTS (ARRAY_SIZE(free_list_sizes) + HEAP_NB_SMALL_FREE_LISTS)
|
|
|
|
typedef struct DECLSPEC_ALIGN(BLOCK_ALIGN) tagSUBHEAP
|
|
{
|
|
SIZE_T __pad[sizeof(SIZE_T) / sizeof(DWORD)];
|
|
SIZE_T block_size;
|
|
SIZE_T data_size;
|
|
struct list entry;
|
|
void *user_value;
|
|
struct block block;
|
|
} SUBHEAP;
|
|
|
|
/* block must be last and aligned */
|
|
C_ASSERT( sizeof(SUBHEAP) == offsetof(SUBHEAP, block) + sizeof(struct block) );
|
|
C_ASSERT( sizeof(SUBHEAP) == 4 * BLOCK_ALIGN );
|
|
|
|
struct heap
|
|
{ /* win32/win64 */
|
|
DWORD_PTR unknown1[2]; /* 0000/0000 */
|
|
DWORD ffeeffee; /* 0008/0010 */
|
|
DWORD auto_flags; /* 000c/0014 */
|
|
DWORD_PTR unknown2[7]; /* 0010/0018 */
|
|
DWORD unknown3[2]; /* 002c/0050 */
|
|
DWORD_PTR unknown4[3]; /* 0034/0058 */
|
|
DWORD flags; /* 0040/0070 */
|
|
DWORD force_flags; /* 0044/0074 */
|
|
/* end of the Windows 10 compatible struct layout */
|
|
|
|
struct list entry; /* Entry in process heap list */
|
|
struct list subheap_list; /* Sub-heap list */
|
|
struct list large_list; /* Large blocks list */
|
|
SIZE_T grow_size; /* Size of next subheap for growing heap */
|
|
SIZE_T min_size; /* Minimum committed size */
|
|
DWORD magic; /* Magic number */
|
|
DWORD pending_pos; /* Position in pending free requests ring */
|
|
struct block **pending_free; /* Ring buffer for pending free requests */
|
|
RTL_CRITICAL_SECTION cs;
|
|
struct entry free_lists[HEAP_NB_FREE_LISTS];
|
|
SUBHEAP subheap;
|
|
};
|
|
|
|
/* subheap must be last and aligned */
|
|
C_ASSERT( sizeof(struct heap) == offsetof(struct heap, subheap) + sizeof(SUBHEAP) );
|
|
C_ASSERT( sizeof(struct heap) % BLOCK_ALIGN == 0 );
|
|
C_ASSERT( offsetof(struct heap, subheap) <= REGION_ALIGN - 1 );
|
|
|
|
#define HEAP_MAGIC ((DWORD)('H' | ('E'<<8) | ('A'<<16) | ('P'<<24)))
|
|
|
|
#define HEAP_DEF_SIZE (0x40000 * BLOCK_ALIGN)
|
|
#define MAX_FREE_PENDING 1024 /* max number of free requests to delay */
|
|
|
|
/* some undocumented flags (names are made up) */
|
|
#define HEAP_PRIVATE 0x00001000
|
|
#define HEAP_ADD_USER_INFO 0x00000100
|
|
#define HEAP_USER_FLAGS_MASK 0x00000f00
|
|
#define HEAP_PAGE_ALLOCS 0x01000000
|
|
#define HEAP_VALIDATE 0x10000000
|
|
#define HEAP_VALIDATE_ALL 0x20000000
|
|
#define HEAP_VALIDATE_PARAMS 0x40000000
|
|
#define HEAP_CHECKING_ENABLED 0x80000000
|
|
|
|
static struct heap *process_heap; /* main process heap */
|
|
|
|
/* check if memory range a contains memory range b */
|
|
static inline BOOL contains( const void *a, SIZE_T a_size, const void *b, SIZE_T b_size )
|
|
{
|
|
const void *a_end = (char *)a + a_size, *b_end = (char *)b + b_size;
|
|
return a <= b && b <= b_end && b_end <= a_end;
|
|
}
|
|
|
|
static inline UINT block_get_flags( const struct block *block )
|
|
{
|
|
return block->block_flags;
|
|
}
|
|
|
|
static inline UINT block_get_type( const struct block *block )
|
|
{
|
|
return block->block_type;
|
|
}
|
|
|
|
static inline void block_set_base( struct block *block, const void *base )
|
|
{
|
|
const char *offset = ROUND_ADDR( block, REGION_ALIGN - 1 );
|
|
block->base_offset = (offset - (char *)base) / REGION_ALIGN;
|
|
}
|
|
|
|
static inline void block_set_type( struct block *block, UINT type )
|
|
{
|
|
block->block_type = type;
|
|
}
|
|
|
|
static inline SUBHEAP *block_get_subheap( const struct heap *heap, const struct block *block )
|
|
{
|
|
char *offset = ROUND_ADDR( block, REGION_ALIGN - 1 );
|
|
void *base = offset - block->base_offset * REGION_ALIGN;
|
|
if (base != (void *)heap) return base;
|
|
else return (SUBHEAP *)&heap->subheap;
|
|
}
|
|
|
|
static inline UINT block_get_overhead( const struct block *block )
|
|
{
|
|
if (block_get_flags( block ) & BLOCK_FLAG_FREE) return sizeof(*block) + sizeof(struct list);
|
|
return sizeof(*block) + block->tail_size;
|
|
}
|
|
|
|
/* return the size of a block, including its header */
|
|
static inline UINT block_get_size( const struct block *block )
|
|
{
|
|
UINT block_size = block->block_size;
|
|
if (block_get_flags( block ) & BLOCK_FLAG_FREE) block_size += (UINT)block->tail_size << 16;
|
|
return block_size * BLOCK_ALIGN;
|
|
}
|
|
|
|
static inline void block_set_size( struct block *block, UINT block_size )
|
|
{
|
|
block_size /= BLOCK_ALIGN;
|
|
if (block_get_flags( block ) & BLOCK_FLAG_FREE) block->tail_size = block_size >> 16;
|
|
block->block_size = block_size;
|
|
}
|
|
|
|
static inline void block_set_flags( struct block *block, BYTE clear, BYTE set )
|
|
{
|
|
UINT block_size = block_get_size( block );
|
|
block->block_flags &= ~clear;
|
|
block->block_flags |= set;
|
|
block_set_size( block, block_size );
|
|
}
|
|
|
|
static inline void *subheap_base( const SUBHEAP *subheap )
|
|
{
|
|
return ROUND_ADDR( subheap, REGION_ALIGN - 1 );
|
|
}
|
|
|
|
static inline SIZE_T subheap_overhead( const SUBHEAP *subheap )
|
|
{
|
|
return (char *)&subheap->block - (char *)subheap_base( subheap );
|
|
}
|
|
|
|
static inline SIZE_T subheap_size( const SUBHEAP *subheap )
|
|
{
|
|
return subheap->block_size + subheap_overhead( subheap );
|
|
}
|
|
|
|
static inline const void *subheap_commit_end( const SUBHEAP *subheap )
|
|
{
|
|
return (char *)(subheap + 1) + subheap->data_size;
|
|
}
|
|
|
|
static void subheap_set_bounds( SUBHEAP *subheap, char *commit_end, char *end )
|
|
{
|
|
subheap->block_size = end - (char *)&subheap->block;
|
|
subheap->data_size = commit_end - (char *)(subheap + 1);
|
|
}
|
|
|
|
static inline void *first_block( const SUBHEAP *subheap )
|
|
{
|
|
return (void *)&subheap->block;
|
|
}
|
|
|
|
static inline const void *last_block( const SUBHEAP *subheap )
|
|
{
|
|
return (char *)subheap_commit_end( subheap ) - sizeof(struct block);
|
|
}
|
|
|
|
static inline struct block *next_block( const SUBHEAP *subheap, const struct block *block )
|
|
{
|
|
const char *data = (char *)(block + 1), *next, *last = last_block( subheap );
|
|
next = (char *)block + block_get_size( block );
|
|
if (!contains( data, last - (char *)data, next, sizeof(*block) )) return NULL;
|
|
return (struct block *)next;
|
|
}
|
|
|
|
static inline BOOL check_subheap( const SUBHEAP *subheap )
|
|
{
|
|
return contains( &subheap->block, subheap->block_size, subheap + 1, subheap->data_size );
|
|
}
|
|
|
|
static BOOL heap_validate( const struct heap *heap );
|
|
|
|
/* mark a block of memory as innacessible for debugging purposes */
|
|
static inline void valgrind_make_noaccess( void const *ptr, SIZE_T size )
|
|
{
|
|
#if defined(VALGRIND_MAKE_MEM_NOACCESS)
|
|
VALGRIND_DISCARD( VALGRIND_MAKE_MEM_NOACCESS( ptr, size ) );
|
|
#elif defined(VALGRIND_MAKE_NOACCESS)
|
|
VALGRIND_DISCARD( VALGRIND_MAKE_NOACCESS( ptr, size ) );
|
|
#endif
|
|
}
|
|
|
|
/* mark a block of memory as initialized for debugging purposes */
|
|
static inline void valgrind_make_readable( void const *ptr, SIZE_T size )
|
|
{
|
|
#if defined(VALGRIND_MAKE_MEM_DEFINED)
|
|
VALGRIND_DISCARD( VALGRIND_MAKE_MEM_DEFINED( ptr, size ) );
|
|
#elif defined(VALGRIND_MAKE_READABLE)
|
|
VALGRIND_DISCARD( VALGRIND_MAKE_READABLE( ptr, size ) );
|
|
#endif
|
|
}
|
|
|
|
/* mark a block of memory as uninitialized for debugging purposes */
|
|
static inline void valgrind_make_writable( void const *ptr, SIZE_T size )
|
|
{
|
|
#if defined(VALGRIND_MAKE_MEM_UNDEFINED)
|
|
VALGRIND_DISCARD( VALGRIND_MAKE_MEM_UNDEFINED( ptr, size ) );
|
|
#elif defined(VALGRIND_MAKE_WRITABLE)
|
|
VALGRIND_DISCARD( VALGRIND_MAKE_WRITABLE( ptr, size ) );
|
|
#endif
|
|
}
|
|
|
|
/* mark a block of memory as free for debugging purposes */
|
|
static inline void mark_block_free( void *ptr, SIZE_T size, DWORD flags )
|
|
{
|
|
if (flags & HEAP_FREE_CHECKING_ENABLED)
|
|
{
|
|
SIZE_T i;
|
|
for (i = 0; i < size / sizeof(DWORD); i++) ((DWORD *)ptr)[i] = BLOCK_FILL_FREE;
|
|
}
|
|
valgrind_make_noaccess( ptr, size );
|
|
}
|
|
|
|
/* mark a block of memory as a tail block */
|
|
static inline void mark_block_tail( struct block *block, DWORD flags )
|
|
{
|
|
char *tail = (char *)block + block_get_size( block ) - block->tail_size;
|
|
if (flags & HEAP_TAIL_CHECKING_ENABLED)
|
|
{
|
|
valgrind_make_writable( tail, BLOCK_ALIGN );
|
|
memset( tail, BLOCK_FILL_TAIL, BLOCK_ALIGN );
|
|
}
|
|
valgrind_make_noaccess( tail, BLOCK_ALIGN );
|
|
if (flags & HEAP_ADD_USER_INFO)
|
|
{
|
|
if (flags & HEAP_TAIL_CHECKING_ENABLED || RUNNING_ON_VALGRIND) tail += BLOCK_ALIGN;
|
|
valgrind_make_writable( tail + sizeof(void *), sizeof(void *) );
|
|
memset( tail + sizeof(void *), 0, sizeof(void *) );
|
|
}
|
|
}
|
|
|
|
/* initialize contents of a newly created block of memory */
|
|
static inline void initialize_block( struct block *block, SIZE_T old_size, SIZE_T size, DWORD flags )
|
|
{
|
|
char *data = (char *)(block + 1);
|
|
|
|
if (size <= old_size) return;
|
|
|
|
if (flags & HEAP_ZERO_MEMORY)
|
|
{
|
|
valgrind_make_writable( data + old_size, size - old_size );
|
|
memset( data + old_size, 0, size - old_size );
|
|
}
|
|
else if (flags & HEAP_FREE_CHECKING_ENABLED)
|
|
{
|
|
valgrind_make_writable( data + old_size, size - old_size );
|
|
memset( data + old_size, BLOCK_FILL_USED, size - old_size );
|
|
}
|
|
}
|
|
|
|
/* notify that a new block of memory has been allocated for debugging purposes */
|
|
static inline void valgrind_notify_alloc( void const *ptr, SIZE_T size, BOOL init )
|
|
{
|
|
#ifdef VALGRIND_MALLOCLIKE_BLOCK
|
|
VALGRIND_MALLOCLIKE_BLOCK( ptr, size, 0, init );
|
|
#endif
|
|
}
|
|
|
|
/* notify that a block of memory has been freed for debugging purposes */
|
|
static inline void valgrind_notify_free( void const *ptr )
|
|
{
|
|
#ifdef VALGRIND_FREELIKE_BLOCK
|
|
VALGRIND_FREELIKE_BLOCK( ptr, 0 );
|
|
#endif
|
|
}
|
|
|
|
static inline void valgrind_notify_resize( void const *ptr, SIZE_T size_old, SIZE_T size_new )
|
|
{
|
|
#ifdef VALGRIND_RESIZEINPLACE_BLOCK
|
|
/* zero is not a valid size */
|
|
VALGRIND_RESIZEINPLACE_BLOCK( ptr, size_old ? size_old : 1, size_new ? size_new : 1, 0 );
|
|
#endif
|
|
}
|
|
|
|
static void valgrind_notify_free_all( SUBHEAP *subheap )
|
|
{
|
|
#ifdef VALGRIND_FREELIKE_BLOCK
|
|
struct block *block;
|
|
|
|
if (!RUNNING_ON_VALGRIND) return;
|
|
if (!check_subheap( subheap )) return;
|
|
|
|
for (block = first_block( subheap ); block; block = next_block( subheap, block ))
|
|
{
|
|
if (block_get_flags( block ) & BLOCK_FLAG_FREE) continue;
|
|
if (block_get_type( block ) == BLOCK_TYPE_USED) valgrind_notify_free( block + 1 );
|
|
}
|
|
#endif
|
|
}
|
|
|
|
/* locate a free list entry of the appropriate size */
|
|
/* size is the size of the whole block including the arena header */
|
|
static inline struct entry *find_free_list( struct heap *heap, SIZE_T block_size, BOOL last )
|
|
{
|
|
struct entry *list, *end = heap->free_lists + ARRAY_SIZE(heap->free_lists);
|
|
unsigned int i;
|
|
|
|
if (block_size <= HEAP_MAX_SMALL_FREE_LIST)
|
|
i = (block_size - HEAP_MIN_BLOCK_SIZE) / BLOCK_ALIGN;
|
|
else for (i = HEAP_NB_SMALL_FREE_LISTS; i < HEAP_NB_FREE_LISTS - 1; i++)
|
|
if (block_size <= free_list_sizes[i - HEAP_NB_SMALL_FREE_LISTS]) break;
|
|
|
|
list = heap->free_lists + i;
|
|
if (last && ++list == end) list = heap->free_lists;
|
|
return list;
|
|
}
|
|
|
|
/* get the memory protection type to use for a given heap */
|
|
static inline ULONG get_protection_type( DWORD flags )
|
|
{
|
|
return (flags & HEAP_CREATE_ENABLE_EXECUTE) ? PAGE_EXECUTE_READWRITE : PAGE_READWRITE;
|
|
}
|
|
|
|
static RTL_CRITICAL_SECTION_DEBUG process_heap_cs_debug =
|
|
{
|
|
0, 0, NULL, /* will be set later */
|
|
{ &process_heap_cs_debug.ProcessLocksList, &process_heap_cs_debug.ProcessLocksList },
|
|
0, 0, { (DWORD_PTR)(__FILE__ ": main process heap section") }
|
|
};
|
|
|
|
static inline ULONG heap_get_flags( const struct heap *heap, ULONG flags )
|
|
{
|
|
if (flags & (HEAP_TAIL_CHECKING_ENABLED | HEAP_FREE_CHECKING_ENABLED)) flags |= HEAP_CHECKING_ENABLED;
|
|
flags &= HEAP_GENERATE_EXCEPTIONS | HEAP_NO_SERIALIZE | HEAP_ZERO_MEMORY | HEAP_REALLOC_IN_PLACE_ONLY | HEAP_CHECKING_ENABLED | HEAP_USER_FLAGS_MASK;
|
|
return heap->flags | flags;
|
|
}
|
|
|
|
static inline void heap_lock( struct heap *heap, ULONG flags )
|
|
{
|
|
if (flags & HEAP_NO_SERIALIZE) return;
|
|
RtlEnterCriticalSection( &heap->cs );
|
|
}
|
|
|
|
static inline void heap_unlock( struct heap *heap, ULONG flags )
|
|
{
|
|
if (flags & HEAP_NO_SERIALIZE) return;
|
|
RtlLeaveCriticalSection( &heap->cs );
|
|
}
|
|
|
|
static void heap_set_status( const struct heap *heap, ULONG flags, NTSTATUS status )
|
|
{
|
|
if (status == STATUS_NO_MEMORY && (flags & HEAP_GENERATE_EXCEPTIONS)) RtlRaiseStatus( status );
|
|
if (status) RtlSetLastWin32ErrorAndNtStatusFromNtStatus( status );
|
|
}
|
|
|
|
static size_t get_free_list_block_size( unsigned int index )
|
|
{
|
|
if (index < HEAP_NB_SMALL_FREE_LISTS) return HEAP_MIN_BLOCK_SIZE + index * BLOCK_ALIGN;
|
|
return free_list_sizes[index - HEAP_NB_SMALL_FREE_LISTS];
|
|
}
|
|
|
|
static void heap_dump( const struct heap *heap )
|
|
{
|
|
const struct block *block;
|
|
const ARENA_LARGE *large;
|
|
const SUBHEAP *subheap;
|
|
unsigned int i;
|
|
|
|
TRACE( "heap: %p\n", heap );
|
|
TRACE( " next %p\n", LIST_ENTRY( heap->entry.next, struct heap, entry ) );
|
|
|
|
TRACE( " free_lists: %p\n", heap->free_lists );
|
|
for (i = 0; i < HEAP_NB_FREE_LISTS; i++)
|
|
TRACE( " %p: size %#8Ix, prev %p, next %p\n", heap->free_lists + i, get_free_list_block_size( i ),
|
|
LIST_ENTRY( heap->free_lists[i].entry.prev, struct entry, entry ),
|
|
LIST_ENTRY( heap->free_lists[i].entry.next, struct entry, entry ) );
|
|
|
|
TRACE( " subheaps: %p\n", &heap->subheap_list );
|
|
LIST_FOR_EACH_ENTRY( subheap, &heap->subheap_list, SUBHEAP, entry )
|
|
{
|
|
SIZE_T free_size = 0, used_size = 0, overhead = 0;
|
|
const char *base = subheap_base( subheap );
|
|
|
|
TRACE( " %p: base %p first %p last %p end %p\n", subheap, base, first_block( subheap ),
|
|
last_block( subheap ), base + subheap_size( subheap ) );
|
|
|
|
if (!check_subheap( subheap )) return;
|
|
|
|
overhead += subheap_overhead( subheap );
|
|
for (block = first_block( subheap ); block; block = next_block( subheap, block ))
|
|
{
|
|
if (block_get_flags( block ) & BLOCK_FLAG_FREE)
|
|
{
|
|
TRACE( " %p: (free) type %#10x, size %#8x, flags %#4x, prev %p, next %p\n", block,
|
|
block_get_type( block ), block_get_size( block ), block_get_flags( block ),
|
|
LIST_ENTRY( ((struct entry *)block)->entry.prev, struct entry, entry ),
|
|
LIST_ENTRY( ((struct entry *)block)->entry.next, struct entry, entry ) );
|
|
|
|
overhead += block_get_overhead( block );
|
|
free_size += block_get_size( block ) - block_get_overhead( block );
|
|
}
|
|
else
|
|
{
|
|
TRACE( " %p: (used) type %#10x, size %#8x, flags %#4x, unused %#4x", block,
|
|
block_get_type( block ), block_get_size( block ), block_get_flags( block ),
|
|
block->tail_size );
|
|
if (!(block_get_flags( block ) & BLOCK_FLAG_PREV_FREE)) TRACE( "\n" );
|
|
else TRACE( ", back %p\n", *((struct block **)block - 1) );
|
|
|
|
overhead += block_get_overhead( block );
|
|
used_size += block_get_size( block ) - block_get_overhead( block );
|
|
}
|
|
}
|
|
|
|
TRACE( " total %#Ix, used %#Ix, free %#Ix, overhead %#Ix (%Iu%%)\n", used_size + free_size + overhead,
|
|
used_size, free_size, overhead, (overhead * 100) / subheap_size( subheap ) );
|
|
}
|
|
|
|
TRACE( " large blocks: %p\n", &heap->large_list );
|
|
LIST_FOR_EACH_ENTRY( large, &heap->large_list, ARENA_LARGE, entry )
|
|
{
|
|
TRACE( " %p: (large) type %#10x, size %#8x, flags %#4x, total_size %#10Ix, alloc_size %#10Ix, prev %p, next %p\n",
|
|
large, block_get_type( &large->block ), block_get_size( &large->block ), block_get_flags( &large->block ), large->block_size,
|
|
large->data_size, LIST_ENTRY( large->entry.prev, ARENA_LARGE, entry ), LIST_ENTRY( large->entry.next, ARENA_LARGE, entry ) );
|
|
}
|
|
|
|
if (heap->pending_free)
|
|
{
|
|
TRACE( " pending blocks: %p\n", heap->pending_free );
|
|
for (i = 0; i < MAX_FREE_PENDING; ++i)
|
|
{
|
|
if (!(block = heap->pending_free[i])) break;
|
|
|
|
TRACE( " %c%p: (pend) type %#10x, size %#8x, flags %#4x, unused %#4x", i == heap->pending_pos ? '*' : ' ',
|
|
block, block_get_type( block ), block_get_size( block ), block_get_flags( block ), block->tail_size );
|
|
if (!(block_get_flags( block ) & BLOCK_FLAG_PREV_FREE)) TRACE( "\n" );
|
|
else TRACE( ", back %p\n", *((struct block **)block - 1) );
|
|
}
|
|
}
|
|
}
|
|
|
|
static const char *debugstr_heap_entry( struct rtl_heap_entry *entry )
|
|
{
|
|
const char *str = wine_dbg_sprintf( "data %p, size %#Ix, overhead %#x, region %#x, flags %#x", entry->lpData,
|
|
entry->cbData, entry->cbOverhead, entry->iRegionIndex, entry->wFlags );
|
|
if (!(entry->wFlags & RTL_HEAP_ENTRY_REGION)) return str;
|
|
return wine_dbg_sprintf( "%s, commit %#lx, uncommit %#lx, first %p, last %p", str, entry->Region.dwCommittedSize,
|
|
entry->Region.dwUnCommittedSize, entry->Region.lpFirstBlock, entry->Region.lpLastBlock );
|
|
}
|
|
|
|
static struct heap *unsafe_heap_from_handle( HANDLE handle, ULONG flags, ULONG *heap_flags )
|
|
{
|
|
struct heap *heap = handle;
|
|
BOOL valid = TRUE;
|
|
|
|
if (!heap || (heap->magic != HEAP_MAGIC))
|
|
{
|
|
ERR( "Invalid handle %p!\n", handle );
|
|
return NULL;
|
|
}
|
|
if (heap->flags & HEAP_VALIDATE_ALL)
|
|
{
|
|
heap_lock( heap, 0 );
|
|
valid = heap_validate( heap );
|
|
heap_unlock( heap, 0 );
|
|
|
|
if (!valid && TRACE_ON(heap))
|
|
{
|
|
heap_dump( heap );
|
|
assert( FALSE );
|
|
}
|
|
}
|
|
|
|
*heap_flags = heap_get_flags( heap, flags );
|
|
return valid ? heap : NULL;
|
|
}
|
|
|
|
|
|
static SUBHEAP *find_subheap( const struct heap *heap, const struct block *block, BOOL heap_walk )
|
|
{
|
|
SUBHEAP *subheap;
|
|
|
|
LIST_FOR_EACH_ENTRY( subheap, &heap->subheap_list, SUBHEAP, entry )
|
|
{
|
|
SIZE_T blocks_size = (char *)last_block( subheap ) - (char *)first_block( subheap );
|
|
if (!check_subheap( subheap )) return NULL;
|
|
if (contains( first_block( subheap ), blocks_size, block, sizeof(*block) )) return subheap;
|
|
/* outside of blocks region, possible corruption or heap_walk */
|
|
if (contains( subheap_base( subheap ), subheap_size( subheap ), block, 0 )) return heap_walk ? subheap : NULL;
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
|
|
static inline BOOL subheap_commit( const struct heap *heap, SUBHEAP *subheap, const struct block *block, SIZE_T block_size )
|
|
{
|
|
const char *end = (char *)subheap_base( subheap ) + subheap_size( subheap ), *commit_end;
|
|
SIZE_T size;
|
|
void *addr;
|
|
|
|
commit_end = (char *)block + block_size + sizeof(struct entry);
|
|
commit_end = ROUND_ADDR( (char *)commit_end + REGION_ALIGN - 1, REGION_ALIGN - 1 );
|
|
|
|
if (commit_end > end) commit_end = end;
|
|
if (commit_end <= (char *)subheap_commit_end( subheap )) return TRUE;
|
|
|
|
addr = (void *)subheap_commit_end( subheap );
|
|
size = commit_end - (char *)addr;
|
|
|
|
if (NtAllocateVirtualMemory( NtCurrentProcess(), &addr, 0, &size, MEM_COMMIT,
|
|
get_protection_type( heap->flags ) ))
|
|
{
|
|
WARN( "Could not commit %#Ix bytes at %p for heap %p\n", size, addr, heap );
|
|
return FALSE;
|
|
}
|
|
|
|
subheap->data_size = (char *)commit_end - (char *)(subheap + 1);
|
|
return TRUE;
|
|
}
|
|
|
|
static inline BOOL subheap_decommit( const struct heap *heap, SUBHEAP *subheap, const void *commit_end )
|
|
{
|
|
char *base = subheap_base( subheap );
|
|
SIZE_T size;
|
|
void *addr;
|
|
|
|
commit_end = ROUND_ADDR( (char *)commit_end + REGION_ALIGN - 1, REGION_ALIGN - 1 );
|
|
if (subheap == &heap->subheap) commit_end = max( (char *)commit_end, (char *)base + heap->min_size );
|
|
if (commit_end >= subheap_commit_end( subheap )) return TRUE;
|
|
|
|
size = (char *)subheap_commit_end( subheap ) - (char *)commit_end;
|
|
addr = (void *)commit_end;
|
|
|
|
if (NtFreeVirtualMemory( NtCurrentProcess(), &addr, &size, MEM_DECOMMIT ))
|
|
{
|
|
WARN( "Could not decommit %#Ix bytes at %p for heap %p\n", size, addr, heap );
|
|
return FALSE;
|
|
}
|
|
|
|
subheap->data_size = (char *)commit_end - (char *)(subheap + 1);
|
|
return TRUE;
|
|
}
|
|
|
|
|
|
static void create_free_block( struct heap *heap, ULONG flags, SUBHEAP *subheap, struct block *block, SIZE_T block_size )
|
|
{
|
|
const char *end = (char *)block + block_size, *commit_end = subheap_commit_end( subheap );
|
|
struct entry *entry = (struct entry *)block, *list;
|
|
struct block *next;
|
|
|
|
valgrind_make_writable( block, sizeof(*entry) );
|
|
block_set_type( block, BLOCK_TYPE_FREE );
|
|
block_set_base( block, subheap_base( subheap ) );
|
|
block_set_flags( block, ~0, BLOCK_FLAG_FREE );
|
|
block_set_size( block, block_size );
|
|
|
|
/* If debugging, erase the freed block content */
|
|
|
|
if (end > commit_end) end = commit_end;
|
|
if (end > (char *)(entry + 1)) mark_block_free( entry + 1, end - (char *)(entry + 1), flags );
|
|
|
|
if ((next = next_block( subheap, block )) && (block_get_flags( next ) & BLOCK_FLAG_FREE))
|
|
{
|
|
/* merge with the next block if it is free */
|
|
struct entry *next_entry = (struct entry *)next;
|
|
list_remove( &next_entry->entry );
|
|
block_size += block_get_size( next );
|
|
block_set_size( block, block_size );
|
|
mark_block_free( next_entry, sizeof(*next_entry), flags );
|
|
}
|
|
|
|
if ((next = next_block( subheap, block )))
|
|
{
|
|
/* set the next block PREV_FREE flag and back pointer */
|
|
block_set_flags( next, 0, BLOCK_FLAG_PREV_FREE );
|
|
valgrind_make_writable( (struct block **)next - 1, sizeof(struct block *) );
|
|
*((struct block **)next - 1) = block;
|
|
}
|
|
|
|
list = find_free_list( heap, block_get_size( block ), !next );
|
|
if (!next) list_add_before( &list->entry, &entry->entry );
|
|
else list_add_after( &list->entry, &entry->entry );
|
|
}
|
|
|
|
|
|
static void free_used_block( struct heap *heap, ULONG flags, struct block *block )
|
|
{
|
|
struct entry *entry;
|
|
SIZE_T block_size;
|
|
SUBHEAP *subheap;
|
|
|
|
if (heap->pending_free)
|
|
{
|
|
struct block *tmp = heap->pending_free[heap->pending_pos];
|
|
heap->pending_free[heap->pending_pos] = block;
|
|
heap->pending_pos = (heap->pending_pos + 1) % MAX_FREE_PENDING;
|
|
block_set_type( block, BLOCK_TYPE_DEAD );
|
|
mark_block_free( block + 1, block_get_size( block ) - sizeof(*block), flags );
|
|
if (!(block = tmp)) return;
|
|
}
|
|
|
|
block_size = block_get_size( block );
|
|
if (block_get_flags( block ) & BLOCK_FLAG_PREV_FREE)
|
|
{
|
|
/* merge with previous block if it is free */
|
|
block = *((struct block **)block - 1);
|
|
block_size += block_get_size( block );
|
|
entry = (struct entry *)block;
|
|
list_remove( &entry->entry );
|
|
}
|
|
else entry = (struct entry *)block;
|
|
|
|
subheap = block_get_subheap( heap, block );
|
|
create_free_block( heap, flags, subheap, block, block_size );
|
|
if (next_block( subheap, block )) return; /* not the last block */
|
|
|
|
if (block == first_block( subheap ) && subheap != &heap->subheap)
|
|
{
|
|
/* free the subheap if it's empty and not the main one */
|
|
void *addr = subheap_base( subheap );
|
|
SIZE_T size = 0;
|
|
|
|
list_remove( &entry->entry );
|
|
list_remove( &subheap->entry );
|
|
NtFreeVirtualMemory( NtCurrentProcess(), &addr, &size, MEM_RELEASE );
|
|
}
|
|
else
|
|
{
|
|
/* keep room for a full committed block as hysteresis */
|
|
subheap_decommit( heap, subheap, (char *)(entry + 1) + REGION_ALIGN );
|
|
}
|
|
}
|
|
|
|
|
|
static inline void shrink_used_block( struct heap *heap, ULONG flags, struct block *block,
|
|
SIZE_T old_block_size, SIZE_T block_size, SIZE_T size )
|
|
{
|
|
SUBHEAP *subheap = block_get_subheap( heap, block );
|
|
|
|
if (old_block_size >= block_size + HEAP_MIN_BLOCK_SIZE)
|
|
{
|
|
block_set_size( block, block_size );
|
|
block->tail_size = block_size - sizeof(*block) - size;
|
|
create_free_block( heap, flags, subheap, next_block( subheap, block ), old_block_size - block_size );
|
|
}
|
|
else
|
|
{
|
|
struct block *next;
|
|
block_set_size( block, old_block_size );
|
|
block->tail_size = old_block_size - sizeof(*block) - size;
|
|
if ((next = next_block( subheap, block ))) block_set_flags( next, BLOCK_FLAG_PREV_FREE, 0 );
|
|
}
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* allocate_large_block
|
|
*/
|
|
static struct block *allocate_large_block( struct heap *heap, DWORD flags, SIZE_T size )
|
|
{
|
|
ARENA_LARGE *arena;
|
|
SIZE_T total_size = ROUND_SIZE( sizeof(*arena) + size, REGION_ALIGN - 1 );
|
|
LPVOID address = NULL;
|
|
struct block *block;
|
|
|
|
if (!(flags & HEAP_GROWABLE)) return NULL;
|
|
if (total_size < size) return NULL; /* overflow */
|
|
if (NtAllocateVirtualMemory( NtCurrentProcess(), &address, 0, &total_size, MEM_COMMIT,
|
|
get_protection_type( heap->flags ) ))
|
|
{
|
|
WARN( "Could not allocate block for %#Ix bytes\n", size );
|
|
return NULL;
|
|
}
|
|
|
|
arena = address;
|
|
block = &arena->block;
|
|
arena->data_size = size;
|
|
arena->block_size = (char *)address + total_size - (char *)block;
|
|
|
|
block_set_type( block, BLOCK_TYPE_LARGE );
|
|
block_set_base( block, arena );
|
|
block_set_flags( block, ~0, BLOCK_FLAG_LARGE | BLOCK_USER_FLAGS( flags ) );
|
|
block_set_size( block, 0 );
|
|
list_add_tail( &heap->large_list, &arena->entry );
|
|
valgrind_make_noaccess( (char *)block + sizeof(*block) + arena->data_size,
|
|
arena->block_size - sizeof(*block) - arena->data_size );
|
|
|
|
return block;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* free_large_block
|
|
*/
|
|
static void free_large_block( struct heap *heap, struct block *block )
|
|
{
|
|
ARENA_LARGE *arena = CONTAINING_RECORD( block, ARENA_LARGE, block );
|
|
LPVOID address = arena;
|
|
SIZE_T size = 0;
|
|
|
|
list_remove( &arena->entry );
|
|
NtFreeVirtualMemory( NtCurrentProcess(), &address, &size, MEM_RELEASE );
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* realloc_large_block
|
|
*/
|
|
static struct block *realloc_large_block( struct heap *heap, DWORD flags, struct block *block, SIZE_T size )
|
|
{
|
|
ARENA_LARGE *arena = CONTAINING_RECORD( block, ARENA_LARGE, block );
|
|
SIZE_T old_size = arena->data_size;
|
|
char *data = (char *)(block + 1);
|
|
|
|
if (arena->block_size - sizeof(*block) >= size)
|
|
{
|
|
/* FIXME: we could remap zero-pages instead */
|
|
valgrind_notify_resize( data, old_size, size );
|
|
initialize_block( block, old_size, size, flags );
|
|
|
|
arena->data_size = size;
|
|
valgrind_make_noaccess( (char *)block + sizeof(*block) + arena->data_size,
|
|
arena->block_size - sizeof(*block) - arena->data_size );
|
|
|
|
return block;
|
|
}
|
|
|
|
if (flags & HEAP_REALLOC_IN_PLACE_ONLY) return NULL;
|
|
if (!(block = allocate_large_block( heap, flags, size )))
|
|
{
|
|
WARN( "Could not allocate block for %#Ix bytes\n", size );
|
|
return NULL;
|
|
}
|
|
|
|
valgrind_notify_alloc( block + 1, size, 0 );
|
|
memcpy( block + 1, data, old_size );
|
|
valgrind_notify_free( data );
|
|
free_large_block( heap, &arena->block );
|
|
|
|
return block;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* find_large_block
|
|
*/
|
|
static BOOL find_large_block( const struct heap *heap, const struct block *block )
|
|
{
|
|
ARENA_LARGE *arena;
|
|
|
|
LIST_FOR_EACH_ENTRY( arena, &heap->large_list, ARENA_LARGE, entry )
|
|
if (block == &arena->block) return TRUE;
|
|
|
|
return FALSE;
|
|
}
|
|
|
|
static BOOL validate_large_block( const struct heap *heap, const struct block *block )
|
|
{
|
|
const ARENA_LARGE *arena = CONTAINING_RECORD( block, ARENA_LARGE, block );
|
|
const char *err = NULL;
|
|
|
|
if (ROUND_ADDR( block, REGION_ALIGN - 1 ) != arena)
|
|
err = "invalid block BLOCK_ALIGN";
|
|
else if (block_get_size( block ))
|
|
err = "invalid block size";
|
|
else if (!(block_get_flags( block ) & BLOCK_FLAG_LARGE))
|
|
err = "invalid block flags";
|
|
else if (block_get_type( block ) != BLOCK_TYPE_LARGE)
|
|
err = "invalid block type";
|
|
else if (!contains( block, arena->block_size, block + 1, arena->data_size ))
|
|
err = "invalid block size";
|
|
|
|
if (err)
|
|
{
|
|
ERR( "heap %p, block %p: %s\n", heap, block, err );
|
|
if (TRACE_ON(heap)) heap_dump( heap );
|
|
}
|
|
|
|
return !err;
|
|
}
|
|
|
|
/***********************************************************************
|
|
* HEAP_CreateSubHeap
|
|
*/
|
|
static SUBHEAP *HEAP_CreateSubHeap( struct heap **heap_ptr, LPVOID address, DWORD flags,
|
|
SIZE_T commitSize, SIZE_T totalSize )
|
|
{
|
|
struct heap *heap = *heap_ptr;
|
|
struct entry *pEntry;
|
|
SIZE_T block_size;
|
|
SUBHEAP *subheap;
|
|
unsigned int i;
|
|
|
|
if (!address)
|
|
{
|
|
if (!commitSize) commitSize = REGION_ALIGN;
|
|
totalSize = min( totalSize, 0xffff0000 ); /* don't allow a heap larger than 4GB */
|
|
if (totalSize < commitSize) totalSize = commitSize;
|
|
commitSize = min( totalSize, (SIZE_T)ROUND_ADDR( commitSize + REGION_ALIGN - 1, REGION_ALIGN - 1 ) );
|
|
|
|
/* allocate the memory block */
|
|
if (NtAllocateVirtualMemory( NtCurrentProcess(), &address, 0, &totalSize, MEM_RESERVE,
|
|
get_protection_type( flags ) ))
|
|
{
|
|
WARN( "Could not allocate %#Ix bytes\n", totalSize );
|
|
return NULL;
|
|
}
|
|
if (NtAllocateVirtualMemory( NtCurrentProcess(), &address, 0, &commitSize, MEM_COMMIT,
|
|
get_protection_type( flags ) ))
|
|
{
|
|
WARN( "Could not commit %#Ix bytes for sub-heap %p\n", commitSize, address );
|
|
return NULL;
|
|
}
|
|
}
|
|
|
|
if (heap)
|
|
{
|
|
/* If this is a secondary subheap, insert it into list */
|
|
|
|
subheap = address;
|
|
subheap_set_bounds( subheap, (char *)address + commitSize, (char *)address + totalSize );
|
|
list_add_head( &heap->subheap_list, &subheap->entry );
|
|
}
|
|
else
|
|
{
|
|
/* If this is a primary subheap, initialize main heap */
|
|
|
|
heap = address;
|
|
heap->ffeeffee = 0xffeeffee;
|
|
heap->auto_flags = (flags & HEAP_GROWABLE);
|
|
heap->flags = (flags & ~HEAP_SHARED);
|
|
heap->magic = HEAP_MAGIC;
|
|
heap->grow_size = max( HEAP_DEF_SIZE, totalSize );
|
|
heap->min_size = commitSize;
|
|
list_init( &heap->subheap_list );
|
|
list_init( &heap->large_list );
|
|
|
|
subheap = &heap->subheap;
|
|
subheap_set_bounds( subheap, (char *)address + commitSize, (char *)address + totalSize );
|
|
list_add_head( &heap->subheap_list, &subheap->entry );
|
|
|
|
/* Build the free lists */
|
|
|
|
list_init( &heap->free_lists[0].entry );
|
|
for (i = 0, pEntry = heap->free_lists; i < HEAP_NB_FREE_LISTS; i++, pEntry++)
|
|
{
|
|
block_set_flags( &pEntry->block, ~0, BLOCK_FLAG_FREE_LINK );
|
|
block_set_size( &pEntry->block, 0 );
|
|
block_set_type( &pEntry->block, BLOCK_TYPE_FREE );
|
|
block_set_base( &pEntry->block, heap );
|
|
if (i) list_add_after( &pEntry[-1].entry, &pEntry->entry );
|
|
}
|
|
|
|
/* Initialize critical section */
|
|
|
|
if (!process_heap) /* do it by hand to avoid memory allocations */
|
|
{
|
|
heap->cs.DebugInfo = &process_heap_cs_debug;
|
|
heap->cs.LockCount = -1;
|
|
heap->cs.RecursionCount = 0;
|
|
heap->cs.OwningThread = 0;
|
|
heap->cs.LockSemaphore = 0;
|
|
heap->cs.SpinCount = 0;
|
|
process_heap_cs_debug.CriticalSection = &heap->cs;
|
|
}
|
|
else
|
|
{
|
|
RtlInitializeCriticalSection( &heap->cs );
|
|
heap->cs.DebugInfo->Spare[0] = (DWORD_PTR)(__FILE__ ": heap.cs");
|
|
}
|
|
}
|
|
|
|
block_size = subheap_size( subheap ) - subheap_overhead( subheap );
|
|
block_size &= ~(BLOCK_ALIGN - 1);
|
|
create_free_block( heap, flags, subheap, first_block( subheap ), block_size );
|
|
|
|
*heap_ptr = heap;
|
|
return subheap;
|
|
}
|
|
|
|
|
|
static struct block *find_free_block( struct heap *heap, ULONG flags, SIZE_T block_size )
|
|
{
|
|
struct list *ptr = &find_free_list( heap, block_size, FALSE )->entry;
|
|
struct entry *entry;
|
|
struct block *block;
|
|
SIZE_T total_size;
|
|
SUBHEAP *subheap;
|
|
|
|
/* Find a suitable free list, and in it find a block large enough */
|
|
|
|
while ((ptr = list_next( &heap->free_lists[0].entry, ptr )))
|
|
{
|
|
entry = LIST_ENTRY( ptr, struct entry, entry );
|
|
block = &entry->block;
|
|
if (block_get_flags( block ) == BLOCK_FLAG_FREE_LINK) continue;
|
|
if (block_get_size( block ) >= block_size)
|
|
{
|
|
if (!subheap_commit( heap, block_get_subheap( heap, block ), block, block_size )) return NULL;
|
|
list_remove( &entry->entry );
|
|
return block;
|
|
}
|
|
}
|
|
|
|
/* If no block was found, attempt to grow the heap */
|
|
|
|
if (!(flags & HEAP_GROWABLE))
|
|
{
|
|
WARN( "Not enough space in heap %p for %#Ix bytes\n", heap, block_size );
|
|
return NULL;
|
|
}
|
|
|
|
/* make sure we can fit the block and a free entry at the end */
|
|
total_size = sizeof(SUBHEAP) + block_size + sizeof(struct entry);
|
|
if (total_size < block_size) return NULL; /* overflow */
|
|
|
|
if ((subheap = HEAP_CreateSubHeap( &heap, NULL, flags, total_size,
|
|
max( heap->grow_size, total_size ) )))
|
|
{
|
|
if (heap->grow_size <= HEAP_MAX_FREE_BLOCK_SIZE / 2) heap->grow_size *= 2;
|
|
}
|
|
else while (!subheap) /* shrink the grow size again if we are running out of space */
|
|
{
|
|
if (heap->grow_size <= total_size || heap->grow_size <= 4 * 1024 * 1024) return NULL;
|
|
heap->grow_size /= 2;
|
|
subheap = HEAP_CreateSubHeap( &heap, NULL, flags, total_size,
|
|
max( heap->grow_size, total_size ) );
|
|
}
|
|
|
|
TRACE( "created new sub-heap %p of %#Ix bytes for heap %p\n", subheap, subheap_size( subheap ), heap );
|
|
|
|
entry = first_block( subheap );
|
|
list_remove( &entry->entry );
|
|
return &entry->block;
|
|
}
|
|
|
|
|
|
static BOOL is_valid_free_block( const struct heap *heap, const struct block *block )
|
|
{
|
|
const SUBHEAP *subheap;
|
|
unsigned int i;
|
|
|
|
if ((subheap = find_subheap( heap, block, FALSE ))) return TRUE;
|
|
for (i = 0; i < HEAP_NB_FREE_LISTS; i++) if (block == &heap->free_lists[i].block) return TRUE;
|
|
return FALSE;
|
|
}
|
|
|
|
static BOOL validate_free_block( const struct heap *heap, const SUBHEAP *subheap, const struct block *block )
|
|
{
|
|
const char *err = NULL, *base = subheap_base( subheap ), *commit_end = subheap_commit_end( subheap );
|
|
const struct entry *entry = (struct entry *)block;
|
|
const struct block *prev, *next;
|
|
DWORD flags = heap->flags;
|
|
|
|
if ((ULONG_PTR)(block + 1) % BLOCK_ALIGN)
|
|
err = "invalid block BLOCK_ALIGN";
|
|
else if (block_get_type( block ) != BLOCK_TYPE_FREE)
|
|
err = "invalid block header";
|
|
else if (!(block_get_flags( block ) & BLOCK_FLAG_FREE) || (block_get_flags( block ) & BLOCK_FLAG_PREV_FREE))
|
|
err = "invalid block flags";
|
|
else if (!contains( base, subheap_size( subheap ), block, block_get_size( block ) ))
|
|
err = "invalid block size";
|
|
else if (!is_valid_free_block( heap, (next = &LIST_ENTRY( entry->entry.next, struct entry, entry )->block) ))
|
|
err = "invalid next free block pointer";
|
|
else if (!(block_get_flags( next ) & BLOCK_FLAG_FREE) || block_get_type( next ) != BLOCK_TYPE_FREE)
|
|
err = "invalid next free block header";
|
|
else if (!is_valid_free_block( heap, (prev = &LIST_ENTRY( entry->entry.prev, struct entry, entry )->block) ))
|
|
err = "invalid previous free block pointer";
|
|
else if (!(block_get_flags( prev ) & BLOCK_FLAG_FREE) || block_get_type( prev ) != BLOCK_TYPE_FREE)
|
|
err = "invalid previous free block header";
|
|
else if ((next = next_block( subheap, block )))
|
|
{
|
|
if (!(block_get_flags( next ) & BLOCK_FLAG_PREV_FREE))
|
|
err = "invalid next block flags";
|
|
if (*((struct block **)next - 1) != block)
|
|
err = "invalid next block back pointer";
|
|
}
|
|
|
|
if (!err && (flags & HEAP_FREE_CHECKING_ENABLED))
|
|
{
|
|
const char *ptr = (char *)(entry + 1), *end = (char *)block + block_get_size( block );
|
|
if (next) end -= sizeof(struct block *);
|
|
if (end > commit_end) end = commit_end;
|
|
while (!err && ptr < end)
|
|
{
|
|
if (*(DWORD *)ptr != BLOCK_FILL_FREE) err = "free block overwritten";
|
|
ptr += sizeof(DWORD);
|
|
}
|
|
}
|
|
|
|
if (err)
|
|
{
|
|
ERR( "heap %p, block %p: %s\n", heap, block, err );
|
|
if (TRACE_ON(heap)) heap_dump( heap );
|
|
}
|
|
|
|
return !err;
|
|
}
|
|
|
|
|
|
static BOOL validate_used_block( const struct heap *heap, const SUBHEAP *subheap, const struct block *block )
|
|
{
|
|
const char *err = NULL, *base = subheap_base( subheap ), *commit_end = subheap_commit_end( subheap );
|
|
DWORD flags = heap->flags;
|
|
const struct block *next;
|
|
int i;
|
|
|
|
if ((ULONG_PTR)(block + 1) % BLOCK_ALIGN)
|
|
err = "invalid block BLOCK_ALIGN";
|
|
else if (block_get_type( block ) != BLOCK_TYPE_USED && block_get_type( block ) != BLOCK_TYPE_DEAD)
|
|
err = "invalid block header";
|
|
else if (block_get_flags( block ) & BLOCK_FLAG_FREE)
|
|
err = "invalid block flags";
|
|
else if (!contains( base, commit_end - base, block, block_get_size( block ) ))
|
|
err = "invalid block size";
|
|
else if (block->tail_size > block_get_size( block ) - sizeof(*block))
|
|
err = "invalid block unused size";
|
|
else if ((next = next_block( subheap, block )) && (block_get_flags( next ) & BLOCK_FLAG_PREV_FREE))
|
|
err = "invalid next block flags";
|
|
else if (block_get_flags( block ) & BLOCK_FLAG_PREV_FREE)
|
|
{
|
|
const struct block *prev = *((struct block **)block - 1);
|
|
if (!is_valid_free_block( heap, prev ))
|
|
err = "invalid previous block pointer";
|
|
else if (!(block_get_flags( prev ) & BLOCK_FLAG_FREE) || block_get_type( prev ) != BLOCK_TYPE_FREE)
|
|
err = "invalid previous block flags";
|
|
if ((char *)prev + block_get_size( prev ) != (char *)block)
|
|
err = "invalid previous block size";
|
|
}
|
|
|
|
if (!err && block_get_type( block ) == BLOCK_TYPE_DEAD)
|
|
{
|
|
const char *ptr = (char *)(block + 1), *end = (char *)block + block_get_size( block );
|
|
while (!err && ptr < end)
|
|
{
|
|
if (*(DWORD *)ptr != BLOCK_FILL_FREE) err = "free block overwritten";
|
|
ptr += sizeof(DWORD);
|
|
}
|
|
}
|
|
else if (!err && (flags & HEAP_TAIL_CHECKING_ENABLED))
|
|
{
|
|
const unsigned char *tail = (unsigned char *)block + block_get_size( block ) - block->tail_size;
|
|
for (i = 0; !err && i < BLOCK_ALIGN; i++) if (tail[i] != BLOCK_FILL_TAIL) err = "invalid block tail";
|
|
}
|
|
|
|
if (err)
|
|
{
|
|
ERR( "heap %p, block %p: %s\n", heap, block, err );
|
|
if (TRACE_ON(heap)) heap_dump( heap );
|
|
}
|
|
|
|
return !err;
|
|
}
|
|
|
|
|
|
static BOOL heap_validate_ptr( const struct heap *heap, const void *ptr )
|
|
{
|
|
const struct block *block = (struct block *)ptr - 1;
|
|
const SUBHEAP *subheap;
|
|
|
|
if (!(subheap = find_subheap( heap, block, FALSE )))
|
|
{
|
|
if (!find_large_block( heap, block ))
|
|
{
|
|
if (WARN_ON(heap)) WARN("heap %p, ptr %p: block region not found\n", heap, ptr );
|
|
return FALSE;
|
|
}
|
|
|
|
return validate_large_block( heap, block );
|
|
}
|
|
|
|
return validate_used_block( heap, subheap, block );
|
|
}
|
|
|
|
static BOOL heap_validate( const struct heap *heap )
|
|
{
|
|
const ARENA_LARGE *large_arena;
|
|
const struct block *block;
|
|
const SUBHEAP *subheap;
|
|
|
|
LIST_FOR_EACH_ENTRY( subheap, &heap->subheap_list, SUBHEAP, entry )
|
|
{
|
|
if (!check_subheap( subheap ))
|
|
{
|
|
ERR( "heap %p, subheap %p corrupted sizes\n", heap, subheap );
|
|
if (TRACE_ON(heap)) heap_dump( heap );
|
|
return FALSE;
|
|
}
|
|
|
|
for (block = first_block( subheap ); block; block = next_block( subheap, block ))
|
|
{
|
|
if (block_get_flags( block ) & BLOCK_FLAG_FREE)
|
|
{
|
|
if (!validate_free_block( heap, subheap, block )) return FALSE;
|
|
}
|
|
else
|
|
{
|
|
if (!validate_used_block( heap, subheap, block )) return FALSE;
|
|
}
|
|
}
|
|
}
|
|
|
|
LIST_FOR_EACH_ENTRY( large_arena, &heap->large_list, ARENA_LARGE, entry )
|
|
if (!validate_large_block( heap, &large_arena->block )) return FALSE;
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
static inline struct block *unsafe_block_from_ptr( struct heap *heap, ULONG flags, const void *ptr )
|
|
{
|
|
struct block *block = (struct block *)ptr - 1;
|
|
const char *err = NULL;
|
|
SUBHEAP *subheap;
|
|
|
|
if (flags & HEAP_VALIDATE)
|
|
{
|
|
heap_lock( heap, flags );
|
|
if (!heap_validate_ptr( heap, ptr )) block = NULL;
|
|
heap_unlock( heap, flags );
|
|
return block;
|
|
}
|
|
|
|
if ((ULONG_PTR)ptr % BLOCK_ALIGN)
|
|
err = "invalid ptr alignment";
|
|
else if ((subheap = block_get_subheap( heap, block )) >= (SUBHEAP *)block)
|
|
err = "invalid base offset";
|
|
else if (block_get_type( block ) == BLOCK_TYPE_USED)
|
|
{
|
|
const char *base = subheap_base( subheap ), *commit_end = subheap_commit_end( subheap );
|
|
if (!contains( base, commit_end - base, block, block_get_size( block ) )) err = "invalid block size";
|
|
}
|
|
else if (block_get_type( block ) == BLOCK_TYPE_LARGE)
|
|
{
|
|
ARENA_LARGE *large = subheap_base( subheap );
|
|
if (block != &large->block) err = "invalid large block";
|
|
}
|
|
else if (block_get_type( block ) == BLOCK_TYPE_DEAD)
|
|
err = "delayed freed block";
|
|
else if (block_get_type( block ) == BLOCK_TYPE_FREE)
|
|
err = "already freed block";
|
|
else
|
|
err = "invalid block type";
|
|
|
|
if (err) WARN( "heap %p, block %p: %s\n", heap, block, err );
|
|
return err ? NULL : block;
|
|
}
|
|
|
|
static DWORD heap_flags_from_global_flag( DWORD flag )
|
|
{
|
|
DWORD ret = 0;
|
|
|
|
if (flag & FLG_HEAP_ENABLE_TAIL_CHECK)
|
|
ret |= HEAP_TAIL_CHECKING_ENABLED;
|
|
if (flag & FLG_HEAP_ENABLE_FREE_CHECK)
|
|
ret |= HEAP_FREE_CHECKING_ENABLED;
|
|
if (flag & FLG_HEAP_VALIDATE_PARAMETERS)
|
|
ret |= HEAP_VALIDATE_PARAMS | HEAP_TAIL_CHECKING_ENABLED | HEAP_FREE_CHECKING_ENABLED;
|
|
if (flag & FLG_HEAP_VALIDATE_ALL)
|
|
ret |= HEAP_VALIDATE_ALL | HEAP_TAIL_CHECKING_ENABLED | HEAP_FREE_CHECKING_ENABLED;
|
|
if (flag & FLG_HEAP_DISABLE_COALESCING)
|
|
ret |= HEAP_DISABLE_COALESCE_ON_FREE;
|
|
if (flag & FLG_HEAP_PAGE_ALLOCS)
|
|
ret |= HEAP_PAGE_ALLOCS;
|
|
|
|
return ret;
|
|
}
|
|
|
|
/***********************************************************************
|
|
* heap_set_debug_flags
|
|
*/
|
|
static void heap_set_debug_flags( HANDLE handle )
|
|
{
|
|
ULONG global_flags = RtlGetNtGlobalFlags();
|
|
DWORD dummy, flags, force_flags;
|
|
struct heap *heap;
|
|
|
|
if (TRACE_ON(heap)) global_flags |= FLG_HEAP_VALIDATE_ALL;
|
|
if (WARN_ON(heap)) global_flags |= FLG_HEAP_VALIDATE_PARAMETERS;
|
|
|
|
heap = unsafe_heap_from_handle( handle, 0, &dummy );
|
|
flags = heap_flags_from_global_flag( global_flags );
|
|
force_flags = (heap->flags | flags) & ~(HEAP_SHARED|HEAP_DISABLE_COALESCE_ON_FREE);
|
|
|
|
if (global_flags & FLG_HEAP_ENABLE_TAGGING) flags |= HEAP_SHARED;
|
|
if (!(global_flags & FLG_HEAP_PAGE_ALLOCS)) force_flags &= ~(HEAP_GROWABLE|HEAP_PRIVATE);
|
|
|
|
if (RUNNING_ON_VALGRIND) flags = 0; /* no sense in validating since Valgrind catches accesses */
|
|
|
|
heap->flags |= flags;
|
|
heap->force_flags |= force_flags;
|
|
|
|
if (flags & (HEAP_FREE_CHECKING_ENABLED | HEAP_TAIL_CHECKING_ENABLED)) /* fix existing blocks */
|
|
{
|
|
struct block *block;
|
|
SUBHEAP *subheap;
|
|
|
|
LIST_FOR_EACH_ENTRY( subheap, &heap->subheap_list, SUBHEAP, entry )
|
|
{
|
|
const char *commit_end = subheap_commit_end( subheap );
|
|
|
|
if (!check_subheap( subheap )) break;
|
|
|
|
for (block = first_block( subheap ); block; block = next_block( subheap, block ))
|
|
{
|
|
if (block_get_flags( block ) & BLOCK_FLAG_FREE)
|
|
{
|
|
char *data = (char *)block + block_get_overhead( block ), *end = (char *)block + block_get_size( block );
|
|
if (next_block( subheap, block )) end -= sizeof(struct block *);
|
|
if (end > commit_end) mark_block_free( data, commit_end - data, flags );
|
|
else mark_block_free( data, end - data, flags );
|
|
}
|
|
else
|
|
{
|
|
if (block_get_type( block ) == BLOCK_TYPE_DEAD) mark_block_free( block + 1, block_get_size( block ) - sizeof(*block), flags );
|
|
else mark_block_tail( block, flags );
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
if ((heap->flags & HEAP_GROWABLE) && !heap->pending_free &&
|
|
((flags & HEAP_FREE_CHECKING_ENABLED) || RUNNING_ON_VALGRIND))
|
|
{
|
|
heap->pending_free = RtlAllocateHeap( handle, HEAP_ZERO_MEMORY,
|
|
MAX_FREE_PENDING * sizeof(*heap->pending_free) );
|
|
heap->pending_pos = 0;
|
|
}
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* RtlCreateHeap (NTDLL.@)
|
|
*
|
|
* Create a new Heap.
|
|
*
|
|
* PARAMS
|
|
* flags [I] HEAP_ flags from "winnt.h"
|
|
* addr [I] Desired base address
|
|
* totalSize [I] Total size of the heap, or 0 for a growable heap
|
|
* commitSize [I] Amount of heap space to commit
|
|
* unknown [I] Not yet understood
|
|
* definition [I] Heap definition
|
|
*
|
|
* RETURNS
|
|
* Success: A HANDLE to the newly created heap.
|
|
* Failure: a NULL HANDLE.
|
|
*/
|
|
HANDLE WINAPI RtlCreateHeap( ULONG flags, PVOID addr, SIZE_T totalSize, SIZE_T commitSize,
|
|
PVOID unknown, PRTL_HEAP_DEFINITION definition )
|
|
{
|
|
struct heap *heap = NULL;
|
|
SUBHEAP *subheap;
|
|
|
|
/* Allocate the heap block */
|
|
|
|
flags &= ~(HEAP_TAIL_CHECKING_ENABLED|HEAP_FREE_CHECKING_ENABLED);
|
|
if (process_heap) flags |= HEAP_PRIVATE;
|
|
if (!process_heap || !totalSize || (flags & HEAP_SHARED)) flags |= HEAP_GROWABLE;
|
|
if (!totalSize) totalSize = HEAP_DEF_SIZE;
|
|
|
|
if (!(subheap = HEAP_CreateSubHeap( &heap, addr, flags, commitSize, totalSize ))) return 0;
|
|
|
|
heap_set_debug_flags( heap );
|
|
|
|
/* link it into the per-process heap list */
|
|
if (process_heap)
|
|
{
|
|
RtlEnterCriticalSection( &process_heap->cs );
|
|
list_add_head( &process_heap->entry, &heap->entry );
|
|
RtlLeaveCriticalSection( &process_heap->cs );
|
|
}
|
|
else if (!addr)
|
|
{
|
|
process_heap = heap; /* assume the first heap we create is the process main heap */
|
|
list_init( &process_heap->entry );
|
|
}
|
|
|
|
return heap;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* RtlDestroyHeap (NTDLL.@)
|
|
*
|
|
* Destroy a Heap created with RtlCreateHeap().
|
|
*
|
|
* PARAMS
|
|
* heap [I] Heap to destroy.
|
|
*
|
|
* RETURNS
|
|
* Success: A NULL HANDLE, if heap is NULL or it was destroyed
|
|
* Failure: The Heap handle, if heap is the process heap.
|
|
*/
|
|
HANDLE WINAPI RtlDestroyHeap( HANDLE handle )
|
|
{
|
|
SUBHEAP *subheap, *next;
|
|
ARENA_LARGE *arena, *arena_next;
|
|
struct block **pending, **tmp;
|
|
struct heap *heap;
|
|
ULONG heap_flags;
|
|
SIZE_T size;
|
|
void *addr;
|
|
|
|
TRACE( "handle %p\n", handle );
|
|
|
|
if (!(heap = unsafe_heap_from_handle( handle, 0, &heap_flags )) && handle &&
|
|
(((struct heap *)handle)->flags & HEAP_VALIDATE_PARAMS) &&
|
|
NtCurrentTeb()->Peb->BeingDebugged)
|
|
{
|
|
DbgPrint( "Attempt to destroy an invalid heap\n" );
|
|
DbgBreakPoint();
|
|
}
|
|
if (!heap) return handle;
|
|
|
|
if ((pending = heap->pending_free))
|
|
{
|
|
heap->pending_free = NULL;
|
|
for (tmp = pending; *tmp && tmp != pending + MAX_FREE_PENDING; ++tmp)
|
|
free_used_block( heap, heap->flags, *tmp );
|
|
RtlFreeHeap( handle, 0, pending );
|
|
}
|
|
|
|
if (heap == process_heap) return handle; /* cannot delete the main process heap */
|
|
|
|
/* remove it from the per-process list */
|
|
RtlEnterCriticalSection( &process_heap->cs );
|
|
list_remove( &heap->entry );
|
|
RtlLeaveCriticalSection( &process_heap->cs );
|
|
|
|
heap->cs.DebugInfo->Spare[0] = 0;
|
|
RtlDeleteCriticalSection( &heap->cs );
|
|
|
|
LIST_FOR_EACH_ENTRY_SAFE( arena, arena_next, &heap->large_list, ARENA_LARGE, entry )
|
|
{
|
|
list_remove( &arena->entry );
|
|
size = 0;
|
|
addr = arena;
|
|
NtFreeVirtualMemory( NtCurrentProcess(), &addr, &size, MEM_RELEASE );
|
|
}
|
|
LIST_FOR_EACH_ENTRY_SAFE( subheap, next, &heap->subheap_list, SUBHEAP, entry )
|
|
{
|
|
if (subheap == &heap->subheap) continue; /* do this one last */
|
|
valgrind_notify_free_all( subheap );
|
|
list_remove( &subheap->entry );
|
|
size = 0;
|
|
addr = ROUND_ADDR( subheap, REGION_ALIGN - 1 );
|
|
NtFreeVirtualMemory( NtCurrentProcess(), &addr, &size, MEM_RELEASE );
|
|
}
|
|
valgrind_notify_free_all( &heap->subheap );
|
|
size = 0;
|
|
addr = heap;
|
|
NtFreeVirtualMemory( NtCurrentProcess(), &addr, &size, MEM_RELEASE );
|
|
return 0;
|
|
}
|
|
|
|
static SIZE_T heap_get_block_size( const struct heap *heap, ULONG flags, SIZE_T size )
|
|
{
|
|
static const ULONG padd_flags = HEAP_VALIDATE | HEAP_VALIDATE_ALL | HEAP_VALIDATE_PARAMS | HEAP_ADD_USER_INFO;
|
|
static const ULONG check_flags = HEAP_TAIL_CHECKING_ENABLED | HEAP_FREE_CHECKING_ENABLED | HEAP_CHECKING_ENABLED;
|
|
SIZE_T overhead, block_size;
|
|
|
|
if ((flags & check_flags)) overhead = BLOCK_ALIGN;
|
|
else overhead = sizeof(struct block);
|
|
|
|
if ((flags & HEAP_TAIL_CHECKING_ENABLED) || RUNNING_ON_VALGRIND) overhead += BLOCK_ALIGN;
|
|
if (flags & padd_flags) overhead += BLOCK_ALIGN;
|
|
|
|
if (size < BLOCK_ALIGN) size = BLOCK_ALIGN;
|
|
block_size = ROUND_SIZE( size + overhead, BLOCK_ALIGN - 1 );
|
|
|
|
if (block_size < size) return ~0U; /* overflow */
|
|
if (block_size < HEAP_MIN_BLOCK_SIZE) block_size = HEAP_MIN_BLOCK_SIZE;
|
|
return block_size;
|
|
}
|
|
|
|
static NTSTATUS heap_allocate( struct heap *heap, ULONG flags, SIZE_T block_size, SIZE_T size, void **ret )
|
|
{
|
|
SIZE_T old_block_size;
|
|
struct block *block;
|
|
|
|
if (block_size >= HEAP_MIN_LARGE_BLOCK_SIZE)
|
|
{
|
|
if (!(block = allocate_large_block( heap, flags, size ))) return STATUS_NO_MEMORY;
|
|
|
|
*ret = block + 1;
|
|
return STATUS_SUCCESS;
|
|
}
|
|
|
|
/* Locate a suitable free block */
|
|
|
|
if (!(block = find_free_block( heap, flags, block_size ))) return STATUS_NO_MEMORY;
|
|
/* read the free block size, changing block type or flags may alter it */
|
|
old_block_size = block_get_size( block );
|
|
|
|
block_set_type( block, BLOCK_TYPE_USED );
|
|
block_set_flags( block, ~0, BLOCK_USER_FLAGS( flags ) );
|
|
shrink_used_block( heap, flags, block, old_block_size, block_size, size );
|
|
initialize_block( block, 0, size, flags );
|
|
mark_block_tail( block, flags );
|
|
|
|
*ret = block + 1;
|
|
return STATUS_SUCCESS;
|
|
}
|
|
|
|
/***********************************************************************
|
|
* RtlAllocateHeap (NTDLL.@)
|
|
*/
|
|
void *WINAPI DECLSPEC_HOTPATCH RtlAllocateHeap( HANDLE handle, ULONG flags, SIZE_T size )
|
|
{
|
|
struct heap *heap;
|
|
SIZE_T block_size;
|
|
void *ptr = NULL;
|
|
ULONG heap_flags;
|
|
NTSTATUS status;
|
|
|
|
if (!(heap = unsafe_heap_from_handle( handle, flags, &heap_flags )))
|
|
status = STATUS_INVALID_HANDLE;
|
|
else if ((block_size = heap_get_block_size( heap, heap_flags, size )) == ~0U)
|
|
status = STATUS_NO_MEMORY;
|
|
else
|
|
{
|
|
heap_lock( heap, heap_flags );
|
|
status = heap_allocate( heap, heap_flags, block_size, size, &ptr );
|
|
heap_unlock( heap, heap_flags );
|
|
}
|
|
|
|
if (!status) valgrind_notify_alloc( ptr, size, flags & HEAP_ZERO_MEMORY );
|
|
|
|
TRACE( "handle %p, flags %#lx, size %#Ix, return %p, status %#lx.\n", handle, flags, size, ptr, status );
|
|
heap_set_status( heap, flags, status );
|
|
return ptr;
|
|
}
|
|
|
|
|
|
static NTSTATUS heap_free( struct heap *heap, ULONG flags, struct block *block )
|
|
{
|
|
if (block_get_flags( block ) & BLOCK_FLAG_LARGE) free_large_block( heap, block );
|
|
else free_used_block( heap, flags, block );
|
|
return STATUS_SUCCESS;
|
|
}
|
|
|
|
/***********************************************************************
|
|
* RtlFreeHeap (NTDLL.@)
|
|
*/
|
|
BOOLEAN WINAPI DECLSPEC_HOTPATCH RtlFreeHeap( HANDLE handle, ULONG flags, void *ptr )
|
|
{
|
|
struct block *block;
|
|
struct heap *heap;
|
|
ULONG heap_flags;
|
|
NTSTATUS status;
|
|
|
|
if (!ptr) return TRUE;
|
|
|
|
valgrind_notify_free( ptr );
|
|
|
|
if (!(heap = unsafe_heap_from_handle( handle, flags, &heap_flags )))
|
|
status = STATUS_INVALID_PARAMETER;
|
|
else if (!(block = unsafe_block_from_ptr( heap, heap_flags, ptr )))
|
|
status = STATUS_INVALID_PARAMETER;
|
|
else
|
|
{
|
|
heap_lock( heap, heap_flags );
|
|
status = heap_free( heap, heap_flags, block );
|
|
heap_unlock( heap, heap_flags );
|
|
}
|
|
|
|
TRACE( "handle %p, flags %#lx, ptr %p, return %u, status %#lx.\n", handle, flags, ptr, !status, status );
|
|
heap_set_status( heap, flags, status );
|
|
return !status;
|
|
}
|
|
|
|
|
|
static NTSTATUS heap_reallocate( struct heap *heap, ULONG flags, struct block *block,
|
|
SIZE_T block_size, SIZE_T size, void **ret )
|
|
{
|
|
SIZE_T old_block_size, old_size;
|
|
NTSTATUS status;
|
|
|
|
if (block_get_flags( block ) & BLOCK_FLAG_LARGE)
|
|
{
|
|
if (!(block = realloc_large_block( heap, flags, block, size ))) return STATUS_NO_MEMORY;
|
|
|
|
*ret = block + 1;
|
|
return STATUS_SUCCESS;
|
|
}
|
|
|
|
/* Check if we need to grow the block */
|
|
|
|
old_block_size = block_get_size( block );
|
|
old_size = old_block_size - block_get_overhead( block );
|
|
if (block_size > old_block_size)
|
|
{
|
|
SUBHEAP *subheap = block_get_subheap( heap, block );
|
|
struct entry *entry;
|
|
struct block *next;
|
|
|
|
if (!(next = next_block( subheap, block )) || !(block_get_flags( next ) & BLOCK_FLAG_FREE) ||
|
|
block_size >= HEAP_MIN_LARGE_BLOCK_SIZE ||
|
|
block_size > old_block_size + block_get_size( next ) || !subheap_commit( heap, subheap, block, block_size ))
|
|
{
|
|
if (flags & HEAP_REALLOC_IN_PLACE_ONLY) return STATUS_NO_MEMORY;
|
|
if ((status = heap_allocate( heap, flags & ~HEAP_ZERO_MEMORY, block_size, size, ret ))) return status;
|
|
valgrind_notify_alloc( *ret, size, 0 );
|
|
memcpy( *ret, block + 1, old_size );
|
|
if (flags & HEAP_ZERO_MEMORY) memset( (char *)*ret + old_size, 0, size - old_size );
|
|
valgrind_notify_free( block + 1 );
|
|
free_used_block( heap, flags, block );
|
|
return STATUS_SUCCESS;
|
|
}
|
|
|
|
/* The next block is free and large enough */
|
|
entry = (struct entry *)next;
|
|
list_remove( &entry->entry );
|
|
old_block_size += block_get_size( next );
|
|
}
|
|
|
|
valgrind_notify_resize( block + 1, old_size, size );
|
|
block_set_flags( block, BLOCK_FLAG_USER_MASK & ~BLOCK_FLAG_USER_INFO, BLOCK_USER_FLAGS( flags ) );
|
|
shrink_used_block( heap, flags, block, old_block_size, block_size, size );
|
|
|
|
initialize_block( block, old_size, size, flags );
|
|
mark_block_tail( block, flags );
|
|
|
|
*ret = block + 1;
|
|
return STATUS_SUCCESS;
|
|
}
|
|
|
|
/***********************************************************************
|
|
* RtlReAllocateHeap (NTDLL.@)
|
|
*/
|
|
void *WINAPI RtlReAllocateHeap( HANDLE handle, ULONG flags, void *ptr, SIZE_T size )
|
|
{
|
|
struct block *block;
|
|
struct heap *heap;
|
|
SIZE_T block_size;
|
|
ULONG heap_flags;
|
|
void *ret = NULL;
|
|
NTSTATUS status;
|
|
|
|
if (!ptr) return NULL;
|
|
|
|
if (!(heap = unsafe_heap_from_handle( handle, flags, &heap_flags )))
|
|
status = STATUS_INVALID_HANDLE;
|
|
else if ((block_size = heap_get_block_size( heap, heap_flags, size )) == ~0U)
|
|
status = STATUS_NO_MEMORY;
|
|
else if (!(block = unsafe_block_from_ptr( heap, heap_flags, ptr )))
|
|
status = STATUS_INVALID_PARAMETER;
|
|
else
|
|
{
|
|
heap_lock( heap, heap_flags );
|
|
status = heap_reallocate( heap, heap_flags, block, block_size, size, &ret );
|
|
heap_unlock( heap, heap_flags );
|
|
}
|
|
|
|
TRACE( "handle %p, flags %#lx, ptr %p, size %#Ix, return %p, status %#lx.\n", handle, flags, ptr, size, ret, status );
|
|
heap_set_status( heap, flags, status );
|
|
return ret;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* RtlCompactHeap (NTDLL.@)
|
|
*
|
|
* Compact the free space in a Heap.
|
|
*
|
|
* PARAMS
|
|
* heap [I] Heap that block was allocated from
|
|
* flags [I] HEAP_ flags from "winnt.h"
|
|
*
|
|
* RETURNS
|
|
* The number of bytes compacted.
|
|
*
|
|
* NOTES
|
|
* This function is a harmless stub.
|
|
*/
|
|
ULONG WINAPI RtlCompactHeap( HANDLE handle, ULONG flags )
|
|
{
|
|
static BOOL reported;
|
|
if (!reported++) FIXME( "handle %p, flags %#lx stub!\n", handle, flags );
|
|
return 0;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* RtlLockHeap (NTDLL.@)
|
|
*
|
|
* Lock a Heap.
|
|
*
|
|
* PARAMS
|
|
* heap [I] Heap to lock
|
|
*
|
|
* RETURNS
|
|
* Success: TRUE. The Heap is locked.
|
|
* Failure: FALSE, if heap is invalid.
|
|
*/
|
|
BOOLEAN WINAPI RtlLockHeap( HANDLE handle )
|
|
{
|
|
struct heap *heap;
|
|
ULONG heap_flags;
|
|
if (!(heap = unsafe_heap_from_handle( handle, 0, &heap_flags ))) return FALSE;
|
|
heap_lock( heap, heap_flags );
|
|
return TRUE;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* RtlUnlockHeap (NTDLL.@)
|
|
*
|
|
* Unlock a Heap.
|
|
*
|
|
* PARAMS
|
|
* heap [I] Heap to unlock
|
|
*
|
|
* RETURNS
|
|
* Success: TRUE. The Heap is unlocked.
|
|
* Failure: FALSE, if heap is invalid.
|
|
*/
|
|
BOOLEAN WINAPI RtlUnlockHeap( HANDLE handle )
|
|
{
|
|
struct heap *heap;
|
|
ULONG heap_flags;
|
|
if (!(heap = unsafe_heap_from_handle( handle, 0, &heap_flags ))) return FALSE;
|
|
heap_unlock( heap, heap_flags );
|
|
return TRUE;
|
|
}
|
|
|
|
|
|
static NTSTATUS heap_size( const struct heap *heap, struct block *block, SIZE_T *size )
|
|
{
|
|
if (block_get_flags( block ) & BLOCK_FLAG_LARGE)
|
|
{
|
|
const ARENA_LARGE *large_arena = CONTAINING_RECORD( block, ARENA_LARGE, block );
|
|
*size = large_arena->data_size;
|
|
}
|
|
else *size = block_get_size( block ) - block_get_overhead( block );
|
|
|
|
return STATUS_SUCCESS;
|
|
}
|
|
|
|
/***********************************************************************
|
|
* RtlSizeHeap (NTDLL.@)
|
|
*/
|
|
SIZE_T WINAPI RtlSizeHeap( HANDLE handle, ULONG flags, const void *ptr )
|
|
{
|
|
SIZE_T size = ~(SIZE_T)0;
|
|
struct block *block;
|
|
struct heap *heap;
|
|
ULONG heap_flags;
|
|
NTSTATUS status;
|
|
|
|
if (!(heap = unsafe_heap_from_handle( handle, flags, &heap_flags )))
|
|
status = STATUS_INVALID_PARAMETER;
|
|
else if (!(block = unsafe_block_from_ptr( heap, heap_flags, ptr )))
|
|
status = STATUS_INVALID_PARAMETER;
|
|
else
|
|
{
|
|
heap_lock( heap, heap_flags );
|
|
status = heap_size( heap, block, &size );
|
|
heap_unlock( heap, heap_flags );
|
|
}
|
|
|
|
TRACE( "handle %p, flags %#lx, ptr %p, return %#Ix, status %#lx.\n", handle, flags, ptr, size, status );
|
|
heap_set_status( heap, flags, status );
|
|
return size;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* RtlValidateHeap (NTDLL.@)
|
|
*/
|
|
BOOLEAN WINAPI RtlValidateHeap( HANDLE handle, ULONG flags, const void *ptr )
|
|
{
|
|
struct heap *heap;
|
|
ULONG heap_flags;
|
|
BOOLEAN ret;
|
|
|
|
if (!(heap = unsafe_heap_from_handle( handle, flags, &heap_flags )))
|
|
ret = FALSE;
|
|
else
|
|
{
|
|
heap_lock( heap, heap_flags );
|
|
if (ptr) ret = heap_validate_ptr( heap, ptr );
|
|
else ret = heap_validate( heap );
|
|
heap_unlock( heap, heap_flags );
|
|
}
|
|
|
|
TRACE( "handle %p, flags %#lx, ptr %p, return %u.\n", handle, flags, ptr, !!ret );
|
|
return ret;
|
|
}
|
|
|
|
|
|
static NTSTATUS heap_walk_blocks( const struct heap *heap, const SUBHEAP *subheap,
|
|
const struct block *block, struct rtl_heap_entry *entry )
|
|
{
|
|
const char *base = subheap_base( subheap ), *commit_end = subheap_commit_end( subheap ), *end = base + subheap_size( subheap );
|
|
const struct block *blocks = first_block( subheap );
|
|
|
|
if (entry->lpData == commit_end) return STATUS_NO_MORE_ENTRIES;
|
|
if (entry->lpData == base) block = blocks;
|
|
else if (!(block = next_block( subheap, block )))
|
|
{
|
|
entry->lpData = (void *)commit_end;
|
|
entry->cbData = end - commit_end;
|
|
entry->cbOverhead = 0;
|
|
entry->iRegionIndex = 0;
|
|
entry->wFlags = RTL_HEAP_ENTRY_UNCOMMITTED;
|
|
return STATUS_SUCCESS;
|
|
}
|
|
|
|
if (block_get_flags( block ) & BLOCK_FLAG_FREE)
|
|
{
|
|
entry->lpData = (char *)block + block_get_overhead( block );
|
|
entry->cbData = block_get_size( block ) - block_get_overhead( block );
|
|
/* FIXME: last free block should not include uncommitted range, which also has its own overhead */
|
|
if (!contains( blocks, commit_end - (char *)blocks, block, block_get_size( block ) ))
|
|
entry->cbData = commit_end - (char *)entry->lpData - 4 * BLOCK_ALIGN;
|
|
entry->cbOverhead = 2 * BLOCK_ALIGN;
|
|
entry->iRegionIndex = 0;
|
|
entry->wFlags = 0;
|
|
}
|
|
else
|
|
{
|
|
entry->lpData = (void *)(block + 1);
|
|
entry->cbData = block_get_size( block ) - block_get_overhead( block );
|
|
entry->cbOverhead = block_get_overhead( block );
|
|
entry->iRegionIndex = 0;
|
|
entry->wFlags = RTL_HEAP_ENTRY_COMMITTED|RTL_HEAP_ENTRY_BLOCK|RTL_HEAP_ENTRY_BUSY;
|
|
}
|
|
|
|
return STATUS_SUCCESS;
|
|
}
|
|
|
|
static NTSTATUS heap_walk( const struct heap *heap, struct rtl_heap_entry *entry )
|
|
{
|
|
const char *data = entry->lpData;
|
|
const ARENA_LARGE *large = NULL;
|
|
const struct block *block;
|
|
const struct list *next;
|
|
const SUBHEAP *subheap;
|
|
NTSTATUS status;
|
|
char *base;
|
|
|
|
if (!data || entry->wFlags & RTL_HEAP_ENTRY_REGION) block = (struct block *)data;
|
|
else if (entry->wFlags & RTL_HEAP_ENTRY_BUSY) block = (struct block *)data - 1;
|
|
else block = (struct block *)(data - sizeof(struct list)) - 1;
|
|
|
|
if (find_large_block( heap, block ))
|
|
{
|
|
large = CONTAINING_RECORD( block, ARENA_LARGE, block );
|
|
next = &large->entry;
|
|
}
|
|
else if ((subheap = find_subheap( heap, block, TRUE )))
|
|
{
|
|
if (!(status = heap_walk_blocks( heap, subheap, block, entry ))) return STATUS_SUCCESS;
|
|
else if (status != STATUS_NO_MORE_ENTRIES) return status;
|
|
next = &subheap->entry;
|
|
}
|
|
else
|
|
{
|
|
if (entry->lpData) return STATUS_INVALID_PARAMETER;
|
|
next = &heap->subheap_list;
|
|
}
|
|
|
|
if (!large && (next = list_next( &heap->subheap_list, next )))
|
|
{
|
|
subheap = LIST_ENTRY( next, SUBHEAP, entry );
|
|
base = subheap_base( subheap );
|
|
entry->lpData = base;
|
|
entry->cbData = subheap_overhead( subheap );
|
|
entry->cbOverhead = 0;
|
|
entry->iRegionIndex = 0;
|
|
entry->wFlags = RTL_HEAP_ENTRY_REGION;
|
|
entry->Region.dwCommittedSize = (char *)subheap_commit_end( subheap ) - base;
|
|
entry->Region.dwUnCommittedSize = subheap_size( subheap ) - entry->Region.dwCommittedSize;
|
|
entry->Region.lpFirstBlock = base + entry->cbData;
|
|
entry->Region.lpLastBlock = base + subheap_size( subheap );
|
|
return STATUS_SUCCESS;
|
|
}
|
|
|
|
if (!next) next = &heap->large_list;
|
|
if ((next = list_next( &heap->large_list, next )))
|
|
{
|
|
large = LIST_ENTRY( next, ARENA_LARGE, entry );
|
|
entry->lpData = (void *)(large + 1);
|
|
entry->cbData = large->data_size;
|
|
entry->cbOverhead = 0;
|
|
entry->iRegionIndex = 64;
|
|
entry->wFlags = RTL_HEAP_ENTRY_COMMITTED|RTL_HEAP_ENTRY_BLOCK|RTL_HEAP_ENTRY_BUSY;
|
|
return STATUS_SUCCESS;
|
|
}
|
|
|
|
return STATUS_NO_MORE_ENTRIES;
|
|
}
|
|
|
|
/***********************************************************************
|
|
* RtlWalkHeap (NTDLL.@)
|
|
*/
|
|
NTSTATUS WINAPI RtlWalkHeap( HANDLE handle, void *entry_ptr )
|
|
{
|
|
struct rtl_heap_entry *entry = entry_ptr;
|
|
struct heap *heap;
|
|
ULONG heap_flags;
|
|
NTSTATUS status;
|
|
|
|
if (!entry) return STATUS_INVALID_PARAMETER;
|
|
|
|
if (!(heap = unsafe_heap_from_handle( handle, 0, &heap_flags )))
|
|
status = STATUS_INVALID_HANDLE;
|
|
else
|
|
{
|
|
heap_lock( heap, heap_flags );
|
|
status = heap_walk( heap, entry );
|
|
heap_unlock( heap, heap_flags );
|
|
}
|
|
|
|
TRACE( "handle %p, entry %p %s, return %#lx\n", handle, entry,
|
|
status ? "<empty>" : debugstr_heap_entry(entry), status );
|
|
return status;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* RtlGetProcessHeaps (NTDLL.@)
|
|
*
|
|
* Get the Heaps belonging to the current process.
|
|
*
|
|
* PARAMS
|
|
* count [I] size of heaps
|
|
* heaps [O] Destination array for heap HANDLE's
|
|
*
|
|
* RETURNS
|
|
* Success: The number of Heaps allocated by the process.
|
|
* Failure: 0.
|
|
*/
|
|
ULONG WINAPI RtlGetProcessHeaps( ULONG count, HANDLE *heaps )
|
|
{
|
|
ULONG total = 1; /* main heap */
|
|
struct list *ptr;
|
|
|
|
RtlEnterCriticalSection( &process_heap->cs );
|
|
LIST_FOR_EACH( ptr, &process_heap->entry ) total++;
|
|
if (total <= count)
|
|
{
|
|
*heaps++ = process_heap;
|
|
LIST_FOR_EACH( ptr, &process_heap->entry )
|
|
*heaps++ = LIST_ENTRY( ptr, struct heap, entry );
|
|
}
|
|
RtlLeaveCriticalSection( &process_heap->cs );
|
|
return total;
|
|
}
|
|
|
|
/***********************************************************************
|
|
* RtlQueryHeapInformation (NTDLL.@)
|
|
*/
|
|
NTSTATUS WINAPI RtlQueryHeapInformation( HANDLE handle, HEAP_INFORMATION_CLASS info_class,
|
|
void *info, SIZE_T size_in, PSIZE_T size_out )
|
|
{
|
|
switch (info_class)
|
|
{
|
|
case HeapCompatibilityInformation:
|
|
if (size_out) *size_out = sizeof(ULONG);
|
|
|
|
if (size_in < sizeof(ULONG))
|
|
return STATUS_BUFFER_TOO_SMALL;
|
|
|
|
*(ULONG *)info = 0; /* standard heap */
|
|
return STATUS_SUCCESS;
|
|
|
|
default:
|
|
FIXME("Unknown heap information class %u\n", info_class);
|
|
return STATUS_INVALID_INFO_CLASS;
|
|
}
|
|
}
|
|
|
|
/***********************************************************************
|
|
* RtlSetHeapInformation (NTDLL.@)
|
|
*/
|
|
NTSTATUS WINAPI RtlSetHeapInformation( HANDLE handle, HEAP_INFORMATION_CLASS info_class, void *info, SIZE_T size )
|
|
{
|
|
FIXME( "handle %p, info_class %d, info %p, size %Id stub!\n", handle, info_class, info, size );
|
|
return STATUS_SUCCESS;
|
|
}
|
|
|
|
/***********************************************************************
|
|
* RtlGetUserInfoHeap (NTDLL.@)
|
|
*/
|
|
BOOLEAN WINAPI RtlGetUserInfoHeap( HANDLE handle, ULONG flags, void *ptr, void **user_value, ULONG *user_flags )
|
|
{
|
|
NTSTATUS status = STATUS_SUCCESS;
|
|
struct block *block;
|
|
struct heap *heap;
|
|
ULONG heap_flags;
|
|
char *tmp;
|
|
|
|
TRACE( "handle %p, flags %#lx, ptr %p, user_value %p, user_flags %p semi-stub!\n",
|
|
handle, flags, ptr, user_value, user_flags );
|
|
|
|
*user_flags = 0;
|
|
|
|
if (!(heap = unsafe_heap_from_handle( handle, flags, &heap_flags )))
|
|
status = STATUS_SUCCESS;
|
|
else if (!(block = unsafe_block_from_ptr( heap, heap_flags, ptr )))
|
|
{
|
|
status = STATUS_INVALID_PARAMETER;
|
|
*user_value = 0;
|
|
}
|
|
else if (!(*user_flags = HEAP_USER_FLAGS(block_get_flags( block ))))
|
|
WARN( "Block %p wasn't allocated with user info\n", ptr );
|
|
else if (block_get_flags( block ) & BLOCK_FLAG_LARGE)
|
|
{
|
|
const ARENA_LARGE *large = CONTAINING_RECORD( block, ARENA_LARGE, block );
|
|
*user_flags = *user_flags & ~HEAP_ADD_USER_INFO;
|
|
*user_value = large->user_value;
|
|
}
|
|
else
|
|
{
|
|
heap_lock( heap, heap_flags );
|
|
|
|
tmp = (char *)block + block_get_size( block ) - block->tail_size + sizeof(void *);
|
|
if ((heap_flags & HEAP_TAIL_CHECKING_ENABLED) || RUNNING_ON_VALGRIND) tmp += BLOCK_ALIGN;
|
|
*user_flags = *user_flags & ~HEAP_ADD_USER_INFO;
|
|
*user_value = *(void **)tmp;
|
|
|
|
heap_unlock( heap, heap_flags );
|
|
}
|
|
|
|
heap_set_status( heap, flags, status );
|
|
return !status;
|
|
}
|
|
|
|
/***********************************************************************
|
|
* RtlSetUserValueHeap (NTDLL.@)
|
|
*/
|
|
BOOLEAN WINAPI RtlSetUserValueHeap( HANDLE handle, ULONG flags, void *ptr, void *user_value )
|
|
{
|
|
struct block *block;
|
|
struct heap *heap;
|
|
ULONG heap_flags;
|
|
BOOLEAN ret;
|
|
char *tmp;
|
|
|
|
TRACE( "handle %p, flags %#lx, ptr %p, user_value %p.\n", handle, flags, ptr, user_value );
|
|
|
|
if (!(heap = unsafe_heap_from_handle( handle, flags, &heap_flags )))
|
|
ret = TRUE;
|
|
else if (!(block = unsafe_block_from_ptr( heap, heap_flags, ptr )))
|
|
ret = FALSE;
|
|
else if (!(block_get_flags( block ) & BLOCK_FLAG_USER_INFO))
|
|
ret = FALSE;
|
|
else if (block_get_flags( block ) & BLOCK_FLAG_LARGE)
|
|
{
|
|
ARENA_LARGE *large = CONTAINING_RECORD( block, ARENA_LARGE, block );
|
|
large->user_value = user_value;
|
|
ret = TRUE;
|
|
}
|
|
else
|
|
{
|
|
heap_lock( heap, heap_flags );
|
|
|
|
tmp = (char *)block + block_get_size( block ) - block->tail_size + sizeof(void *);
|
|
if ((heap_flags & HEAP_TAIL_CHECKING_ENABLED) || RUNNING_ON_VALGRIND) tmp += BLOCK_ALIGN;
|
|
*(void **)tmp = user_value;
|
|
ret = TRUE;
|
|
|
|
heap_unlock( heap, heap_flags );
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
/***********************************************************************
|
|
* RtlSetUserFlagsHeap (NTDLL.@)
|
|
*/
|
|
BOOLEAN WINAPI RtlSetUserFlagsHeap( HANDLE handle, ULONG flags, void *ptr, ULONG clear, ULONG set )
|
|
{
|
|
struct block *block;
|
|
struct heap *heap;
|
|
ULONG heap_flags;
|
|
BOOLEAN ret;
|
|
|
|
TRACE( "handle %p, flags %#lx, ptr %p, clear %#lx, set %#lx.\n", handle, flags, ptr, clear, set );
|
|
|
|
if ((clear | set) & ~(0xe00))
|
|
{
|
|
SetLastError( ERROR_INVALID_PARAMETER );
|
|
return FALSE;
|
|
}
|
|
|
|
if (!(heap = unsafe_heap_from_handle( handle, flags, &heap_flags )))
|
|
ret = TRUE;
|
|
else if (!(block = unsafe_block_from_ptr( heap, heap_flags, ptr )))
|
|
ret = FALSE;
|
|
else if (!(block_get_flags( block ) & BLOCK_FLAG_USER_INFO))
|
|
ret = FALSE;
|
|
else
|
|
{
|
|
block_set_flags( block, BLOCK_USER_FLAGS( clear ), BLOCK_USER_FLAGS( set ) );
|
|
ret = TRUE;
|
|
}
|
|
|
|
return ret;
|
|
}
|