Lines Matching refs:old_size

2509 			 INTERNAL_SIZE_T old_size, size_t minsize,
2516 size = ALIGN_UP (size + old_size, pagesize);
2550 INTERNAL_SIZE_T old_size; /* its size */
2607 old_size = chunksize (old_top);
2608 old_end = (char *) (chunk_at_offset (old_top, old_size));
2613 If not the first time through, we require old_size to be
2617 assert ((old_top == initial_top (av) && old_size == 0) ||
2618 ((unsigned long) (old_size) >= MINSIZE &&
2623 assert ((unsigned long) (old_size) < (unsigned long) (nb + MINSIZE));
2634 if ((long) (MINSIZE + nb - old_size) > 0
2635 && grow_heap (old_heap, MINSIZE + nb - old_size) == 0)
2656 old_size = (old_size - MINSIZE) & ~MALLOC_ALIGN_MASK;
2657 set_head (chunk_at_offset (old_top, old_size + CHUNK_HDR_SZ),
2659 if (old_size >= MINSIZE)
2661 set_head (chunk_at_offset (old_top, old_size),
2663 set_foot (chunk_at_offset (old_top, old_size), CHUNK_HDR_SZ);
2664 set_head (old_top, old_size | PREV_INUSE | NON_MAIN_ARENA);
2669 set_head (old_top, (old_size + CHUNK_HDR_SZ) | PREV_INUSE);
2670 set_foot (old_top, (old_size + CHUNK_HDR_SZ));
2696 size -= old_size;
2747 mbrk = sysmalloc_mmap_fallback (&size, nb, old_size,
2752 mbrk = sysmalloc_mmap_fallback (&size, nb, old_size, pagesize,
2773 set_head (old_top, (size + old_size) | PREV_INUSE);
2775 else if (contiguous (av) && old_size && brk < old_end)
2809 if (old_size)
2834 correction += old_size;
2908 if (old_size != 0)
2915 old_size = (old_size - 2 * CHUNK_HDR_SZ) & ~MALLOC_ALIGN_MASK;
2916 set_head (old_top, old_size | PREV_INUSE);
2920 old_top when old_size was previously MINSIZE. This is
2924 set_head (chunk_at_offset (old_top, old_size),
2927 old_size + CHUNK_HDR_SZ),
2931 if (old_size >= MINSIZE)