Lines Matching refs:oldp
3426 const mchunkptr oldp = mem2chunk (oldmem);
3428 const INTERNAL_SIZE_T oldsize = chunksize (oldp);
3430 if (chunk_is_mmapped (oldp))
3435 ar_ptr = arena_for_chunk (oldp);
3442 if ((__builtin_expect ((uintptr_t) oldp > (uintptr_t) -oldsize, 0)
3443 || __builtin_expect (misaligned_chunk (oldp), 0)))
3452 if (chunk_is_mmapped (oldp))
3457 newp = mremap_chunk (oldp, nb);
3479 munmap_chunk (oldp);
3485 newp = _int_realloc (ar_ptr, oldp, oldsize, nb);
3494 newp = _int_realloc (ar_ptr, oldp, oldsize, nb);
3507 size_t sz = memsize (oldp);
3509 (void) tag_region (chunk2mem (oldp), sz);
3510 _int_free (ar_ptr, oldp, 0);
4816 _int_realloc (mstate av, mchunkptr oldp, INTERNAL_SIZE_T oldsize,
4823 mchunkptr next; /* next contiguous chunk after oldp */
4829 if (__builtin_expect (chunksize_nomask (oldp) <= CHUNK_HDR_SZ, 0)
4833 check_inuse_chunk (av, oldp);
4836 assert (!chunk_is_mmapped (oldp));
4838 next = chunk_at_offset (oldp, oldsize);
4847 newp = oldp;
4858 set_head_size (oldp, nb | (av != &main_arena ? NON_MAIN_ARENA : 0));
4859 av->top = chunk_at_offset (oldp, nb);
4861 check_inuse_chunk (av, oldp);
4862 return tag_new_usable (chunk2mem (oldp));
4871 newp = oldp;
4886 Avoid copy if newp is next chunk after oldp.
4891 newp = oldp;
4895 void *oldmem = chunk2mem (oldp);
4896 size_t sz = memsize (oldp);
4900 _int_free (av, oldp, 1);