Reputation: 59269
std::realloc
is dangerous in c++ if the malloc'd memory contains non-pod types. It seems the only problem is that std::realloc
wont call the type destructors if it cannot grow the memory in situ.
A trivial work around would be a try_realloc
function. Instead of malloc'ing new memory if it cannot be grown in situ, it would simply return false. In which case new memory could be allocated, the objects copied (or moved) to the new memory, and finally the old memory freed.
This seems supremely useful. std::vector
could make great use of this, possibly avoiding all copies/reallocations.
preemptive flame retardant: Technically, that is same Big-O performance, but if vector growth is a bottle neck in your application a x2 speed up is nice even if the Big-O remains unchanged.
BUT, I cannot find any c api that works like a try_realloc
.
Am I missing something? Is try_realloc
not as useful as I imagine? Is there some hidden bug that makes try_realloc
unusable?
Better yet, Is there some less documented API that performs like try_realloc
?
NOTE: I'm obviously, in library/platform specific code here. I'm not worried as try_realloc
is inherently an optimization.
Update:
Following Steve Jessops comment's on whether vector
would be more efficient using realloc I wrote up a proof of concept to test. The realloc-vector
simulates a vector's growth pattern but has the option to realloc instead. I ran the program up to a million elements in the vector.
For comparison a vector
must allocate 19 times while growing to a million elements.
The results, if the realloc-vector
is the only thing using the heap the results are awesome, 3-4 allocation while growing to the size of million bytes.
If the realloc-vector
is used alongside a vector
that grows at 66% the speed of the realloc-vector
The results are less promising, allocating 8-10 times during growth.
Finally, if the realloc-vector
is used alongside a vector
that grows at the same rate, the realloc-vector
allocates 17-18 times. Barely saving one allocation over the standard vector behavior.
I don't doubt that a hacker could game allocation sizes to improve the savings, but I agree with Steve that the tremendous effort to write and maintain such an allocator isn't work the gain.
Upvotes: 17
Views: 2773
Reputation: 215193
realloc
in C is hardly more than a convenience function; it has very little benefit for performance/reducing copies. The main exception I can think of is code that allocates a big array then reduces the size once the size needed is known - but even this might require moving data on some malloc
implementations (ones which segregate blocks strictly by size) so I consider this usage of realloc
really bad practice.
As long as you don't constantly reallocate your array every time you add an element, but instead grow the array exponentially (e.g. by 25%, 50%, or 100%) whenever you run out of space, just manually allocating new memory, copying, and freeing the old will yield roughly the same (and identical, in case of memory fragmentation) performance to using realloc
. This is surely the approach that C++ STL implementations use, so I think your whole concern is unfounded.
Edit: The one (rare but not unheard-of) case where realloc
is actually useful is for giant blocks on systems with virtual memory, where the C library interacts with the kernel to relocate whole pages to new addresses. The reason I say this is rare is because you need to be dealing with very big blocks (at least several hundred kB) before most implementations will even enter the realm of dealing with page-granularity allocation, and probably much larger (several MB maybe) before entering and exiting kernelspace to rearrange virtual memory is cheaper than simply doing the copy. Of course try_realloc
would not be useful here, since the whole benefit comes from actually doing the move inexpensively.
Upvotes: 2
Reputation: 279225
vector
generally grows in large increments. You can't do that repeatedly without relocating, unless you carefully arrange things so that there's a large extent of free addresses just above the internal buffer of the vector (which in effect requires assigning whole pages, because obviously you can't have other allocations later on the same page).
So I think that in order to get a really good optimization here, you need more than a "trivial workaround" that does a cheap reallocation if possible - you have to somehow do some preparation to make it possible, and that preparation costs you address space. If you only do it for certain vectors, ones that indicate they're going to become big, then it's fairly pointless, because they can indicate with reserve()
that they're going to become big. You can only do it automatically for all vectors if you have a vast address space, so that you can "waste" a big chunk of it on every vector.
As I understand it, the reason that the Allocator
concept has no reallocation function is to keep it simple. If std::allocator
had a try_realloc
function, then either every Allocator would have to have one (which in most cases couldn't be implemented, and would just have to return false always), or else every standard container would have to be specialized for std::allocator
to take advantage of it. Neither option is a great Allocator interface, although I suppose it wouldn't be a huge effort for implementers of almost all Allocator classes just to add a do-nothing try_realloc
function.
If vector
is slow due to re-allocation, deque
might be a good replacement.
Upvotes: 11
Reputation: 88711
You could implement something like the try_realloc
you proposed, using mmap
with MAP_ANONYMOUS
and MAP_FIXED
and mremap
with MREMAP_FIXED
.
Edit: just noticed that the man page for mremap even says:
mremap() uses the Linux page table scheme. mremap() changes the mapping between virtual addresses and memory pages. This can be used to implement a very efficient realloc(3).
Upvotes: 4