In our application (a network daemon), there are roughly three uses of heap allocated memory.
Memory allocated on startup to hold the result of parsing the application's global configuration.
Memory allocated for thread-specific data as threads are created (and freed as they're destroyed).
Memory allocated when servicing requests and bound to the lifetime of the request.
We use talloc to manage memory in all three cases.
We've recently run into some memory corruption issues where bad pointer values have meant one or more of the threads are writing to the global configuration and causing crashes.
Because of the way the application is structured, nothing should ever write to the memory allocated in case 1) after the application starts processing requests.
Is there a way of marking the memory allocated in case 1) as read-only?
In the POSIX specification there's a function, mprotect.
mprotect
allows the permissions (read/write/execute) on individual pages of memory to be changed.The problem with using
mprotect
to mark up parts of the heap as read-only, is the fact that the highest granularity is a single page, which is usually 4k (dependent on OS/architecture). Padding all heap allocated structures to a multiple of 4k would cause massive memory bloat, boo.So in order to use
mprotect
for case 1) we need to get it all the data we want to protect in one contiguous area of memory.Talloc can help out here. talloc pools are a type of slab allocation that can give large performance gains when used correctly, and (if of sufficient size), allow all allocations within the pool to be done in a single contiguous memory area.
Great! Problem solved, allocate a talloc memory pool, do all the instantiation and parsing work, use
mprotect
to mark the pool as read-only, done! Unfortunately, it's not quite that simple...There are three additional issues to solve:
mprotect
needs memory to be a multiple of the page size.mprotect
needs the start address to be page aligned.Problem 1 is simple enough, we just need to round up to a multiple of the page size (which can be conveniently retrieved with
getpagesize
).Problem 2 it turns out is also pretty easy. If we allocate a single byte within the pool we can predict where the first 'real' allocation will occur. We can also subtract the pool's address from the allocation's address to figure out how much memory talloc is going to use for the chunk headers.
With this information, we can (if needed), perform a second allocation to pad the pool memory to the next page, ensuring 'real' allocations occur within the protected region. We can then return the address of the next page for use in
mprotect
. The only slight issue here, is we need to over-allocate the pool by one page to ensure there's enough memory.Problem 3 is annoying, and solutions are, unfortunately, application specific. If there are no side effects from performing all the instantiation in case 1), and the amount of memory used is consistent, a two-pass approach could be used to figure out how much memory to allocate to the pool. Pass 1 would use
talloc_init
to get the top level chunk andtalloc_total_size
to reveal how much memory was in use, pass 2 would allocate a pool of an appropriate size.For our specific use-case, we just allow the user to determine the pool size. This is because we're using protected memory as a debugging feature, so the user is also a developer, and allocating 1G of memory to ensure there's enough for the configuration is not a problem.
So what does all this look like? Well here's the function I came up with:
The above also uses
talloc_set_memlimit
to ensure no allocations can occur outside of the contiguous region.When there is an errant write into protected memory on macOS you'll see a SEGV, and if running under lldb, you'll get a full backtrace showing exactly where the bad write was.