зеркало из https://github.com/mozilla/pjs.git
561 строка
18 KiB
C
561 строка
18 KiB
C
/* -*- Mode: C; tab-width: 8; indent-tabs-mode: nil; c-basic-offset: 4 -*-
|
|
*
|
|
* The contents of this file are subject to the Netscape Public
|
|
* License Version 1.1 (the "License"); you may not use this file
|
|
* except in compliance with the License. You may obtain a copy of
|
|
* the License at http://www.mozilla.org/NPL/
|
|
*
|
|
* Software distributed under the License is distributed on an "AS
|
|
* IS" basis, WITHOUT WARRANTY OF ANY KIND, either express oqr
|
|
* implied. See the License for the specific language governing
|
|
* rights and limitations under the License.
|
|
*
|
|
* The Original Code is Mozilla Communicator client code, released
|
|
* March 31, 1998.
|
|
*
|
|
* The Initial Developer of the Original Code is Netscape
|
|
* Communications Corporation. Portions created by Netscape are
|
|
* Copyright (C) 1998 Netscape Communications Corporation. All
|
|
* Rights Reserved.
|
|
*
|
|
* Contributor(s):
|
|
*
|
|
* Alternatively, the contents of this file may be used under the
|
|
* terms of the GNU Public License (the "GPL"), in which case the
|
|
* provisions of the GPL are applicable instead of those above.
|
|
* If you wish to allow use of your version of this file only
|
|
* under the terms of the GPL and not to allow others to use your
|
|
* version of this file under the NPL, indicate your decision by
|
|
* deleting the provisions above and replace them with the notice
|
|
* and other provisions required by the GPL. If you do not delete
|
|
* the provisions above, a recipient may use your version of this
|
|
* file under either the NPL or the GPL.
|
|
*/
|
|
|
|
/*
|
|
* Lifetime-based fast allocation, inspired by much prior art, including
|
|
* "Fast Allocation and Deallocation of Memory Based on Object Lifetimes"
|
|
* David R. Hanson, Software -- Practice and Experience, Vol. 20(1).
|
|
*/
|
|
#include "jsstddef.h"
|
|
#include <stdlib.h>
|
|
#include <string.h>
|
|
#include "jstypes.h"
|
|
#include "jsbit.h"
|
|
#include "jsarena.h" /* Added by JSIFY */
|
|
#include "jsutil.h" /* Added by JSIFY */
|
|
#include "jslock.h"
|
|
|
|
static JSArena *arena_freelist;
|
|
|
|
#ifdef JS_THREADSAFE
|
|
static JSLock *arena_freelist_lock;
|
|
#endif
|
|
|
|
#ifdef JS_ARENAMETER
|
|
static JSArenaStats *arena_stats_list;
|
|
|
|
#define COUNT(pool,what) (pool)->stats.what++
|
|
#else
|
|
#define COUNT(pool,what) /* nothing */
|
|
#endif
|
|
|
|
#define JS_ARENA_DEFAULT_ALIGN sizeof(double)
|
|
|
|
JS_PUBLIC_API(void)
|
|
JS_InitArenaPool(JSArenaPool *pool, const char *name, size_t size, size_t align)
|
|
{
|
|
#ifdef JS_THREADSAFE
|
|
/* Must come through here once in primordial thread to init safely! */
|
|
if (!arena_freelist_lock) {
|
|
arena_freelist_lock = JS_NEW_LOCK();
|
|
JS_ASSERT(arena_freelist_lock);
|
|
}
|
|
#endif
|
|
if (align == 0)
|
|
align = JS_ARENA_DEFAULT_ALIGN;
|
|
pool->mask = JS_BITMASK(JS_CeilingLog2(align));
|
|
pool->first.next = NULL;
|
|
pool->first.base = pool->first.avail = pool->first.limit =
|
|
JS_ARENA_ALIGN(pool, &pool->first + 1);
|
|
pool->current = &pool->first;
|
|
pool->arenasize = size;
|
|
#ifdef JS_ARENAMETER
|
|
memset(&pool->stats, 0, sizeof pool->stats);
|
|
pool->stats.name = strdup(name);
|
|
pool->stats.next = arena_stats_list;
|
|
arena_stats_list = &pool->stats;
|
|
#endif
|
|
}
|
|
|
|
/*
|
|
* An allocation that consumes more than pool->arenasize also has a header
|
|
* pointing back to its previous arena's next member. This header is not
|
|
* included in [a->base, a->limit), so its space can't be wrongly claimed.
|
|
*
|
|
* As the header is a pointer, it must be well-aligned. If pool->mask is
|
|
* greater than or equal to POINTER_MASK, the header just preceding a->base
|
|
* for an oversized arena a is well-aligned, because a->base is well-aligned.
|
|
* However, we may need to add more space to pad the JSArena ** back-pointer
|
|
* so that it lies just behind a->base, because a might not be aligned such
|
|
* that (jsuword)(a + 1) is on a pointer boundary.
|
|
*
|
|
* By how much must we pad? Let M be the alignment modulus for pool and P
|
|
* the modulus for a pointer. Given M >= P, the greatest distance between a
|
|
* pointer aligned on an M boundary and one aligned on a P boundary is M-P.
|
|
* If M and P are powers of two, then M-P = (pool->mask - POINTER_MASK).
|
|
*
|
|
* How much extra padding might spill over unused into the remainder of the
|
|
* allocation, in the worst case (where M > P)?
|
|
*
|
|
* If we add M-P to the nominal back-pointer address and then round down to
|
|
* align on a P boundary, we will use at most M-P bytes of padding, and at
|
|
* least P (M > P => M >= 2P; M == 2P gives the least padding, P). So if we
|
|
* use P bytes of padding, then we will overallocate a by P+M-1 bytes, as we
|
|
* also add M-1 to the estimated size in case malloc returns an odd pointer.
|
|
* a->limit must include this overestimation to satisfy a->avail in [a->base,
|
|
* a->limit].
|
|
*
|
|
* Similarly, if pool->mask is less than POINTER_MASK, we must include enough
|
|
* space in the header size to align the back-pointer on a P boundary so that
|
|
* it can be found by subtracting P from a->base. This means a->base must be
|
|
* on a P boundary, even though subsequent allocations from a may be aligned
|
|
* on a lesser (M) boundary. Given powers of two M and P as above, the extra
|
|
* space needed when P > M is P-M or POINTER_MASK - pool->mask.
|
|
*
|
|
* The size of a header including padding is given by the HEADER_SIZE macro,
|
|
* below, for any pool (for any value of M).
|
|
*
|
|
* The mask to align a->base for any pool is (pool->mask | POINTER_MASK), or
|
|
* HEADER_BASE_MASK(pool).
|
|
*
|
|
* PTR_TO_HEADER computes the address of the back-pointer, given an oversized
|
|
* allocation at p. By definition, p must be a->base for the arena a that
|
|
* contains p. GET_HEADER and SET_HEADER operate on an oversized arena a, in
|
|
* the case of SET_HEADER with back-pointer ap.
|
|
*/
|
|
#define POINTER_MASK ((jsuword)(JS_ALIGN_OF_POINTER - 1))
|
|
#define HEADER_SIZE(pool) (sizeof(JSArena **) \
|
|
+ (((pool)->mask < POINTER_MASK) \
|
|
? POINTER_MASK - (pool)->mask \
|
|
: (pool)->mask - POINTER_MASK))
|
|
#define HEADER_BASE_MASK(pool) ((pool)->mask | POINTER_MASK)
|
|
#define PTR_TO_HEADER(pool,p) (JS_ASSERT(((jsuword)(p) \
|
|
& HEADER_BASE_MASK(pool)) \
|
|
== 0), \
|
|
(JSArena ***)(p) - 1)
|
|
#define GET_HEADER(pool,a) (*PTR_TO_HEADER(pool, (a)->base))
|
|
#define SET_HEADER(pool,a,ap) (*PTR_TO_HEADER(pool, (a)->base) = (ap))
|
|
|
|
JS_PUBLIC_API(void *)
|
|
JS_ArenaAllocate(JSArenaPool *pool, size_t nb)
|
|
{
|
|
JSArena **ap, **bp, *a, *b;
|
|
jsuword extra, hdrsz, gross, sz;
|
|
void *p;
|
|
|
|
/* Search pool from current forward till we find or make enough space. */
|
|
JS_ASSERT((nb & pool->mask) == 0);
|
|
for (a = pool->current; a->avail + nb > a->limit; pool->current = a) {
|
|
ap = &a->next;
|
|
if (!*ap) {
|
|
/* Not enough space in pool -- try to reclaim a free arena. */
|
|
extra = (nb > pool->arenasize) ? HEADER_SIZE(pool) : 0;
|
|
hdrsz = sizeof *a + extra + pool->mask;
|
|
gross = hdrsz + JS_MAX(nb, pool->arenasize);
|
|
bp = &arena_freelist;
|
|
JS_ACQUIRE_LOCK(arena_freelist_lock);
|
|
while ((b = *bp) != NULL) {
|
|
/*
|
|
* Insist on exact arenasize match if nb is not greater than
|
|
* arenasize. Otherwise take any arena big enough, but not by
|
|
* more than gross + arenasize.
|
|
*/
|
|
sz = JS_UPTRDIFF(b->limit, b);
|
|
if (extra
|
|
? sz >= gross && sz <= gross + pool->arenasize
|
|
: sz == gross) {
|
|
*bp = b->next;
|
|
JS_RELEASE_LOCK(arena_freelist_lock);
|
|
b->next = NULL;
|
|
COUNT(pool, nreclaims);
|
|
goto claim;
|
|
}
|
|
bp = &b->next;
|
|
}
|
|
|
|
/* Nothing big enough on the freelist, so we must malloc. */
|
|
JS_RELEASE_LOCK(arena_freelist_lock);
|
|
b = (JSArena *) malloc(gross);
|
|
if (!b)
|
|
return 0;
|
|
b->next = NULL;
|
|
b->limit = (jsuword)b + gross;
|
|
JS_COUNT_ARENA(pool,++);
|
|
COUNT(pool, nmallocs);
|
|
|
|
claim:
|
|
/* If oversized, store ap in the header, just before a->base. */
|
|
*ap = a = b;
|
|
JS_ASSERT(gross <= JS_UPTRDIFF(a->limit, a));
|
|
if (extra) {
|
|
a->base = a->avail =
|
|
((jsuword)a + hdrsz) & ~HEADER_BASE_MASK(pool);
|
|
SET_HEADER(pool, a, ap);
|
|
} else {
|
|
a->base = a->avail = JS_ARENA_ALIGN(pool, a + 1);
|
|
}
|
|
continue;
|
|
}
|
|
a = *ap; /* move to next arena */
|
|
}
|
|
|
|
p = (void *)a->avail;
|
|
a->avail += nb;
|
|
JS_ASSERT(a->base <= a->avail && a->avail <= a->limit);
|
|
return p;
|
|
}
|
|
|
|
JS_PUBLIC_API(void *)
|
|
JS_ArenaRealloc(JSArenaPool *pool, void *p, size_t size, size_t incr)
|
|
{
|
|
JSArena **ap, *a, *b;
|
|
jsuword boff, aoff, extra, hdrsz, gross;
|
|
|
|
/*
|
|
* Use the oversized-single-allocation header to avoid searching for ap.
|
|
* See JS_ArenaAllocate, the SET_HEADER call.
|
|
*/
|
|
if (size > pool->arenasize) {
|
|
ap = *PTR_TO_HEADER(pool, p);
|
|
a = *ap;
|
|
} else {
|
|
ap = &pool->first.next;
|
|
while ((a = *ap) != pool->current)
|
|
ap = &a->next;
|
|
}
|
|
|
|
JS_ASSERT(a->base == (jsuword)p);
|
|
boff = JS_UPTRDIFF(a->base, a);
|
|
aoff = size + incr;
|
|
JS_ASSERT(aoff > pool->arenasize);
|
|
extra = HEADER_SIZE(pool); /* oversized header holds ap */
|
|
hdrsz = sizeof *a + extra + pool->mask; /* header and alignment slop */
|
|
gross = hdrsz + aoff;
|
|
a = (JSArena *) realloc(a, gross);
|
|
if (!a)
|
|
return NULL;
|
|
#ifdef JS_ARENAMETER
|
|
pool->stats.nreallocs++;
|
|
#endif
|
|
|
|
if (a != *ap) {
|
|
/* Oops, realloc moved the allocation: update other pointers to a. */
|
|
if (pool->current == *ap)
|
|
pool->current = a;
|
|
b = a->next;
|
|
if (b && b->avail - b->base > pool->arenasize) {
|
|
JS_ASSERT(GET_HEADER(pool, b) == &(*ap)->next);
|
|
SET_HEADER(pool, b, &a->next);
|
|
}
|
|
|
|
/* Now update *ap, the next link of the arena before a. */
|
|
*ap = a;
|
|
}
|
|
|
|
a->base = ((jsuword)a + hdrsz) & ~HEADER_BASE_MASK(pool);
|
|
a->limit = (jsuword)a + gross;
|
|
a->avail = JS_ARENA_ALIGN(pool, a->base + aoff);
|
|
JS_ASSERT(a->base <= a->avail && a->avail <= a->limit);
|
|
|
|
/* Check whether realloc aligned differently, and copy if necessary. */
|
|
if (boff != JS_UPTRDIFF(a->base, a))
|
|
memmove((void *)a->base, (char *)a + boff, size);
|
|
|
|
/* Store ap in the oversized-load arena header. */
|
|
SET_HEADER(pool, a, ap);
|
|
return (void *)a->base;
|
|
}
|
|
|
|
JS_PUBLIC_API(void *)
|
|
JS_ArenaGrow(JSArenaPool *pool, void *p, size_t size, size_t incr)
|
|
{
|
|
void *newp;
|
|
|
|
/*
|
|
* If p points to an oversized allocation, it owns an entire arena, so we
|
|
* can simply realloc the arena.
|
|
*/
|
|
if (size > pool->arenasize)
|
|
return JS_ArenaRealloc(pool, p, size, incr);
|
|
|
|
JS_ARENA_ALLOCATE(newp, pool, size + incr);
|
|
if (newp)
|
|
memcpy(newp, p, size);
|
|
return newp;
|
|
}
|
|
|
|
/*
|
|
* Free tail arenas linked after head, which may not be the true list head.
|
|
* Reset pool->current to point to head in case it pointed at a tail arena.
|
|
*/
|
|
static void
|
|
FreeArenaList(JSArenaPool *pool, JSArena *head, JSBool reallyFree)
|
|
{
|
|
JSArena **ap, *a;
|
|
|
|
ap = &head->next;
|
|
a = *ap;
|
|
if (!a)
|
|
return;
|
|
|
|
#ifdef DEBUG
|
|
do {
|
|
JS_ASSERT(a->base <= a->avail && a->avail <= a->limit);
|
|
a->avail = a->base;
|
|
JS_CLEAR_UNUSED(a);
|
|
} while ((a = a->next) != NULL);
|
|
a = *ap;
|
|
#endif
|
|
|
|
if (reallyFree) {
|
|
do {
|
|
*ap = a->next;
|
|
JS_CLEAR_ARENA(a);
|
|
JS_COUNT_ARENA(pool,--);
|
|
free(a);
|
|
} while ((a = *ap) != NULL);
|
|
} else {
|
|
/* Insert the whole arena chain at the front of the freelist. */
|
|
do {
|
|
ap = &(*ap)->next;
|
|
} while (*ap);
|
|
JS_ACQUIRE_LOCK(arena_freelist_lock);
|
|
*ap = arena_freelist;
|
|
arena_freelist = a;
|
|
JS_RELEASE_LOCK(arena_freelist_lock);
|
|
head->next = NULL;
|
|
}
|
|
|
|
pool->current = head;
|
|
}
|
|
|
|
JS_PUBLIC_API(void)
|
|
JS_ArenaRelease(JSArenaPool *pool, char *mark)
|
|
{
|
|
JSArena *a;
|
|
|
|
for (a = &pool->first; a; a = a->next) {
|
|
JS_ASSERT(a->base <= a->avail && a->avail <= a->limit);
|
|
|
|
if (JS_UPTRDIFF(mark, a->base) <= JS_UPTRDIFF(a->avail, a->base)) {
|
|
a->avail = JS_ARENA_ALIGN(pool, mark);
|
|
JS_ASSERT(a->avail <= a->limit);
|
|
FreeArenaList(pool, a, JS_TRUE);
|
|
return;
|
|
}
|
|
}
|
|
}
|
|
|
|
JS_PUBLIC_API(void)
|
|
JS_ArenaFreeAllocation(JSArenaPool *pool, void *p, size_t size)
|
|
{
|
|
JSArena **ap, *a, *b;
|
|
jsuword q;
|
|
|
|
/*
|
|
* If the allocation is oversized, it consumes an entire arena, and it has
|
|
* a header just before the allocation pointing back to its predecessor's
|
|
* next member. Otherwise, we have to search pool for a.
|
|
*/
|
|
if (size > pool->arenasize) {
|
|
ap = *PTR_TO_HEADER(pool, p);
|
|
a = *ap;
|
|
} else {
|
|
q = (jsuword)p + size;
|
|
q = JS_ARENA_ALIGN(pool, q);
|
|
ap = &pool->first.next;
|
|
while ((a = *ap) != NULL) {
|
|
JS_ASSERT(a->base <= a->avail && a->avail <= a->limit);
|
|
|
|
if (a->avail == q) {
|
|
/*
|
|
* If a is consumed by the allocation at p, we can free it to
|
|
* the malloc heap.
|
|
*/
|
|
if (a->base == (jsuword)p)
|
|
break;
|
|
|
|
/*
|
|
* We can't free a, but we can "retract" its avail cursor --
|
|
* whether there are others after it in pool.
|
|
*/
|
|
a->avail = (jsuword)p;
|
|
return;
|
|
}
|
|
ap = &a->next;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* At this point, a is doomed, so ensure that pool->current doesn't point
|
|
* at it. What's more, force future allocations to scavenge all arenas on
|
|
* pool, in case some have free space.
|
|
*/
|
|
if (pool->current == a)
|
|
pool->current = &pool->first;
|
|
|
|
/*
|
|
* This is a non-LIFO deallocation, so take care to fix up a->next's back
|
|
* pointer in its header, if a->next is oversized.
|
|
*/
|
|
*ap = b = a->next;
|
|
if (b && b->avail - b->base > pool->arenasize) {
|
|
JS_ASSERT(GET_HEADER(pool, b) == &a->next);
|
|
SET_HEADER(pool, b, ap);
|
|
}
|
|
JS_CLEAR_ARENA(a);
|
|
JS_COUNT_ARENA(pool,--);
|
|
free(a);
|
|
}
|
|
|
|
JS_PUBLIC_API(void)
|
|
JS_FreeArenaPool(JSArenaPool *pool)
|
|
{
|
|
FreeArenaList(pool, &pool->first, JS_FALSE);
|
|
COUNT(pool, ndeallocs);
|
|
}
|
|
|
|
JS_PUBLIC_API(void)
|
|
JS_FinishArenaPool(JSArenaPool *pool)
|
|
{
|
|
FreeArenaList(pool, &pool->first, JS_TRUE);
|
|
#ifdef JS_ARENAMETER
|
|
{
|
|
JSArenaStats *stats, **statsp;
|
|
|
|
if (pool->stats.name)
|
|
free(pool->stats.name);
|
|
for (statsp = &arena_stats_list; (stats = *statsp) != 0;
|
|
statsp = &stats->next) {
|
|
if (stats == &pool->stats) {
|
|
*statsp = stats->next;
|
|
return;
|
|
}
|
|
}
|
|
}
|
|
#endif
|
|
}
|
|
|
|
JS_PUBLIC_API(void)
|
|
JS_ArenaFinish()
|
|
{
|
|
JSArena *a, *next;
|
|
|
|
JS_ACQUIRE_LOCK(arena_freelist_lock);
|
|
a = arena_freelist;
|
|
arena_freelist = NULL;
|
|
JS_RELEASE_LOCK(arena_freelist_lock);
|
|
for (; a; a = next) {
|
|
next = a->next;
|
|
free(a);
|
|
}
|
|
}
|
|
|
|
JS_PUBLIC_API(void)
|
|
JS_ArenaShutDown(void)
|
|
{
|
|
#ifdef JS_THREADSAFE
|
|
/* Must come through here once in the process's last thread! */
|
|
if (arena_freelist_lock) {
|
|
JS_DESTROY_LOCK(arena_freelist_lock);
|
|
arena_freelist_lock = NULL;
|
|
}
|
|
#endif
|
|
}
|
|
|
|
#ifdef JS_ARENAMETER
|
|
JS_PUBLIC_API(void)
|
|
JS_ArenaCountAllocation(JSArenaPool *pool, size_t nb)
|
|
{
|
|
pool->stats.nallocs++;
|
|
pool->stats.nbytes += nb;
|
|
if (nb > pool->stats.maxalloc)
|
|
pool->stats.maxalloc = nb;
|
|
pool->stats.variance += nb * nb;
|
|
}
|
|
|
|
JS_PUBLIC_API(void)
|
|
JS_ArenaCountInplaceGrowth(JSArenaPool *pool, size_t size, size_t incr)
|
|
{
|
|
pool->stats.ninplace++;
|
|
}
|
|
|
|
JS_PUBLIC_API(void)
|
|
JS_ArenaCountGrowth(JSArenaPool *pool, size_t size, size_t incr)
|
|
{
|
|
pool->stats.ngrows++;
|
|
pool->stats.nbytes += incr;
|
|
pool->stats.variance -= size * size;
|
|
size += incr;
|
|
if (size > pool->stats.maxalloc)
|
|
pool->stats.maxalloc = size;
|
|
pool->stats.variance += size * size;
|
|
}
|
|
|
|
JS_PUBLIC_API(void)
|
|
JS_ArenaCountRelease(JSArenaPool *pool, char *mark)
|
|
{
|
|
pool->stats.nreleases++;
|
|
}
|
|
|
|
JS_PUBLIC_API(void)
|
|
JS_ArenaCountRetract(JSArenaPool *pool, char *mark)
|
|
{
|
|
pool->stats.nfastrels++;
|
|
}
|
|
|
|
#include <math.h>
|
|
#include <stdio.h>
|
|
|
|
JS_PUBLIC_API(void)
|
|
JS_DumpArenaStats(FILE *fp)
|
|
{
|
|
JSArenaStats *stats;
|
|
uint32 nallocs, nbytes;
|
|
double mean, variance, sigma;
|
|
|
|
for (stats = arena_stats_list; stats; stats = stats->next) {
|
|
nallocs = stats->nallocs;
|
|
if (nallocs != 0) {
|
|
nbytes = stats->nbytes;
|
|
mean = (double)nbytes / nallocs;
|
|
variance = stats->variance * nallocs - nbytes * nbytes;
|
|
if (variance < 0 || nallocs == 1)
|
|
variance = 0;
|
|
else
|
|
variance /= nallocs * (nallocs - 1);
|
|
sigma = sqrt(variance);
|
|
} else {
|
|
mean = variance = sigma = 0;
|
|
}
|
|
|
|
fprintf(fp, "\n%s allocation statistics:\n", stats->name);
|
|
fprintf(fp, " number of arenas: %u\n", stats->narenas);
|
|
fprintf(fp, " number of allocations: %u\n", stats->nallocs);
|
|
fprintf(fp, " number of free arena reclaims: %u\n", stats->nreclaims);
|
|
fprintf(fp, " number of malloc calls: %u\n", stats->nmallocs);
|
|
fprintf(fp, " number of deallocations: %u\n", stats->ndeallocs);
|
|
fprintf(fp, " number of allocation growths: %u\n", stats->ngrows);
|
|
fprintf(fp, " number of in-place growths: %u\n", stats->ninplace);
|
|
fprintf(fp, " number of realloc'ing growths: %u\n", stats->nreallocs);
|
|
fprintf(fp, "number of released allocations: %u\n", stats->nreleases);
|
|
fprintf(fp, " number of fast releases: %u\n", stats->nfastrels);
|
|
fprintf(fp, " total bytes allocated: %u\n", stats->nbytes);
|
|
fprintf(fp, " mean allocation size: %g\n", mean);
|
|
fprintf(fp, " standard deviation: %g\n", sigma);
|
|
fprintf(fp, " maximum allocation size: %u\n", stats->maxalloc);
|
|
}
|
|
}
|
|
#endif /* JS_ARENAMETER */
|