2000-07-09 12:53:46 +04:00
|
|
|
/* -*- Mode: C; tab-width: 8; indent-tabs-mode: nil; c-basic-offset: 4 -*-
|
|
|
|
*
|
|
|
|
* The contents of this file are subject to the Mozilla Public
|
|
|
|
* License Version 1.1 (the "License"); you may not use this file
|
|
|
|
* except in compliance with the License. You may obtain a copy of
|
|
|
|
* the License at http://www.mozilla.org/MPL/
|
|
|
|
*
|
|
|
|
* Software distributed under the License is distributed on an "AS
|
|
|
|
* IS" basis, WITHOUT WARRANTY OF ANY KIND, either express oqr
|
|
|
|
* implied. See the License for the specific language governing
|
|
|
|
* rights and limitations under the License.
|
|
|
|
*
|
|
|
|
* The Original Code is tmreader.h/tmreader.c code, released
|
|
|
|
* July 7, 2000.
|
|
|
|
*
|
|
|
|
* The Initial Developer of the Original Code is Netscape
|
|
|
|
* Communications Corporation. Portions created by Netscape are
|
|
|
|
* Copyright (C) 2000 Netscape Communications Corporation. All
|
|
|
|
* Rights Reserved.
|
|
|
|
*
|
|
|
|
* Contributor(s):
|
|
|
|
* Brendan Eich, 7-July-2000
|
|
|
|
*
|
|
|
|
* Alternatively, the contents of this file may be used under the
|
|
|
|
* terms of the GNU Public License (the "GPL"), in which case the
|
|
|
|
* provisions of the GPL are applicable instead of those above.
|
|
|
|
* If you wish to allow use of your version of this file only
|
|
|
|
* under the terms of the GPL and not to allow others to use your
|
|
|
|
* version of this file under the MPL, indicate your decision by
|
|
|
|
* deleting the provisions above and replace them with the notice
|
|
|
|
* and other provisions required by the GPL. If you do not delete
|
|
|
|
* the provisions above, a recipient may use your version of this
|
|
|
|
* file under either the MPL or the GPL.
|
|
|
|
*/
|
|
|
|
#include <stdio.h>
|
|
|
|
#include <stdlib.h>
|
|
|
|
#include <string.h>
|
|
|
|
#include <errno.h> /* XXX push error reporting out to clients? */
|
|
|
|
#include <unistd.h>
|
|
|
|
#include "prlog.h"
|
|
|
|
#include "plhash.h"
|
|
|
|
#include "nsTraceMalloc.h"
|
|
|
|
#include "tmreader.h"
|
|
|
|
|
|
|
|
static int accum_byte(FILE *fp, uint32 *uip)
|
|
|
|
{
|
|
|
|
int c = getc(fp);
|
|
|
|
if (c == EOF)
|
|
|
|
return 0;
|
|
|
|
*uip = (*uip << 8) | c;
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int get_uint32(FILE *fp, uint32 *uip)
|
|
|
|
{
|
|
|
|
int c;
|
|
|
|
uint32 ui;
|
|
|
|
|
|
|
|
c = getc(fp);
|
|
|
|
if (c == EOF)
|
|
|
|
return 0;
|
|
|
|
ui = 0;
|
|
|
|
if (c & 0x80) {
|
|
|
|
c &= 0x7f;
|
|
|
|
if (c & 0x40) {
|
|
|
|
c &= 0x3f;
|
|
|
|
if (c & 0x20) {
|
|
|
|
c &= 0x1f;
|
|
|
|
if (c & 0x10) {
|
|
|
|
if (!accum_byte(fp, &ui))
|
|
|
|
return 0;
|
|
|
|
} else {
|
|
|
|
ui = (uint32) c;
|
|
|
|
}
|
|
|
|
if (!accum_byte(fp, &ui))
|
|
|
|
return 0;
|
|
|
|
} else {
|
|
|
|
ui = (uint32) c;
|
|
|
|
}
|
|
|
|
if (!accum_byte(fp, &ui))
|
|
|
|
return 0;
|
|
|
|
} else {
|
|
|
|
ui = (uint32) c;
|
|
|
|
}
|
|
|
|
if (!accum_byte(fp, &ui))
|
|
|
|
return 0;
|
|
|
|
} else {
|
|
|
|
ui = (uint32) c;
|
|
|
|
}
|
|
|
|
*uip = ui;
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
static char *get_string(FILE *fp)
|
|
|
|
{
|
|
|
|
char *cp;
|
|
|
|
int c;
|
|
|
|
static char buf[256];
|
|
|
|
static char *bp = buf, *ep = buf + sizeof buf;
|
|
|
|
static size_t bsize = sizeof buf;
|
|
|
|
|
|
|
|
cp = bp;
|
|
|
|
do {
|
|
|
|
c = getc(fp);
|
|
|
|
if (c == EOF)
|
|
|
|
return 0;
|
|
|
|
if (cp == ep) {
|
|
|
|
if (bp == buf) {
|
|
|
|
bp = malloc(2 * bsize);
|
|
|
|
if (bp)
|
|
|
|
memcpy(bp, buf, bsize);
|
|
|
|
} else {
|
|
|
|
bp = realloc(bp, 2 * bsize);
|
|
|
|
}
|
|
|
|
if (!bp)
|
|
|
|
return 0;
|
|
|
|
cp = bp + bsize;
|
|
|
|
bsize *= 2;
|
|
|
|
ep = bp + bsize;
|
|
|
|
}
|
|
|
|
*cp++ = c;
|
|
|
|
} while (c != '\0');
|
|
|
|
return strdup(bp);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int get_tmevent(FILE *fp, tmevent *event)
|
|
|
|
{
|
|
|
|
int c;
|
|
|
|
char *s;
|
|
|
|
|
|
|
|
c = getc(fp);
|
|
|
|
if (c == EOF)
|
|
|
|
return 0;
|
|
|
|
event->type = (char) c;
|
|
|
|
if (!get_uint32(fp, &event->serial))
|
|
|
|
return 0;
|
|
|
|
switch (c) {
|
2000-07-22 00:42:15 +04:00
|
|
|
case TM_EVENT_LIBRARY:
|
2000-07-09 12:53:46 +04:00
|
|
|
s = get_string(fp);
|
|
|
|
if (!s)
|
|
|
|
return 0;
|
|
|
|
event->u.libname = s;
|
|
|
|
break;
|
|
|
|
|
2000-07-22 00:42:15 +04:00
|
|
|
case TM_EVENT_METHOD:
|
2000-07-09 12:53:46 +04:00
|
|
|
if (!get_uint32(fp, &event->u.method.library))
|
|
|
|
return 0;
|
|
|
|
s = get_string(fp);
|
|
|
|
if (!s)
|
|
|
|
return 0;
|
|
|
|
event->u.method.name = s;
|
|
|
|
break;
|
|
|
|
|
2000-07-22 00:42:15 +04:00
|
|
|
case TM_EVENT_CALLSITE:
|
2000-07-09 12:53:46 +04:00
|
|
|
if (!get_uint32(fp, &event->u.site.parent))
|
|
|
|
return 0;
|
|
|
|
if (!get_uint32(fp, &event->u.site.method))
|
|
|
|
return 0;
|
|
|
|
if (!get_uint32(fp, &event->u.site.offset))
|
|
|
|
return 0;
|
|
|
|
break;
|
|
|
|
|
2000-07-22 00:42:15 +04:00
|
|
|
case TM_EVENT_MALLOC:
|
|
|
|
case TM_EVENT_CALLOC:
|
|
|
|
case TM_EVENT_FREE:
|
2000-07-09 12:53:46 +04:00
|
|
|
event->u.alloc.oldsize = 0;
|
|
|
|
if (!get_uint32(fp, &event->u.alloc.size))
|
|
|
|
return 0;
|
|
|
|
break;
|
|
|
|
|
2000-07-22 00:42:15 +04:00
|
|
|
case TM_EVENT_REALLOC:
|
2000-07-09 12:53:46 +04:00
|
|
|
if (!get_uint32(fp, &event->u.alloc.oldsize))
|
|
|
|
return 0;
|
|
|
|
if (!get_uint32(fp, &event->u.alloc.size))
|
|
|
|
return 0;
|
|
|
|
break;
|
|
|
|
|
2000-07-22 00:42:15 +04:00
|
|
|
case TM_EVENT_STATS:
|
2000-07-09 12:53:46 +04:00
|
|
|
if (!get_uint32(fp, &event->u.stats.tmstats.calltree_maxstack))
|
|
|
|
return 0;
|
|
|
|
if (!get_uint32(fp, &event->u.stats.tmstats.calltree_maxdepth))
|
|
|
|
return 0;
|
|
|
|
if (!get_uint32(fp, &event->u.stats.tmstats.calltree_parents))
|
|
|
|
return 0;
|
|
|
|
if (!get_uint32(fp, &event->u.stats.tmstats.calltree_maxkids))
|
|
|
|
return 0;
|
|
|
|
if (!get_uint32(fp, &event->u.stats.tmstats.calltree_kidhits))
|
|
|
|
return 0;
|
|
|
|
if (!get_uint32(fp, &event->u.stats.tmstats.calltree_kidmisses))
|
|
|
|
return 0;
|
|
|
|
if (!get_uint32(fp, &event->u.stats.tmstats.calltree_kidsteps))
|
|
|
|
return 0;
|
|
|
|
if (!get_uint32(fp, &event->u.stats.tmstats.callsite_recurrences))
|
|
|
|
return 0;
|
|
|
|
if (!get_uint32(fp, &event->u.stats.tmstats.backtrace_calls))
|
|
|
|
return 0;
|
|
|
|
if (!get_uint32(fp, &event->u.stats.tmstats.backtrace_failures))
|
|
|
|
return 0;
|
|
|
|
if (!get_uint32(fp, &event->u.stats.tmstats.btmalloc_failures))
|
|
|
|
return 0;
|
|
|
|
if (!get_uint32(fp, &event->u.stats.tmstats.dladdr_failures))
|
|
|
|
return 0;
|
|
|
|
if (!get_uint32(fp, &event->u.stats.tmstats.malloc_calls))
|
|
|
|
return 0;
|
|
|
|
if (!get_uint32(fp, &event->u.stats.tmstats.malloc_failures))
|
|
|
|
return 0;
|
|
|
|
if (!get_uint32(fp, &event->u.stats.tmstats.calloc_calls))
|
|
|
|
return 0;
|
|
|
|
if (!get_uint32(fp, &event->u.stats.tmstats.calloc_failures))
|
|
|
|
return 0;
|
|
|
|
if (!get_uint32(fp, &event->u.stats.tmstats.realloc_calls))
|
|
|
|
return 0;
|
|
|
|
if (!get_uint32(fp, &event->u.stats.tmstats.realloc_failures))
|
|
|
|
return 0;
|
|
|
|
if (!get_uint32(fp, &event->u.stats.tmstats.free_calls))
|
|
|
|
return 0;
|
|
|
|
if (!get_uint32(fp, &event->u.stats.tmstats.null_free_calls))
|
|
|
|
return 0;
|
|
|
|
if (!get_uint32(fp, &event->u.stats.calltree_maxkids_parent))
|
|
|
|
return 0;
|
|
|
|
if (!get_uint32(fp, &event->u.stats.calltree_maxstack_top))
|
|
|
|
return 0;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void *generic_alloctable(void *pool, PRSize size)
|
|
|
|
{
|
|
|
|
return malloc(size);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void generic_freetable(void *pool, void *item)
|
|
|
|
{
|
|
|
|
free(item);
|
|
|
|
}
|
|
|
|
|
|
|
|
static PLHashEntry *callsite_allocentry(void *pool, const void *key)
|
|
|
|
{
|
|
|
|
return malloc(sizeof(tmcallsite));
|
|
|
|
}
|
|
|
|
|
|
|
|
static PLHashEntry *graphnode_allocentry(void *pool, const void *key)
|
|
|
|
{
|
|
|
|
tmgraphnode *node = (tmgraphnode*) malloc(sizeof(tmgraphnode));
|
|
|
|
if (!node)
|
|
|
|
return NULL;
|
|
|
|
node->in = node->out = NULL;
|
|
|
|
node->up = node->down = node->next = NULL;
|
|
|
|
node->low = 0;
|
- Remove <unistd.h> include from bloatblame, it's no longer needed and the link(2) prototype it drags in was causing warnings whenever I used a local variable named 'link'.
- Use uint32 for all quantities, and double for all differences, so that we can overflow without breaking deltas such as (busy == allocs - frees). This doesn't help sorting, however: if one of the sort keys has overflowed past 0 just a bit, while the other is a very large unsigned number, beware.
- Separate graph link (half an edge, structurally speaking -- no per-edge stats) from graph edge, so that an edge is two links and some stats. This avoids bloat and copying in connect_nodes (which is soon to become generic and move to tmreader.[ch]).
- Factor data structures better: we now have {allocs,frees} x {bytes,calls} x {direct, total}, declaring the second set as struct tmallcounts and the third as tmcounts. So, for example, total number of calls to allocators would be allocs.calls.total; bytes freed directly by a graphnode (library, component, or method) would be frees.bytes.direct.
- Teach tmreader_eventloop about 'F' (TM_EVENT_FREE) events: it now updates the direct free byte and call counts for a method, its component, and its library when it reads the event. Of course, bloatblame ignores this info, because it is concerned only with bloat (total memory allocated).
- Right-align numbers in the first (trace-malloc stats) table.
2000-07-26 04:24:08 +04:00
|
|
|
node->allocs.bytes.direct = node->allocs.bytes.total = 0;
|
|
|
|
node->allocs.calls.direct = node->allocs.calls.total = 0;
|
|
|
|
node->frees.bytes.direct = node->frees.bytes.total = 0;
|
|
|
|
node->frees.calls.direct = node->frees.calls.total = 0;
|
2000-07-09 12:53:46 +04:00
|
|
|
node->sqsum = 0;
|
|
|
|
node->sort = -1;
|
|
|
|
return &node->entry;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void graphnode_freeentry(void *pool, PLHashEntry *he, PRUintn flag)
|
|
|
|
{
|
|
|
|
/* Always free the value, which points to a strdup'd string. */
|
|
|
|
free(he->value);
|
|
|
|
|
|
|
|
/* Free the whole thing if we're told to. */
|
|
|
|
if (flag == HT_FREE_ENTRY)
|
|
|
|
free((void*) he);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void component_freeentry(void *pool, PLHashEntry *he, PRUintn flag)
|
|
|
|
{
|
|
|
|
if (flag == HT_FREE_ENTRY) {
|
|
|
|
tmgraphnode *comp = (tmgraphnode*) he;
|
|
|
|
|
|
|
|
/* Free the key, which was strdup'd (N.B. value also points to it). */
|
|
|
|
free((void*) tmcomponent_name(comp));
|
|
|
|
free((void*) comp);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static PLHashAllocOps callsite_hashallocops = {
|
|
|
|
generic_alloctable, generic_freetable,
|
|
|
|
callsite_allocentry, graphnode_freeentry
|
|
|
|
};
|
|
|
|
|
|
|
|
static PLHashAllocOps graphnode_hashallocops = {
|
|
|
|
generic_alloctable, generic_freetable,
|
|
|
|
graphnode_allocentry, graphnode_freeentry
|
|
|
|
};
|
|
|
|
|
|
|
|
static PLHashAllocOps component_hashallocops = {
|
|
|
|
generic_alloctable, generic_freetable,
|
|
|
|
graphnode_allocentry, component_freeentry
|
|
|
|
};
|
|
|
|
|
|
|
|
static PLHashNumber hash_serial(const void *key)
|
|
|
|
{
|
|
|
|
return (PLHashNumber) key;
|
|
|
|
}
|
|
|
|
|
|
|
|
tmreader *tmreader_new(const char *program, void *data)
|
|
|
|
{
|
|
|
|
tmreader *tmr;
|
|
|
|
|
|
|
|
tmr = calloc(1, sizeof *tmr);
|
|
|
|
if (!tmr)
|
|
|
|
return NULL;
|
|
|
|
tmr->program = program;
|
|
|
|
tmr->data = data;
|
|
|
|
|
|
|
|
tmr->libraries = PL_NewHashTable(100, hash_serial, PL_CompareValues,
|
|
|
|
PL_CompareStrings, &graphnode_hashallocops,
|
|
|
|
NULL);
|
|
|
|
tmr->components = PL_NewHashTable(10000, PL_HashString, PL_CompareStrings,
|
|
|
|
PL_CompareValues, &component_hashallocops,
|
|
|
|
NULL);
|
|
|
|
tmr->methods = PL_NewHashTable(10000, hash_serial, PL_CompareValues,
|
|
|
|
PL_CompareStrings, &graphnode_hashallocops,
|
|
|
|
NULL);
|
|
|
|
tmr->callsites = PL_NewHashTable(200000, hash_serial, PL_CompareValues,
|
|
|
|
PL_CompareValues, &callsite_hashallocops,
|
|
|
|
NULL);
|
|
|
|
tmr->calltree_root.entry.value = (void*) strdup("root");
|
|
|
|
|
|
|
|
if (!tmr->libraries || !tmr->components || !tmr->methods ||
|
|
|
|
!tmr->callsites || !tmr->calltree_root.entry.value) {
|
|
|
|
tmreader_destroy(tmr);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
return tmr;
|
|
|
|
}
|
|
|
|
|
|
|
|
void tmreader_destroy(tmreader *tmr)
|
|
|
|
{
|
|
|
|
if (tmr->libraries)
|
|
|
|
PL_HashTableDestroy(tmr->libraries);
|
|
|
|
if (tmr->components)
|
|
|
|
PL_HashTableDestroy(tmr->components);
|
|
|
|
if (tmr->methods)
|
|
|
|
PL_HashTableDestroy(tmr->methods);
|
|
|
|
if (tmr->callsites)
|
|
|
|
PL_HashTableDestroy(tmr->callsites);
|
|
|
|
free(tmr);
|
|
|
|
}
|
|
|
|
|
2000-07-22 00:42:15 +04:00
|
|
|
int tmreader_eventloop(tmreader *tmr, const char *filename,
|
|
|
|
tmeventhandler eventhandler)
|
2000-07-09 12:53:46 +04:00
|
|
|
{
|
|
|
|
FILE *fp;
|
|
|
|
char buf[NS_TRACE_MALLOC_MAGIC_SIZE];
|
|
|
|
tmevent event;
|
|
|
|
static const char magic[] = NS_TRACE_MALLOC_MAGIC;
|
|
|
|
|
|
|
|
if (strcmp(filename, "-") == 0) {
|
|
|
|
fp = stdin;
|
|
|
|
} else {
|
|
|
|
fp = fopen(filename, "r");
|
|
|
|
if (!fp) {
|
|
|
|
fprintf(stderr, "%s: can't open %s: %s.\n",
|
|
|
|
tmr->program, filename, strerror(errno));
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (read(fileno(fp), buf, sizeof buf) != sizeof buf ||
|
|
|
|
strncmp(buf, magic, sizeof buf) != 0) {
|
|
|
|
fprintf(stderr, "%s: bad magic string %s at start of %s.\n",
|
|
|
|
tmr->program, buf, filename);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
while (get_tmevent(fp, &event)) {
|
|
|
|
switch (event.type) {
|
2000-07-22 00:42:15 +04:00
|
|
|
case TM_EVENT_LIBRARY: {
|
2000-07-09 12:53:46 +04:00
|
|
|
const void *key;
|
|
|
|
PLHashNumber hash;
|
|
|
|
PLHashEntry **hep, *he;
|
|
|
|
|
|
|
|
key = (const void*) event.serial;
|
|
|
|
hash = hash_serial(key);
|
|
|
|
hep = PL_HashTableRawLookup(tmr->libraries, hash, key);
|
|
|
|
he = *hep;
|
|
|
|
PR_ASSERT(!he);
|
|
|
|
if (he) exit(2);
|
|
|
|
|
|
|
|
he = PL_HashTableRawAdd(tmr->libraries, hep, hash, key,
|
|
|
|
event.u.libname);
|
|
|
|
if (!he) {
|
|
|
|
perror(tmr->program);
|
2000-07-22 00:42:15 +04:00
|
|
|
return -1;
|
2000-07-09 12:53:46 +04:00
|
|
|
}
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2000-07-22 00:42:15 +04:00
|
|
|
case TM_EVENT_METHOD: {
|
2000-07-09 12:53:46 +04:00
|
|
|
const void *key;
|
|
|
|
PLHashNumber hash;
|
|
|
|
PLHashEntry **hep, *he;
|
|
|
|
char *name, *head, *mark, save;
|
|
|
|
tmgraphnode *meth, *comp, *lib;
|
|
|
|
|
|
|
|
key = (const void*) event.serial;
|
|
|
|
hash = hash_serial(key);
|
|
|
|
hep = PL_HashTableRawLookup(tmr->methods, hash, key);
|
|
|
|
he = *hep;
|
|
|
|
PR_ASSERT(!he);
|
|
|
|
if (he) exit(2);
|
|
|
|
|
|
|
|
name = event.u.method.name;
|
|
|
|
he = PL_HashTableRawAdd(tmr->methods, hep, hash, key, name);
|
|
|
|
if (!he) {
|
|
|
|
perror(tmr->program);
|
2000-07-22 00:42:15 +04:00
|
|
|
return -1;
|
2000-07-09 12:53:46 +04:00
|
|
|
}
|
|
|
|
meth = (tmgraphnode*) he;
|
|
|
|
|
|
|
|
head = name;
|
|
|
|
mark = strchr(name, ':');
|
|
|
|
if (!mark) {
|
|
|
|
mark = name;
|
|
|
|
while (*mark != '\0' && *mark == '_')
|
|
|
|
mark++;
|
|
|
|
head = mark;
|
|
|
|
mark = strchr(head, '_');
|
|
|
|
if (!mark) {
|
|
|
|
mark = strchr(head, '+');
|
|
|
|
if (!mark)
|
|
|
|
mark = head + strlen(head);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
save = *mark;
|
|
|
|
*mark = '\0';
|
|
|
|
hash = PL_HashString(head);
|
|
|
|
hep = PL_HashTableRawLookup(tmr->components, hash, head);
|
|
|
|
he = *hep;
|
|
|
|
if (he) {
|
|
|
|
comp = (tmgraphnode*) he;
|
|
|
|
} else {
|
|
|
|
head = strdup(head);
|
|
|
|
if (head) {
|
|
|
|
he = PL_HashTableRawAdd(tmr->components, hep, hash, head,
|
|
|
|
head);
|
|
|
|
}
|
|
|
|
if (!he) {
|
|
|
|
perror(tmr->program);
|
2000-07-22 00:42:15 +04:00
|
|
|
return -1;
|
2000-07-09 12:53:46 +04:00
|
|
|
}
|
|
|
|
comp = (tmgraphnode*) he;
|
|
|
|
|
|
|
|
key = (const void*) event.u.method.library;
|
|
|
|
hash = hash_serial(key);
|
|
|
|
lib = (tmgraphnode*)
|
|
|
|
*PL_HashTableRawLookup(tmr->libraries, hash, key);
|
|
|
|
if (lib) {
|
|
|
|
comp->up = lib;
|
|
|
|
comp->next = lib->down;
|
|
|
|
lib->down = comp;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
*mark = save;
|
|
|
|
|
|
|
|
meth->up = comp;
|
|
|
|
meth->next = comp->down;
|
|
|
|
comp->down = meth;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2000-07-22 00:42:15 +04:00
|
|
|
case TM_EVENT_CALLSITE: {
|
2000-07-09 12:53:46 +04:00
|
|
|
const void *key, *mkey;
|
|
|
|
PLHashNumber hash, mhash;
|
|
|
|
PLHashEntry **hep, *he;
|
|
|
|
tmcallsite *site, *parent;
|
|
|
|
tmgraphnode *meth;
|
|
|
|
|
|
|
|
key = (const void*) event.serial;
|
|
|
|
hash = hash_serial(key);
|
|
|
|
hep = PL_HashTableRawLookup(tmr->callsites, hash, key);
|
|
|
|
he = *hep;
|
|
|
|
PR_ASSERT(!he);
|
|
|
|
if (he) exit(2);
|
|
|
|
|
|
|
|
if (event.u.site.parent == 0) {
|
|
|
|
parent = &tmr->calltree_root;
|
|
|
|
} else {
|
2000-07-22 00:42:15 +04:00
|
|
|
parent = tmreader_callsite(tmr, event.u.site.parent);
|
2000-07-09 12:53:46 +04:00
|
|
|
if (!parent) {
|
|
|
|
fprintf(stderr, "%s: no parent for %lu (%lu)!\n",
|
|
|
|
tmr->program, (unsigned long) event.serial,
|
|
|
|
(unsigned long) event.u.site.parent);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
he = PL_HashTableRawAdd(tmr->callsites, hep, hash, key, NULL);
|
|
|
|
if (!he) {
|
|
|
|
perror(tmr->program);
|
2000-07-22 00:42:15 +04:00
|
|
|
return -1;
|
2000-07-09 12:53:46 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
site = (tmcallsite*) he;
|
|
|
|
site->parent = parent;
|
|
|
|
site->siblings = parent->kids;
|
|
|
|
parent->kids = site;
|
|
|
|
site->kids = NULL;
|
|
|
|
|
|
|
|
mkey = (const void*) event.u.site.method;
|
|
|
|
mhash = hash_serial(mkey);
|
|
|
|
meth = (tmgraphnode*)
|
|
|
|
*PL_HashTableRawLookup(tmr->methods, mhash, mkey);
|
|
|
|
site->method = meth;
|
|
|
|
site->offset = event.u.site.offset;
|
- Remove <unistd.h> include from bloatblame, it's no longer needed and the link(2) prototype it drags in was causing warnings whenever I used a local variable named 'link'.
- Use uint32 for all quantities, and double for all differences, so that we can overflow without breaking deltas such as (busy == allocs - frees). This doesn't help sorting, however: if one of the sort keys has overflowed past 0 just a bit, while the other is a very large unsigned number, beware.
- Separate graph link (half an edge, structurally speaking -- no per-edge stats) from graph edge, so that an edge is two links and some stats. This avoids bloat and copying in connect_nodes (which is soon to become generic and move to tmreader.[ch]).
- Factor data structures better: we now have {allocs,frees} x {bytes,calls} x {direct, total}, declaring the second set as struct tmallcounts and the third as tmcounts. So, for example, total number of calls to allocators would be allocs.calls.total; bytes freed directly by a graphnode (library, component, or method) would be frees.bytes.direct.
- Teach tmreader_eventloop about 'F' (TM_EVENT_FREE) events: it now updates the direct free byte and call counts for a method, its component, and its library when it reads the event. Of course, bloatblame ignores this info, because it is concerned only with bloat (total memory allocated).
- Right-align numbers in the first (trace-malloc stats) table.
2000-07-26 04:24:08 +04:00
|
|
|
site->allocs.bytes.direct = site->allocs.bytes.total = 0;
|
|
|
|
site->allocs.calls.direct = site->allocs.calls.total = 0;
|
|
|
|
site->frees.bytes.direct = site->frees.bytes.total = 0;
|
|
|
|
site->frees.calls.direct = site->frees.calls.total = 0;
|
2000-07-09 12:53:46 +04:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2000-07-22 00:42:15 +04:00
|
|
|
case TM_EVENT_MALLOC:
|
|
|
|
case TM_EVENT_CALLOC:
|
|
|
|
case TM_EVENT_REALLOC: {
|
2000-07-09 12:53:46 +04:00
|
|
|
tmcallsite *site;
|
- Remove <unistd.h> include from bloatblame, it's no longer needed and the link(2) prototype it drags in was causing warnings whenever I used a local variable named 'link'.
- Use uint32 for all quantities, and double for all differences, so that we can overflow without breaking deltas such as (busy == allocs - frees). This doesn't help sorting, however: if one of the sort keys has overflowed past 0 just a bit, while the other is a very large unsigned number, beware.
- Separate graph link (half an edge, structurally speaking -- no per-edge stats) from graph edge, so that an edge is two links and some stats. This avoids bloat and copying in connect_nodes (which is soon to become generic and move to tmreader.[ch]).
- Factor data structures better: we now have {allocs,frees} x {bytes,calls} x {direct, total}, declaring the second set as struct tmallcounts and the third as tmcounts. So, for example, total number of calls to allocators would be allocs.calls.total; bytes freed directly by a graphnode (library, component, or method) would be frees.bytes.direct.
- Teach tmreader_eventloop about 'F' (TM_EVENT_FREE) events: it now updates the direct free byte and call counts for a method, its component, and its library when it reads the event. Of course, bloatblame ignores this info, because it is concerned only with bloat (total memory allocated).
- Right-align numbers in the first (trace-malloc stats) table.
2000-07-26 04:24:08 +04:00
|
|
|
uint32 size, oldsize;
|
|
|
|
double delta, sqdelta, sqszdelta;
|
2000-07-09 12:53:46 +04:00
|
|
|
tmgraphnode *meth, *comp, *lib;
|
|
|
|
|
2000-07-22 00:42:15 +04:00
|
|
|
site = tmreader_callsite(tmr, event.serial);
|
2000-07-09 12:53:46 +04:00
|
|
|
if (!site) {
|
|
|
|
fprintf(stderr, "%s: no callsite for '%c' (%lu)!\n",
|
|
|
|
tmr->program, event.type, (unsigned long) event.serial);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
- Remove <unistd.h> include from bloatblame, it's no longer needed and the link(2) prototype it drags in was causing warnings whenever I used a local variable named 'link'.
- Use uint32 for all quantities, and double for all differences, so that we can overflow without breaking deltas such as (busy == allocs - frees). This doesn't help sorting, however: if one of the sort keys has overflowed past 0 just a bit, while the other is a very large unsigned number, beware.
- Separate graph link (half an edge, structurally speaking -- no per-edge stats) from graph edge, so that an edge is two links and some stats. This avoids bloat and copying in connect_nodes (which is soon to become generic and move to tmreader.[ch]).
- Factor data structures better: we now have {allocs,frees} x {bytes,calls} x {direct, total}, declaring the second set as struct tmallcounts and the third as tmcounts. So, for example, total number of calls to allocators would be allocs.calls.total; bytes freed directly by a graphnode (library, component, or method) would be frees.bytes.direct.
- Teach tmreader_eventloop about 'F' (TM_EVENT_FREE) events: it now updates the direct free byte and call counts for a method, its component, and its library when it reads the event. Of course, bloatblame ignores this info, because it is concerned only with bloat (total memory allocated).
- Right-align numbers in the first (trace-malloc stats) table.
2000-07-26 04:24:08 +04:00
|
|
|
size = event.u.alloc.size;
|
|
|
|
oldsize = event.u.alloc.oldsize;
|
|
|
|
delta = (double)size - (double)oldsize;
|
|
|
|
site->allocs.bytes.direct += delta;
|
2000-07-22 00:42:15 +04:00
|
|
|
if (event.type != TM_EVENT_REALLOC)
|
- Remove <unistd.h> include from bloatblame, it's no longer needed and the link(2) prototype it drags in was causing warnings whenever I used a local variable named 'link'.
- Use uint32 for all quantities, and double for all differences, so that we can overflow without breaking deltas such as (busy == allocs - frees). This doesn't help sorting, however: if one of the sort keys has overflowed past 0 just a bit, while the other is a very large unsigned number, beware.
- Separate graph link (half an edge, structurally speaking -- no per-edge stats) from graph edge, so that an edge is two links and some stats. This avoids bloat and copying in connect_nodes (which is soon to become generic and move to tmreader.[ch]).
- Factor data structures better: we now have {allocs,frees} x {bytes,calls} x {direct, total}, declaring the second set as struct tmallcounts and the third as tmcounts. So, for example, total number of calls to allocators would be allocs.calls.total; bytes freed directly by a graphnode (library, component, or method) would be frees.bytes.direct.
- Teach tmreader_eventloop about 'F' (TM_EVENT_FREE) events: it now updates the direct free byte and call counts for a method, its component, and its library when it reads the event. Of course, bloatblame ignores this info, because it is concerned only with bloat (total memory allocated).
- Right-align numbers in the first (trace-malloc stats) table.
2000-07-26 04:24:08 +04:00
|
|
|
site->allocs.calls.direct++;
|
2000-07-09 12:53:46 +04:00
|
|
|
meth = site->method;
|
|
|
|
if (meth) {
|
- Remove <unistd.h> include from bloatblame, it's no longer needed and the link(2) prototype it drags in was causing warnings whenever I used a local variable named 'link'.
- Use uint32 for all quantities, and double for all differences, so that we can overflow without breaking deltas such as (busy == allocs - frees). This doesn't help sorting, however: if one of the sort keys has overflowed past 0 just a bit, while the other is a very large unsigned number, beware.
- Separate graph link (half an edge, structurally speaking -- no per-edge stats) from graph edge, so that an edge is two links and some stats. This avoids bloat and copying in connect_nodes (which is soon to become generic and move to tmreader.[ch]).
- Factor data structures better: we now have {allocs,frees} x {bytes,calls} x {direct, total}, declaring the second set as struct tmallcounts and the third as tmcounts. So, for example, total number of calls to allocators would be allocs.calls.total; bytes freed directly by a graphnode (library, component, or method) would be frees.bytes.direct.
- Teach tmreader_eventloop about 'F' (TM_EVENT_FREE) events: it now updates the direct free byte and call counts for a method, its component, and its library when it reads the event. Of course, bloatblame ignores this info, because it is concerned only with bloat (total memory allocated).
- Right-align numbers in the first (trace-malloc stats) table.
2000-07-26 04:24:08 +04:00
|
|
|
meth->allocs.bytes.direct += delta;
|
2000-07-09 12:53:46 +04:00
|
|
|
sqdelta = delta * delta;
|
2000-07-22 00:42:15 +04:00
|
|
|
if (event.type == TM_EVENT_REALLOC) {
|
2000-07-09 12:53:46 +04:00
|
|
|
sqszdelta = ((double)size * size)
|
|
|
|
- ((double)oldsize * oldsize);
|
|
|
|
meth->sqsum += sqszdelta;
|
|
|
|
} else {
|
|
|
|
meth->sqsum += sqdelta;
|
- Remove <unistd.h> include from bloatblame, it's no longer needed and the link(2) prototype it drags in was causing warnings whenever I used a local variable named 'link'.
- Use uint32 for all quantities, and double for all differences, so that we can overflow without breaking deltas such as (busy == allocs - frees). This doesn't help sorting, however: if one of the sort keys has overflowed past 0 just a bit, while the other is a very large unsigned number, beware.
- Separate graph link (half an edge, structurally speaking -- no per-edge stats) from graph edge, so that an edge is two links and some stats. This avoids bloat and copying in connect_nodes (which is soon to become generic and move to tmreader.[ch]).
- Factor data structures better: we now have {allocs,frees} x {bytes,calls} x {direct, total}, declaring the second set as struct tmallcounts and the third as tmcounts. So, for example, total number of calls to allocators would be allocs.calls.total; bytes freed directly by a graphnode (library, component, or method) would be frees.bytes.direct.
- Teach tmreader_eventloop about 'F' (TM_EVENT_FREE) events: it now updates the direct free byte and call counts for a method, its component, and its library when it reads the event. Of course, bloatblame ignores this info, because it is concerned only with bloat (total memory allocated).
- Right-align numbers in the first (trace-malloc stats) table.
2000-07-26 04:24:08 +04:00
|
|
|
meth->allocs.calls.direct++;
|
2000-07-09 12:53:46 +04:00
|
|
|
}
|
|
|
|
comp = meth->up;
|
|
|
|
if (comp) {
|
- Remove <unistd.h> include from bloatblame, it's no longer needed and the link(2) prototype it drags in was causing warnings whenever I used a local variable named 'link'.
- Use uint32 for all quantities, and double for all differences, so that we can overflow without breaking deltas such as (busy == allocs - frees). This doesn't help sorting, however: if one of the sort keys has overflowed past 0 just a bit, while the other is a very large unsigned number, beware.
- Separate graph link (half an edge, structurally speaking -- no per-edge stats) from graph edge, so that an edge is two links and some stats. This avoids bloat and copying in connect_nodes (which is soon to become generic and move to tmreader.[ch]).
- Factor data structures better: we now have {allocs,frees} x {bytes,calls} x {direct, total}, declaring the second set as struct tmallcounts and the third as tmcounts. So, for example, total number of calls to allocators would be allocs.calls.total; bytes freed directly by a graphnode (library, component, or method) would be frees.bytes.direct.
- Teach tmreader_eventloop about 'F' (TM_EVENT_FREE) events: it now updates the direct free byte and call counts for a method, its component, and its library when it reads the event. Of course, bloatblame ignores this info, because it is concerned only with bloat (total memory allocated).
- Right-align numbers in the first (trace-malloc stats) table.
2000-07-26 04:24:08 +04:00
|
|
|
comp->allocs.bytes.direct += delta;
|
2000-07-22 00:42:15 +04:00
|
|
|
if (event.type == TM_EVENT_REALLOC) {
|
2000-07-09 12:53:46 +04:00
|
|
|
comp->sqsum += sqszdelta;
|
|
|
|
} else {
|
|
|
|
comp->sqsum += sqdelta;
|
- Remove <unistd.h> include from bloatblame, it's no longer needed and the link(2) prototype it drags in was causing warnings whenever I used a local variable named 'link'.
- Use uint32 for all quantities, and double for all differences, so that we can overflow without breaking deltas such as (busy == allocs - frees). This doesn't help sorting, however: if one of the sort keys has overflowed past 0 just a bit, while the other is a very large unsigned number, beware.
- Separate graph link (half an edge, structurally speaking -- no per-edge stats) from graph edge, so that an edge is two links and some stats. This avoids bloat and copying in connect_nodes (which is soon to become generic and move to tmreader.[ch]).
- Factor data structures better: we now have {allocs,frees} x {bytes,calls} x {direct, total}, declaring the second set as struct tmallcounts and the third as tmcounts. So, for example, total number of calls to allocators would be allocs.calls.total; bytes freed directly by a graphnode (library, component, or method) would be frees.bytes.direct.
- Teach tmreader_eventloop about 'F' (TM_EVENT_FREE) events: it now updates the direct free byte and call counts for a method, its component, and its library when it reads the event. Of course, bloatblame ignores this info, because it is concerned only with bloat (total memory allocated).
- Right-align numbers in the first (trace-malloc stats) table.
2000-07-26 04:24:08 +04:00
|
|
|
comp->allocs.calls.direct++;
|
2000-07-09 12:53:46 +04:00
|
|
|
}
|
|
|
|
lib = comp->up;
|
|
|
|
if (lib) {
|
- Remove <unistd.h> include from bloatblame, it's no longer needed and the link(2) prototype it drags in was causing warnings whenever I used a local variable named 'link'.
- Use uint32 for all quantities, and double for all differences, so that we can overflow without breaking deltas such as (busy == allocs - frees). This doesn't help sorting, however: if one of the sort keys has overflowed past 0 just a bit, while the other is a very large unsigned number, beware.
- Separate graph link (half an edge, structurally speaking -- no per-edge stats) from graph edge, so that an edge is two links and some stats. This avoids bloat and copying in connect_nodes (which is soon to become generic and move to tmreader.[ch]).
- Factor data structures better: we now have {allocs,frees} x {bytes,calls} x {direct, total}, declaring the second set as struct tmallcounts and the third as tmcounts. So, for example, total number of calls to allocators would be allocs.calls.total; bytes freed directly by a graphnode (library, component, or method) would be frees.bytes.direct.
- Teach tmreader_eventloop about 'F' (TM_EVENT_FREE) events: it now updates the direct free byte and call counts for a method, its component, and its library when it reads the event. Of course, bloatblame ignores this info, because it is concerned only with bloat (total memory allocated).
- Right-align numbers in the first (trace-malloc stats) table.
2000-07-26 04:24:08 +04:00
|
|
|
lib->allocs.bytes.direct += delta;
|
2000-07-22 00:42:15 +04:00
|
|
|
if (event.type == TM_EVENT_REALLOC) {
|
2000-07-09 12:53:46 +04:00
|
|
|
lib->sqsum += sqszdelta;
|
|
|
|
} else {
|
|
|
|
lib->sqsum += sqdelta;
|
- Remove <unistd.h> include from bloatblame, it's no longer needed and the link(2) prototype it drags in was causing warnings whenever I used a local variable named 'link'.
- Use uint32 for all quantities, and double for all differences, so that we can overflow without breaking deltas such as (busy == allocs - frees). This doesn't help sorting, however: if one of the sort keys has overflowed past 0 just a bit, while the other is a very large unsigned number, beware.
- Separate graph link (half an edge, structurally speaking -- no per-edge stats) from graph edge, so that an edge is two links and some stats. This avoids bloat and copying in connect_nodes (which is soon to become generic and move to tmreader.[ch]).
- Factor data structures better: we now have {allocs,frees} x {bytes,calls} x {direct, total}, declaring the second set as struct tmallcounts and the third as tmcounts. So, for example, total number of calls to allocators would be allocs.calls.total; bytes freed directly by a graphnode (library, component, or method) would be frees.bytes.direct.
- Teach tmreader_eventloop about 'F' (TM_EVENT_FREE) events: it now updates the direct free byte and call counts for a method, its component, and its library when it reads the event. Of course, bloatblame ignores this info, because it is concerned only with bloat (total memory allocated).
- Right-align numbers in the first (trace-malloc stats) table.
2000-07-26 04:24:08 +04:00
|
|
|
lib->allocs.calls.direct++;
|
2000-07-09 12:53:46 +04:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
- Remove <unistd.h> include from bloatblame, it's no longer needed and the link(2) prototype it drags in was causing warnings whenever I used a local variable named 'link'.
- Use uint32 for all quantities, and double for all differences, so that we can overflow without breaking deltas such as (busy == allocs - frees). This doesn't help sorting, however: if one of the sort keys has overflowed past 0 just a bit, while the other is a very large unsigned number, beware.
- Separate graph link (half an edge, structurally speaking -- no per-edge stats) from graph edge, so that an edge is two links and some stats. This avoids bloat and copying in connect_nodes (which is soon to become generic and move to tmreader.[ch]).
- Factor data structures better: we now have {allocs,frees} x {bytes,calls} x {direct, total}, declaring the second set as struct tmallcounts and the third as tmcounts. So, for example, total number of calls to allocators would be allocs.calls.total; bytes freed directly by a graphnode (library, component, or method) would be frees.bytes.direct.
- Teach tmreader_eventloop about 'F' (TM_EVENT_FREE) events: it now updates the direct free byte and call counts for a method, its component, and its library when it reads the event. Of course, bloatblame ignores this info, because it is concerned only with bloat (total memory allocated).
- Right-align numbers in the first (trace-malloc stats) table.
2000-07-26 04:24:08 +04:00
|
|
|
case TM_EVENT_FREE: {
|
|
|
|
tmcallsite *site;
|
|
|
|
uint32 size;
|
|
|
|
tmgraphnode *meth, *comp, *lib;
|
|
|
|
|
|
|
|
site = tmreader_callsite(tmr, event.serial);
|
|
|
|
if (!site) {
|
|
|
|
fprintf(stderr, "%s: no callsite for '%c' (%lu)!\n",
|
|
|
|
tmr->program, event.type, (unsigned long) event.serial);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
size = event.u.alloc.size;
|
|
|
|
site->frees.bytes.direct += size;
|
|
|
|
site->frees.calls.direct++;
|
|
|
|
meth = site->method;
|
|
|
|
if (meth) {
|
|
|
|
meth->frees.bytes.direct += size;
|
|
|
|
meth->frees.calls.direct++;
|
|
|
|
comp = meth->up;
|
|
|
|
if (comp) {
|
|
|
|
comp->frees.bytes.direct += size;
|
|
|
|
comp->frees.calls.direct++;
|
|
|
|
lib = comp->up;
|
|
|
|
if (lib) {
|
|
|
|
lib->frees.bytes.direct += size;
|
|
|
|
lib->frees.calls.direct++;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2000-07-09 12:53:46 +04:00
|
|
|
break;
|
- Remove <unistd.h> include from bloatblame, it's no longer needed and the link(2) prototype it drags in was causing warnings whenever I used a local variable named 'link'.
- Use uint32 for all quantities, and double for all differences, so that we can overflow without breaking deltas such as (busy == allocs - frees). This doesn't help sorting, however: if one of the sort keys has overflowed past 0 just a bit, while the other is a very large unsigned number, beware.
- Separate graph link (half an edge, structurally speaking -- no per-edge stats) from graph edge, so that an edge is two links and some stats. This avoids bloat and copying in connect_nodes (which is soon to become generic and move to tmreader.[ch]).
- Factor data structures better: we now have {allocs,frees} x {bytes,calls} x {direct, total}, declaring the second set as struct tmallcounts and the third as tmcounts. So, for example, total number of calls to allocators would be allocs.calls.total; bytes freed directly by a graphnode (library, component, or method) would be frees.bytes.direct.
- Teach tmreader_eventloop about 'F' (TM_EVENT_FREE) events: it now updates the direct free byte and call counts for a method, its component, and its library when it reads the event. Of course, bloatblame ignores this info, because it is concerned only with bloat (total memory allocated).
- Right-align numbers in the first (trace-malloc stats) table.
2000-07-26 04:24:08 +04:00
|
|
|
}
|
2000-07-09 12:53:46 +04:00
|
|
|
|
2000-07-22 00:42:15 +04:00
|
|
|
case TM_EVENT_STATS:
|
2000-07-09 12:53:46 +04:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
eventhandler(tmr, &event);
|
|
|
|
}
|
|
|
|
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2000-07-22 00:42:15 +04:00
|
|
|
tmgraphnode *tmreader_library(tmreader *tmr, uint32 serial)
|
|
|
|
{
|
|
|
|
const void *key;
|
|
|
|
PLHashNumber hash;
|
|
|
|
|
|
|
|
key = (const void*) serial;
|
|
|
|
hash = hash_serial(key);
|
|
|
|
return (tmgraphnode*) *PL_HashTableRawLookup(tmr->libraries, hash, key);
|
|
|
|
}
|
|
|
|
|
|
|
|
tmgraphnode *tmreader_component(tmreader *tmr, const char *name)
|
|
|
|
{
|
|
|
|
PLHashNumber hash;
|
|
|
|
|
|
|
|
hash = PL_HashString(name);
|
|
|
|
return (tmgraphnode*) *PL_HashTableRawLookup(tmr->components, hash, name);
|
|
|
|
}
|
|
|
|
|
|
|
|
tmgraphnode *tmreader_method(tmreader *tmr, uint32 serial)
|
|
|
|
{
|
|
|
|
const void *key;
|
|
|
|
PLHashNumber hash;
|
|
|
|
|
|
|
|
key = (const void*) serial;
|
|
|
|
hash = hash_serial(key);
|
|
|
|
return (tmgraphnode*) *PL_HashTableRawLookup(tmr->methods, hash, key);
|
|
|
|
}
|
|
|
|
|
|
|
|
tmcallsite *tmreader_callsite(tmreader *tmr, uint32 serial)
|
2000-07-09 12:53:46 +04:00
|
|
|
{
|
|
|
|
const void *key;
|
|
|
|
PLHashNumber hash;
|
|
|
|
|
|
|
|
key = (const void*) serial;
|
|
|
|
hash = hash_serial(key);
|
|
|
|
return (tmcallsite*) *PL_HashTableRawLookup(tmr->callsites, hash, key);
|
|
|
|
}
|
2000-07-26 06:19:27 +04:00
|
|
|
|
|
|
|
int tmgraphnode_connect(tmgraphnode *from, tmgraphnode *to, tmcallsite *site)
|
|
|
|
{
|
|
|
|
tmgraphlink *outlink;
|
|
|
|
tmgraphedge *edge;
|
|
|
|
|
|
|
|
for (outlink = from->out; outlink; outlink = outlink->next) {
|
|
|
|
if (outlink->node == to) {
|
|
|
|
/*
|
|
|
|
* Say the stack looks like this: ... => JS => js => JS => js.
|
|
|
|
* We must avoid overcounting JS=>js because the first edge total
|
|
|
|
* includes the second JS=>js edge's total (which is because the
|
|
|
|
* lower site's total includes all its kids' totals).
|
|
|
|
*/
|
|
|
|
edge = TM_LINK_TO_EDGE(outlink, TM_EDGE_OUT_LINK);
|
|
|
|
if (!to->low || to->low < from->low) {
|
|
|
|
/* Add the direct and total counts to edge->allocs. */
|
|
|
|
edge->allocs.bytes.direct += site->allocs.bytes.direct;
|
|
|
|
edge->allocs.bytes.total += site->allocs.bytes.total;
|
|
|
|
edge->allocs.calls.direct += site->allocs.calls.direct;
|
|
|
|
edge->allocs.calls.total += site->allocs.calls.total;
|
|
|
|
|
|
|
|
/* Now update the free counts. */
|
|
|
|
edge->frees.bytes.direct += site->frees.bytes.direct;
|
|
|
|
edge->frees.bytes.total += site->frees.bytes.total;
|
|
|
|
edge->frees.calls.direct += site->frees.calls.direct;
|
|
|
|
edge->frees.calls.total += site->frees.calls.total;
|
|
|
|
}
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
edge = (tmgraphedge*) malloc(sizeof(tmgraphedge));
|
|
|
|
if (!edge)
|
|
|
|
return 0;
|
|
|
|
edge->links[TM_EDGE_OUT_LINK].node = to;
|
|
|
|
edge->links[TM_EDGE_OUT_LINK].next = from->out;
|
|
|
|
from->out = &edge->links[TM_EDGE_OUT_LINK];
|
|
|
|
edge->links[TM_EDGE_IN_LINK].node = from;
|
|
|
|
edge->links[TM_EDGE_IN_LINK].next = to->in;
|
|
|
|
to->in = &edge->links[TM_EDGE_IN_LINK];
|
|
|
|
edge->allocs = site->allocs;
|
|
|
|
edge->frees = site->frees;
|
|
|
|
return 1;
|
|
|
|
}
|