"dbz.c: Can't malloc" during expire

Elizabeth Zaenger liz at eecs.umich.edu
Tue Apr 16 13:25:05 UTC 2002

> From: davidsen at tmr.com (bill davidsen)
> Newsgroups: mail.inn-workers
> Date: 16 Apr 2002 13:11:42 GMT
> In article <72bsctguf4.fsf at nd1.eng.demon.net>,
> Alex Kiernan  <alexk at demon.net> wrote:
> | 
> | Elizabeth Zaenger <liz at eecs.umich.edu> writes:
> | 
> | > Hi Alex,
> | > 
> | > Thanks, but what does this mean?  Or rather, what's the fix?
> | > 
> | 
> | Larger ulimit, more swap, more physical RAM, move to 64 bit
> | architecure (probably in that order).
> Actually, after setting ulimit higher, if you can't add more physical
> memory I would consider rebuilding with tagged-hash on. Having run a few
> too-small machines I believe that the performance hit will be less than
> using a lot of swap and paging heavily.

Except that I'm running --with-largefiles, and I don't think the two
are compatible.  My history file is currently 2.4G.

Btw, I sent a post that the problem is fixed.  I tweaked a variable in 
the kernel, and upped the datasize limit, and expire started running again.

But while I'm on the subject, just how many people out there are running
--with-largefiles?  I often feel like there's just me and Joe with a 
history file bigger than 2G....(Er, Joe, your history file *is* bigger
than 2G, isn't it?)


Elizabeth Zaenger, Unix Support
Departmental Computing Organization
University of Michigan, Electrical Engineering and Computer Science Dept.

More information about the inn-workers mailing list