[torquedev] saving inodes with lots of TM tasks

David Singleton David.Singleton at anu.edu.au
Wed Feb 21 16:44:56 MST 2007

Garrick Staples wrote:
> On Sat, Feb 17, 2007 at 09:52:37AM +0100, ?ke Sandgren alleged:
>> On Fri, 2007-02-16 at 20:33 -0700, Garrick Staples wrote:
>>> I have an immediate requirement to support hundreds of thousands of TM
>>> tasks within a single job.  The limiter at this point is that pbs_mom
>>> saves a TM struct in a file per task.  After a few hundred thousand
>>> tasks, the filesystem runs out of inodes!
>>> The purpose of the disk files are only to read in on a pbs_mom restart.
>>> In normal use, they are never read.
>>> So the solution seems simple: use 1 file.  I've attached a patch that
>>> implements this idea by opening 1 file and seeking (sizeof(struct) *
>>> task_id) bytes into the file before writing.  I didn't bother trying to
>>> be backwards compatible at this point because I think this will only be
>>> for 2.2.x.
>> Don't you have to make a copy of task_save in the read loop?
>> while (read(fds,(char *)&task_save,sizeof(task_save)) ==
>> sizeof(task_save)) {
>> ...
>> pt->ti_qs = task_save;
>> }
>> You overwrite task_save each round...
> Hrm, certainly looks like it, doesn't it.  But I didn't actually change
> that.  That's the way it has always been.

I think it's OK - it's a struct copy, not a pointer assignment.


More information about the torquedev mailing list