[Mauiusers] Problem with MAUI 3.2.6p16 and GRES

alexandre babin alexandre.babin at gmail.com
Fri Aug 11 03:36:28 MDT 2006


Hello,
I use MAUI 3.2.6p16 and torque 2.1.2.

I want to declare GRES on node in maui.cfg:

SERVERHOST            linux-ci

# primary admin must be first in list
ADMIN1                maui

# Resource Manager Definition

RMCFG[LINUX-CI] TYPE=PBS HOST=linux-ci
SUBMITCMD=/usr/local/torque/1.2.0p6/bin/qsub

RMPOLLINTERVAL        00:00:30

# Allocation Manager Definition

AMCFG[bank]  TYPE=NONE

# full parameter docs at
http://clusterresources.com/mauidocs/a.fparameters.html
# use the 'schedctl -l' command to display current configuration

SERVERPORT            42559
SERVERMODE            NORMAL

# Admin: http://clusterresources.com/mauidocs/a.esecurity.html

LOGFILE               maui.log
LOGFILEMAXSIZE        10000000
LOGLEVEL              3
LOGFILEROLLDEPTH      5

# Job Priority:
http://clusterresources.com/mauidocs/5.1jobprioritization.html

QUEUETIMEWEIGHT 250
XFACTORWEIGHT   100
RESWEIGHT       100
NODEWEIGHT      100
PROCWEIGHT      200
PEWEIGHT        50
FSWEIGHT        1000# FairShare:
http://clusterresources.com/mauidocs/6.3fairshare.html

FSPOLICY              DEDICATEDPS
FSDEPTH               7
FSINTERVAL            86400
FSDECAY               0.60

# Throttling Policies:
http://clusterresources.com/mauidocs/6.2throttlingpolicies.html

# NONE SPECIFIED

# Backfill: http://clusterresources.com/mauidocs/8.2backfill.html

BACKFILLPOLICY        FIRSTFIT
BACKFILLMETRIC        PROCSECONDS

RESERVATIONPOLICY     CURRENTHIGHEST
RESERVATIONDEPTH      1

# Node Allocation:
http://clusterresources.com/mauidocs/5.2nodeallocation.html

NODEALLOCATIONPOLICY    CPULOAD
NODEAVAILABILITYPOLICY  COMBINED
NODEPOLLFREQUENCY       0
NODEACCESSPOLICY        SHARED# Standing Reservations:
http://clusterresources.com/mauidocs/7.1.3standingreservations.html

SRCFG[0] NAME=inter_ope HOSTLIST=l6-ci PERIOD=INFINITY CLASSLIST=q_ope_shell
SRCFG[1] NAME=inter_gen HOSTLIST=l1-ci,l2-ci,l3-ci,l4-ci,l5-ci
PERIOD=INFINITY CLASSLIST=q_gen_shell
#SRCFG[2] NAME=batch_ope HOSTLIST=l16-ci PERIOD=INFINITY
CLASSLIST=q_ope_15m,q_ope_4h,q_ope_max,q_gen_15m-
#SRCFG[3] NAME=batch_gen
HOSTLIST=l7-ci,l8-ci,l9-ci,l10-ci,l11-ci,l12-ci,l13-ci,l14-ci,l15-ci
PERIOD=INFINITY CLASSLIST=q_gen_15m,q_gen_4h,q_gen_max

# SRSTARTTIME[test] 8:00:00
# SRENDTIME[test]   17:00:00
# SRDAYS[test]      MON TUE WED THU FRI
# SRTASKCOUNT[test] 20
# SRMAXTIME[test]   0:30:00

# Creds: http://clusterresources.com/mauidocs/6.1fairnessoverview.html

USERCFG[DEFAULT]        FSTARGET=10.0-
GROUPCFG[DEFAULT]       FSTARGET=20.0-
# USERCFG[john]         PRIORITY=100  FSTARGET=10.0-
# GROUPCFG[staff]       PRIORITY=1000 QLIST=hi:low QDEF=hi
# CLASSCFG[batch]       FLAGS=PREEMPTEE
# CLASSCFG[interactive] FLAGS=PREEMPTOR

# Temps d'attente avant remise en queue
DEFERTIME       00:15:00
#DEFERTIME 0

# Autoriser les requetes a ressources multiples
ENABLEMULTINODEJOBS     TRUE
ENABLEMULTIREQJOBS      FALSE
#ENABLEMULTIREQJOBS     TRUE

# Annulation des jobs depassant le temps ou les ressources imparties
JOBMAXOVERRUN           00:00:00
#JOBMAXSTARTTIME                01:00:00:00
RESOURCELIMITPOLICY     PROC:ALWAYS:CANCEL# Nodes Definition

NODECFG[DEFAULT] SPEED=2.2
#NODECFG[l9-ci]
#NODECFG[l13-ci]
#NODECFG[l14-ci]
#NODECFG[l15-ci]
#NODECFG[l16-ci]
NODECFG[l8-ci]  GRES=matlab:2,tape:2
NODECFG[l9-ci]  GRES=matlab:2
#NODECFG[l10-ci]  GRES=tape:5
#NODECFG[l7-ci] FEATURES=matlab:2




RESOURCELIMITPOLICY     SWAP:ALWAYS:CANCEL
RESOURCELIMITPOLICY     DISK:ALWAYS:CANCEL
RESOURCELIMITPOLICY     MEM:ALWAYS:CANCEL

#JOBNODEMATCHPOLICY      EXACTNODE

# QOS: http://clusterresources.com/mauidocs/7.3qos.html

# QOSCFG[hi]  PRIORITY=100 XFTARGET=100 FLAGS=PREEMPTOR:IGNMAXJOB
# QOSCFG[low] PRIORITY=-1000 FLAGS=PREEMPTEE


FSUSERWEIGHT    10
FSGROUPWEIGHT   1

But with this config, I can't run job with gres=matalab, because the
reschecking node l9-ci

State:      Idle  (in current state for 00:15:14)
Configured Resources: PROCS: 2  MEM: 277M  SWAP: 738M  DISK: 1011M  matlab:
2
Utilized   Resources: DISK: 83M  matlab: 2
<<<<<<<<<<!!!!!!!!!!!!!!!!!!!!!!!!
Dedicated  Resources: [NONE]
Opsys:         linux  Arch:      [NONE]
Speed:      1.00  Load:       0.000
Network:    [DEFAULT]
Features:   [genbatch][short]
Attributes: [Batch]
Classes:    [q_gen_15m 2:2][q_gen_shell 2:2][q_ope_max 2:2][q_ope_4h
2:2][q_ope_batch 2:2][q_gen_max 2:2][q_gen_batch 2:2][q_gen_4h
2:2][q_ope_15m 2:2][q_ope_shell 2:2]

Total Time: 5:46:02  Up: 5:44:05 (99.44%)  Active: 00:10:56 (3.16%)

Reservations:
NOTE:  no reservations on node

[maui at linux-ci ~]$ checknode l8-ci


checking node l8-ci

State:      Idle  (in current state for 00:15:14)
Configured Resources: PROCS: 2  MEM: 277M  SWAP: 734M  DISK: 1011M  matlab:
2
Utilized   Resources: DISK: 115M  matlab: 2
<<<<<<<<<<<<<<<<<<<!!!!!!!!!!!!!!!!!!!!!!!!!!
Dedicated  Resources: [NONE]
Opsys:         linux  Arch:      [NONE]
Speed:      1.00  Load:       0.000
Network:    [DEFAULT]
Features:   [genbatch][short]
Attributes: [Batch]
Classes:    [q_gen_15m 2:2][q_gen_shell 2:2][q_ope_max 2:2][q_ope_4h
2:2][q_ope_batch 2:2][q_gen_max 2:2][q_gen_batch 2:2][q_gen_4h
2:2][q_ope_15m 2:2][q_ope_shell 2:2]

Total Time: 5:46:02  Up: 4:54:18 (85.05%)  Active: 00:06:20 (1.83%)

Reservations:
NOTE:  no reservations on node
ource is utilized...

Thanks for your help...
-------------- next part --------------
An HTML attachment was scrubbed...
URL: http://www.supercluster.org/pipermail/mauiusers/attachments/20060811/280d4542/attachment-0001.html


More information about the mauiusers mailing list