bqueues

displays information about queues

Synopsis

bqueues [-alloc] [-w | -l | -r] [-m host_name | -m host_group | -m cluster_name | -m all] [-u user_name |-u user_group | -u all] [queue_name ...]
bqueues [-h | -V]

Description

Displays information about queues.

By default, returns the following information about all queues: queue name, queue priority, queue status, task statistics, and job state statistics.

When a resizable job has a resize allocation request, bqueues displays pending requests. When LSF adds more resources to a running resizable job, bqueues decreases job PEND counts and displays the added resources. When LSF removes resources from a running resizable job, bqueues displays the updated resources.

In MultiCluster, returns the information about all queues in the local cluster.

Returns job slot statistics if -alloc option is used.

Batch queue names and characteristics are set up by the LSF administrator in lsb.queues.

CPU time is normalized.

CPU time output is not consistent with bacct

bacct displays the sum of CPU time consumed by all past jobs in event files, regardless of the execution host type and run time (unless you indicate a begin and end time.) For a specified job, bacct and bhist have the same result.

Because the value of CPU time for bqueues is used by mbatchd to calculate fairshare priority, it does not display the actual CPU time for the queue, but a CPU time normalized by CPU factor. This results in a different CPU time output in bacct and bqueues.

Options

Start of change-allocEnd of change
Start of changeShows counters for slots in RUN, SSUSP, USUSP, and RSV. The slot allocation will be different depending on whether the job is an exclusive job or not.End of change
-l

Displays queue information in a long multiline format. The -l option displays the following additional information: queue description, queue characteristics and statistics, scheduling parameters, resource usage limits, scheduling policies, users, hosts, associated commands, dispatch and run windows, success exit values, host limits per parallel job, and job controls

Also displays user shares.

If you specified an administrator comment with the -C option of the queue control commands qclose, qopen, qact, and qinact, qhist displays the comment text.

Displays absolute priority scheduling (APS) information for queues configured with APS_PRIORITY.

-r

Displays the same information as the -l option. In addition, if fairshare is defined for the queue, displays recursively the share account tree of the fairshare queue. When queue-based fairshare is used along with bsub -G and LSB_SACCT_ONE_UG=Y in lsf.conf, share accounts are only created for active users and for the default user group (if defined).

If you installed Fix Pack #329767 for LSF 9.1.3, which is available from IBM Passport Advantage, -r also displays the global fairshare policy name for the participating queue. Displays remote share load (REMOTE_LOAD column) for each share account in the queue. For example:

$ bqueues -r admin

SCHEDULING POLICIES:  GLOBAL_FAIRSHARE(admin)  NO_INTERACTIVE
USER_SHARES:  [ugroup1, 100]

SHARE_INFO_FOR: admin/
 USER/GROUP   SHARES  PRIORITY  STARTED  RESERVED  CPU_TIME  RUN_TIME   ADJUST  REMOTE_LOAD
 ugroup1       100       0.128      0        0         0.0        0      0.000      776.443

SHARE_INFO_FOR: admin/ugroup1/
 USER/GROUP   SHARES  PRIORITY  STARTED  RESERVED  CPU_TIME  RUN_TIME   ADJUST  REMOTE_LOAD
 user3          10       0.038      0        0         0.0        0      0.000      258.814
 user2          10       0.038      0        0         0.0        0      0.000      258.814
 user1          10       0.038      0        0         0.0        0      0.000      258.814

USERS: all
HOSTS:  all
-w

Displays queue information in a wide format. Fields are displayed without truncation.

-m host_name | -m host_group | -m cluster_name | -m all

Displays the queues that can run jobs on the specified host. If the keyword all is specified, displays the queues that can run jobs on all hosts.

If a host group is specified, displays the queues that include that group in their configuration. For a list of host groups see bmgroup(1).

In MultiCluster, if the all keyword is specified, displays the queues that can run jobs on all hosts in the local cluster. If a cluster name is specified, displays all queues in the specified cluster.

-u user_name | -u user_group | -u all

Displays the queues that can accept jobs from the specified user. If the keyword all is specified, displays the queues that can accept jobs from all users.

If a user group is specified, displays the queues that include that group in their configuration. For a list of user groups see bugroup(1)).

queue_name ...

Displays information about the specified queues.

-h

Prints command usage to stderr and exits.

-V

Prints LSF release version to stderr and exits.

Default Output

Displays the following fields:

QUEUE_NAME

The name of the queue. Queues are named to correspond to the type of jobs usually submitted to them, or to the type of services they provide.

lost_and_found

If the LSF administrator removes queues from the system, LSF creates a queue called lost_and_found and places the jobs from the removed queues into the lost_and_found queue. Jobs in the lost_and_found queue are not started unless they are switched to other queues (see bswitch).

PRIO

The priority of the queue. The larger the value, the higher the priority. If job priority is not configured, determines the queue search order at job dispatch, suspension and resumption time. Jobs from higher priority queues are dispatched first (this is contrary to UNIX process priority ordering), and jobs from lower priority queues are suspended first when hosts are overloaded.

STATUS

The current status of the queue. The possible values are:

Open

The queue is able to accept jobs.

Closed

The queue is not able to accept jobs.

Active

Jobs in the queue may be started.

Inactive

Jobs in the queue cannot be started for the time being.

At any moment, each queue is either Open or Closed, and is either Active or Inactive. The queue can be opened, closed, inactivated and re-activated by the LSF administrator using badmin (see badmin(8)).

Jobs submitted to a queue that is later closed are still dispatched as long as the queue is active. The queue can also become inactive when either its dispatch window is closed or its run window is closed (see DISPATCH_WINDOWS in the “Output for the -l Option” section). In this case, the queue cannot be activated using badmin. The queue is re-activated by LSF when one of its dispatch windows and one of its run windows are open again. The initial state of a queue at LSF boot time is set to open, and either active or inactive depending on its windows.

MAX

The maximum number of job slots that can be used by the jobs from the queue. These job slots are used by dispatched jobs that have not yet finished, and by pending jobs that have slots reserved for them.

A sequential job uses one job slot when it is dispatched to a host, while a parallel job uses as many job slots as is required by bsub -n when it is dispatched. See bsub(1) for details. If a dash (-) is displayed, there is no limit.

JL/U

The maximum number of job slots each user can use for jobs in the queue. These job slots are used by your dispatched jobs that have not yet finished, and by pending jobs that have slots reserved for them. If a dash (-) is displayed, there is no limit.

JL/P

The maximum number of job slots a processor can process from the queue. This includes job slots of dispatched jobs that have not yet finished, and job slots reserved for some pending jobs. The job slot limit per processor (JL/P) controls the number of jobs sent to each host. This limit is configured per processor so that multiprocessor hosts are automatically allowed to run more jobs. If a dash (-) is displayed, there is no limit.

JL/H

The maximum number of job slots a host can allocate from this queue. This includes the job slots of dispatched jobs that have not yet finished, and those reserved for some pending jobs. The job slot limit per host (JL/H) controls the number of jobs sent to each host, regardless of whether a host is a uniprocessor host or a multiprocessor host. If a dash (-) is displayed, there is no limit.

NJOBS

The total number of tasks for jobs in the queue. This includes tasks in pending, running, and suspended jobs. See bjobs(1) for an explanation of batch job states.

If -alloc is used, total will be the sum of the RUN, SSUSP, USUSP, and RSV counters.

PEND

The total number of tasks for all pending jobs in the queue. If used with -alloc, total will be "0".

RUN

The total number of tasks for all running jobs in the queue. If -alloc is used, total will be allocated slots for the jobs in the queue.

SUSP

The total number of tasks for all suspended jobs in the queue.

Long Output (-l)

In addition to the above fields, the -l option displays the following:

Description

A description of the typical use of the queue.

Default queue indication

Indicates that this is the default queue.

PARAMETERS/ STATISTICS
NICE

The nice value at which jobs in the queue are run. This is the UNIX nice value for reducing the process priority (see nice(1)).

STATUS
Inactive

The long format for the -l option gives the possible reasons for a queue to be inactive:

Inact_Win

The queue is out of its dispatch window or its run window.

Inact_Adm

The queue has been inactivated by the LSF administrator.

SSUSP

The number of tasks for all jobs in the queue that are suspended by LSF because of load levels or run windows. If -alloc is used, total will be allocated slots for the jobs in the queue.

USUSP

The number of tasks for all jobs in the queue that are suspended by the job submitter or by the LSF administrator. If -alloc is used, total will be allocated slots for the jobs in the queue.

RSV

The number of tasks reserving slots that are reserved by LSF for pending jobs in the queue. If -alloc is used, total will be allocated slots for the jobs in the queue.

Migration threshold

The length of time in seconds that a job dispatched from the queue remains suspended by the system before LSF attempts to migrate the job to another host. See the MIG parameter in lsb.queues and lsb.hosts.

Schedule delay for a new job

The delay time in seconds for scheduling after a new job is submitted. If the schedule delay time is zero, a new scheduling session is started as soon as the job is submitted to the queue. See the NEW_JOB_SCHED_DELAY parameter in lsb.queues.

Interval for a host to accept two jobs

The length of time in seconds to wait after dispatching a job to a host before dispatching a second job to the same host. If the job accept interval is zero, a host may accept more than one job in each dispatching interval. See the JOB_ACCEPT_INTERVAL parameter in lsb.queues and lsb.params.

RESOURCE LIMITS

The hard resource usage limits that are imposed on the jobs in the queue (see getrlimit(2) and lsb.queues(5)). These limits are imposed on a per-job and a per-process basis.

The possible per-job limits are:

CPULIMIT

The maximum CPU time a job can use, in minutes, relative to the CPU factor of the named host. CPULIMIT is scaled by the CPU factor of the execution host so that jobs are allowed more time on slower hosts.

When the job-level CPULIMIT is reached, a SIGXCPU signal is sent to all processes belonging to the job. If the job has no signal handler for SIGXCPU, the job is killed immediately. If the SIGXCPU signal is handled, blocked, or ignored by the application, then after the grace period expires, LSF sends SIGINT, SIGTERM, and SIGKILL to the job to kill it.

Start of changeTASKLIMITEnd of change
Start of change

The maximum number of tasks allocated to a job. Jobs that have fewer tasks than the minimum TASKLIMIT or more tasks than the maximum TASKLIMIT are rejected. If the job requests minimum and maximum tasks, the maximum tasks requested cannot be less than the minimum TASKLIMIT, and the minimum tasks requested cannot be more than the maximum TASKLIMIT.

End of change
MEMLIMIT

The maximum running set size (RSS) of a process. If a process uses more memory than the limit allows, its priority is reduced so that other processes are more likely to be paged in to available memory. This limit is enforced by the setrlimit system call if it supports the RLIMIT_RSS option.

By default, the limit is shown in KB. Use LSF_UNIT_FOR_LIMITS in lsf.conf to specify a larger unit for display (MB, GB, TB, PB, or EB).

SWAPLIMIT

The swap space limit that a job may use. If SWAPLIMIT is reached, the system sends the following signals in sequence to all processes in the job: SIGINT, SIGTERM, and SIGKILL.

By default, the limit is shown in KB. Use LSF_UNIT_FOR_LIMITS in lsf.conf to specify a larger unit for display (MB, GB, TB, PB, or EB).

PROCESSLIMIT

The maximum number of concurrent processes allocated to a job. If PROCESSLIMIT is reached, the system sends the following signals in sequence to all processes belonging to the job: SIGINT, SIGTERM, and SIGKILL.

THREADLIMIT

The maximum number of concurrent threads allocated to a job. If THREADLIMIT is reached, the system sends the following signals in sequence to all processes belonging to the job: SIGINT, SIGTERM, and SIGKILL.

RUNLIMIT

The maximum wall clock time a process can use, in minutes. RUNLIMIT is scaled by the CPU factor of the execution host. When a job has been in the RUN state for a total of RUNLIMIT minutes, LSF sends a SIGUSR2 signal to the job. If the job does not exit within 5 minutes, LSF sends a SIGKILL signal to kill the job.

FILELIMIT

The maximum file size a process can create, in KB. This limit is enforced by the UNIX setrlimit system call if it supports the RLIMIT_FSIZE option, or the ulimit system call if it supports the UL_SETFSIZE option.

DATALIMIT

The maximum size of the data segment of a process, in KB. This restricts the amount of memory a process can allocate. DATALIMIT is enforced by the setrlimit system call if it supports the RLIMIT_DATA option, and unsupported otherwise.

STACKLIMIT

The maximum size of the stack segment of a process. This limit restricts the amount of memory a process can use for local variables or recursive function calls. STACKLIMIT is enforced by the setrlimit system call if it supports the RLIMIT_STACK option.

By default, the limit is shown in KB. Use LSF_UNIT_FOR_LIMITS in lsf.conf to specify a larger unit for display (MB, GB, TB, PB, or EB).

CORELIMIT

The maximum size of a core file. This limit is enforced by the setrlimit system call if it supports the RLIMIT_CORE option.

If a job submitted to the queue has any of these limits specified (see bsub(1)), then the lower of the corresponding job limits and queue limits are used for the job.

If no resource limit is specified, the resource is assumed to be unlimited.

By default, the limit is shown in KB. Use LSF_UNIT_FOR_LIMITS in lsf.conf to specify a larger unit for display (MB, GB, TB, PB, or EB).

HOSTLIMIT_PER_JOB

The maximum number of hosts that a job in this queue can use. LSF verifies the host limit during the allocation phase of scheduling. If the number of hosts requested for a parallel job exceeds this limit and LSF cannot satisfy the minimum number of request slots, the parallel job will pend.

SCHEDULING PARAMETERS

The scheduling and suspending thresholds for the queue.

The scheduling threshold loadSched and the suspending threshold loadStop are used to control batch job dispatch, suspension, and resumption. The queue thresholds are used in combination with the thresholds defined for hosts (see bhosts(1) and lsb.hosts(5)). If both queue level and host level thresholds are configured, the most restrictive thresholds are applied.

The loadSched and loadStop thresholds have the following fields:

r15s

The 15-second exponentially averaged effective CPU run queue length.

r1m

The 1-minute exponentially averaged effective CPU run queue length.

r15m

The 15-minute exponentially averaged effective CPU run queue length.

ut

The CPU utilization exponentially averaged over the last minute, expressed as a percentage between 0 and 1.

pg

The memory paging rate exponentially averaged over the last minute, in pages per second.

io

The disk I/O rate exponentially averaged over the last minute, in KB per second.

ls

The number of current login users.

it

On UNIX, the idle time of the host (keyboard not touched on all logged in sessions), in minutes.

On Windows, the it index is based on the time a screen saver has been active on a particular host.

tmp

The amount of free space in /tmp, in MB.

swp

The amount of currently available swap space. By default, swap space is shown in MB. Use LSF_UNIT_FOR_LIMITS in lsf.conf to specify a larger unit for display (MB, GB, TB, PB, or EB).

mem

The amount of currently available memory. By default, memory is shown in MB. Use LSF_UNIT_FOR_LIMITS in lsf.conf to specify a larger unit for display (MB, GB, TB, PB, or EB).

cpuspeed

The speed of each individual cpu, in megahertz (MHz).

bandwidth

The maximum bandwidth requirement, in megabits per second (Mbps).

In addition to these internal indices, external indices are also displayed if they are defined in lsb.queues (see lsb.queues(5)).

The loadSched threshold values specify the job dispatching thresholds for the corresponding load indices. If a dash (-) is displayed as the value, it means the threshold is not applicable. Jobs in the queue may be dispatched to a host if the values of all the load indices of the host are within (below or above, depending on the meaning of the load index) the corresponding thresholds of the queue and the host. The same conditions are used to resume jobs dispatched from the queue that have been suspended on this host.

Similarly, the loadStop threshold values specify the thresholds for job suspension. If any of the load index values on a host go beyond the corresponding threshold of the queue, jobs in the queue are suspended.

JOB EXCEPTION PARAMETERS

Configured job exception thresholds and number of jobs in each exception state for the queue.

Threshold and NumOfJobs have the following fields:

overrun

Configured threshold in minutes for overrun jobs, and the number of jobs in the queue that have triggered an overrun job exception by running longer than the overrun threshold

underrun

Configured threshold in minutes for underrun jobs, and the number of jobs in the queue that have triggered an underrun job exception by finishing sooner than the underrun threshold

idle

Configured threshold (CPU time/runtime) for idle jobs, and the number of jobs in the queue that have triggered an overrun job exception by having a job idle factor less than the threshold

SCHEDULING POLICIES

Scheduling policies of the queue. Optionally, one or more of the following policies may be configured:

APS_PRIORITY

Absolute Priority Scheduling is enabled. Pending jobs in the queue are ordered according to the calculated APS value.

FAIRSHARE

Queue-level fairshare scheduling is enabled. Jobs in this queue are scheduled based on a fairshare policy instead of the first-come, first-served (FCFS) policy.

BACKFILL

A job in a backfill queue can use the slots reserved by other jobs if the job can run to completion before the slot-reserving jobs start.

Backfilling does not occur on queue limits and user limit but only on host based limits. That is, backfilling is only supported when MXJ, JL/U, JL/P, PJOB_LIMIT, and HJOB_LIMIT are reached. Backfilling is not supported when MAX_JOBS, QJOB_LIMIT, and UJOB_LIMIT are reached.

IGNORE_DEADLINE

If IGNORE_DEADLINE is set to Y, starts all jobs regardless of the run limit.

EXCLUSIVE

Jobs dispatched from an exclusive queue can run exclusively on a host if the user so specifies at job submission time (see bsub(1)). Exclusive execution means that the job is sent to a host with no other batch job running there, and no further job, batch or interactive, is dispatched to that host while the job is running. The default is not to allow exclusive jobs.

NO_INTERACTIVE

This queue does not accept batch interactive jobs. (see the -I, -Is, and -Ip options of bsub(1)). The default is to accept both interactive and non-interactive jobs.

ONLY_INTERACTIVE

This queue only accepts batch interactive jobs. Jobs must be submitted using the -I, -Is, and -Ip options of bsub(1). The default is to accept both interactive and non-interactive jobs.

SLA_GUARANTEES_IGNORE

This queue is allowed to ignore SLA resource guarantees when scheduling jobs.

FAIRSHARE_QUEUES

Lists queues participating in cross-queue fairshare. The first queue listed is the master queue—the queue in which fairshare is configured; all other queues listed inherit the fairshare policy from the master queue. Fairshare information applies to all the jobs running in all the queues in the master-slave set.

QUEUE_GROUP

Lists queues participating in an absolute priority scheduling (APS) queue group.

If both FAIRSHARE and APS_PRIORITY are enabled in the same queue, the FAIRSHARE_QUEUES are not displayed. These queues are instead displayed as QUEUE_GROUP.

DISPATCH_ORDER

DISPATCH_ORDER=QUEUE is set in the master queue. Jobs from this queue are dispatched according to the order of queue priorities first, then user fairshare priority. Within the queue, dispatch order is based on user share quota. This avoids having users with higher fairshare priority getting jobs dispatched from low-priority queues.

USER_SHARES

A list of [user_name, share] pairs. user_name is either a user name or a user group name. share is the number of shares of resources assigned to the user or user group. A party receives a portion of the resources proportional to that party's share divided by the sum of the shares of all parties specified in this queue.

DEFAULT HOST SPECIFICATION

The default host or host model that is used to normalize the CPU time limit of all jobs.

If you want to view a list of the CPU factors defined for the hosts in your cluster, see lsinfo(1). The CPU factors are configured in lsf.shared(5).

The appropriate CPU scaling factor of the host or host model is used to adjust the actual CPU time limit at the execution host (see CPULIMIT in lsb.queues(5)). The DEFAULT_HOST_SPEC parameter in lsb.queues overrides the system DEFAULT_HOST_SPEC parameter in lsb.params (see lsb.params(5)). If a user explicitly gives a host specification when submitting a job using bsub -c cpu_limit[/host_name | /host_model], the user specification overrides the values defined in both lsb.params and lsb.queues.

RUN_WINDOWS

The time windows in a week during which jobs in the queue may run.

When a queue is out of its window or windows, no job in this queue is dispatched. In addition, when the end of a run window is reached, any running jobs from this queue are suspended until the beginning of the next run window, when they are resumed. The default is no restriction, or always open.

DISPATCH_WINDOWS

Dispatch windows are the time windows in a week during which jobs in the queue may be dispatched.

When a queue is out of its dispatch window or windows, no job in this queue is dispatched. Jobs already dispatched are not affected by the dispatch windows. The default is no restriction, or always open (that is, twenty-four hours a day, seven days a week). Note that such windows are only applicable to batch jobs. Interactive jobs scheduled by LIM are controlled by another set of dispatch windows (see lshosts(1)). Similar dispatch windows may be configured for individual hosts (see bhosts(1)).

A window is displayed in the format begin_timeend_time. Time is specified in the format [day:]hour[:minute], where all fields are numbers in their respective legal ranges: 0(Sunday)-6 for day, 0-23 for hour, and 0-59 for minute. The default value for minute is 0 (on the hour). The default value for day is every day of the week. The begin_time and end_time of a window are separated by a dash (-), with no blank characters (SPACE and TAB) in between. Both begin_time and end_time must be present for a window. Windows are separated by blank characters.

USERS

A list of users allowed to submit jobs to this queue. LSF administrators can submit jobs to the queue even if they are not listed here.

User group names have a slash (/) added at the end of the group name. See bugroup(1).

If the fairshare scheduling policy is enabled, users cannot submit jobs to the queue unless they also have a share assignment. This also applies to LSF administrators.

HOSTS

A list of hosts where jobs in the queue can be dispatched.

Host group names have a slash (/) added at the end of the group name. See bmgroup(1).

NQS DESTINATION QUEUES

A list of NQS destination queues to which this queue can dispatch jobs.

When you submit a job using bsub -q queue_name, and the specified queue is configured to forward jobs to the NQS system, LSF routes your job to one of the NQS destination queues. The job runs on an NQS batch server host, which is not a member of the LSF cluster. Although running on an NQS system outside the LSF cluster, the job is still managed by LSF in almost the same way as jobs running inside the LSF cluster. Thus, you may have your batch jobs transparently sent to an NQS system to run and then get the results of your jobs back. You may use any supported user interface, including LSF commands and NQS commands (see lsnqs(1)) to submit, monitor, signal and delete your batch jobs that are running in an NQS system. See lsb.queues(5) and bsub(1) for more information.

ADMINISTRATORS

A list of queue administrators. The users whose names are specified here are allowed to operate on the jobs in the queue and on the queue itself. See lsb.queues(5) for more information.

PRE_EXEC

The job-based pre-execution command for the queue. The PRE_EXEC command runs on the execution host before the job associated with the queue is dispatched to the execution host (or to the first host selected for a parallel batch job).

POST_EXEC

The job-based post-execution command for the queue. The POST_EXEC command runs on the execution host after the job finishes.

HOST_PRE_EXEC

The host-based pre-execution command for the queue. The HOST_PRE_EXEC command runs on all execution hosts before the job associated with the queue is dispatched to the execution hosts. If job based pre-execution PRE_EXEC was defined at the queue-level/application-level/job-level, the HOST_PRE_EXEC command runs before PRE_EXEC of any level. The host-based pre-execution command cannot be executed on Windows platforms.

HOST_POST_EXEC

The host-based post-execution command for the queue. The HOST_POST_EXEC command runs on all execution hosts after the job finishes. If job based post-execution POST_EXEC was defined at the queue-level/application-level/job-level, the HOST_POST_EXEC command runs after POST_EXEC of any level. The host-based post-execution command cannot be executed on Windows platforms.

Start of changeLOCAL_MAX_PREEXEC_RETRY_ACTIONEnd of change
Start of change

The action to take on a job when the number of times to attempt its pre-execution command on the local cluster (LOCAL_MAX_PREEXEC_RETRY) is reached.

End of change
REQUEUE_EXIT_VALUES

Jobs that exit with these values are automatically requeued. See lsb.queues(5) for more information.

RES_REQ

Resource requirements of the queue. Only the hosts that satisfy these resource requirements can be used by the queue.

RESRSV_LIMIT

Resource requirement limits of the queue. Queue-level RES_REQ rusage values (set in lsb.queues) must be in the range set by RESRSV_LIMIT, or the queue-level RES_REQ is ignored. Merged RES_REQ rusage values from the job and application levels must be in the range of RESRSV_LIMIT, or the job is rejected.

Maximum slot reservation time

The maximum time in seconds a slot is reserved for a pending job in the queue. See the SLOT_RESERVE=MAX_RESERVE_TIME[n] parameter in lsb.queues.

RESUME_COND

The conditions that must be satisfied to resume a suspended job on a host. See lsb.queues(5) for more information.

STOP_COND

The conditions that determine whether a job running on a host should be suspended. See lsb.queues(5) for more information.

JOB_STARTER

An executable file that runs immediately prior to the batch job, taking the batch job file as an input argument. All jobs submitted to the queue are run via the job starter, which is generally used to create a specific execution environment before processing the jobs themselves. See lsb.queues(5) for more information.

CHUNK_JOB_SIZE

Chunk jobs only. Specifies the maximum number of jobs allowed to be dispatched together in a chunk job. All of the jobs in the chunk are scheduled and dispatched as a unit rather than individually. The ideal candidates for job chunking are jobs that typically takes 1 to 2 minutes to run.

SEND_JOBS_TO

MultiCluster. List of remote queue names to which the queue forwards jobs.

RECEIVE_JOBS_FROM

MultiCluster. List of remote cluster names from which the queue receives jobs.

PREEMPTION
PREEMPTIVE

The queue is preemptive. Jobs in this queue may preempt running jobs from lower-priority queues, even if the lower-priority queues are not specified as preemptive.

PREEMPTABLE

The queue is preemptable. Running jobs in this queue may be preempted by jobs in higher-priority queues, even if the higher-priority queues are not specified as preemptive.

RERUNNABLE

If the RERUNNABLE field displays yes, jobs in the queue are rerunnable. That is, jobs in the queue are automatically restarted or rerun if the execution host becomes unavailable. However, a job in the queue is not restarted if you remove the rerunnable option from the job.

CHECKPOINT

If the CHKPNTDIR field is displayed, jobs in the queue are checkpointable. Jobs use the default checkpoint directory and period unless you specify other values. Note that a job in the queue is not checkpointed if you remove the checkpoint option from the job.

CHKPNTDIR

Specifies the checkpoint directory using an absolute or relative path name.

CHKPNTPERIOD

Specifies the checkpoint period in seconds.

Although the output of bqueues reports the checkpoint period in seconds, the checkpoint period is defined in minutes (the checkpoint period is defined through the bsub -k "checkpoint_dir []" option, or in lsb.queues).

JOB CONTROLS

The configured actions for job control. See JOB_CONTROLS parameter in lsb.queues.

The configured actions are displayed in the format [action_type, command] where action_type is either SUSPEND, RESUME, or TERMINATE.

ADMIN ACTION COMMENT

If the LSF administrator specified an administrator comment with the -C option of the queue control commands qclose, qopen, qact, and qinact, qhist the comment text is displayed.

SLOT_SHARE

Share of job slots for queue-based fairshare. Represents the percentage of running jobs (job slots) in use from the queue. SLOT_SHARE must be greater than zero.

The sum of SLOT_SHARE for all queues in the pool does not need to be 100%. It can be more or less, depending on your needs.

SLOT_POOL

Name of the pool of job slots the queue belongs to for queue-based fairshare. A queue can only belong to one pool. All queues in the pool must share the same set of hosts.

MAX_SLOTS_IN_POOL

Maximum number of job slots available in the slot pool the queue belongs to for queue-based fairshare. Defined in the first queue of the slot pool.

USE_PRIORITY_IN_POOL

Queue-based fairshare only. After job scheduling occurs for each queue, this parameter enables LSF to dispatch jobs to any remaining slots in the pool in first-come first-served order across queues.

NO_PREEMPT_INTERVAL

The uninterrupted running time (minutes) that must pass before preemption is permitted. Configured in lsb.queues.

MAX_TOTAL_TIME_PREEMPT

The maximum total preemption time (minutes) above which preemption is not permitted. Configured in lsb.queues.

SHARE_INFO_FOR

User shares and dynamic priority information based on the scheduling policy in place for the queue.

USER/GROUP

Name of users or user groups who have access to the queue.

SHARES

Number of shares of resources assigned to each user or user group in this queue, as configured in the file lsb.queues. The shares affect dynamic user priority for when fairshare scheduling is configured at the queue level.

PRIORITY

Dynamic user priority for the user or user group. Larger values represent higher priorities. Jobs belonging to the user or user group with the highest priority are considered first for dispatch.

In general, users or user groups with larger SHARES, fewer STARTED and RESERVED, and a lower CPU_TIME and RUN_TIME have higher PRIORITY.

STARTED

Number of job slots used by running or suspended jobs owned by users or user groups in the queue.

RESERVED

Number of job slots reserved by the jobs owned by users or user groups in the queue.

CPU_TIME

Cumulative CPU time used by jobs of users or user groups executed in the queue. Measured in seconds, to one decimal place.

LSF calculates the cumulative CPU time using the actual (not normalized) CPU time and a decay factor such that 1 hour of recently-used CPU time decays to 0.1 hours after an interval of time specified by HIST_HOURS in lsb.params (5 hours by default).

RUN_TIME

Wall-clock run time plus historical run time of jobs of users or user groups that are executed in the queue. Measured in seconds.

LSF calculates the historical run time using the actual run time of finished jobs and a decay factor such that 1 hour of recently-used run time decays to 0.1 hours after an interval of time specified by HIST_HOURS in lsb.params (5 hours by default). Wall-clock run time is the run time of running jobs.

ADJUST

Dynamic priority calculation adjustment made by the user-defined fairshare plugin(libfairshareadjust.*).

The fairshare adjustment is enabled and weighted by the parameter FAIRSHARE_ADJUSTMENT_FACTOR in lsb.params.

RUN_TIME_FACTOR

The weighting parameter for run_time within the dynamic priority calculation. If not defined for the queue, the cluster-wide value defined in lsb.params is used.

CPU_TIME_FACTOR

The dynamic priority calculation weighting parameter for CPU time. If not defined for the queue, the cluster-wide value defined in lsb.params is used.

ENABLE_HIST_RUN_TIME

Enables the use of historic run time (run time for completed jobs) in the dynamic priority calculation. If not defined for the queue, the cluster-wide value defined in lsb.params is used.

RUN_TIME_DECAY

Enables the decay of run time in the dynamic priority calculation. The decay rate is set by the parameter HIST_HOURS (set for the queue in lsb.queues or set for the cluster in lsb.params). If not defined for the queue, the cluster-wide value defined in lsb.params is used.

HIST_HOURS

Decay parameter for CPU time, run time, and historic run time. If not defined for the queue, the cluster-wide value defined in lsb.params is used.

FAIRSHARE_ADJUSTMENT_FACTOR

Enables and weights the dynamic priority calculation adjustment made by the user-defined fairshare plugin(libfairshareadjust.*).If not defined for the queue, the cluster-wide value defined in lsb.params is used.

RUN_JOB_FACTOR

The dynamic priority calculation weighting parameter for the number of job slots reserved and in use by a user. If not defined for the queue, the cluster-wide value defined in lsb.params is used.

COMMITTED_RUN_TIME_FACTOR

The dynamic priority calculation weighting parameter for committed run time. If not defined for the queue, the cluster-wide value defined in lsb.params is used.

Start of changeJOB_SIZE_LISTEnd of change
Start of change

A list of job sizes (number of tasks) allowed on this queue, including the default job size that is assigned if the job submission does not request a job size. Configured in lsb.queues.

End of change

Recursive Share Tree Output (-r)

In addition to the fields displayed for the -l option, the -r option displays the following:

SCHEDULING POLICIES
FAIRSHARE

The -r option causes bqueues to recursively display the entire share information tree associated with the queue.

See also

bugroup, nice, getrlimit, lsb.queues, bsub, bjobs, bhosts, badmin, mbatchd