summaryrefslogtreecommitdiffstats
path: root/doc/admin-guide/en-US/markdown/admin_monitoring_workload.md
diff options
context:
space:
mode:
Diffstat (limited to 'doc/admin-guide/en-US/markdown/admin_monitoring_workload.md')
-rw-r--r--doc/admin-guide/en-US/markdown/admin_monitoring_workload.md893
1 files changed, 0 insertions, 893 deletions
diff --git a/doc/admin-guide/en-US/markdown/admin_monitoring_workload.md b/doc/admin-guide/en-US/markdown/admin_monitoring_workload.md
deleted file mode 100644
index c3ac0609b99..00000000000
--- a/doc/admin-guide/en-US/markdown/admin_monitoring_workload.md
+++ /dev/null
@@ -1,893 +0,0 @@
-#Monitoring your GlusterFS Workload
-
-You can monitor the GlusterFS volumes on different parameters.
-Monitoring volumes helps in capacity planning and performance tuning
-tasks of the GlusterFS volume. Using these information, you can identify
-and troubleshoot issues.
-
-You can use Volume Top and Profile commands to view the performance and
-identify bottlenecks/hotspots of each brick of a volume. This helps
-system administrators to get vital performance information whenever
-performance needs to be probed.
-
-You can also perform statedump of the brick processes and nfs server
-process of a volume, and also view volume status and volume information.
-
-##Running GlusterFS Volume Profile Command
-
-GlusterFS Volume Profile command provides an interface to get the
-per-brick I/O information for each File Operation (FOP) of a volume. The
-per brick information helps in identifying bottlenecks in the storage
-system.
-
-This section describes how to run GlusterFS Volume Profile command by
-performing the following operations:
-
-- [Start Profiling](#start-profiling)
-- [Displaying the I/0 Information](#displaying-io)
-- [Stop Profiling](#stop-profiling)
-
-<a name="start-profiling" />
-###Start Profiling
-
-You must start the Profiling to view the File Operation information for
-each brick.
-
-To start profiling, use following command:
-
-`# gluster volume profile start `
-
-For example, to start profiling on test-volume:
-
- # gluster volume profile test-volume start
- Profiling started on test-volume
-
-When profiling on the volume is started, the following additional
-options are displayed in the Volume Info:
-
- diagnostics.count-fop-hits: on
- diagnostics.latency-measurement: on
-
-<a name="displaying-io" />
-###Displaying the I/0 Information
-
-You can view the I/O information of each brick by using the following command:
-
-`# gluster volume profile info`
-
-For example, to see the I/O information on test-volume:
-
- # gluster volume profile test-volume info
- Brick: Test:/export/2
- Cumulative Stats:
-
- Block 1b+ 32b+ 64b+
- Size:
- Read: 0 0 0
- Write: 908 28 8
-
- Block 128b+ 256b+ 512b+
- Size:
- Read: 0 6 4
- Write: 5 23 16
-
- Block 1024b+ 2048b+ 4096b+
- Size:
- Read: 0 52 17
- Write: 15 120 846
-
- Block 8192b+ 16384b+ 32768b+
- Size:
- Read: 52 8 34
- Write: 234 134 286
-
- Block 65536b+ 131072b+
- Size:
- Read: 118 622
- Write: 1341 594
-
-
- %-latency Avg- Min- Max- calls Fop
- latency Latency Latency
- ___________________________________________________________
- 4.82 1132.28 21.00 800970.00 4575 WRITE
- 5.70 156.47 9.00 665085.00 39163 READDIRP
- 11.35 315.02 9.00 1433947.00 38698 LOOKUP
- 11.88 1729.34 21.00 2569638.00 7382 FXATTROP
- 47.35 104235.02 2485.00 7789367.00 488 FSYNC
-
- ------------------
-
- ------------------
-
- Duration : 335
-
- BytesRead : 94505058
-
- BytesWritten : 195571980
-
-<a name="stop-profiling" />
-###Stop Profiling
-
-You can stop profiling the volume, if you do not need profiling
-information anymore.
-
-Stop profiling using the following command:
-
- `# gluster volume profile stop`
-
-For example, to stop profiling on test-volume:
-
- `# gluster volume profile stop`
-
- `Profiling stopped on test-volume`
-
-##Running GlusterFS Volume TOP Command
-
-GlusterFS Volume Top command allows you to view the glusterfs bricks’
-performance metrics like read, write, file open calls, file read calls,
-file write calls, directory open calls, and directory real calls. The
-top command displays up to 100 results.
-
-This section describes how to run and view the results for the following
-GlusterFS Top commands:
-
-- [Viewing Open fd Count and Maximum fd Count](#open-fd-count)
-- [Viewing Highest File Read Calls](#file-read)
-- [Viewing Highest File Write Calls](#file-write)
-- [Viewing Highest Open Calls on Directories](#open-dir)
-- [Viewing Highest Read Calls on Directory](#read-dir)
-- [Viewing List of Read Performance on each Brick](#read-perf)
-- [Viewing List of Write Performance on each Brick](#write-perf)
-
-<a name="open-fd-count" />
-###Viewing Open fd Count and Maximum fd Count
-
-You can view both current open fd count (list of files that are
-currently the most opened and the count) on the brick and the maximum
-open fd count (count of files that are the currently open and the count
-of maximum number of files opened at any given point of time, since the
-servers are up and running). If the brick name is not specified, then
-open fd metrics of all the bricks belonging to the volume will be
-displayed.
-
-- View open fd count and maximum fd count using the following command:
-
- `# gluster volume top open [brick ] [list-cnt ]`
-
- For example, to view open fd count and maximum fd count on brick
- server:/export of test-volume and list top 10 open calls:
-
- `# gluster volume top open brick list-cnt `
-
- `Brick: server:/export/dir1 `
-
- `Current open fd's: 34 Max open fd's: 209 `
-
- ==========Open file stats========
-
- open file name
- call count
-
- 2 /clients/client0/~dmtmp/PARADOX/
- COURSES.DB
-
- 11 /clients/client0/~dmtmp/PARADOX/
- ENROLL.DB
-
- 11 /clients/client0/~dmtmp/PARADOX/
- STUDENTS.DB
-
- 10 /clients/client0/~dmtmp/PWRPNT/
- TIPS.PPT
-
- 10 /clients/client0/~dmtmp/PWRPNT/
- PCBENCHM.PPT
-
- 9 /clients/client7/~dmtmp/PARADOX/
- STUDENTS.DB
-
- 9 /clients/client1/~dmtmp/PARADOX/
- STUDENTS.DB
-
- 9 /clients/client2/~dmtmp/PARADOX/
- STUDENTS.DB
-
- 9 /clients/client0/~dmtmp/PARADOX/
- STUDENTS.DB
-
- 9 /clients/client8/~dmtmp/PARADOX/
- STUDENTS.DB
-
-<a name="file-read" />
-###Viewing Highest File Read Calls
-
-You can view highest read calls on each brick. If brick name is not
-specified, then by default, list of 100 files will be displayed.
-
-- View highest file Read calls using the following command:
-
- `# gluster volume top read [brick ] [list-cnt ] `
-
- For example, to view highest Read calls on brick server:/export of
- test-volume:
-
- `# gluster volume top read brick list-cnt `
-
- `Brick:` server:/export/dir1
-
- ==========Read file stats========
-
- read filename
- call count
-
- 116 /clients/client0/~dmtmp/SEED/LARGE.FIL
-
- 64 /clients/client0/~dmtmp/SEED/MEDIUM.FIL
-
- 54 /clients/client2/~dmtmp/SEED/LARGE.FIL
-
- 54 /clients/client6/~dmtmp/SEED/LARGE.FIL
-
- 54 /clients/client5/~dmtmp/SEED/LARGE.FIL
-
- 54 /clients/client0/~dmtmp/SEED/LARGE.FIL
-
- 54 /clients/client3/~dmtmp/SEED/LARGE.FIL
-
- 54 /clients/client4/~dmtmp/SEED/LARGE.FIL
-
- 54 /clients/client9/~dmtmp/SEED/LARGE.FIL
-
- 54 /clients/client8/~dmtmp/SEED/LARGE.FIL
-
-<a name="file-write" />
-###Viewing Highest File Write Calls
-
-You can view list of files which has highest file write calls on each
-brick. If brick name is not specified, then by default, list of 100
-files will be displayed.
-
-- View highest file Write calls using the following command:
-
- `# gluster volume top write [brick ] [list-cnt ] `
-
- For example, to view highest Write calls on brick server:/export of
- test-volume:
-
- `# gluster volume top write brick list-cnt `
-
- `Brick: server:/export/dir1 `
-
- ==========Write file stats========
- write call count filename
-
- 83 /clients/client0/~dmtmp/SEED/LARGE.FIL
-
- 59 /clients/client7/~dmtmp/SEED/LARGE.FIL
-
- 59 /clients/client1/~dmtmp/SEED/LARGE.FIL
-
- 59 /clients/client2/~dmtmp/SEED/LARGE.FIL
-
- 59 /clients/client0/~dmtmp/SEED/LARGE.FIL
-
- 59 /clients/client8/~dmtmp/SEED/LARGE.FIL
-
- 59 /clients/client5/~dmtmp/SEED/LARGE.FIL
-
- 59 /clients/client4/~dmtmp/SEED/LARGE.FIL
-
- 59 /clients/client6/~dmtmp/SEED/LARGE.FIL
-
- 59 /clients/client3/~dmtmp/SEED/LARGE.FIL
-
-<a name="open-dir" />
-###Viewing Highest Open Calls on Directories
-
-You can view list of files which has highest open calls on directories
-of each brick. If brick name is not specified, then the metrics of all
-the bricks belonging to that volume will be displayed.
-
-- View list of open calls on each directory using the following
- command:
-
- `# gluster volume top opendir [brick ] [list-cnt ] `
-
- For example, to view open calls on brick server:/export/ of
- test-volume:
-
- `# gluster volume top opendir brick list-cnt `
-
- `Brick: server:/export/dir1 `
-
- ==========Directory open stats========
-
- Opendir count directory name
-
- 1001 /clients/client0/~dmtmp
-
- 454 /clients/client8/~dmtmp
-
- 454 /clients/client2/~dmtmp
-
- 454 /clients/client6/~dmtmp
-
- 454 /clients/client5/~dmtmp
-
- 454 /clients/client9/~dmtmp
-
- 443 /clients/client0/~dmtmp/PARADOX
-
- 408 /clients/client1/~dmtmp
-
- 408 /clients/client7/~dmtmp
-
- 402 /clients/client4/~dmtmp
-
-<a name="read-dir" />
-###Viewing Highest Read Calls on Directory
-
-You can view list of files which has highest directory read calls on
-each brick. If brick name is not specified, then the metrics of all the
-bricks belonging to that volume will be displayed.
-
-- View list of highest directory read calls on each brick using the
- following command:
-
- `# gluster volume top readdir [brick ] [list-cnt ] `
-
- For example, to view highest directory read calls on brick
- server:/export of test-volume:
-
- `# gluster volume top readdir brick list-cnt `
-
- `Brick: `
-
- ==========Directory readdirp stats========
-
- readdirp count directory name
-
- 1996 /clients/client0/~dmtmp
-
- 1083 /clients/client0/~dmtmp/PARADOX
-
- 904 /clients/client8/~dmtmp
-
- 904 /clients/client2/~dmtmp
-
- 904 /clients/client6/~dmtmp
-
- 904 /clients/client5/~dmtmp
-
- 904 /clients/client9/~dmtmp
-
- 812 /clients/client1/~dmtmp
-
- 812 /clients/client7/~dmtmp
-
- 800 /clients/client4/~dmtmp
-
-<a name="read-perf" />
-###Viewing List of Read Performance on each Brick
-
-You can view the read throughput of files on each brick. If brick name
-is not specified, then the metrics of all the bricks belonging to that
-volume will be displayed. The output will be the read throughput.
-
- ==========Read throughput file stats========
-
- read filename Time
- through
- put(MBp
- s)
-
- 2570.00 /clients/client0/~dmtmp/PWRPNT/ -2011-01-31
- TRIDOTS.POT 15:38:36.894610
- 2570.00 /clients/client0/~dmtmp/PWRPNT/ -2011-01-31
- PCBENCHM.PPT 15:38:39.815310
- 2383.00 /clients/client2/~dmtmp/SEED/ -2011-01-31
- MEDIUM.FIL 15:52:53.631499
-
- 2340.00 /clients/client0/~dmtmp/SEED/ -2011-01-31
- MEDIUM.FIL 15:38:36.926198
-
- 2299.00 /clients/client0/~dmtmp/SEED/ -2011-01-31
- LARGE.FIL 15:38:36.930445
-
- 2259.00 /clients/client0/~dmtmp/PARADOX/ -2011-01-31
- COURSES.X04 15:38:40.549919
-
- 2221.00 /clients/client0/~dmtmp/PARADOX/ -2011-01-31
- STUDENTS.VAL 15:52:53.298766
-
- 2221.00 /clients/client3/~dmtmp/SEED/ -2011-01-31
- COURSES.DB 15:39:11.776780
-
- 2184.00 /clients/client3/~dmtmp/SEED/ -2011-01-31
- MEDIUM.FIL 15:39:10.251764
-
- 2184.00 /clients/client5/~dmtmp/WORD/ -2011-01-31
- BASEMACH.DOC 15:39:09.336572
-
-This command will initiate a dd for the specified count and block size
-and measures the corresponding throughput.
-
-- View list of read performance on each brick using the following
- command:
-
- `# gluster volume top read-perf [bs count ] [brick ] [list-cnt ]`
-
- For example, to view read performance on brick server:/export/ of
- test-volume, 256 block size of count 1, and list count 10:
-
- `# gluster volume top read-perf bs 256 count 1 brick list-cnt `
-
- `Brick: server:/export/dir1 256 bytes (256 B) copied, Throughput: 4.1 MB/s `
-
- ==========Read throughput file stats========
-
- read filename Time
- through
- put(MBp
- s)
-
- 2912.00 /clients/client0/~dmtmp/PWRPNT/ -2011-01-31
- TRIDOTS.POT 15:38:36.896486
-
- 2570.00 /clients/client0/~dmtmp/PWRPNT/ -2011-01-31
- PCBENCHM.PPT 15:38:39.815310
-
- 2383.00 /clients/client2/~dmtmp/SEED/ -2011-01-31
- MEDIUM.FIL 15:52:53.631499
-
- 2340.00 /clients/client0/~dmtmp/SEED/ -2011-01-31
- MEDIUM.FIL 15:38:36.926198
-
- 2299.00 /clients/client0/~dmtmp/SEED/ -2011-01-31
- LARGE.FIL 15:38:36.930445
-
- 2259.00 /clients/client0/~dmtmp/PARADOX/ -2011-01-31
- COURSES.X04 15:38:40.549919
-
- 2221.00 /clients/client9/~dmtmp/PARADOX/ -2011-01-31
- STUDENTS.VAL 15:52:53.298766
-
- 2221.00 /clients/client8/~dmtmp/PARADOX/ -2011-01-31
- COURSES.DB 15:39:11.776780
-
- 2184.00 /clients/client3/~dmtmp/SEED/ -2011-01-31
- MEDIUM.FIL 15:39:10.251764
-
- 2184.00 /clients/client5/~dmtmp/WORD/ -2011-01-31
- BASEMACH.DOC 15:39:09.336572
-
-<a name="write-perf" />
-###Viewing List of Write Performance on each Brick
-
-You can view list of write throughput of files on each brick. If brick
-name is not specified, then the metrics of all the bricks belonging to
-that volume will be displayed. The output will be the write throughput.
-
-This command will initiate a dd for the specified count and block size
-and measures the corresponding throughput. To view list of write
-performance on each brick:
-
-- View list of write performance on each brick using the following
- command:
-
- `# gluster volume top write-perf [bs count ] [brick ] [list-cnt ] `
-
- For example, to view write performance on brick server:/export/ of
- test-volume, 256 block size of count 1, and list count 10:
-
- `# gluster volume top write-perf bs 256 count 1 brick list-cnt `
-
- `Brick`: server:/export/dir1
-
- `256 bytes (256 B) copied, Throughput: 2.8 MB/s `
-
- ==========Write throughput file stats========
-
- write filename Time
- throughput
- (MBps)
-
- 1170.00 /clients/client0/~dmtmp/SEED/ -2011-01-31
- SMALL.FIL 15:39:09.171494
-
- 1008.00 /clients/client6/~dmtmp/SEED/ -2011-01-31
- LARGE.FIL 15:39:09.73189
-
- 949.00 /clients/client0/~dmtmp/SEED/ -2011-01-31
- MEDIUM.FIL 15:38:36.927426
-
- 936.00 /clients/client0/~dmtmp/SEED/ -2011-01-31
- LARGE.FIL 15:38:36.933177
- 897.00 /clients/client5/~dmtmp/SEED/ -2011-01-31
- MEDIUM.FIL 15:39:09.33628
-
- 897.00 /clients/client6/~dmtmp/SEED/ -2011-01-31
- MEDIUM.FIL 15:39:09.27713
-
- 885.00 /clients/client0/~dmtmp/SEED/ -2011-01-31
- SMALL.FIL 15:38:36.924271
-
- 528.00 /clients/client5/~dmtmp/SEED/ -2011-01-31
- LARGE.FIL 15:39:09.81893
-
- 516.00 /clients/client6/~dmtmp/ACCESS/ -2011-01-31
- FASTENER.MDB 15:39:01.797317
-
-##Displaying Volume Information
-
-You can display information about a specific volume, or all volumes, as
-needed.
-
-- Display information about a specific volume using the following
- command:
-
- `# gluster volume info ``VOLNAME`
-
- For example, to display information about test-volume:
-
- # gluster volume info test-volume
- Volume Name: test-volume
- Type: Distribute
- Status: Created
- Number of Bricks: 4
- Bricks:
- Brick1: server1:/exp1
- Brick2: server2:/exp2
- Brick3: server3:/exp3
- Brick4: server4:/exp4
-
-- Display information about all volumes using the following command:
-
- `# gluster volume info all`
-
- # gluster volume info all
-
- Volume Name: test-volume
- Type: Distribute
- Status: Created
- Number of Bricks: 4
- Bricks:
- Brick1: server1:/exp1
- Brick2: server2:/exp2
- Brick3: server3:/exp3
- Brick4: server4:/exp4
-
- Volume Name: mirror
- Type: Distributed-Replicate
- Status: Started
- Number of Bricks: 2 X 2 = 4
- Bricks:
- Brick1: server1:/brick1
- Brick2: server2:/brick2
- Brick3: server3:/brick3
- Brick4: server4:/brick4
-
- Volume Name: Vol
- Type: Distribute
- Status: Started
- Number of Bricks: 1
- Bricks:
- Brick: server:/brick6
-
-##Performing Statedump on a Volume
-
-Statedump is a mechanism through which you can get details of all
-internal variables and state of the glusterfs process at the time of
-issuing the command.You can perform statedumps of the brick processes
-and nfs server process of a volume using the statedump command. The
-following options can be used to determine what information is to be
-dumped:
-
-- **mem** - Dumps the memory usage and memory pool details of the
- bricks.
-
-- **iobuf** - Dumps iobuf details of the bricks.
-
-- **priv** - Dumps private information of loaded translators.
-
-- **callpool** - Dumps the pending calls of the volume.
-
-- **fd** - Dumps the open fd tables of the volume.
-
-- **inode** - Dumps the inode tables of the volume.
-
-**To display volume statedump**
-
-- Display statedump of a volume or NFS server using the following
- command:
-
- `# gluster volume statedump [nfs] [all|mem|iobuf|callpool|priv|fd|inode]`
-
- For example, to display statedump of test-volume:
-
- # gluster volume statedump test-volume
- Volume statedump successful
-
- The statedump files are created on the brick servers in the` /tmp`
- directory or in the directory set using `server.statedump-path`
- volume option. The naming convention of the dump file is
- `<brick-path>.<brick-pid>.dump`.
-
-- By defult, the output of the statedump is stored at
- ` /tmp/<brickname.PID.dump>` file on that particular server. Change
- the directory of the statedump file using the following command:
-
- `# gluster volume set server.statedump-path `
-
- For example, to change the location of the statedump file of
- test-volume:
-
- # gluster volume set test-volume server.statedump-path /usr/local/var/log/glusterfs/dumps/
- Set volume successful
-
- You can view the changed path of the statedump file using the
- following command:
-
- `# gluster volume info `
-
-##Displaying Volume Status
-
-You can display the status information about a specific volume, brick or
-all volumes, as needed. Status information can be used to understand the
-current status of the brick, nfs processes, and overall file system.
-Status information can also be used to monitor and debug the volume
-information. You can view status of the volume along with the following
-details:
-
-- **detail** - Displays additional information about the bricks.
-
-- **clients** - Displays the list of clients connected to the volume.
-
-- **mem** - Displays the memory usage and memory pool details of the
- bricks.
-
-- **inode** - Displays the inode tables of the volume.
-
-- **fd** - Displays the open fd (file descriptors) tables of the
- volume.
-
-- **callpool** - Displays the pending calls of the volume.
-
-**To display volume status**
-
-- Display information about a specific volume using the following
- command:
-
- `# gluster volume status [all| []] [detail|clients|mem|inode|fd|callpool]`
-
- For example, to display information about test-volume:
-
- # gluster volume status test-volume
- STATUS OF VOLUME: test-volume
- BRICK PORT ONLINE PID
- --------------------------------------------------------
- arch:/export/1 24009 Y 22445
- --------------------------------------------------------
- arch:/export/2 24010 Y 22450
-
-- Display information about all volumes using the following command:
-
- `# gluster volume status all`
-
- # gluster volume status all
- STATUS OF VOLUME: volume-test
- BRICK PORT ONLINE PID
- --------------------------------------------------------
- arch:/export/4 24010 Y 22455
-
- STATUS OF VOLUME: test-volume
- BRICK PORT ONLINE PID
- --------------------------------------------------------
- arch:/export/1 24009 Y 22445
- --------------------------------------------------------
- arch:/export/2 24010 Y 22450
-
-- Display additional information about the bricks using the following
- command:
-
- `# gluster volume status detail`
-
- For example, to display additional information about the bricks of
- test-volume:
-
- # gluster volume status test-volume details
- STATUS OF VOLUME: test-volume
- -------------------------------------------
- Brick : arch:/export/1
- Port : 24009
- Online : Y
- Pid : 16977
- File System : rootfs
- Device : rootfs
- Mount Options : rw
- Disk Space Free : 13.8GB
- Total Disk Space : 46.5GB
- Inode Size : N/A
- Inode Count : N/A
- Free Inodes : N/A
-
- Number of Bricks: 1
- Bricks:
- Brick: server:/brick6
-
-- Display the list of clients accessing the volumes using the
- following command:
-
- `# gluster volume status clients`
-
- For example, to display the list of clients connected to
- test-volume:
-
- # gluster volume status test-volume clients
- Brick : arch:/export/1
- Clients connected : 2
- Hostname Bytes Read BytesWritten
- -------- --------- ------------
- 127.0.0.1:1013 776 676
- 127.0.0.1:1012 50440 51200
-
-- Display the memory usage and memory pool details of the bricks using
- the following command:
-
- `# gluster volume status mem`
-
- For example, to display the memory usage and memory pool details of
- the bricks of test-volume:
-
- Memory status for volume : test-volume
- ----------------------------------------------
- Brick : arch:/export/1
- Mallinfo
- --------
- Arena : 434176
- Ordblks : 2
- Smblks : 0
- Hblks : 12
- Hblkhd : 40861696
- Usmblks : 0
- Fsmblks : 0
- Uordblks : 332416
- Fordblks : 101760
- Keepcost : 100400
-
- Mempool Stats
- -------------
- Name HotCount ColdCount PaddedSizeof AllocCount MaxAlloc
- ---- -------- --------- ------------ ---------- --------
- test-volume-server:fd_t 0 16384 92 57 5
- test-volume-server:dentry_t 59 965 84 59 59
- test-volume-server:inode_t 60 964 148 60 60
- test-volume-server:rpcsvc_request_t 0 525 6372 351 2
- glusterfs:struct saved_frame 0 4096 124 2 2
- glusterfs:struct rpc_req 0 4096 2236 2 2
- glusterfs:rpcsvc_request_t 1 524 6372 2 1
- glusterfs:call_stub_t 0 1024 1220 288 1
- glusterfs:call_stack_t 0 8192 2084 290 2
- glusterfs:call_frame_t 0 16384 172 1728 6
-
-- Display the inode tables of the volume using the following command:
-
- `# gluster volume status inode`
-
- For example, to display the inode tables of the test-volume:
-
- # gluster volume status test-volume inode
- inode tables for volume test-volume
- ----------------------------------------------
- Brick : arch:/export/1
- Active inodes:
- GFID Lookups Ref IA type
- ---- ------- --- -------
- 6f3fe173-e07a-4209-abb6-484091d75499 1 9 2
- 370d35d7-657e-44dc-bac4-d6dd800ec3d3 1 1 2
-
- LRU inodes:
- GFID Lookups Ref IA type
- ---- ------- --- -------
- 80f98abe-cdcf-4c1d-b917-ae564cf55763 1 0 1
- 3a58973d-d549-4ea6-9977-9aa218f233de 1 0 1
- 2ce0197d-87a9-451b-9094-9baa38121155 1 0 2
-
-- Display the open fd tables of the volume using the following
- command:
-
- `# gluster volume status fd`
-
- For example, to display the open fd tables of the test-volume:
-
- # gluster volume status test-volume fd
-
- FD tables for volume test-volume
- ----------------------------------------------
- Brick : arch:/export/1
- Connection 1:
- RefCount = 0 MaxFDs = 128 FirstFree = 4
- FD Entry PID RefCount Flags
- -------- --- -------- -----
- 0 26311 1 2
- 1 26310 3 2
- 2 26310 1 2
- 3 26311 3 2
-
- Connection 2:
- RefCount = 0 MaxFDs = 128 FirstFree = 0
- No open fds
-
- Connection 3:
- RefCount = 0 MaxFDs = 128 FirstFree = 0
- No open fds
-
-- Display the pending calls of the volume using the following command:
-
- `# gluster volume status callpool`
-
- Each call has a call stack containing call frames.
-
- For example, to display the pending calls of test-volume:
-
- # gluster volume status test-volume
-
- Pending calls for volume test-volume
- ----------------------------------------------
- Brick : arch:/export/1
- Pending calls: 2
- Call Stack1
- UID : 0
- GID : 0
- PID : 26338
- Unique : 192138
- Frames : 7
- Frame 1
- Ref Count = 1
- Translator = test-volume-server
- Completed = No
- Frame 2
- Ref Count = 0
- Translator = test-volume-posix
- Completed = No
- Parent = test-volume-access-control
- Wind From = default_fsync
- Wind To = FIRST_CHILD(this)->fops->fsync
- Frame 3
- Ref Count = 1
- Translator = test-volume-access-control
- Completed = No
- Parent = repl-locks
- Wind From = default_fsync
- Wind To = FIRST_CHILD(this)->fops->fsync
- Frame 4
- Ref Count = 1
- Translator = test-volume-locks
- Completed = No
- Parent = test-volume-io-threads
- Wind From = iot_fsync_wrapper
- Wind To = FIRST_CHILD (this)->fops->fsync
- Frame 5
- Ref Count = 1
- Translator = test-volume-io-threads
- Completed = No
- Parent = test-volume-marker
- Wind From = default_fsync
- Wind To = FIRST_CHILD(this)->fops->fsync
- Frame 6
- Ref Count = 1
- Translator = test-volume-marker
- Completed = No
- Parent = /export/1
- Wind From = io_stats_fsync
- Wind To = FIRST_CHILD(this)->fops->fsync
- Frame 7
- Ref Count = 1
- Translator = /export/1
- Completed = No
- Parent = test-volume-server
- Wind From = server_fsync_resume
- Wind To = bound_xl->fops->fsync
-
-