gluster(8) - Gluster Console Manager (command line utility)



  • Gluster(8)				 Gluster Inc.				   Gluster(8)
    
    
    
    NAME
           gluster - Gluster Console Manager (command line utility)
    
    SYNOPSIS
           gluster
    
           To run the program and display gluster prompt:
    
           gluster [--xml]
    
           (or)
    
           To specify a command directly:
    
           gluster [commands] [options] [--xml]
    
    
    DESCRIPTION
           The  Gluster  Console Manager is a command line utility for elastic volume management.
           You can run the gluster command on any export server. The command enables  administra‐
           tors to perform cloud operations, such as creating, expanding, shrinking, rebalancing,
           and migrating volumes without needing to schedule server downtime.
    
    COMMANDS
       Volume Commands
    	volume info [all|<VOLNAME>]
    	      Display information about all volumes, or the specified volume.
    
    	volume create <NEW-VOLNAME> [stripe <COUNT>] [replica <COUNT>]	[disperse  [<COUNT>]]
           [redundancy <COUNT>] [transport <tcp|rdma|tcp,rdma>] <NEW-BRICK> ...
    	      Create a new volume of the specified type using the specified bricks and trans‐
    	      port type (the default transport type is tcp).  To create a  volume  with	 both
    	      transports (tcp and rdma), give 'transport tcp,rdma' as an option.
    
    	volume delete <VOLNAME>
    	      Delete the specified volume.
    
    	volume start <VOLNAME>
    	      Start the specified volume.
    
    	volume stop <VOLNAME> [force]
    	      Stop the specified volume.
    
    	volume set <VOLNAME> <OPTION> <PARAMETER> [<OPTION> <PARAMETER>] ...
    	      Set the volume options.
    
    	volume get <VOLNAME> <OPTION/all>
    	      Get the volume options.
    
    	volume help
    	      Display help for the volume command.
    
       Brick Commands
    	volume add-brick <VOLNAME> <NEW-BRICK> ...
    	      Add the specified brick to the specified volume.
    
    	volume remove-brick <VOLNAME> <BRICK> ...
    	      Remove the specified brick from the specified volume.
    
    	      Note: If you remove the brick, the data stored in that brick will not be avail‐
    	      able. You can migrate data  from	one  brick  to	another	 using	replace-brick
    	      option.
    
    	volume replace-brick <VOLNAME> <SOURCE-BRICK> <NEW-BRICK> commit force
    	      Replace the specified source brick with a new brick.
    
    	volume rebalance <VOLNAME> start
    	      Start rebalancing the specified volume.
    
    	volume rebalance <VOLNAME> stop
    	      Stop rebalancing the specified volume.
    
    	volume rebalance <VOLNAME> status
    	      Display the rebalance status of the specified volume.
    
       Log Commands
    	volume log filename <VOLNAME> [BRICK] <DIRECTORY>
    	      Set the log directory for the corresponding volume/brick.
    
    	volume log locate <VOLNAME> [BRICK]
    	      Locate the log file for corresponding volume/brick.
    
    	volume log rotate <VOLNAME> [BRICK]
    	      Rotate the log file for corresponding volume/brick.
    
       Peer Commands
    	peer probe <HOSTNAME>
    	      Probe  the  specified  peer. In case the <HOSTNAME> given belongs to an already
    	      probed peer, the peer probe command will	add  the  hostname  to	the  peer  if
    	      required.
    
    	peer detach <HOSTNAME>
    	      Detach the specified peer.
    
    	peer status
    	      Display the status of peers.
    
    	peer help
    	      Display help for the peer command.
    
       Tier Commands
    	volume tier <VOLNAME> attach [<replica COUNT>] <NEW-BRICK>...
    	      Attach  to  an  existing	volume	a  tier of specified type using the specified
    	      bricks.
    
    	volume tier <VOLNAME> status
    	      Display statistics on data migration between the hot and cold tiers.
    
    	volume tier <VOLNAME> detach start
    	      Begin detaching the hot tier from the volume. Data will be moved from  the  hot
    	      tier to the cold tier.
    
    	volume tier <VOLNAME> detach commit [force]
    	      Commit  detaching	 the  hot tier from the volume. The volume will revert to its
    	      original state before the hot tier was attached.
    
    	volume tier <VOLNAME> detach status
    	      Check status of data movement from the hot to cold tier.
    
    	volume tier <VOLNAME> detach stop
    	      Stop detaching the hot tier from the volume.
    
       Geo-replication Commands
    	Note: password-less ssh, from the master node (where these commands are executed)  to
           the slave node <SLAVE_HOST>, is a prerequisite for the geo-replication commands.
    
    	system:: execute gsec_create
    	      Generates pem keys which are required for push-pem
    
    	volume	 geo-replication  <MASTER_VOL>	<SLAVE_HOST>::<SLAVE_VOL>  create  [push-pem]
           [force]
    	      Create a new geo-replication session from	 <MASTER_VOL>  to  <SLAVE_HOST>	 host
    	      machine having <SLAVE_VOL>.  Use push-pem to push the keys automatically.
    
    	volume geo-replication <MASTER_VOL> <SLAVE_HOST>::<SLAVE_VOL> {start|stop} [force]
    	      Start/stop  the  geo-replication session from <MASTER_VOL> to <SLAVE_HOST> host
    	      machine having <SLAVE_VOL>.
    
    	volume geo-replication [<MASTER_VOL> [<SLAVE_HOST>::<SLAVE_VOL>]] status [detail]
    	      Query status of the geo-replication session from <MASTER_VOL>  to	 <SLAVE_HOST>
    	      host machine having <SLAVE_VOL>.
    
    	volume geo-replication <MASTER_VOL> <SLAVE_HOST>::<SLAVE_VOL> {pause|resume} [force]
    	      Pause/resume the geo-replication session from <MASTER_VOL> to <SLAVE_HOST> host
    	      machine having <SLAVE_VOL>.
    
    	volume geo-replication	<MASTER_VOL>  <SLAVE_HOST>::<SLAVE_VOL>	 delete	 [reset-sync-
           time]
    	      Delete  the  geo-replication  session  from  <MASTER_VOL>	 to <SLAVE_HOST> host
    	      machine having <SLAVE_VOL>.  Optionally you can also reset  the  sync  time  in
    	      case you need to resync the entire volume on session recreate.
    
    	volume	geo-replication	 <MASTER_VOL>  <SLAVE_HOST>::<SLAVE_VOL> config [[!]<options>
           [<value>]]
    	      View (when no option provided) or set configuration  for	this  geo-replication
    	      session.	Use "!<OPTION>" to reset option <OPTION> to default value.
    
       Bitrot Commands
    	volume bitrot <VOLNAME> {enable|disable}
    	      Enable/disable bitrot for volume <VOLNAME>
    
    	volume bitrot <VOLNAME> scrub-throttle {lazy|normal|aggressive}
    	      Scrub-throttle  value  is a measure of how fast or slow the scrubber scrubs the
    	      filesystem for volume <VOLNAME>
    
    	volume bitrot <VOLNAME> scrub-frequency {daily|weekly|biweekly|monthly}
    	      Scrub frequency for volume <VOLNAME>
    
    	volume bitrot <VOLNAME> scrub {pause|resume|status|ondemand}
    	      Pause/Resume scrub. Upon resume, scrubber continues where it left	 off.  status
    	      option  shows  the statistics of scrubber. ondemand option starts the scrubbing
    	      immediately if the scrubber is not paused or already running.
    
    
    	  Snapshot Commands
    
    	snapshot  create  <snapname>  <volname>	 [no-timestamp]	 [description  <description>]
           [force]
    	      Creates  a  snapshot  of a GlusterFS volume. User can provide a snap-name and a
    	      description to identify the snap. Snap will be created by	 appending  timestamp
    	      in  GMT.	User  can  override  this  behaviour using "no-timestamp" option. The
    	      description cannot be more than 1024 characters. To be able to take a snapshot,
    	      volume should be present and it should be in started state.
    
    	snapshot restore <snapname>
    	      Restores	an  already taken snapshot of a GlusterFS volume. Snapshot restore is
    	      an offline activity therefore if the volume is online (in started	 state)	 then
    	      the  restore  operation will fail. Once the snapshot is restored it will not be
    	      available in the list of snapshots.
    
    	snapshot clone <clonename> <snapname>
    	      Create a clone of a snapshot volume, the resulting  volume  will	be  GlusterFS
    	      volume.  User  can  provide  a clone-name. To be able to take a clone, snapshot
    	      should be present and it should be in activated state.
    
    	snapshot delete ( all | <snapname> | volume <volname> )
    	      If snapname is specified then mentioned snapshot	is  deleted.  If  volname  is
    	      specified then all snapshots belonging to that particular volume is deleted. If
    	      keyword *all* is used then all snapshots belonging to the system is deleted.
    
    	snapshot list [volname]
    	      Lists all snapshots taken. If volname is	provided,  then	 only  the  snapshots
    	      belonging to that particular volume is listed.
    
    	snapshot info [snapname | (volume <volname>)]
    	      This  command  gives  information such as snapshot name, snapshot UUID, time at
    	      which snapshot was created, and it lists down the snap-volume-name,  number  of
    	      snapshots	 already  taken and number of snapshots still available for that par‐
    	      ticular volume, and the state of the snapshot. If snapname  is  specified	 then
    	      info  of	the  mentioned	snapshot is  displayed.	 If volname is specified then
    	      info of all snapshots belonging to that volume is displayed.  If	 both	snap‐
    	      name  and	 volname  is  not specified then info of all the snapshots present in
    	      the system are displayed.
    
    	snapshot status [snapname | (volume <volname>)]
    	      This command gives status of the snapshot. The details  included	are  snapshot
    	      brick  path,  volume  group(LVM details), status of the snapshot bricks, PID of
    	      the bricks, data percentage filled for that particular volume  group  to	which
    	      the snapshots belong to, and total size of the logical volume.
    
    	      If snapname is specified then status of the mentioned snapshot is displayed. If
    	      volname is specified then status of all snapshots belonging to that  volume  is
    	      displayed. If both snapname and volname is not specified then status of all the
    	      snapshots present in the system are displayed.
    
    	snapshot config [volname] ([snap-max-hard-limit <count>]  [snap-max-soft-limit	<per‐
           cent>]) | ([auto-delete <enable|disable>]) | ([activate-on-create <enable|disable>])
    	      Displays and sets the snapshot config values.
    
    	      snapshot config without any keywords displays the snapshot config values of all
    	      volumes in the system. If volname is provided, then the snapshot config  values
    	      of that volume is displayed.
    
    	      Snapshot	config command along with keywords can be used to change the existing
    	      config values. If volname is provided then  config  value	 of  that  volume  is
    	      changed, else it will set/change the system limit.
    
    	      snap-max-soft-limit  and auto-delete are global options, that will be inherited
    	      by all volumes in the system and cannot be set to individual volumes.
    
    	      snap-max-hard-limit can be set globally, as well	as  per	 volume.  The  lowest
    	      limit  between  the  global system limit and the volume specific limit, becomes
    	      the "Effective snap-max-hard-limit" for a volume.
    
    	      snap-max-soft-limit is a percentage value, which is applied on  the  "Effective
    	      snap-max-hard-limit" to get the "Effective snap-max-soft-limit".
    
    	      When  auto-delete	 feature  is enabled, then upon reaching the "Effective snap-
    	      max-soft-limit", with every successful snapshot creation, the  oldest  snapshot
    	      will be deleted.
    
    	      When  auto-delete	 feature is disabled, then upon reaching the "Effective snap-
    	      max-soft-limit", the user gets a warning with every  successful  snapshot	 cre‐
    	      ation.
    
    	      When  auto-delete	 feature is disabled, then upon reaching the "Effective snap-
    	      max-hard-limit", further	snapshot  creations  will not be allowed.
    
    	      activate-on-create is disabled by default. If  you  enable  activate-on-create,
    	      then further snapshot will be activated during the time of snapshot creation.
    
    	snapshot activate <snapname>
    	      Activates the mentioned snapshot.
    
    	      Note : By default the snapshot is activated during snapshot creation.
    
    	snapshot deactivate <snapname>
    	      Deactivates the mentioned snapshot.
    
    	snapshot help
    	      Display help for the snapshot commands.
    
       Self-heal Commands
    	volume heal <VOLNAME>
    	      Triggers index self heal for the files that need healing.
    
    
    	volume heal  <VOLNAME> [enable | disable]
    	      Enable/disable self-heal-daemon for volume <VOLNAME>.
    
    
    	volume heal <VOLNAME> full
    	      Triggers self heal on all the files.
    
    
    	volume heal <VOLNAME> info
    	      Lists the files that need healing.
    
    
    	volume heal <VOLNAME> info split-brain
    	      Lists the files which are in split-brain state.
    
    
    	volume heal <VOLNAME> statistics
    	      Lists the crawl statistics.
    
    
    	volume heal <VOLNAME> statistics heal-count
    	      Displays the count of files to be healed.
    
    
    	volume heal <VOLNAME> statistics heal-count replica <HOSTNAME:BRICKNAME>
    	      Displays	the  number of files to be healed from a particular replica subvolume
    	      to which the brick <HOSTNAME:BRICKNAME> belongs.
    
    
    	volume heal <VOLNAME> split-brain bigger-file <FILE>
    	      Performs healing of <FILE> which is in split-brain by choosing the bigger	 file
    	      in the replica as source.
    
    
    	volume heal <VOLNAME> split-brain source-brick <HOSTNAME:BRICKNAME>
    	      Selects <HOSTNAME:BRICKNAME> as the source for all the files that are in split-
    	      brain in that replica and heals them.
    
    
    	volume heal <VOLNAME> split-brain source-brick <HOSTNAME:BRICKNAME> <FILE>
    	      Selects the split-brained <FILE> present in <HOSTNAME:BRICKNAME> as source  and
    	      completes heal.
    
       Other Commands
    	get-state [<daemon>] [odir </path/to/output/dir/>] [file <filename>]
    	      Get  local  state representation of mentioned daemon and store data in provided
    	      path information
    
    	help  Display the command options.
    
    	quit  Exit the gluster command line interface.
    
    
    FILES
           /var/lib/glusterd/*
    
    SEE ALSO
           fusermount(1), mount.glusterfs(8), glusterfs(8), glusterd(8)
    
    COPYRIGHT
           Copyright(c) 2006-2011  Gluster, Inc.  <http://www.gluster.com>
    
    
    
    07 March 2011			 Gluster command line utility			   Gluster(8)
    

Log in to reply
 

© Lightnetics 2024