Provided by: s3fs_1.86-1_amd64 bug

NAME

       S3FS - FUSE-based file system backed by Amazon S3

SYNOPSIS

   mounting
       s3fs bucket[:/path] mountpoint  [options]

       s3fs mountpoint  [options (must specify bucket= option)]

   unmounting
       umount mountpoint
              For root.

       fusermount -u mountpoint
              For unprivileged user.

   utility mode (remove interrupted multipart uploading objects)
       s3fs --incomplete-mpu-list (-u) bucket

       s3fs --incomplete-mpu-abort[=all | =<expire date format>] bucket

DESCRIPTION

       s3fs  is  a  FUSE  filesystem  that  allows  you  to  mount an Amazon S3 bucket as a local
       filesystem. It stores files natively and transparently in S3  (i.e.,  you  can  use  other
       programs to access the same files).

AUTHENTICATION

       s3fs         supports        the        standard        AWS        credentials        file
       (https://docs.aws.amazon.com/cli/latest/userguide/cli-config-files.html)     stored     in
       `${HOME}/.aws/credentials`.   Alternatively,  s3fs supports a custom passwd file. Only AWS
       credentials file format can be used when AWS session token is required.  The s3fs password
       file has this format (use this format if you have only one set of credentials):
           accessKeyId:secretAccessKey

       If you have more than one set of credentials, this syntax is also recognized:
           bucketName:accessKeyId:secretAccessKey

       Password files can be stored in two locations:
            /etc/passwd-s3fs     [0640]
            $HOME/.passwd-s3fs   [0600]

       s3fs also recognizes the AWSACCESSKEYID and AWSSECRETACCESSKEY environment variables.

OPTIONS

   general options
       -h   --help
              print help

            --version
              print version

       -f     FUSE foreground option - do not run as daemon.

       -s     FUSE singlethreaded option (disables multi-threaded operation)

   mount options
       All s3fs options must given in the form where "opt" is:
               <option_name>=<option_value>

       -o bucket
              if  it  is  not specified bucket name (and path) in command line, must specify this
              option after -o option for bucket name.

       -o default_acl (default="private")
              the default canned acl to  apply  to  all  written  s3  objects,  e.g.,  "private",
              "public-read".        see      https://docs.aws.amazon.com/AmazonS3/latest/dev/acl-
              overview.html#canned-acl for the full list of canned acls.

       -o retries (default="5")
              number of times to retry a failed S3 transaction.

       -o use_cache (default="" which means disabled)
              local folder to use for local file cache.

       -o check_cache_dir_exist (default is disable)
              If use_cache is set, check if the cache directory exists.  If this  option  is  not
              specified, it will be created at runtime when the cache directory does not exist.

       -o del_cache - delete local file cache
              delete local file cache when s3fs starts and exits.

       -o storage_class (default="standard")
              store  object  with  specified  storage class.  this option replaces the old option
              use_rrs.  Possible values: standard, standard_ia,  onezone_ia,  reduced_redundancy,
              and intelligent_tiering.

       -o use_rrs (default is disable)
              use  Amazon's  Reduced  Redundancy  Storage.  this option can not be specified with
              use_sse.  (can specify use_rrs=1 for old version) this option has been replaced  by
              new storage_class option.

       -o use_sse (default is disable)
              Specify  three type Amazon's Server-Site Encryption: SSE-S3, SSE-C or SSE-KMS. SSE-
              S3 uses Amazon S3-managed encryption keys, SSE-C uses customer-provided  encryption
              keys, and SSE-KMS uses the master key which you manage in AWS KMS.  You can specify
              "use_sse" or "use_sse=1" enables SSE-S3 type (use_sse=1  is  old  type  parameter).
              Case  of  setting  SSE-C, you can specify "use_sse=custom", "use_sse=custom:<custom
              key file path>" or "use_sse=<custom key file path>" (only <custom  key  file  path>
              specified  is old type parameter).  You can use "c" for short "custom".  The custom
              key file must be 600 permission. The file can have some lines,  each  line  is  one
              SSE-C key.  The first line in file is used as Customer-Provided Encryption Keys for
              uploading and changing headers etc.  If there are some keys after first line, those
              are used downloading object which are encrypted by not first key.  So that, you can
              keep all SSE-C keys in file, that is SSE-C key history.  If  you  specify  "custom"
              ("c")  without  file  path,  you  need  to  set  custom key by load_sse_c option or
              AWSSSECKEYS environment. (AWSSSECKEYS environment has  some  SSE-C  keys  with  ":"
              separator.)   This  option  is  used to decide the SSE type.  So that if you do not
              want to encrypt a object at uploading, but you need to decrypt encrypted object  at
              downloading,  you  can  use  load_sse_c option instead of this option.  For setting
              SSE-KMS, specify "use_sse=kmsid" or "use_sse=kmsid:<kms id>".  You can use "k"  for
              short  "kmsid".  If you san specify SSE-KMS type with your <kms id> in AWS KMS, you
              can set it after "kmsid:" (or "k:").  If you specify only "kmsid" ("k"),  you  need
              to  set AWSSSEKMSID environment which value is <kms id>.  You must be careful about
              that you can not use the KMS id which is not same EC2 region.

       -o load_sse_c - specify SSE-C keys
              Specify  the  custom-provided  encryption  keys  file  path   for   decrypting   at
              downloading.   If  you  use  the  custom-provided  encryption key at uploading, you
              specify with "use_sse=custom".  The file has many lines, one line means one  custom
              key.   So  that  you  can  keep  all SSE-C keys in file, that is SSE-C key history.
              AWSSSECKEYS environment is as same as this file contents.

       -o passwd_file (default="")
              specify the path to the password  file,  which  which  takes  precedence  over  the
              password in $HOME/.passwd-s3fs and /etc/passwd-s3fs

       -o ahbe_conf (default="" which means disabled)
              This option specifies the configuration file path which file is the additional HTTP
              header by file (object) extension.
               The configuration file format is below:
               -----------
               line         = [file suffix or regex] HTTP-header [HTTP-values]
               file  suffix   =  file  (object)  suffix,  if  this  field  is  empty,  it   means
              "reg:(.*)".(=all object).
               regex         =  regular  expression  to  match  the file (object) path. this type
              starts with "reg:" prefix.
               HTTP-header  = additional HTTP header name
               HTTP-values  = additional HTTP header value
               -----------
               Sample:
               -----------
               .gz                    Content-Encoding  gzip
               .Z                     Content-Encoding  compress
               reg:^/MYDIR/(.*)[.]t2$ Content-Encoding  text2
               -----------
               A sample configuration file is uploaded in "test" directory.  If you specify  this
              option for set "Content-Encoding" HTTP header, please take care for RFC 2616.

       -o profile (default="default")
              Choose  a  profile  from ${HOME}/.aws/credentials to authenticate against S3.  Note
              that this format matches the AWS CLI  format  and  differs  from  the  s3fs  passwd
              format.

       -o public_bucket (default="" which means disabled)
              anonymously mount a public bucket when set to 1, ignores the $HOME/.passwd-s3fs and
              /etc/passwd-s3fs files.  S3 does not allow copy object  api  for  anonymous  users,
              then  s3fs  sets  nocopyapi  option  automatically  when  public_bucket=1 option is
              specified.

       -o connect_timeout (default="300" seconds)
              time to wait for connection before giving up.

       -o readwrite_timeout (default="120" seconds)
              time to wait between read/write activity before giving up.

       -o list_object_max_keys (default="1000")
              specify the maximum number of keys returned by S3 list object API. The  default  is
              1000. you can set this value to 1000 or more.

       -o max_stat_cache_size (default="100,000" entries (about 40MB))
              maximum number of entries in the stat cache and symbolic link cache.

       -o stat_cache_expire (default is no expire)
              specify  expire  time  (seconds)  for  entries  in the stat cache and symbolic link
              cache. This expire time indicates the time since cached.

       -o stat_cache_interval_expire (default is no expire)
              specify expire time (seconds) for entries in  the  stat  cache  and  symbolic  link
              cache.  This  expire  time  is based on the time from the last access time of those
              cache.   This  option  is  exclusive  with  stat_cache_expire,  and  is  left   for
              compatibility with older versions.

       -o enable_noobj_cache (default is disable)
              enable cache entries for the object which does not exist.  s3fs always has to check
              whether file (or sub directory) exists under object  (path)  when  s3fs  does  some
              command,  since  s3fs has recognized a directory which does not exist and has files
              or sub directories  under  itself.   It  increases  ListBucket  request  and  makes
              performance  bad.   You  can specify this option for performance, s3fs memorizes in
              stat cache that the object (file or directory) does not exist.

       -o no_check_certificate (by default this option is disabled)
              do not check ssl certificate.  server certificate  won't  be  checked  against  the
              available certificate authorities.

       -o ssl_verify_hostname (default="2")
              When 0, do not verify the SSL certificate against the hostname.

       -o nodnscache - disable dns cache.
              s3fs is always using dns cache, this option make dns cache disable.

       -o nosscache - disable ssl session cache.
              s3fs is always using ssl session cache, this option make ssl session cache disable.

       -o multireq_max (default="20")
              maximum number of parallel request for listing objects.

       -o parallel_count (default="5")
              number  of  parallel  request for uploading big objects.  s3fs uploads large object
              (over 20MB) by multipart post request, and sends parallel  requests.   This  option
              limits  parallel request count which s3fs requests at once.  It is necessary to set
              this value depending on a CPU and a network band.

       -o multipart_size (default="10")
              part size, in MB, for each multipart request.  The minimum value is 5  MB  and  the
              maximum value is 5 GB.

       -o ensure_diskfree (default 0)
              sets  MB  to  ensure disk free space. This option means the threshold of free space
              size on disk which is used for the  cache  file  by  s3fs.   s3fs  makes  file  for
              downloading,  uploading  and caching files.  If the disk free space is smaller than
              this value, s3fs do not use diskspace as possible in exchange for the performance.

       -o singlepart_copy_limit (default="512")
              maximum size, in MB, of a single-part copy before trying multipart copy.

       -o host (default="https://s3.amazonaws.com")
              Set a non-Amazon host, e.g., https://example.com.

       -o sevicepath (default="/")
              Set a service path when the non-Amazon host requires a prefix.

       -o url (default="https://s3.amazonaws.com")
              sets the url to use to access Amazon S3. If you want to use HTTP, then you can  set
              "url=http://s3.amazonaws.com".   If  you  do  not use https, please specify the URL
              with the url option.

       -o endpoint (default="us-east-1")
              sets the endpoint to use on signature version 4.  If this option is not  specified,
              s3fs  uses "us-east-1" region as the default.  If the s3fs could not connect to the
              region specified by this option, s3fs could not run.  But if  you  do  not  specify
              this option, and if you can not connect with the default region, s3fs will retry to
              automatically connect to the other region.  So s3fs can  know  the  correct  region
              name, because s3fs can find it in an error from the S3 server.

       -o sigv2 (default is signature version 4)
              sets signing AWS requests by using Signature Version 2.

       -o mp_umask (default is "0000")
              sets  umask  for the mount point directory.  If allow_other option is not set, s3fs
              allows access to the mount point only to the owner.   In  the  opposite  case  s3fs
              allows  access  to  all  users as the default.  But if you set the allow_other with
              this option, you can control the permissions of the mount point by this option like
              umask.

       -o umask (default is "0000")
              sets  umask  for  files  under the mountpoint.  This can allow users other than the
              mounting user to read and write to files that they did not create.

       -o nomultipart - disable multipart uploads

       -o enable_content_md5 (default is disable)
              Allow S3 server to check data integrity of  uploads  via  the  Content-MD5  header.
              This can add CPU overhead to transfers.

       -o ecs (default is disable)
              This  option  instructs s3fs to query the ECS container credential metadata address
              instead of the instance metadata address.

       -o iam_role (default is no IAM role)
              This option requires the IAM role name or "auto". If you specify "auto", s3fs  will
              automatically  use  the  IAM role names that are set to an instance. If you specify
              this option without any argument, it is the same as that  you  have  specified  the
              "auto".

       -o ibm_iam_auth (default is not using IBM IAM authentication)
              This  option  instructs  s3fs  to  use  IBM  IAM  authentication. In this mode, the
              AWSAccessKey and AWSSecretKey will be used as IBM's Service-Instance-ID and APIKey,
              respectively.

       -o ibm_iam_endpoint (default is https://iam.bluemix.net)
              Sets the URL to use for IBM IAM authentication.

       -o use_xattr (default is not handling the extended attribute)
              Enable  to handle the extended attribute (xattrs).  If you set this option, you can
              use the extended attribute.  For example, encfs and ecryptfs need  to  support  the
              extended  attribute.   Notice: if s3fs handles the extended attribute, s3fs can not
              work to copy command with preserve=mode.

       -o noxmlns - disable registering xml name space.
              disable  registering  xml  name  space  for  response   of   ListBucketResult   and
              ListVersionsResult    etc.    Default    name    space    is    looked    up   from
              "http://s3.amazonaws.com/doc/2006-03-01".  This option should not be specified now,
              because s3fs looks up xmlns automatically after v1.66.

       -o nomixupload - disable copy in multipart uploads.
              Disable  to  use  PUT  (copy  api) when multipart uploading large size objects.  By
              default, when doing multipart upload, the range of  unchanged  data  will  use  PUT
              (copy  api)  whenever possible.  When nocopyapi or norenameapi is specified, use of
              PUT (copy api) is invalidated even if this option is not specified.

       -o nocopyapi - for other incomplete compatibility object storage.
              For a distributed object storage which is compatibility S3 API  without  PUT  (copy
              api).   If  you set this option, s3fs do not use PUT with "x-amz-copy-source" (copy
              api). Because traffic is increased 2-3 times by this option, we  do  not  recommend
              this.

       -o norenameapi - for other incomplete compatibility object storage.
              For  a  distributed  object storage which is compatibility S3 API without PUT (copy
              api).  This option is a subset of nocopyapi option. The nocopyapi option  does  not
              use  copy-api  for  all command (ex. chmod, chown, touch, mv, etc), but this option
              does not use copy-api for  only  rename  command  (ex.  mv).   If  this  option  is
              specified with nocopyapi, then s3fs ignores it.

       -o use_path_request_style (use legacy API calling style)
              Enable  compatibility  with  S3-like  APIs  which  do  not support the virtual-host
              request style, by using the older path request style.

       -o noua (suppress User-Agent header)
              Usually s3fs outputs of the User-Agent  in  "s3fs/<version>  (commit  hash  <hash>;
              <using  ssl  library  name>)" format.  If this option is specified, s3fs suppresses
              the output of the User-Agent.

       -o cipher_suites
              Customize the list of TLS cipher suites. Expects a colon separated list  of  cipher
              suite  names.  A list of available cipher suites, depending on your TLS engine, can
              be  found  on  the  CURL  library   documentation:   https://curl.haxx.se/docs/ssl-
              ciphers.html

       -o instance_name
              The  instance  name  of  the  current  s3fs mountpoint.  This name will be added to
              logging messages and user agent headers sent by s3fs.

       -o complement_stat (complement lack of file/directory mode)
              s3fs complements lack of information about file/directory  mode  if  a  file  or  a
              directory  object  does not have x-amz-meta-mode header.  As default, s3fs does not
              complements stat information for a object, then the object will not be able  to  be
              allowed to list/modify.

       -o notsup_compat_dir (not support compatibility directory types)
              As  a  default, s3fs supports objects of the directory type as much as possible and
              recognizes them as directories.   Objects  that  can  be  recognized  as  directory
              objects are "dir/", "dir", "dir_$folder$", and there is a file object that does not
              have a directory object but contains that directory  path.   s3fs  needs  redundant
              communication  to  support  all these directory types.  The object as the directory
              created by s3fs is "dir/".  By restricting s3fs  to  recognize  only  "dir/"  as  a
              directory,  communication traffic can be reduced.  This option is used to give this
              restriction to s3fs.  However, if there is a directory object other than "dir/"  in
              the  bucket,  specifying  this  option is not recommended.  s3fs may not be able to
              recognize the object correctly if an object created by s3fs exists in  the  bucket.
              Please use this option when the directory in the bucket is only "dir/" object.

       -o use_wtf8 - support arbitrary file system encoding.
              S3  requires  all object names to be valid utf-8. But some clients, notably Windows
              NFS clients, use their own encoding.  This option re-encodes invalid  utf-8  object
              names  into valid utf-8 by mapping offending codes into a 'private' codepage of the
              Unicode set.  Useful on clients not using utf-8 as their file system encoding.

       -o use_session_token - indicate that session token should be provided.
              If credentials are provided by environment variables this  switch  forces  presence
              check of AWSSESSIONTOKEN variable.  Otherwise an error is returned.

       -o requester_pays (default is disable)
              This  option instructs s3fs to enable requests involving Requester Pays buckets (It
              includes the 'x-amz-request-payer=requester' entry in the request header).

       -o dbglevel (default="crit")
              Set the debug message level. set  value  as  crit  (critical),  err  (error),  warn
              (warning),  info (information) to debug level. default debug level is critical.  If
              s3fs run with "-d" option, the debug level is set information.  When s3fs catch the
              signal SIGUSR2, the debug level is bumpup.

       -o curldbg - put curl debug message
              Put the debug message from libcurl when this option is specified.

   utility mode options
       -u or --incomplete-mpu-list
              Lists multipart incomplete objects uploaded to the specified bucket.

       --incomplete-mpu-abort all or date format (default="24H")
              Delete  the multipart incomplete object uploaded to the specified bucket.  If "all"
              is specified for this option, all multipart incomplete objects will be deleted.  If
              you  specify  no  argument  as an option, objects older than 24 hours (24H) will be
              deleted (This is the default value).  You can specify an optional date format.   It
              can  be specified as year, month, day, hour, minute, second, and it is expressed as
              "Y", "M", "D", "h", "m", "s" respectively.  For example, "1Y6M10D12h30m30s".

FUSE/MOUNT OPTIONS

       Most of the generic mount options described in 'man mount' are supported  (ro,  rw,  suid,
       nosuid,  dev,  nodev, exec, noexec, atime, noatime, sync async, dirsync).  Filesystems are
       mounted with '-onodev,nosuid' by default, which can only be  overridden  by  a  privileged
       user.

       There  are  many  FUSE specific mount options that can be specified. e.g. allow_other. See
       the FUSE README for the full set.

NOTES

       The maximum size of objects that s3fs can handle depends on Amazon S3. For example, up  to
       5  GB  when using single PUT API. And up to 5 TB is supported when Multipart Upload API is
       used.

       If enabled via the "use_cache" option, s3fs automatically maintains a local cache of files
       in  the  folder specified by use_cache. Whenever s3fs needs to read or write a file on S3,
       it first downloads the entire file locally  to  the  folder  specified  by  use_cache  and
       operates  on  it.  When fuse_release() is called, s3fs will re-upload the file to S3 if it
       has been changed. s3fs uses md5 checksums to minimize downloads from S3.

       The folder specified by use_cache is just a local cache. It can be deleted  at  any  time.
       s3fs rebuilds it on demand.

       Local file caching works by calculating and comparing md5 checksums (ETag HTTP header).

       s3fs  leverages  /etc/mime.types  to "guess" the "correct" content-type based on file name
       extension. This means that you can copy a website to S3 and serve it up directly  from  S3
       with correct content-types!

SEE ALSO

       fuse(8), mount(8), fusermount(1), fstab(5)

BUGS

       Due  to  S3's  "eventual consistency" limitations, file creation can and will occasionally
       fail. Even after a successful create, subsequent reads can fail for an indeterminate time,
       even  after  one  or  more  successful  reads.  Create  and read enough files and you will
       eventually encounter this failure. This is not a flaw in s3fs and it is  not  something  a
       FUSE  wrapper  like  s3fs can work around. The retries option does not address this issue.
       Your application must either tolerate or compensate for these  failures,  for  example  by
       retrying creates or reads.

AUTHOR

       s3fs has been written by Randy Rizun <rrizun@gmail.com>.