bionic (3) Net::Amazon::S3.3pm.gz

Provided by: libnet-amazon-s3-perl_0.80-1_all bug

NAME

       Net::Amazon::S3 - Use the Amazon S3 - Simple Storage Service

VERSION

       version 0.80

SYNOPSIS

         use Net::Amazon::S3;
         my $aws_access_key_id     = 'fill me in';
         my $aws_secret_access_key = 'fill me in too';

         my $s3 = Net::Amazon::S3->new(
             {   aws_access_key_id     => $aws_access_key_id,
                 aws_secret_access_key => $aws_secret_access_key,
                 # or use an IAM role.
                 use_iam_role          => 1

                 retry                 => 1,
             }
         );

         # a bucket is a globally-unique directory
         # list all buckets that i own
         my $response = $s3->buckets;
         foreach my $bucket ( @{ $response->{buckets} } ) {
             print "You have a bucket: " . $bucket->bucket . "\n";
         }

         # create a new bucket
         my $bucketname = 'acmes_photo_backups';
         my $bucket = $s3->add_bucket( { bucket => $bucketname } )
             or die $s3->err . ": " . $s3->errstr;

         # or use an existing bucket
         $bucket = $s3->bucket($bucketname);

         # store a file in the bucket
         $bucket->add_key_filename( '1.JPG', 'DSC06256.JPG',
             { content_type => 'image/jpeg', },
         ) or die $s3->err . ": " . $s3->errstr;

         # store a value in the bucket
         $bucket->add_key( 'reminder.txt', 'this is where my photos are backed up' )
             or die $s3->err . ": " . $s3->errstr;

         # list files in the bucket
         $response = $bucket->list_all
             or die $s3->err . ": " . $s3->errstr;
         foreach my $key ( @{ $response->{keys} } ) {
             my $key_name = $key->{key};
             my $key_size = $key->{size};
             print "Bucket contains key '$key_name' of size $key_size\n";
         }

         # fetch file from the bucket
         $response = $bucket->get_key_filename( '1.JPG', 'GET', 'backup.jpg' )
             or die $s3->err . ": " . $s3->errstr;

         # fetch value from the bucket
         $response = $bucket->get_key('reminder.txt')
             or die $s3->err . ": " . $s3->errstr;
         print "reminder.txt:\n";
         print "  content length: " . $response->{content_length} . "\n";
         print "    content type: " . $response->{content_type} . "\n";
         print "            etag: " . $response->{content_type} . "\n";
         print "         content: " . $response->{value} . "\n";

         # delete keys
         $bucket->delete_key('reminder.txt') or die $s3->err . ": " . $s3->errstr;
         $bucket->delete_key('1.JPG')        or die $s3->err . ": " . $s3->errstr;

         # and finally delete the bucket
         $bucket->delete_bucket or die $s3->err . ": " . $s3->errstr;

DESCRIPTION

       This module provides a Perlish interface to Amazon S3. From the developer blurb: "Amazon S3 is storage
       for the Internet. It is designed to make web-scale computing easier for developers. Amazon S3 provides a
       simple web services interface that can be used to store and retrieve any amount of data, at any time,
       from anywhere on the web. It gives any developer access to the same highly scalable, reliable, fast,
       inexpensive data storage infrastructure that Amazon uses to run its own global network of web sites. The
       service aims to maximize benefits of scale and to pass those benefits on to developers".

       To find out more about S3, please visit: http://s3.amazonaws.com/

       To use this module you will need to sign up to Amazon Web Services and provide an "Access Key ID" and "
       Secret Access Key". If you use this module, you will incurr costs as specified by Amazon. Please check
       the costs. If you use this module with your Access Key ID and Secret Access Key you must be responsible
       for these costs.

       I highly recommend reading all about S3, but in a nutshell data is stored in values. Values are
       referenced by keys, and keys are stored in buckets. Bucket names are global.

       Note: This is the legacy interface, please check out Net::Amazon::S3::Client instead.

       Development of this code happens here: http://github.com/pfig/net-amazon-s3/

       Homepage for the project (just started) is at http://pfig.github.com/net-amazon-s3/

METHODS

   new
       Create a new S3 client object. Takes some arguments:

       aws_access_key_id
           Use your Access Key ID as the value of the AWSAccessKeyId parameter in requests you send to Amazon
           Web Services (when required). Your Access Key ID identifies you as the party responsible for the
           request.

       aws_secret_access_key
           Since your Access Key ID is not encrypted in requests to AWS, it could be discovered and used by
           anyone. Services that are not free require you to provide additional information, a request
           signature, to verify that a request containing your unique Access Key ID could only have come from
           you.

           DO NOT INCLUDE THIS IN SCRIPTS OR APPLICATIONS YOU DISTRIBUTE. YOU'LL BE SORRY

       aws_session_token
           If you are using temporary credentials provided by the AWS Security Token Service, set the token
           here, and it will be added to the request in order to authenticate it.

       use_iam_role
           If you'd like to use IAM provided temporary credentials, pass this option with a true value.

       secure
           Set this to 1 if you want to use SSL-encrypted connections when talking to S3. Defaults to 0.

       timeout
           How many seconds should your script wait before bailing on a request to S3? Defaults to 30.

       retry
           If this library should retry upon errors. This option is recommended.  This uses exponential backoff
           with retries after 1, 2, 4, 8, 16, 32 seconds, as recommended by Amazon. Defaults to off.

       host
           The S3 host endpoint to use. Defaults to 's3.amazonaws.com'. This allows you to connect to any
           S3-compatible host.

   buckets
       Returns undef on error, else hashref of results

   add_bucket
       Takes a hashref:

       bucket
           The name of the bucket you want to add

       acl_short (optional)
           See the set_acl subroutine for documenation on the acl_short options

       location_constraint (option)
           Sets the location constraint of the new bucket. If left unspecified, the default S3 datacenter
           location will be used. Otherwise, you can set it to 'EU' for a European data center - note that costs
           are different.

       Returns 0 on failure, Net::Amazon::S3::Bucket object on success

   bucket BUCKET
       Takes a scalar argument, the name of the bucket you're creating

       Returns an (unverified) bucket object from an account. Does no network access.

   delete_bucket
       Takes either a Net::Amazon::S3::Bucket object or a hashref containing

       bucket
           The name of the bucket to remove

       Returns false (and fails) if the bucket isn't empty.

       Returns true if the bucket is successfully deleted.

   list_bucket
       List all keys in this bucket.

       Takes a hashref of arguments:

       MANDATORY

       bucket
           The name of the bucket you want to list keys on

       OPTIONAL

       prefix
           Restricts the response to only contain results that begin with the specified prefix. If you omit this
           optional argument, the value of prefix for your query will be the empty string. In other words, the
           results will be not be restricted by prefix.

       delimiter
           If this optional, Unicode string parameter is included with your request, then keys that contain the
           same string between the prefix and the first occurrence of the delimiter will be rolled up into a
           single result element in the CommonPrefixes collection. These rolled-up keys are not returned
           elsewhere in the response.  For example, with prefix="USA/" and delimiter="/", the matching keys
           "USA/Oregon/Salem" and "USA/Oregon/Portland" would be summarized in the response as a single
           "USA/Oregon" element in the CommonPrefixes collection. If an otherwise matching key does not contain
           the delimiter after the prefix, it appears in the Contents collection.

           Each element in the CommonPrefixes collection counts as one against the MaxKeys limit. The rolled-up
           keys represented by each CommonPrefixes element do not.  If the Delimiter parameter is not present in
           your request, keys in the result set will not be rolled-up and neither the CommonPrefixes collection
           nor the NextMarker element will be present in the response.

       max-keys
           This optional argument limits the number of results returned in response to your query. Amazon S3
           will return no more than this number of results, but possibly less. Even if max-keys is not
           specified, Amazon S3 will limit the number of results in the response.  Check the IsTruncated flag to
           see if your results are incomplete.  If so, use the Marker parameter to request the next page of
           results.  For the purpose of counting max-keys, a 'result' is either a key in the 'Contents'
           collection, or a delimited prefix in the 'CommonPrefixes' collection. So for delimiter requests, max-
           keys limits the total number of list results, not just the number of keys.

       marker
           This optional parameter enables pagination of large result sets.  "marker" specifies where in the
           result set to resume listing. It restricts the response to only contain results that occur
           alphabetically after the value of marker. To retrieve the next page of results, use the last key from
           the current page of results as the marker in your next request.

           See also "next_marker", below.

           If "marker" is omitted,the first page of results is returned.

       Returns undef on error and a hashref of data on success:

       The hashref looks like this:

         {
               bucket          => $bucket_name,
               prefix          => $bucket_prefix,
               common_prefixes => [$prefix1,$prefix2,...]
               marker          => $bucket_marker,
               next_marker     => $bucket_next_available_marker,
               max_keys        => $bucket_max_keys,
               is_truncated    => $bucket_is_truncated_boolean
               keys            => [$key1,$key2,...]
          }

       Explanation of bits of that:

       common_prefixes
           If list_bucket was requested with a delimiter, common_prefixes will contain a list of prefixes
           matching that delimiter.  Drill down into these prefixes by making another request with the prefix
           parameter.

       is_truncated
           B flag that indicates whether or not all results of your query were returned in this response. If
           your results were truncated, you can make a follow-up paginated request using the Marker parameter to
           retrieve the rest of the results.

       next_marker
           A convenience element, useful when paginating with delimiters. The value of "next_marker", if
           present, is the largest (alphabetically) of all key names and all CommonPrefixes prefixes in the
           response.  If the "is_truncated" flag is set, request the next page of results by setting "marker" to
           the value of "next_marker". This element is only present in the response if the "delimiter" parameter
           was sent with the request.

       Each key is a hashref that looks like this:

            {
               key           => $key,
               last_modified => $last_mod_date,
               etag          => $etag, # An MD5 sum of the stored content.
               size          => $size, # Bytes
               storage_class => $storage_class # Doc?
               owner_id      => $owner_id,
               owner_displayname => $owner_name
           }

   list_bucket_all
       List all keys in this bucket without having to worry about 'marker'. This is a convenience method, but
       may make multiple requests to S3 under the hood.

       Takes the same arguments as list_bucket.

   add_key
       DEPRECATED. DO NOT USE

   get_key
       DEPRECATED. DO NOT USE

   head_key
       DEPRECATED. DO NOT USE

   delete_key
       DEPRECATED. DO NOT USE

LICENSE

       This module contains code modified from Amazon that contains the following notice:

         #  This software code is made available "AS IS" without warranties of any
         #  kind.  You may copy, display, modify and redistribute the software
         #  code either by itself or as incorporated into your code; provided that
         #  you do not remove any proprietary notices.  Your use of this software
         #  code is at your own risk and you waive any claim against Amazon
         #  Digital Services, Inc. or its affiliates with respect to your use of
         #  this software code. (c) 2006 Amazon Digital Services, Inc. or its
         #  affiliates.

TESTING

       Testing S3 is a tricky thing. Amazon wants to charge you a bit of money each time you use their service.
       And yes, testing counts as using.  Because of this, the application's test suite skips anything
       approaching a real test unless you set these three environment variables:

       AMAZON_S3_EXPENSIVE_TESTS
           Doesn't matter what you set it to. Just has to be set

       AWS_ACCESS_KEY_ID
           Your AWS access key

       AWS_ACCESS_KEY_SECRET
           Your AWS sekkr1t passkey. Be forewarned that setting this environment variable on a shared system
           might leak that information to another user. Be careful.

AUTHOR

       Leon Brocard <acme@astray.com> and unknown Amazon Digital Services programmers.

       Brad Fitzpatrick <brad@danga.com> - return values, Bucket object

       Pedro Figueiredo <me@pedrofigueiredo.org> - since 0.54

SEE ALSO

       Net::Amazon::S3::Bucket

AUTHOR

       Rusty Conover <rusty@luckydinosaur.com>

       This software is copyright (c) 2015 by Amazon Digital Services, Leon Brocard, Brad Fitzpatrick, Pedro
       Figueiredo, Rusty Conover.

       This is free software; you can redistribute it and/or modify it under the same terms as the Perl 5
       programming language system itself.