s3mkbucket(1) Create Amazon AWS S3 buckets


s3mkbucket [options] [bucket ...]

--access-key AWS Access Key ID
--secret-key AWS Secret Access Key
--acl-short private|public-read|public-read-write|authenticated-read


Print a brief help message and exits.
Prints the manual page and exits.
Print a message for each created bucket.
--access-key and --secret-key
Specify the ``AWS Access Key Identifiers'' for the AWS account. --access-key is the ``Access Key ID'', and --secret-key is the ``Secret Access Key''. These are effectively the ``username'' and ``password'' to the AWS account, and should be kept confidential.

The access keys MUST be specified, either via these command line parameters, or via the AWS_ACCESS_KEY_ID and AWS_ACCESS_KEY_SECRET environment variables.

Specifying them on the command line overrides the environment variables.

Uses SSL/TLS HTTPS to communicate with the AWS service, instead of HTTP.
Apply a ``canned ACL'' to the bucket when it is created. To set a more complex ACL, use the "s3acl" tool after the bucket is created.

The following canned ACLs are currently defined by S3:

Owner gets "FULL_CONTROL". No one else has any access rights. This is the default.
Owner gets "FULL_CONTROL". The anonymous principal is granted "READ" access.
Owner gets "FULL_CONTROL". The anonymous principal is granted "READ" and "WRITE" access. This is a useful policy to apply to a bucket, if you intend for any anonymous user to PUT objects into the bucket.
Owner gets "FULL_CONTROL" . Any principal authenticated as a registered Amazon S3 user is granted "READ" access.
One or more bucket names. As many as possible will be created.

A user may have no more than 100 buckets.

Bucket names must be between 3 and 255 characters long, and can only contain alphanumeric characters, underscore, period, and dash. Bucket names are case sensitive. Buckets with names containing uppercase characters or underscores are not accessible using the virtual hosting method.

Buckets are unique in a global namespace. That means if someone has created a bucket with a given name, someone else cannot create another bucket with the same name.

If a bucket name begins with one or more dashes, it might be mistaken for a command line option. If this is the case, separate the command line options from the bucket names with two dashes, like so:

  s3mkbucket --verbose -- --bucketname


Specify the ``AWS Access Key Identifiers'' for the AWS account. AWS_ACCESS_KEY_ID contains the ``Access Key ID'', and AWS_ACCESS_KEY_SECRET contains the ``Secret Access Key''. These are effectively the ``username'' and ``password'' to the AWS service, and should be kept confidential.

The access keys MUST be specified, either via these environment variables, or via the --access-key and --secret-key command line parameters.

If the command line parameters are set, they override these environment variables.


The configuration options will be read from the file "~/.s3-tools" if it exists. The format is the same as the command line options with one option per line. For example, the file could contain:

    --access-key <AWS access key>
    --secret-key <AWS secret key>

This example configuration file would specify the AWS access keys and that a secure connection using HTTPS should be used for all communications.


Create buckets in the Amazon Simple Storage Service (S3).


Report bugs to Mark Atwood [email protected]

Making a bucket that already exists and is owned by the user does not fail. It is unclear whether this is a bug or not.

Occasionally the S3 service will randomly fail for no externally apparent reason. When that happens, this tool should retry, with a delay and a backoff.

Access to the S3 service can be authenticated with a X.509 certificate, instead of via the ``AWS Access Key Identifiers''. This tool should support that.

It might be useful to be able to specify the ``AWS Access Key Identifiers'' in the user's "~/.netrc" file. This tool should support that.

Errors and warnings are very ``Perl-ish'', and can be confusing.


Written by Mark Atwood [email protected]

Many thanks to Wotan LLC <http://wotanllc.com>, for supporting the development of these S3 tools.

Many thanks to the Amazon AWS engineers for developing S3.