Understanding Your AWS Billing and Usage Reports for Amazon S3
Amazon S3 billing and usage reports use codes and abbreviations. For example, for
usage type,
which is defined in the following table, region
is replaced
with one of the following abbreviations:
-
APN1: Asia Pacific (Tokyo)
-
APN2: Asia Pacific (Seoul)
-
APS1: Asia Pacific (Singapore)
-
APS2: Asia Pacific (Sydney)
-
APS3: Asia Pacific (Mumbai)
-
CAN1: Canada (Central)
-
EUC1: EU (Frankfurt)
-
EU: EU (Ireland)
-
EUW2: EU (London)
-
SAE1: South America (São Paulo)
-
UGW1: AWS GovCloud (US)
-
USE1 (or no prefix): US East (N. Virginia)
-
USE2: US East (Ohio)
-
USW1: US West (N. California)
-
USW2: US West (Oregon)
For information about pricing by region, see Amazon S3 Pricing.
The first column in the following table lists usage types that appear in your billing and usage reports.
Usage Types
Usage Type | Units | Granularity | Description |
---|---|---|---|
|
Bytes |
Hourly |
The amount of data transferred in to AWS Region1 from AWS Region2 |
|
Bytes |
Hourly |
The amount of data transferred from AWS Region1 to AWS Region2 |
|
Bytes |
Hourly |
The amount of data transferred from Amazon S3 to Amazon EC2 within the same AWS Region |
|
Bytes |
Hourly |
The amount of data transferred into Amazon S3 from Amazon EC2 within the same AWS Region |
|
Bytes |
Hourly |
The amount of data transferred from Amazon S3 to transition objects to GLACIER storage |
|
Bytes |
Hourly |
The amount of data transferred into Amazon S3 to restore objects from GLACIER storage |
|
Bytes |
Hourly |
The amount of data transferred from Amazon S3 to AWS resources within the same AWS Region |
StorageObjectCount |
Count |
Daily |
The number of objects stored within a given bucket |
|
Bytes |
Hourly |
The amount of data transferred into an AWS Region from a CloudFront distribution |
|
Bytes |
Hourly |
The amount of data transferred from an AWS Region to a CloudFront distribution |
|
Byte-Hours1 |
Hourly |
Prorated storage usage for objects deleted from GLACIER storage before the 90-day minimum commitment ends2 |
|
Byte-Hours |
Hourly |
Prorated storage usage for objects deleted from STANDARD_IA before the 30-day minimum commitment ends3 |
|
Byte-Hours |
Hourly |
Prorated storage usage for small objects (smaller than 128KB) that were deleted from STANDARD_IA before the 30-day minimum commitment ended4 |
|
Objects |
Hourly |
The number of objects listed for an object group (objects are grouped by bucket or prefix) with an inventory list |
|
Count |
Hourly |
The number of PUT, COPY, POST, or LIST requests on STANDARD_IA objects |
|
Count |
Hourly |
The number of GET and all other non-SIA-Tier1 requests on STANDARD_IA objects |
|
Count |
Hourly |
The number of PUT, COPY, POST, or LIST requests for STANDARD, RRS, and tags |
|
Count |
Hourly |
The number of GET and all other non-Tier1 requests |
|
Count |
Hourly |
The number of GLACIER archive requests and standard restore requests |
|
Count |
Hourly |
The number of lifecycle transitions to STANDARD_IA storage |
|
Count |
Hourly |
The number of Bulk GLACIER restore requests |
|
Count |
Hourly |
The number of Expedited GLACIER restore requests |
|
Bytes |
Hourly |
The number of bytes of data retrieved with Bulk GLACIER requests |
|
Bytes |
Hourly |
The number of bytes of data retrieved with Expedited GLACIER requests |
|
Bytes |
Hourly |
The number of bytes of data retrieved with standard GLACIER requests |
|
Bytes |
Hourly |
The number of bytes of data retrieved from STANDARD_IA storage |
|
Objects |
Hourly |
The number of unique objects in each object group (where objects are grouped by bucket or prefix) tracked by storage analytics |
|
Tag-Hours |
Daily |
The total of tags on all objects in the bucket reported by hour |
|
Byte-Hours |
Daily |
The number of byte-hours that data was stored in STANDARD storage |
|
Byte-Hours |
Daily |
The number of byte-hours that data was stored in GLACIER storage |
|
Byte-Hours |
Daily |
The number of byte-hours that data was stored in Reduced Redundancy Storage (RRS) storage |
|
Byte-Hours |
Daily |
The number of byte-hours that data was stored in STANDARD_IA storage |
|
Byte-Hours |
Daily |
The number of byte-hours that small objects (smaller than 128KB) were stored in STANDARD_IA storage |
Notes:
-
For more information on the byte-hours unit, see Converting Usage Byte-Hours to Billed GB-Months.
-
For objects that are archived to the GLACIER storage class, when they are deleted prior to 90 days there is a prorated charge per gigabyte for the remaining days.
-
For objects that are in STANDARD_IA storage, when they are deleted, overwritten, or transitioned to a different storage class prior to 30 days there is a prorated charge per gigabyte for the remaining days.
-
For small objects that are in STANDARD_IA storage, when they are deleted, overwritten, or transitioned to a different storage class prior to 30 days there is a prorated charge per gigabyte for the remaining days.
Tracking Operations in Your Usage Reports
Operations describe the action taken on your AWS object or bucket by the specified
usage
type. Operations are indicated by self-explanatory codes, such as
PutObject
or ListBucket
. To see which actions on your
bucket generated a specific type of usage, use these codes. When you create a
usage
report, you can choose to include All Operations, or a specific
operation, for example, GetObject, to report on.
Converting Usage Byte-Hours to Billed GB-Months
The volume of storage that we bill you for each month is based on the average amount of storage you used throughout the month. You are billed for all of the object data and metadata stored in buckets that you created under your AWS account. For more information about metadata, see Object Key and Metadata.
We measure your storage usage in TimedStorage-ByteHrs, which are totaled up at the end of the month to generate your monthly charges. The usage report reports your storage usage in byte-hours and the billing reports report storage usage in GB-months. To correlate your usage report to your billing reports you need to convert byte-hours into GB-months.
For example, if you store 100 GB (107,374,182,400 bytes) of Standard Amazon S3 storage data in your bucket for the first 15 days in March, and 100 TB (109,951,162,777,600 bytes) of Standard Amazon S3 storage data for the final 16 days in March, you will have used 42,259,901,212,262,400 byte-hours.
First, calculate the total byte-hour usage:
[107,374,182,400 bytes x 15 days x (24 hours/day)] + [109,951,162,777,600 bytes x 16 days x (24 hours/day)] = 42,259,901,212,262,400 byte-hours
Then convert the byte-hours to GB-Months:
42,259,901,212,262,400 byte-hours/1,073,741,824 bytes per GB/24 hours per day /31 days in March =52,900 GB-Months