BTS of Amazon S3

Accredian Publication
5 min readFeb 23, 2022


by Pronay Ghosh and Hiren Rupchandani

  • In the previous article, we covered a high-level overview of Amazon S3.
  • We understood what exactly S3 is, why should one choose S3, and the top 3 features of Amazon S3.
  • Here, in this article, we will learn about the working of Amazon S3.
  • After that, we will learn how to build, train and deploy a machine learning model with the help of Amazon Sagemaker.
  • So as we know, Amazon Simple Storage Service (S3) is an object storage-based storage service that is massively scalable.
  • It has a very high level of durability, as well as high availability and performance.
  • Data can be accessed from anywhere in the world using the Internet using the Amazon Console, and the powerful S3 API.

Attributes of Amazon S3

  • Buckets: Data can be organized into buckets. Unstructured data can be stored in an infinite number of buckets.
  • Elastic scalability: S3 does not have a storage limit. Individual objects can have a maximum size of 5TB.
  • Flexible data structure: Each object is identified by a unique key, and metadata can be used to organize data in a variety of ways.
  • Downloading data: Data can be easily shared with anyone inside or outside your organization, and data can be downloaded via the Internet.
  • Permissions: To ensure that only authorized users can access data, assign permissions at the bucket or object level.
  • APIs: The S3 API, has become an industry standard and is integrated into a wide range of existing tools.

How Does S3 Storage Work?

  • Objects are used to store data in Amazon S3.
  • This method enables highly scalable cloud storage.
  • Objects can be stored on a wide range of physical disc drives located throughout the data center.
  • To provide true elastic scalability, Amazon data centers employ specialized hardware, software, and distributed file systems.
  • Amazon uses block storage methods to provide redundancy and version control.

Data is automatically stored in multiple locations, across multiple discs, and, in some cases, across multiple availability zones or regions.

  • The Amazon S3 service checks the integrity of the data on a regular basis by examining the control hash value.
  • If there is data corruption, redundant data is used to restore the object.

How to use Amazon S3?

  • The first step would be to create an Amazon S3 account.
S3 Dashboard
  • The user can then build a bucket, add objects to the bucket, view objects, move objects, and delete objects and buckets.
  • Data is stored in Amazon S3 as objects in buckets.
  • An object is made up of a file and any metadata that pertains to that file.
  • To store an object in Amazon S3, the user must first upload the file to be stored in the bucket.
  • If we need to transmit a big amount of data, Amazon provides Import/Export, which allows us to upload and download data in S3.
  • A bucket with objects will be available on Amazon S3.
  • Because the bucket name is always a component of the URL, it should be unique across all Amazon accounts.
  • The concept of folders is also included in Amazon S3’s administration console.
  • A bucket cannot be contained within a bucket, although a folder can be contained within a bucket (grouping of multiple objects).
  • The public URL will be visible whenever a user submits an object.

The URL can be one of two types:

— (virtual-hosted-style) or

— (path style).

Typical behaviors marked in S3

  • When a process writes an item to Amazon S3 and then tries to read it all at the same time.
  • Amazon S3 may report “key does not exist” before the update is fully propagated.
  • A procedure creates a new object in Amazon S3 and lists the keys in its bucket right away.
  • The object may not appear in the list until the modification is fully propagated.
  • A process substitutes an existing object and tries to read it right away.
  • Amazon S3 may return the previous data until the update is fully propagated.
  • A procedure deletes an existing object and then tries to read it right away.
  • Amazon S3 may return removed data until the deletion is fully propagated.
  • A process deletes an existing object and lists the keys within its bucket immediately.
  • Amazon S3 may list the removed object until the deletion is fully propagated.

S3 Functionality

  • The user can create, read, and delete objects ranging in size from one byte to five terabytes.
  • There is no limit to the number of things that can be saved.
  • A developer-assigned key can also be used to obtain the objects.
  • The authentication measures are in place to keep data secure against unauthorized access.
  • Objects can be made private or public, and certain users can be granted rights.
  • Uses REST and SOAP (https) interfaces that are meant to interact with any Internet-based toolkit.
  • The BitTorrent Protocol is also supported.
  • Reduced Redundancy Storage (S3) provides less durability at a cheaper cost.


  • So far in this article, we covered a high-level overview of how does Amazon S3 works.
  • In the next article, we will learn about Amazon Sagemaker.
  • After that, we will learn how to build, train and deploy a machine learning model with the help of Amazon Sagemaker.

Follow us for more upcoming future articles related to Data Science, Machine Learning, and Artificial Intelligence.

Also, Do give us a Clap👏 if you find this article useful as your encouragement catalyzes inspiration for and helps to create more cool stuff like this.



Accredian Publication

One of India’s leading institutions providing world-class Data Science & AI programs for working professionals with a mission to groom Data leaders of tomorrow!