You are architecting an accessibility application using AWS textract and Poly , as part of the process you are creating MP3 files and a text file (you should keep them unless users download it )for each Adobe form that is coming as an input. What kind of data storage you will use for the best possible solution?
1. Keep the Mp3 files in EBS volume and text files are in S3 as S3 is used to store only text files
2. Keep both the files in S3
3. Keep both the files in EBS and make sure you take perioding snap shot of it
4. Use dynamodb to store the textfiles and S3 to store in EBS as this will give you the maximum scaling
Explanation
Correct answer Keep both the files in S3
S3 supports storing of any files and it scales infinitely and has the durability of highest standard and is much cheaper than EBS
Storing textfiles in dynamodb is much costlier and is not the best practice
You are building an application which has potentials for thousands of users and you have decided on using S3 for storing the files (sizes in Megabytes ) which needs to be served to the users. You are not sure which documents and how frequently it will be used . What will be the storage class that you will use in this case for best cost and availability
1. S3 Standard for general-purpose storage of frequently accessed data
2. S3 Intelligent-Tiering for data with unknown or changing access patterns
3. Amazon S3 Glacier Deep Archive for long-term archive and digital preservation
4. S3 One Zone-IAExplanation
Explanation
Correct answer S3 Intelligent-Tiering
The last sentence of the question is important it tells best cost and availability
S3 Intelligent-Tiering charges a small tiering fee and has a minimum eligible object size of 128KB for auto-tiering. Smaller objects may be stored but will always be charged at the Frequent Access tier rates. See the Amazon S3 Pricing for more information
standard will cost more as we know the size is of MBs if the document are accessed frequently and infrequent access
Use glacier only when you know it is used for archival process
Because S3 One Zone-IA stores data in a single AWS Availability Zone, data stored in this storage class will be lost in the event of Availability Zone destruction.
https://aws.amazon.com/s3/storage-classes/
Your application sees certain spike when it serves document to the different users .Few users suddenly starts to download big files using torrent , while suddenly many concurrent users come in while some promotion is there what kind of architecture will you be using for storing the data for best solution
1. Use EFS and mount that across multiple EC2 instance this will be able to handle sudden spike
2. Use compute optimized instance types and EBS
3. Use S3
4. Use memory optimized instance
Explanation
Correct answer S3 can handel multi thread and almost infinite load and sudden spike
EFS can definitely be the option but it is much expensive but This distributed data storage design means that multithreaded applications and applications that concurrently access data from multiple Amazon EC2 instances can drive substantial levels of aggregate throughput and IOPS. Big data and analytics workloads, media processing workflows, content management, and web serving are examples of these applications.
Memory and compute optimized solution does not fit this requirement
You have got a version enabled bucket and assume that you have 31 day months . On the first day you have created an object of size 4 gb and on the 16 th day you perform a put object of 6 gb using the same key in the same bucket . What will be the charge for this bucket
1. 4 gb X 31 days
2. 6 gb X 15 days
3. 6 gb X 15 days + 4 gb X 31 days
4. 4 gb X 15 days
Explanation
Correct answer 6 gb X 15 days + 4 gb X 31 days
When analyzing the storage costs of the above operations, please note that the 4 GB object from Day 1 is not deleted from the bucket when the 6 GB object is written on Day 16. Instead, the 4 GB object is preserved as an older version and the 6 GB object becomes the most recently written version of the object within your bucket.
https://aws.amazon.com/s3/faqs/#s3ta
you as an AWS architect need to design a solution where you manage multiple aws account .In one of the account there is an S3 bucket which holds large set of web crawling data which is accessed by other account users in a regular basis. You need to make sure each account should be charged as much they use the bucket .What should be your solution
1 Configure requester pay by using aws console
2. Nothing is required to be done be default aws will charge the account which uses most
3. All account will be charged equally and you cannot change it
4. you need to turn on aws logs for that bucket and find out which account used how much data is consumed by each account
Explanation
Configure requester pay by using aws console
A bucket owner can configure for requester pay back and aws will charge the different account if other accounts access the bucket
https://docs.aws.amazon.com/AmazonS3/latest/dev/RequesterPaysBuckets.html
As part of the latest security requirement you need to secure the data stored in S3 at rest what will be the possible solution [Choose 2]
1. By default all data in S3 are encrypted
2. Use SSE-S3
3. Use a third party encryption library
4. You can securely upload/download your data to Amazon S3 via SSL endpoints using the HTTPS protocol.
Explanation
Use SSE-S3 or third party encryption library
S3 data is not encrypted by default HTTPS protocol is not going to secure the data at rest
< View All polls