Practice Exam | AWS Certified Database Specialty
Preparing for AWS Certified Database Specialty DBS-C01? This is THE practice exam course to give you the winning edge.
These practice exams have been co-authored by Stephane Maarek and Abhishek Singh who bring their collective experience of passing 18 AWS Certifications to the table.
The tone and tenor of the questions mimic the real exam. Along with the detailed description and “exam alert” provided within the explanations, we have also extensively referenced AWS documentation to get you up to speed on all domain areas being tested for the DBS-C01 exam.
We want you to think of this course as the final pit-stop so that you can cross the winning line with absolute confidence and get AWS Certified! Trust our process, you are in good hands.
All questions have been written from scratch! And more questions are being added over time!
Quality speaks for itself...
A gaming company maintains a staging environment for its flagship application which uses a DynamoDB table to keep track of gaming history for the players. This data needs to be kept for only a week and then it can be deleted. The IT manager has noticed that the table has several months of data in the table. The company wants to implement a cost-effective solution to keep only the latest week's data in the table.
Which of the following solutions requires the MINIMUM development effort and on-going maintenance?
Add a new attribute in the table to track the expiration time and set up a Glue job to delete items that are more than a week old
Add a created_at attribute in the table and then use a cron job on EC2 instance to invoke a Python script daily. The script deletes items older than a week on the basis of this attribute
Add a created_at attribute in the table and then use a CloudWatch Events rule to invoke a Lambda function daily. The Lambda function deletes items older than a week on the basis of this attribute
Add a new attribute in the table to track the expiration time and enable time to live (TTL) on the table
What's your guess? Scroll below for the answer...
Add a new attribute in the table to track the expiration time and enable time to live (TTL) on the table"
You can use DynamoDB Time to Live (TTL) to determine when an item is no longer needed. TTL uses a per-item timestamp to determine items for expiration. Shortly after the date and time of the specified timestamp, DynamoDB deletes the item from your table without consuming any write throughput. After you enable TTL on a table, a per-partition scanner background process automatically and continuously evaluates the expiry status of items in the table.
< diagrams included >
"Add a new attribute in the table to track the expiration time and set up a Glue job to delete items that are more than a week old" - Although this solution is theoretically possible, but it would require a significant development effort to build and maintain a Glue job to trim the DynamoDB table. There is an additional cost of running the Glue job on a daily basis as well.
"Add a created_at attribute in the table and then use a cron job on EC2 instance to invoke a Python script daily. The script deletes items older than a week on the basis of this attribute" - Provisioning an EC2 instance and developing Python script involves significant development effort, maintenance and additional cost. So this option is not correct.
"Add a created_at attribute in the table and then use a CloudWatch Events rule to invoke a Lambda function daily. The Lambda function deletes items older than a week on the basis of this attribute" - Custom code needs to be written for the Lambda function to implement the functionality to trim the DynamoDB table. Lambda function execution is also charged separately. So this option is not correct.
< reference links included >
My name is Stephane Maarek, and I'll be your instructor in this practice exam. I teach about AWS certifications with my focus always on helping my students improve their professional proficiencies in AWS. I am also the author of some of the most highly-rated & best-selling courses on AWS Lambda, AWS CloudFormation & AWS EC2.
Throughout my career in designing and delivering these certifications and courses, I have already taught 1,000,000+ students and gotten 350,000+ reviews!
I'm delighted to welcome Abhishek Singh as my co-instructor for these practice exams!
Welcome to the best practice exams to help you prepare for your AWS Certified Database Specialty exam.
You can retake the exams as many times as you want
This is an original question bank
You get support from instructors if you have questions
Each question has a detailed explanation
Mobile-compatible with the Udemy app
30-days money-back guarantee if you're not satisfied
We hope that by now you're convinced!... And there are a lot more questions inside the course.
Happy learning and best of luck for your AWS Certified Database Specialty exam!
Who this course is for:
- Anyone preparing for the AWS Certified Database Specialty DBS-C01
Stephane is a solutions architect, consultant and software developer that has a particular interest in all things related to Big Data, Cloud & API. He's also a many-times best seller instructor on Udemy for his courses in Apache Kafka and AWS.
[See FAQ below to see in which order you can take my courses]
Stéphane is recognized as an AWS Hero and is an AWS Certified Solutions Architect Professional & AWS Certified DevOps Professional. He loves to teach people how to use the AWS properly, to get them ready for their AWS certifications, and most importantly for the real world.
He also loves Apache Kafka. He sits on the 2019 Program Committee organizing the Kafka Summit in New York, London and San Francisco. He is also an active member of the Apache Kafka community, authoring blogs on Medium and a guest blog for Confluent.
During his spare time he enjoys cooking, practicing yoga, surfing, watching TV shows, and traveling to awesome destinations!
FAQ: In which order should you learn?...
AWS Cloud: Start with AWS Certified Solutions Architect Associate, then move on to AWS Certified Developer Associate and then AWS Certified SysOps Administrator. Afterwards you can either do AWS Certified Solutions Architect Professional or AWS Certified DevOps Professional, or a specialty certification of your choosing. Once ready, you can learn AWS Lambda and AWS CloudFormation in depth, or do the AWS Big Data certification.
Apache Kafka: Start with Apache Kafka for Beginners, then you can learn Connect, Streams and Schema Registry if you're a developer, and Setup and Monitoring courses if you're an admin. Both tracks are needed to pass the Confluent Kafka certification.
gRPC: First do the protocol buffers course, then move on to gRPC Java or gRPC Golang course.
In order to share knowledge I am not able to respond to private messages on Udemy. If you're a student, please ask questions inside the course. Thanks for understanding!
Abhishek is an AWS veteran and has built successful SaaS and consumer solutions using AWS services since 2012. Over the course of his professional career, Abhishek has interviewed and mentored hundreds of candidates for entry-level and lateral positions for Cloud based IT solutions development. Abhishek is passionate about sharing his knowledge on AWS Cloud, Machine Learning and Big Data. He wants to help his fellow IT Professionals level-up their skills to ace the AWS Certifications and above all, get ready for the real world AWS ecosystem.
He is an AWS Certified Solutions Architect Professional, AWS Certified DevOps Engineer Professional, AWS Certified Machine Learning Specialist, AWS Certified Big Data Specialist and AWS Certified Database Specialist.
Overall, Abhishek has over 14 years of experience working on a diverse range of Enterprise Technologies based on ML, Big Data and Analytics. He runs a successful ML and Big Data Consultancy advocating solutions on AWS Cloud and has advised multiple clients in the US to architect and implement their ML and Big Data solutions using the AWS suite of services.