Exams > Amazon > AWS Certified Database - Specialty
AWS Certified Database - Specialty
Page 3 out of 27 pages Questions 21-30 out of 262 questions
Question#21

A company wants to automate the creation of secure test databases with random credentials to be stored safely for later use. The credentials should have sufficient information about each test database to initiate a connection and perform automated credential rotations. The credentials should not be logged or stored anywhere in an unencrypted form.
Which steps should a Database Specialist take to meet these requirements using an AWS CloudFormation template?

  • A. Create the database with the MasterUserName and MasterUserPassword properties set to the default values. Then, create the secret with the user name and password set to the same default values. Add a Secret Target Attachment resource with the SecretId and TargetId properties set to the Amazon Resource Names (ARNs) of the secret and the database. Finally, update the secret's password value with a randomly generated string set by the GenerateSecretString property.
  • B. Add a Mapping property from the database Amazon Resource Name (ARN) to the secret ARN. Then, create the secret with a chosen user name and a randomly generated password set by the GenerateSecretString property. Add the database with the MasterUserName and MasterUserPassword properties set to the user name of the secret.
  • C. Add a resource of type AWS::SecretsManager::Secret and specify the GenerateSecretString property. Then, define the database user name in the SecureStringTemplate template. Create a resource for the database and reference the secret string for the MasterUserName and MasterUserPassword properties. Then, add a resource of type AWS::SecretsManagerSecretTargetAttachment with the SecretId and TargetId properties set to the Amazon Resource Names (ARNs) of the secret and the database.
  • D. Create the secret with a chosen user name and a randomly generated password set by the GenerateSecretString property. Add an SecretTargetAttachment resource with the SecretId property set to the Amazon Resource Name (ARN) of the secret and the TargetId property set to a parameter value matching the desired database ARN. Then, create a database with the MasterUserName and MasterUserPassword properties set to the previously created values in the secret.
Discover Answer Hide Answer

C
Reference:
https://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/aws-resource-secretsmanager-secrettargetattachment.html

Question#22

A company is going to use an Amazon Aurora PostgreSQL DB cluster for an application backend. The DB cluster contains some tables with sensitive data. A
Database Specialist needs to control the access privileges at the table level.
How can the Database Specialist meet these requirements?

  • A. Use AWS IAM database authentication and restrict access to the tables using an IAM policy.
  • B. Configure the rules in a NACL to restrict outbound traffic from the Aurora DB cluster.
  • C. Execute GRANT and REVOKE commands that restrict access to the tables containing sensitive data.
  • D. Define access privileges to the tables containing sensitive data in the pg_hba.conf file.
Discover Answer Hide Answer

C
Reference:
https://aws.amazon.com/blogs/database/managing-postgresql-users-and-roles/

Question#23

A Database Specialist is working with a company to launch a new website built on Amazon Aurora with several Aurora Replicas. This new website will replace an on-premises website connected to a legacy relational database. Due to stability issues in the legacy database, the company would like to test the resiliency of
Aurora.
Which action can the Database Specialist take to test the resiliency of the Aurora DB cluster?

  • A. Stop the DB cluster and analyze how the website responds
  • B. Use Aurora fault injection to crash the master DB instance
  • C. Remove the DB cluster endpoint to simulate a master DB instance failure
  • D. Use Aurora Backtrack to crash the DB cluster
Discover Answer Hide Answer

B
Reference:
https://docs.aws.amazon.com/AmazonRDS/latest/AuroraUserGuide/AuroraMySQL.Managing.FaultInjectionQueries.html

Question#24

A company just migrated to Amazon Aurora PostgreSQL from an on-premises Oracle database. After the migration, the company discovered there is a period of time every day around 3:00 PM where the response time of the application is noticeably slower. The company has narrowed down the cause of this issue to the database and not the application.
Which set of steps should the Database Specialist take to most efficiently find the problematic PostgreSQL query?

  • A. Create an Amazon CloudWatch dashboard to show the number of connections, CPU usage, and disk space consumption. Watch these dashboards during the next slow period.
  • B. Launch an Amazon EC2 instance, and install and configure an open-source PostgreSQL monitoring tool that will run reports based on the output error logs.
  • C. Modify the logging database parameter to log all the queries related to locking in the database and then check the logs after the next slow period for this information.
  • D. Enable Amazon RDS Performance Insights on the PostgreSQL database. Use the metrics to identify any queries that are related to spikes in the graph during the next slow period.
Discover Answer Hide Answer

D

Question#25

A company has a web-based survey application that uses Amazon DynamoDB. During peak usage, when survey responses are being collected, a Database
Specialist sees the ProvisionedThroughputExceededException error.
What can the Database Specialist do to resolve this error? (Choose two.)

  • A. Change the table to use Amazon DynamoDB Streams
  • B. Purchase DynamoDB reserved capacity in the affected Region
  • C. Increase the write capacity units for the specific table
  • D. Change the table capacity mode to on-demand
  • E. Change the table type to throughput optimized
Discover Answer Hide Answer

CE
Reference:
https://forums.aws.amazon.com/thread.jspa?threadID=174315

Question#26

A company is running a two-tier ecommerce application in one AWS account. The web server is deployed using an Amazon RDS for MySQL Multi-AZ DB instance. A Developer mistakenly deleted the database in the production environment. The database has been restored, but this resulted in hours of downtime and lost revenue.
Which combination of changes in existing IAM policies should a Database Specialist make to prevent an error like this from happening in the future? (Choose three.)

  • A. Grant least privilege to groups, users, and roles
  • B. Allow all users to restore a database from a backup that will reduce the overall downtime to restore the database
  • C. Enable multi-factor authentication for sensitive operations to access sensitive resources and API operations
  • D. Use policy conditions to restrict access to selective IP addresses
  • E. Use AccessList Controls policy type to restrict users for database instance deletion
  • F. Enable AWS CloudTrail logging and Enhanced Monitoring
Discover Answer Hide Answer

ACD

Question#27

A company is building a new web platform where user requests trigger an AWS Lambda function that performs an insert into an Amazon Aurora MySQL DB cluster. Initial tests with less than 10 users on the new platform yielded successful execution and fast response times. However, upon more extensive tests with the actual target of 3,000 concurrent users, Lambda functions are unable to connect to the DB cluster and receive too many connections errors.
Which of the following will resolve this issue?

  • A. Edit the my.cnf file for the DB cluster to increase max_connections
  • B. Increase the instance size of the DB cluster
  • C. Change the DB cluster to Multi-AZ
  • D. Increase the number of Aurora Replicas
Discover Answer Hide Answer

B

Question#28

A company is developing a multi-tier web application hosted on AWS using Amazon Aurora as the database. The application needs to be deployed to production and other non-production environments. A Database Specialist needs to specify different MasterUsername and MasterUserPassword properties in the AWS
CloudFormation templates used for automated deployment. The CloudFormation templates are version controlled in the company's code repository. The company also needs to meet compliance requirement by routinely rotating its database master password for production.
What is most secure solution to store the master password?

  • A. Store the master password in a parameter file in each environment. Reference the environment-specific parameter file in the CloudFormation template.
  • B. Encrypt the master password using an AWS KMS key. Store the encrypted master password in the CloudFormation template.
  • C. Use the secretsmanager dynamic reference to retrieve the master password stored in AWS Secrets Manager and enable automatic rotation.
  • D. Use the ssm dynamic reference to retrieve the master password stored in the AWS Systems Manager Parameter Store and enable automatic rotation.
Discover Answer Hide Answer

C

Question#29

A company is writing a new survey application to be used with a weekly televised game show. The application will be available for 2 hours each week. The company expects to receive over 500,000 entries every week, with each survey asking 2-3 multiple choice questions of each user. A Database Specialist needs to select a platform that is highly scalable for a large number of concurrent writes to handle the anticipated volume.
Which AWS services should the Database Specialist consider? (Choose two.)

  • A. Amazon DynamoDB
  • B. Amazon Redshift
  • C. Amazon Neptune
  • D. Amazon Elasticsearch Service
  • E. Amazon ElastiCache
Discover Answer Hide Answer

AE

Question#30

A company has migrated a single MySQL database to Amazon Aurora. The production data is hosted in a DB cluster in VPC_PROD, and 12 testing environments are hosted in VPC_TEST using the same AWS account. Testing results in minimal changes to the test data. The Development team wants each environment refreshed nightly so each test database contains fresh production data every day.
Which migration approach will be the fastest and most cost-effective to implement?

  • A. Run the master in Amazon Aurora MySQL. Create 12 clones in VPC_TEST, and script the clones to be deleted and re-created nightly.
  • B. Run the master in Amazon Aurora MySQL. Take a nightly snapshot, and restore it into 12 databases in VPC_TEST using Aurora Serverless.
  • C. Run the master in Amazon Aurora MySQL. Create 12 Aurora Replicas in VPC_TEST, and script the replicas to be deleted and re-created nightly.
  • D. Run the master in Amazon Aurora MySQL using Aurora Serverless. Create 12 clones in VPC_TEST, and script the clones to be deleted and re-created nightly.
Discover Answer Hide Answer

A

chevron rightPrevious Nextchevron right