Summer Sale Special Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: v4s65

Data-Architect Exam Dumps - Salesforce Certified Data Architect (SP25)

Go to page:
Question # 41

Universals Containers’ system administrators have been complaining that they are not able to make changes to its users’ record, including moving them to new territories without getting “unable to lock row” errors. This is causing the system admins to spend hours updating user records every day.

What should the data architect do to prevent the error?

A.

Reduce number of users updated concurrently.

B.

Enable granular locking.

C.

Analyze Splunk query to spot offending records.

D.

Increase CPU for the Salesforce org.

Full Access
Question # 42

Universal Containers (UC) is migrating from an on-premise homegrown customer relationship management (CRM) system- During analysis, UC users highlight a pain point that there are multiple versions of many customers.

What should the data architect do for a successful migration to mitigate the pain point?

A.

Hire an intern manually de-duplicate the records after migrating to Salesforce.

B.

Migrate the data as is, and use Salesforce's de-duplicating feature.

C.

Have the users manually clean the data in the old system prior to migration.

D.

Store the data in a staging database, and de-duplicate identical records.

Full Access
Question # 43

Universal Containers (UC) has built a custom application on Salesforce to help track shipments around the world. A majority of the shipping records are stored on premise in an external data source. UC needs shipment details to be exposed to the custom application, and the data needs to be accessible in real time. The external data source is not OData enabled, and UC does not own a middleware tool.

Which Salesforce Connect procedure should a data architect use to ensure UC's requirements are met?

A.

Write an Apex class that makes a REST callout to the external API.

B.

Develop a process that calls an inviable web service method.

C.

Migrate the data to Heroku and register Postgres as a data source.

D.

Write a custom adapter with the Apex Connector Framework.

Full Access
Question # 44

Northern Trail Outfitters has these simple requirements for a data export process:

File format should be in CSV.

Process should be scheduled and run once per week.

The expert should be configurable through the Salesforce UI.

Which tool should a data architect leverage to accomplish these requirements?

A.

Bulk API

B.

Data export wizard

C.

Third-party ETL tool

D.

Data loader

Full Access
Question # 45

Northern Trail Outfitters needs to implement an archive solution for Salesforce data. This archive solution needs to help NTO do the following:

1. Remove outdated Information not required on a day-to-day basis.

2. Improve Salesforce performance.

Which solution should be used to meet these requirements?

A.

Identify a location to store archived data and use scheduled batch jobs to migrate and purge the aged data on a nightly basis,

B.

Identify a location to store archived data, and move data to the location using a time-based workflow.

C.

Use a formula field that shows true when a record reaches a defined age and use that field to run a report and export a report into SharePoint.

D.

Create a full copy sandbox, and use it as a source for retaining archived data.

Full Access
Question # 46

DreamHouse Realty has an integration that creates records in a Salesforce Custom Object. The Custom Object has a field marked as required on the page layout.

DreamHouse Realty has noticed that many of the records coming from the external system are missing data in this field.

The Architect needs to ensure this field always contains data coming from the source system.

Which two approaches should the Architect take? Choose 2 answers

A.

Set up a Validation Rule to prevent blank values.

B.

Create a Workflow to default a value into this field.

C.

Mark the field required in setup at the field level.

D.

Blame the customer's external system for bad data.

Full Access
Question # 47

Universal containers is implementing Salesforce lead management.  UC Procure lead data from multiple sources and would like to make sure lead data as company profile and location information. Which solution should a data architect recommend to make sure lead data has both profile and location information? Option

A.

Ask sales people to search for populating company profile and location data

B.

Run reports to identify records which does not have company profile and location data

C.

Leverage external data providers populate company profile and location data

D.

Export data out of Salesforce and send to another team to populate company profile and location data

Full Access
Question # 48

UC recently migrated 1 Billion customer related records from a legacy data store to Heroku Postgres. A subset of the data need to be synchronized with salesforce so that service agents are able to support customers directly within the service console. The remaining non- synchronized set of data will need to be accessed by salesforce at any point in time, but UC management is concerned about storage limitations.

What should a data architect recommend to meet these requirements with minimal effort?

A.

Virtualize the remaining set of data with salesforce connect and external objects.

B.

Use Heroku connect to bi-directional, sync all data between systems.

C.

As needed, make call outs into Heroku postgres and persist the data in salesforce.

D.

Migrate the data to big objects and leverage async SOQL with custom objects.

Full Access
Go to page: