Large Data Volumes Flashcards Preview

SFDC Data Architecture & Management > Large Data Volumes > Flashcards

Flashcards in Large Data Volumes Deck (21)
Loading flashcards...
1
Q

What technique can be used to reduce the amount of data in Salesforce?

A

By using Mashups:

Maintain large data sets in a different application, and then make that application available to Salesforce as needed.

2
Q

What are Mashups?

A

Mashups use Salesforce presentation to display Salesforce-hosted data and externally hosted data

3
Q

Provide two mashup designs supported by Salesforce

A
  1. External website

2. Callouts

4
Q

Name two advantages of Using Mashups

A
  1. Data is never stale

2. No proprietary method needs to be developed to integrate the two systems

5
Q

Name two disadvantages of Using Mashups

A
  1. Accessing data takes more time
  2. Functionality is reduced. For example, reporting and workflow do not work on the external data

In addition: Because of their real-time restrictions, mashups are limited to short interactions and small amounts of data

6
Q

What is the impact of soft deleted records?

A

While the data is soft deleted, it still affects database performance because the data is still resident, and deleted records have to be excluded from any queries.

7
Q

How long does data stay in the recycle bin?

A

For 15 days, or until the recycle bin grows to a specific size.

8
Q

How can you hard delete records?

A

Use the Bulk API’s hard delete function to delete large data volumes.

9
Q

What is the best practice when you want to improve the performance of loading data from the API?

A

Use the Salesforce Bulk API when you have more than a few hundred thousand records

10
Q

What is the best practice when you want to delete large volumes of data

A

When deleting large volumes of data, a process that involves deleting one million or more records, use the hard delete option of the Bulk API.

Deleting large volumes of data might take significant time due to the complexity of the deletion process

11
Q

What is the best practice when you want to make the data deletion process more efficient

A

When deleting records that have many children, delete the children first

12
Q

What is the best practice when you want to avoid sharing computations

A

Avoid having any user own more than 10,000 records

13
Q

What is the best practice when you want to improve performance when you have a large amount of data

A

Use a data-tiering strategy that spreads data across multiple objects and brings in data on demand from another object or external store

14
Q

What is the best practice when you want to reduce the time it takes to create full copies of production sandboxes with large data volumes

A

When creating copies of production sandboxes, exclude field history if it isn’t required, and don’t change a lot of data until the sandbox copy is created

15
Q

Provide a solution for the following situation:

The customer designed a custom integration to synchronize Salesforce data with external customer applications.

The integration process involved:

  • Querying Salesforce for all data in a given object
  • Loading this data into the external systems
  • Querying Salesforce again to get IDs of all the data so the integration process could determine what data has been deleted from Salesforce

The objects contained several million records. The integration also used a specific API user that was part of the sharing hierarchy to limit the records retrieved. The queries were taking minutes to complete

A

The solution was to give the query access to all the data and then to use selective filters to get the appropriate records.

For example, using an administrator as the API user would have provided access to all of the data and prevented sharing from being considered in the query.

An additional solution would have been to create a delta extraction, lowering the volume of data that needed to be processed.

16
Q

If you have a shopping portal where about 700K folk are trading, 2M registered consumers and roughly 100K orders per day. There are 1K support reps and 100 support managers that handles roughly 10K cases every day.

What kind of licenses would you recommend?

What is a possible data model for the above consumers & traders?

A

Knowledge Feature license for all knowledge authors and contributors

Community Cloud License for 2M registered users and 700K traders

Service Cloud License for 1K support reps and 100 support managers

Consumers modeled as Person accounts with High-Volume consumer user license
Traders uses a Business Account with trader users as contacts. Also use a High-Customer Community License.

Use sharing sets to efficiently share

  • Account-level orders and cases to traders
  • Users’ purchase history and cases to consumers

Have different profiles between traders and consumers

Archive cases and orders older than 3 years.

17
Q

Scenario: You have 10 million subscribers. Account, Contact and Case data are stored in Salesforce, Policy and Policy related parties data are stored outside of Salesforce, and disbursement details are available in an existing web application.

What solution / data model would work for them (including a way to see case trends)

A
  1. Continue to use Account, Contact and Case data in SFDC
  2. Have Policy and Policy Related Parties as external objects (linked to contacts within Salesforce)
  3. Use Canvas to show the Disbursement data within Salesforce
  4. Generate a case trend snapshot as follows
    - Create a new custom report that includes the fields to load as records into a target object
    - Create the new custom object in which to store the records
    - Create fields on the target object that will receive the source report’s results when the reporting snapshot runs
18
Q

For reporting snapshots, what is the max number of new records that can be added to a target object? What happens with the remaining ones?

A

2000 records
If the source report generates more than 2K, an error is displayed for the additional records in the Row Failures related list. This can be accessed via the Run History section of a reporting snapshot detail page.

19
Q

What is an ideal scenario for using external objects?

A

When you have a large amount of data that you don’t want to store in your org, and you only need to use a small amount of that data at any one time

20
Q

Which of these is true when it comes to truncating records?
- It’s not as fast as batch-deleting records and re-creating the object

  • All of the object’s records are removed permanently, but the object’s definition remains
  • You can truncate standard objects or custom objects that are on the master side of a master-detail relationship
  • You can truncate custom objects when your org has reached its limit on allowed custom objects
A

Al of the object’s records are removed permanently, but the object’s definition remains

21
Q

What are the main areas of the application that are impacted by differing (or sub optimal architectures in implementations with large data volumes?

A
  1. The loading or updating of large number of records, either directly or with integrations
  2. Extracting records using reports, list views , or queries