The Client identifies that a segment of customers need to see some products on the site that other customers should not be able to access. All products are maintained within one catalog but in separate categories. A custom attribute will be used on the Profile system object to identify customers that belong to this special segment. A customer group will be made that is qualified for by this Profile custom attribute. The storefront will be customized to include navigation to relevant categories for this customer group.
Unfortunately during technical review the Client points out that the business teams have raised a concern with maintenance and want to use a shared navigation within the catalog and not use separate categories.
Which item should the Architect suggest to efficiently fulfil this new requirement while maintaining scalability?
The optimal solution for this requirement is to modify the search results dynamically based on the user's customer group membership. This approach:
Ensures that all users can navigate the same catalog and categories without seeing separate categories for special segments.
Dynamically includes or excludes products from search results based on the user's membership in the special segment, effectively using existing catalog structures while personalizing product visibility.
Maintains scalability by leveraging existing catalog and category infrastructure without needing additional custom attributes for navigation purposes.
This method aligns with best practices for creating personalized customer experiences in B2C Commerce without complicating catalog management.
A client receives multiple feeds from third parties on the same SFTP location:
* Product prices (sftp: prod/prices)
* Stores information (sftp: prod/stores;
* Product information (sftp: prod/catalog)
* Categories information (sftp: prod/marketing)
* Content (sftp: prod/marketing)
Some of the feeds are placed on sftp multiple times a day, as the information is updated in the source system.
The Architect decides to have only two jobs:
* One that checks and downloads available feeds every hour
* One that imports the files from Webdav once a day before the data replication, using the standards steps available in the Job Framework
Which design is correctfor the import Job, taking the steps scope in consideration?
This design maximizes efficiency and concurrency. By having the jobs that import products, stores, prices, and content run in parallel, the system can handle multiple data streams simultaneously, reducing total processing time. The sequential execution of importing categories followed by reindexing ensures that all new and updated information is properly indexed and available for site use, following the completion of the import of more frequently updated data. This order respects dependencies between steps and aligns with best practices for handling complex data workflows in B2C Commerce environments.
During a review of the most recent release notes, the Architect finds that Salesforce has deprecated an API that is used throughout the site. After reviewing the deprecated API usage in Business Manager, the Architect narrows down the usage of that API to a particular LINK integration cartridge. The cartridge was integrated when the site was first launched and is heavily customized for the Client.
What is the recommended way for the Architect to remove the deprecated API so the LINK integration continues to work without interruptions, and lowest level of effort'
When facing a deprecated API that is used in a LINK integration cartridge, the recommended approach is to check for an updated version of the cartridge that may have replaced the deprecated API with a supported one. If an update is available:
Integrate the updated cartridge into the site, ensuring compatibility with the current site configuration.
Re-apply customizations that were made to the original cartridge to maintain functional consistency.
Thorough testing should be conducted to ensure that the integration works seamlessly without causing disruptions in the site's functionality.
This approach minimizes effort by leveraging updates provided by the cartridge vendor while ensuring the site remains functional and compliant with current API standards.
A company that is a shoe-producer is doing Salesforce B2C Commerce implementation. In their Enterprise Resource Warning (ERP) system, the products are marked as being one of three types: boots, sandals, and sneakers. The business requirements based on the type are:
* The messaging on Product Detail page is different
* Customers are able to filler their Product Search Results
The customer's operations team asks about the format in which to send this value in the catalog.
Which data type should the Architect specify for this attribute In the Data Mapping document?
For the shoe producer implementing Salesforce B2C Commerce, the attribute used to specify the type of shoes (boots, sandals, sneakers) should be set as a single selectable value, hence the use of an enum-of-string type. This data type allows for the selection of one option from a predefined list, which is ideal for cases where each product can only be categorized into one type. This ensures that data integrity is maintained while providing clear and distinct categorization that can easily be used for filtering on the Product Search Results page and for customized messaging on the Product Detail page.
A Client has automated builds that deploy the code; however, recent builds started failing with compilation issues, which are not reproducible in developer's environment.
Whetfirst step needs to be taken to identify and fix the issue?
When facing build failures that are not reproducible in developers' environments, a common and effective first step is to clean up the build server workspace (Answer A). This action clears any residual files, settings, or data that might be influencing the build process and causing discrepancies between local and server builds. By resetting the workspace to a clean state, it helps ensure that the build process is starting from a consistent base, potentially resolving issues caused by stale data or corrupted build artifacts. This step often resolves hidden configuration or environment-specific issues that are not immediately apparent in the code or module versions.
Stefania
13 days agoMoira
2 months agoAhmed
3 months agoLoren
4 months agoHollis
5 months agoShonda
5 months agoJules
5 months agoKyoko
6 months agoCrista
6 months agoVeronika
6 months agoMicheal
7 months agoLettie
7 months agoMartin
7 months agoNu
8 months agoBrigette
8 months agoKayleigh
9 months agoKiera
9 months agoEmily
9 months agoRaul
9 months agoAdell
10 months agoSolange
10 months agoFranchesca
10 months agoTwana
11 months agoTamekia
11 months agoTheron
12 months agoKyoko
1 years agoShawna
1 years agoEileen
1 years ago