Wednesday, 9 June 2021

AEM As Cloud deployment fails with dispatcher image issue

During my initial POCs with sample WKND code, I have noticed a behavior on AEM As Cloud service.

When ever I go for the pipeline and trigger it, the deployment fails with reason - dispatcher image failed. Some analysis was done around it from the logs and was able to resolve it as below.

I was doing my code checkout on windows and while unzipping I was using a software which was breaking the symlinks within dispatcher package. I got some warning during the unzip which I did not give much attention that point of time

To debug this, first I have commented out the dispatcher section in parent pom and deployed and tested the build. It worked on cloud.



Next I used another unzip tool (I remember Winzip was the tool which helped me. 7 Zip was removing symlinks. If it doesn't work, try other way around).

This time the symlink warning did not come durng my unzip and I have redeployed the code again. It worked perfectly.

Another option to fix this, may be, we can change those symlinks to real links if you know how to do that.

As a final option, still if you are unable to find out the reason, my recommendation is to reach out to spp help of Adobe, so that they can verify whether the envirnment was setup correctly.

 

Video Tutorial on AEM Cloud Deployment 




Friday, 4 June 2021

Site Template in AEM FAQ

 Site Template in AEM FAQ

What is Site template (known as Sites 30)?

Site template is a predefined template which can be the starting point of any website creation. It contains basic theming, page templates, configurations and sample content. Instead of going to the archetype way of creating step by step approach of developing a website, Site template helps to utilize the out of the box components and templates and setup a website in shorter period of time. Once the site is up, the developers parallely work towards achieving the brand styles and customization (utilizing CSS, JS)

This is the new recommended Adobe way of developing an AEM site

Can I get it installed on my local AEM instance?
As of now the Site Template is enabled as part of AEM as A Cloud Service ONLY. We will have to contact Adobe to get this feature enabled in our AEM AS CLOUD.

Where I can get the templates?
This is an open-sourced project like Core Components. It can be downloaded from Basic Site Template project on GitHub, but always ensure you use the latest release.
https://github.com/adobe/aem-site-template-basic

What is themes in Sites?

Themes used to depict styles. We can modify theme sources of an Adobe Experience Manager Site to apply brand specific styles.

Where I can find more details?

We can always get demo and tutorials on below link.
https://experienceleague.adobe.com/docs/experience-manager-learn/getting-started-wknd-tutorial-develop/site-template/overview.html?lang=en

Wednesday, 31 March 2021

Crx2Oak Migration tool Concepts and Demo

 Crx2Oak helps migrate data from older CQ versions based on Apache Jackrabbit 2 to Oak, and it can also be used to copy data between Oak repositories.


 

You may also read: AEM As A Cloud Video Tutorials

The tool can be used for:

  • Migrating from older CQ 5 versions to AEM 6
  • Copying data between multiple Oak repositories
  • Converting data between different Oak MicroKernel implementations. (S3DataStore to FileDataStore)


Some of its features are

  • The migration can be interrupted at any time, with the possibility to resume it afterwards.
  • Custom Java logic can also be implemented using CommitHooks.
  • CRX2Oak also supports memory mapped operations by default. Memory mapping greatly improves performance


Parameters
Node Store Options
--cache: Cache size in MB (default is 256)

--mmap: Enable memory mapped file access for Segment Store

--src-password: Password for the source RDB database

--src-user: User for the source RDB

--user: User for the targed RDB

--password: Password for the target RDB.

Version Store Options
--copy-orphaned-versions: Skips copying orphaned versions. Parameters supported are: true, false and yyyy-mm-dd. Defaults to true.

--copy-versions: Copies the version storage. Parameters: true, false, yyyy-mm-dd. Defaults to true.


Path Options
--include-paths: Comma-separated list of paths to include during copy
--merge-paths: Comma-separated list of paths to merge during copy
--exclude-paths: Comma-separated list of paths to exclude during copy.


Source Blob Store Options
--src-datastore: The datastore directory to be used as a source FileDataStore

--src-fileblobstore: The datastore directory to be used as a source FileBlobStore

--src-s3datastore: The datastore directory to be used for the source S3DataStore

--src-s3config: The configuration file for the source S3DataStore.


Destination BlobStore Options
--datastore: The datastore directory to be used as a target FileDataStore

--fileblobstore: The datastore directory to be used as a target FileBlobStore

--s3datastore: The datastore directory to be used for the target S3DataStore

--s3config: The configuration file for the target S3DataStore.

Help Options
-?, -h, --help: Shows help information.

Debugging
--log-level TRACE or --log-level DEBUG 

Demo Video Series

What are the option for asset migration between AEM instances?


There are cases where we need to move assets across AEM instances. This may occur when we have multiple asset repositories and as part of merging assets we may have to move them into one single instance. While moving assets we have to consider various factors like asset metadata, asset versions, Tags w.r.t assets all moved and validated.

The overall size of asset is another consideration while selecting any tools to migrate assets. Below given few tools which can be included in our consideration while we go for asset migration.


-Replication Agent
By configuring replication agent, we can migrate assets across AEM Instances

-Vault Remote Copy

Jackrabbit vault offers a simple method to copy nodes between repositories.

This can be used for bulk assets.

References:
https://jackrabbit.apache.org/filevault/rcp.html
https://experienceleague.adobe.com/docs/experience-manager-65/assets/administer/assets-migration-guide.html?lang=en#migrating-between-aem-instances
License: Jackrabbit Oak and any of its parts are licensed according to the terms listed in the Apache License, Version 2.0

-Grabbit

Grabbit provides a fast and reliable way of copying content from one Sling (specifically Adobe AEM) instance to another.
Grabbit creates a stream of data using Google’s Protocol Buffers aka "ProtoBuf". Protocol Buffers are an extremely efficient (in terms of CPU, memory and wire size) binary protocol that includes compression.

This is one of the Adobe recommended solution & can be considered for bulk assets movement.

Ref:
https://github.com/TWCable/grabbit
https://relevancelab.com/2019/07/04/get-moving-with-aem-and-grabbit/
License: Licensed under the Apache License, Version 2.0 https://github.com/TWCable/grabbit/blob/master/docs/LicenseInfo.adoc


-Recap

Crx sync option based on vlt rcp

Ref:
http://adamcin.net/net.adamcin.recap/

-Crx2Oak

Crx sync option  - tools provided by Adobe while upgrading AEM or for migration of crx data.
This is one of the Adobe recommended & can be used for bulk Assets

-S3 Asset Ingestor

This is part of ACS AEM Commons .It pulls files from an Amazon S3 bucket instead of the local file system. You can load a directory of assets into AEM very easily with this tool. Because of the ability to overload a server with assets, this tool only appears for the “admin” user right now.

Ref:
https://adobe-consulting-services.github.io/acs-aem-commons/features/mcp-tools/asset-ingestion/s3-asset-ingestor/index.html





What is AEP? Adobe Experience Platform FAQ



What is AEP(Adobe Experience Platform)
AEP helps to capture the customer journey. In AEP, data from various sources are stitched together using a schema. Thus an identity graph is built which is unique to a customer. AEP has a data lake where data from various sources are streamlined and fed into. This will be used to create profile data for a customer.




AEP is basically a combination of Real time customer profile + AI & machine Learning + Open Ecosystem

What are all the various data sets defined in AEP?
AEP has various data sets like,
Attributes: Characters like customer name, email, gender etc.
Identities: Unique identity info like ECID(experience cloud id), Email, membership id, phone no etc.
Segments : Categories like  online shoppers, gender, Location. [one such use case is, these segments can be exported to utilize in an email campaign]
Behaviors: Like login to the website, installed appication, added an item to cart etc.

What AEP Solves?
AEP solves below concerns.

  1. Disconnected identities.
  2. Slow and vulnerable data transfer.
  3. AI & ML operates in silos usually. Extraction of data is tough in such cases.
  4. Data governance is not enforced usually(CCPA, GDPR etc)
  5. Centralization of multiple features.


Capabilities or major AEP Features:

  • Create Actionable, real time intelligent customer profiles.
  • Enrich data and derive more insights with AI & ML & Data queries(SQL).
  • Innovate with open & composable components (Open APIS etc.)
  • Enhance delivery
  • Privacy and data protection(Privacy framework, consent offering, security )


AEP integration into other Adobe Cloud Applications
Adobe Experince cloud applications (Marketing cloud, Analytics cloud, Advertising cloud, commerce cloud) can be easily integrated to  AEP.

All customer attributes are fed into AEP from different applications.
For eg.
Adobe Analytics send data(when ever a data point is captured, immediately it goes into AEP), Adobe Target send data(decision made, content presented kind of data), Audience (send trace and audience), Campaign(profile and event data) can be easily fed into AEP via launch.

AEP uses Launch & websdk to import data directly into AEP from various applications.

How AEM or Forms utilizes AEP?
AEM can use this AEP data to personalize content on pages or forms.

Various AEP Implementation Phases & Roles responsible for.

Plan - (Leads and an enterprise architect will plan referring business goals and document it by defining KPIs)
Implement - (Data architects and engineers create data lakes(create model, schemas) and make available), ensure data integrity by query services.
Use - Marketrs, data analysts, data scientists(uses query service), application developer interacts with UI & start working towards integrating with other adobe applications(campaign,target, analytics)
Grow - People highlights the growth to the initial set of team (Enterprise architects, company lead etc)

Basic architecture of AEP

Through data ingestion (either third party ETL, ERP, Sales or Adobe applications via launch), we can ingest data into AEP data lake. The data resides as batches and files. Any data getting pushed also placed in Experience platform pipeline. Any data gets into AEP traverse to identity graph and profile store quickly.

The controls native to AEP are
Access control: specific permission rights to data & users
Data governance: to ensure data integrity
Experience data model systems: common data model, which cn be extended based on needs
Query service: SQL way of accessing data - it has connectors, so other sql tools can connect to this query service
Data science workspace: allows data scientists to create data models build train and deploy.
Intelligent services: like Attribution AI or customer AI - prebuild models can be configured to operate on data
Segmentation capabilities: for categorization. it includes streams and batches

Application Services
-Customer journey analytics - Combine all data from every channel. It has analysis workspace/ interface on top of AEP, helps visualize and explore all data from data lake.
-Real time customer data platform(CDP) - rich real time customer profiles, actionable insights, data governance. CDP has segmentation capabilities.
-Journey orchestration -  enables orchestration of triggered interactions like registration confirmations, location based information
-Intelligent Services -  Utilizing the AI & ML Capabilities to intelligently predict customer behavior.
-Offer decisioning - Build offer, apply decisioning & then deploy the right offer.

Use cases of AEP

1) Real time customer data platform - Stitch known and unknown data to activate customer profiles
2) Customer journey intelligence - Utilize the data driven methodologies, best practices AI & ML to enable real time decisions and actions
3) Delivery and cross channel experience - capability to deliver consistent and personalized experiences across all channels with the combination of platform and experience cloud products.
4) Customer experience application development - AEP gives an open and extensible platform for low latency access to profiles decisions and insights to create new customer experience applications.


Saturday, 2 January 2021

All you need to know about Adobe Client Data Layer

What is Adobe Client Data Layer

A data layer in general consists of a JavaScript client side event driven data store that can be used on webpages, to collect data about what the visitors experience on the web page & to communicate this data to digital analytics and reporting servers(e.g. Adobe analytics or Adobe target)

What does it do?

The Adobe client data layer is a java script store for data and events happening on a page within the scope of a request. It reduce the effort to instrument websites by providing a standardized method to expose and access any kind of data for any script. 

It provides API to,

1) Register data that is to be merged into the data layer state
2) Trigger events that relate to the data stored in the data layer
3) Get the current data layer state of all merged data
4) Register listeners that are called for specific events or data changes

Steps to set up a data layer

1) Loading the data layer script
2) Declare the adobeDatalayer array

Once above steps has been configred, we can work on various push menthods(  Push the Data object/ Push Event Object/Push functions) Events (registering & unregistering)

The AEM Core Components are availed with Adobe Client Data Layer(Disabled by default - we need to enable it if we have plans to use it).

To push the data from website to Analytics, we need Adobe Client Data Layer extension & Adobe Analytics, Core Extensions to be configured.

Some of the use cases
1) Retrieving Analytics data & using it for Personalization
2) Trigger an update event on page when the stock market value changes

 Reference URLs:

 Adobe client data layer - https://github.com/adobe/adobe-client-data-layer

Wiki - https://github.com/adobe/adobe-client-data-layer/wiki

All you need to know about Project Firefly

 What is Project Firefly? 

Project Firefly is a run time framework for building 3rd party cloud native applications that extend the functionality of Adobe experience Platform(AEP) and Adobe experience cloud.

It provides everything we need to develop an application, even this is extendable - which grows with the needs.

What it contains?

  • Adobe I/O runtime - which is a server-less foundation for running 3rd party custom code on Adobe infra. It provides scaling in /out etc.
  • CLI & SDK - Enables local development, CI/CD. Streamlined way for developers to interact with core Adobe services and automated process.
  • Spectrum(Adobes design language) UI Framework - A React based UI framework for creating experiences that feels native comfort
  • Custom events - Publish and consume custom events with support for webhooks and journaling
  • Cloud services - a range of services for running managing and optimizing custom digital experiences.(Cloud storage, blob storage,CDN etc)
  • Set of Developer tools - Has UX modeling tools, IDE plugins(code completion) and other tools to aid in testing, debugging  and deploying custom experiences.


Whom it will be helpful?

  • System integration developers - who are typically specialized on integrating and extending Adobe enterprise solutions(AEM, Campaign, Marketo, Magento etc)
  • Enterprise developers - who works with enterprise customers to create business sue case demos etc.


What is the difference between I/O Runtime & project Firefly

Project Firefly is a complete app framework to build custom cloud native Adobe Apps, where in Adobe I/O runtime is a server-less platform for running custom code.

Project Firefly is built on top of Adobe I/O Runtime.

What are all the features of Project Firefly?

  • Storage services - We get all relevant storage services to work with a 3rd party app
  • Debugging - Provides various debugging options
  • Logging - Provides evident logging mechanisms
  • Action templates - Project starter templates
  • UI templates - React spectrum templates to help developers
  • Security - Firefly is highly secured


Some of the typical use cases of Firefly
1) Asset migration in AEM(External DAM to AEM) which are in Gigabytes(bulk uploads) - the task can be offloaded to a headless firefly app that uses content fragment API, so  that normal AEM tasks run without issues.

2) Offloading analytics data to Firefly - Firefly can pull the data from Adobe analytics and save it into DB or make it available for Target

3) Data ingestion monitoring - There are cases where we need to import Huge data into AEP and there could be errors in the process. We can use Firefly to monitor the Adobe i/O events and trigger the other system when something goes wrong to take corrective actions

4) Campaign stand dashboard for monitoring- unlike campaign classic, campaign standard doesnt provide a dashboard to monitor workflows. Using Firefly, we can create SPA that display the status of all workflows in Campaign standard - which helps marketers