DronaBlog

Showing posts with label Master Data Management. Show all posts
Showing posts with label Master Data Management. Show all posts

Monday, December 6, 2021

What are the phases in BES External call in Informatica MDM ?

                 Are you looking for the details about various phases in BES External calls? Are you also interested in knowing what are categories under which these phases come? If so, then you reached the right place. In this article, we will learn about BES  External calls and their phases.

A) What BES External calls?

                BES External calls are the web service configured in the provisioning tool in order to achieve customization in Entity 360 or Customer 360 application. we need to develop a custom web service and deploy it on the application server. Once it is deployed, we can register the endpoint URL in the provisioning tool in Informatica MDM.





B) Categories of phases of BES External calls? 

                 The phases of BES External calls are categorized as -

                      a) Read calls

                     b) Write calls 

                     c) Merge calls  

               a)  Read calls: This category is used for read and search operations.

              b)  Write calls: This category uses write co for insert, update and delete operations.

              c)  Merge calls: This category uses preview merge co which is designed for merge-operations.

C)  Phases of External calls 

                   Let's understand detailed phases for the above categories 

              i) Read call phases : 

                        The phases for read calls are

                      * Read Co.Before Everything 

                      * Read Co.After Everything 

                      * Read view.Before Everything

                      * Read view.After Everything






             ii) Write call Phases

                      * Write Co.Before Everything

                      * Write Co.Before Validate

                      * Write Co. After Validate

                      * Write Co.After Everything  

                      * Writeview. Before Everything 

                      * Writeview. Before Vadidate

                      * Writeview. After Vadidate

                      * Writeview. After Everything


               iii) Merge call phases

                      * Perview Merge Co. Before Everything

                      * Preview Merge Co.After Everything

                      * Merge Co. Before Everything

                      * Merge Co. After Everything


                 Learn more about Informatica here 



            

                   

                

Tuesday, November 30, 2021

What is Dynamic Data Masking?

                        Are you looking for details about Dynamic Data Masking? Are you also interested in knowing what are the things we need to consider for implementing Dynamic Data Masking also known as DDM? If so, then you reached the right place. In this article, we will explore various aspects of Dynamic Data Masking.

A) What is Dynamic Data Masking (DDM)? 

                       Dynamic Data Masking is a technology using which we can mask production data in real-time. Dynamic Data Masking also called DDM does not change data physically. DDM  just changes the data stream in order to mask the sensitive data when the requestor request such information.





B) What are the Dynamic Data Masking tools?

                       Various vendors provide Dynamic Data Masking functionalities and these are

                 1. Microsoft Azure SQL Database 

                 2. Oracle Enterprise Manager 

                 3. iMask 

                 4. Informatica Dynamic Data Masking

                 5. Imperva Data Masking 

                 6. Infosphere Optim Data Privacy by IBM

                 7. K2 view Data Masking 

                 8. Mentis

C) What are data masking Rules?

                        The rules contain various conditions and actions that rule engines use in order to process the request.

              e.g. 

                     1) Connection rules : It process application connection requests.

                     2) Security rules :  It process SQL statements.

                      Here are important points about rules 

                     a) We can define and create rules in order to process SQL requests that are executed by an application against the database.

                    b) DDM rule uses two techniques i.e connection criteria and masking techniques.

                    c) In order to forward the requests the rule Engine uses connection criteria.

                    d) In order to mask the data the masking technique is used.





D) What are DDM rule components? 

                         The DDM  rule components are as below

                    a) Matcher: It defines the criteria for the rule engine to identify the match.

                   b) Action: It defines action which will be applied by the rule engine to request.

                  c) Processing Action: The rule engine applies specific action to the request after applying the rule.


              Learn more about Dynamic Data Masking here 



 

                



Friday, November 19, 2021

What are the Batch Processes in Informatica MDM?

Are you looking for an article that explains various processes in Informatica Master Data Management (MDM)? If so, then you reached the right place. In this article, we will learn about various processes through which records are loaded to the MDM system. Let's start.


Informatica MDM contains the various processes and those are -


Step 1: The land process transfers data from a source system via ETL jobs to landing tables in the MDM ORS (Operational Reference Store).

Step 2: The stage process (Stage Job) reads the data from the landing table, cleanses the data if applicable, and moves the cleansed data into a staging table via mapping in HUB Console.

Step 3: The load process (Load Job) loads data from the staging table into the corresponding base object in MDM ORS.

Step 4: The tokenize process (Tokenization Job) generates match tokens based on match columns that are used subsequently by the match process to identify candidate base object records for matching.

Step 5: The match process (Match Job) compares two records for points of similarity. If sufficient points of similarity are found to indicate that the two records are probably duplicates of each other, then Informatica MDM Hub flags those records for consolidation.





Step 6: The consolidate process (Merge Job) merges duplicate records into a single record after duplicate records have been identified in the match process.

Step 7: Publish or distribution process is the main outbound flow for Informatica MDM Hub. Hub integrates with external systems or DB Schemas to share the consolidated (Golden) Records.


Learn more about Informatica MDM here,





Friday, October 22, 2021

What are differences between multimerge and merge API in Informatica MDM

                Are you interested in knowing what is the use of multimerge and merge APIs? Are you also would like to know the difference between merge and multimerge API? If so, then you reached the right place. In this article, we will learn about these APIs in detail.


A) What is Multimerge API? 

                 The Multimerge API is used to merge the list of records together. Multimerge is the generic form of merge API.






B) What is Merge API? 

                The merge API is used to merge two base object records that are identified as the same base object record.


C) What are the differences between Multimerge and Merge API? 

          1) Number of records to merge : 

              a) Merge API allows only two records to merge 

              b) Multimerge API allows more than two records to merge.

         2) Parameters to request : 

             a) Merge API accepts sourceRecord key and targetRecord key as parameters in the input

             b) Multimerge API accepts multiple record key lists as parameters in the request.





         3) Consolidated records : 

             a) Merge API allows records irrespective of the value of consolidation indicator 

             b) Multimerge API allows merging of unconsolidated records only i.e. consolidation indicator                   !=1

         4) Final value for consolidation indicator : 

            a) The final value for consolidation indicator after performing merge API operation is 1 i.e                           consolidated state  

            b) Multimerge API does not change consolidation indicator value for surviving records.

        5) Surviving Record : 

             a) The surviving record is specified in merge API with targetRecordkey as the parameter.

             b) For Multimerge API, the surviving record will be determined based on survivorship rules of the XREF that are participating in the merge process.


                 Learn more about Informatica MDM survivorship rules here 



   

Sunday, September 5, 2021

How to design landing table in the Informatica MDM ?

          Are you planning to implement Informatica MDM in your project and starting designing a landing table? Are you also interested in knowing the types of landing table designs? If so, then you have reached the right place. In this article, we will explore factors that need to consider while designing a landing table in Informatica MDM.


A) what is the landing table in Informatica MDM?

             Are landing tables are the tables where data from the source is loaded in order to process the data and sent through a stage process to cleanse and standardize it. For the stage process, the landing tables act as source and stage tables as targets.






B) Factors to be considered for landing table design

               We need to consider the following factors

While designing landing tables in MDM

             1. How many source systems are involved

             2. What is the volume from each source system

             3. Impact of development timelines

             4. Maintenance requirements

             5. Partition requirements


C) Type of landing table designs

             Based on the Information capture in the previous section, we can design landing tables in two ways

             1. One landing table for each source

             2. One landing table for multiple sources





1. One landing table for each source

             If each source is having different types of data ( e.g.one source is customer-centric and another source is Account centric), if the volume in each source is almost equal, or if we have good development and maintenance bandwidth then we can design one landing table for each source.









2. One landing table for multiple sources

             If multiple sources are having similar attributes and data types and volume in each source system is low and if we need to expertise the development time then we can design one landing table for multiple source systems. 



                                 



Learn more about MDM landing table here







Saturday, August 21, 2021

What are the log Files in Informatica MDM ?

        Are you trying to analyze the issue in Informatica MDM? Are you looking for the details of the log files which are generated during various Processes in the Informatica MDM? This article will explore more about the log files, their locations, and when to use those.





A) Introduction

         Informatica MDM has various components such as application server, database, business process management tool, Application user interface such as Entity 360 or customer 360. Each of these components generates logs throughout its processing.

         Here we will understand various types of log files and these are

         1. Hub server logs

         2. Process or cleanse server logs

         3. E360 logs

         4. Provisioning logs

         5. Post Installation logs

         6. Elastic search logs

         7. Application server logs

         8. Database logs

1. Hub server logs

         Informatica MDM has two core components: hub server and process server earlier it was called cleanse server. The hub server is used to initiate the jobs, managing and controlling the threads in short hub server is master component in Informatica. The logs are generated when we access the Administration section of the MDM hub. Especially when we validate the ORS . These logs are captured in Hub server logs.

          Location :- <Informatica MDM install folder>/hub|server|log

        e.g./abc|hub|server|logs|cmxserver.log

2. Process server logs

           When we execute the jobs such as Stage, load, tokenization, match and merge jobs, the logs are captured in the process server logs.

            Location :- <Informatica MDM install folder>/hub|Cleanse/logs

             e.g . /abc/hub|cleanse/logs|cmxcleanse.log

3. E360 logs

              We can configure the user Interface using the provisioning tool. The User Interface is called Entity 360 application. When we access the application the logs are generated.

              Location:- <Informatica MDM install folder>/hub|server|logs 

              e.g. /abc/hub|server|logs|entity360view.log

4. Provisioning logs

             We use provisioning to configure business entities, transformations, views, tasks, and E360 applications. When we use the provisioning tool the logs are generated and stored below location.

             Location :- < Informatica MDM install folder>/hub|server/logs 

             e.g. /abc/hub|server|logs|provisioning.logs





5. Post Install logs

               The post-install logs are generated when we install Informatica MDM as well as when we apply EBF or upgrade.

               Location :- <Informatica MDM install folder>/hub|server|logs

              e.g. /abc/hub|server|logs|postInstallSetup.log

6. Elastic search logs

                If you are using Elastic search in your Informatica MDM then you may need to use Elastic search logs.

                Location:- <Elastic search folder>/logs 

                e.g. /aqr|logs|elastic search.log

7. Application server logs

                Application server logs are located as below

                a) Jboss

                 <Jboss home>/standalone/log/server.log

                  b) Weblogic

                   < Weblogic home >/domains/domain name/servers/server name/log/abc.out

                  c) websphere

                   <websphere home>/Appserver/profiles/profile name/logs/server name/systemout.log

8. Database logs

                  Database logs are not directly accessible. You need to reach out to your DBA to get database logs.


           Learn more about Informatica MDM here







        



Friday, August 20, 2021

What is difference between 'Remove from match list ' and 'Not a match' in IDD ?

             Are you looking for an article on Informatica Data Director which explains what is the difference between 'Remove from match list' and 'Not a match' options which are available in IDD application in Informatica Master Data Management? If so, then you reached the right place. Let's understand these two options here.


A) What is the match process?

                Informatica MDM comes with a process named match process. With help of this process, we can determine potential matching records. In other words, we can remove duplicate records from the system. Informatica Data Director application uses a match engine that comes with MDM in order to achieve it.

                   IDD application uses a match engine at the time of processing manual match records as well as at the time of creating a new record. This requires some specific to be mode using IDD configuration manager.






B) Where in the IDD application we can find the match feature?

                IDD application is used by data stewards or business users to manage the data. In order to manage the data, records need to be first searched and then opened. Once the record is opened we can see data, Xref, timeline, history, and match sections. The match section shows potential matching consolidate to the given record.


C)  Difference between ' Remove from match list ' and ' Not a match' in IDD

                  As discussed in the earlier section, potential matching records are available in the match section. If uses are working on a manual merge queue using this match section the users either can merge the record or can perform one of the below actions on the merge task.

                  1. Remove from match list

                  2. Not a match





                 The remove from match list removes the record from the match view in the IDD application. If the user logins again the record will be shown again on the screen.

              On the other hand, if the user selects the ' Not a match ' action then the matching entry will be deleted from match table. The record will not be shown in the IDD view anymore. This will also delete the merge task.

              Learn more about the merge process here -

 


Thursday, August 5, 2021

What is lifecycle of consolidation_ind ?

    Are you looking for details about consolidation indicator in Informatica MDM? Are you also interested in knowing what are valid values for the consolidation indicator column? If so then you are right place. In this article, we will also explore the lifecycle of consolidation indicator for a record.


A) What is a consolidation indicator?

          The consolidation indicator is a column in the base object table. The consolidation process is also known as the merge process updates the consolidation indicator value for a record based on the record's process.






B) What are valid values for consolidation_ind?

          When we execute load jobs, Match job and merge jobs record goes through various processes because which value of consolidation indicator column goes through the values below

        4 : New record

        3 : Record queued for the match process

        2 : Record has gone through the match process

        1 : Consolidation record  

a) Consolidation_ind = 4

           The actions below cause record to have consolidation_ind = 4

          1. Inserting a new record in the Informatica hub 

          2. Queue a record as the new using data manager 

          3. When we unmerge the record either through API or E360 application

b) Consolidation_ind = 3

          The actions below cause record to have the consolidation_ind=3

           1. When we queue a record for a match using data manager

           2. If match job fails, the records that were picked up for matching but the match did not complete for that record will have value for consolidation indicator as 3

C) Consolidation_ind = 2

         The actions below cause record to have the consolidation_ind =2

         1. When the match process completes for the record

         2. When the record is queued for merge using data manager or any API or application.





d) Consolidation_ind = 1 

          The actions below cause record to have consolidation_ind =1 

         1. For the Golden source system the records will be loaded with consolidation_ind=1

         2. When the record is accepted as unique 

         3. If Accept Record as unique is set to yes and match process does not find matching record.

         4. If Accept Record as unique is set to yes and after merging records, it does not have any more matches.

e) Consolidation_ind =9

         If a business user puts the record on hold.

 

      Learn more about consolidation indicator in Informatica MDM here -



Tuesday, July 20, 2021

Top 10 things you need to know before implementing Informatica MDM?

           Are you planning to implement Informatica Master Data Management aka MDM? Are you not sure what are the things you need to consider before considering MDM solution? If so, then you reached the right place. In this article, we will see the top 10 things which you need to know before implementing Informatica Master Data management in your organization. 

 1.Data Quality Measurement 
            You need to know how you are currently measuring data quality not only in a single project but also across the enterprise. This will give you two benefits one, you will know better options for quality of data measurement, and second, a baseline to measure the quality of data after MDM implementation.





 2.MDM and Data Quality
              Is there a relationship between master Data management and Data Quality? Can MDM help in improving data quality? The answer is Yes. However MDM and Data Quality are two distinct processes in any organization. You need to know what is the relationship between Master Data Management and Data Quality.

 3.Returns of Improved data quality?
               We initiative various projects for improvement in the processes and to achieve better results on Investment. You need to know what is the return you will be getting after improving data quality.

 4.Data Governance 
              Data governance is a crucial part of the business. Are you aware of how is data governance is implemented in your enterprise? You need to have proper data governance to get optimum benefits from MDM implementation.

 5.Data for business strategy 
               It is not new that this Era of data. We are in data 4.0 where the majority of the businesses are data-driven. You need to plan your business strategies based on data that is of great quality and well maintained.





 6.Data enrichment
               Why do you need data enrichment? One may ask this question. The answer is to make better decisions and recommendations we need to take important steps towards data enrichment.

 7.Privacy regulations 
               These rules and regulations we need to follow a business. We need to be fully aware of those rules and regulations and consider those implementing any MDM solutions.

 8.Customer satisfaction 
                Is your customer satisfied with your services? What are your customer's preferences and how are you managing these? How are addressing your customer's concerns and feedback? These are important questions you need to answer so that you improve those with MDM.

 9.Risk measurement and assessment 
                Informatica MDM defiantly plays a vital role in risk assessment and measurement. However, you need to know your current solutions and look for better opportunities to improve those.

 10.Future Perspective
             While implementing Informatica MDM, you need to look for long-term benefits instead of short-term MDM with great benefits for the long run.

Learn more about information MDM here

       

Wednesday, July 7, 2021

How to achieve better stage job performance in Informatica MDM

      Are you working on MDM and want to understand its different fields, then this is the right place in this article we are going to know about what are the product recommendation, thread setting properties, and database recommendations of MDM stage job performance


 A. Thread setting for MDM Stage Job Performance

       In this article, we will see what are the different reasons for issues and their solutions regarding stage job performance.

     1) Post/ Pre Stage UE

          The reason for this issue is it we are going to write the query regarding inserting or updating and record, it will create locks and because of this it slows down the performance to avoid this there is a need to put logger statement in UE  code. If doable rerun the jobs by removing the UE code.





     2) Cleanse Function 

            Each in every record will have to go through a cleansing process so here need to check that if any cleanse function is taking more time to process.

           Depends on which type of cleanse function you are using there is a need to check network latency also we need to check of IDQ  end.

     3) Directory 

            The main reason for the directory issue is having a shared directory within 2 different instances of the process server.

           To avoid file locks each process should have its own directory.

     4) Tables  

            The system will reduce the performance if the tables like RAW, REJ, C_REPOS_JOB_CONTROL table contain a huge amount of data.

           We can keep the important data and archive old data. 

   5) Log level

           We will have to keep the logging level to INFO.


B. Thread setting for MDM Stage Job Performance

          The configuration that can be updated :

       1) Threads for cleanse operation (HUB Console)

          Max value = ( Number of CPU cores -1 )

        2) In cmxcleanse properties

            (cmx.server.cleanse.min_size_for_distribution)

            Default size is 1000





        3) com.informatica.mdm.batchcontroller.Batchjob.min_rec_for_multithreading

            The default size is 1000. We can decrease the size if multithreading is enabled and no. of records are lesser than 1000.


C. DB Recommendations for MDM Stage Job Performance.

        1) collect AWR reports.

        2) For the DB performance, collect TESTIO results


D. Appserver Recommendation for MDM Stage Job Performance

        1) when we run the processing server and DB server jobs, check that the CPU is going high.

        2) Have to check basic java arguments such as - xmx value.

           


 

Monday, July 5, 2021

What is Business Entity Services in Informatica MDM?

    Are you looking for details about the Business Entity Services aka BES in the Informatica MDM? Are you also interested in knowing how to use Business Entity Services in your project? If so, then you reached the right place. In this article, we will explore more about Business Entity Services. Let's start.


A) What are Business Entity Services in MDM?

         Informatica MDM is used to master the data from various source systems. We onboard the data from sources to the MDM landing table. The data is moved to the staging table from the MDM landing table and then to the base object tables by executing stage and load jobs. Once we consolidate the data in the MDM, it will be ready to consume by consuming systems. Consuming can consume a golden copy of the records by consuming ETL or JMS Queue or Business Entity Services. Business Entity Services are web services that can be used for real-time integration of external systems with Informatica MDM.

B) What are operations can be performed using Business Entity Services?

      The various operations such as Read, Write, the search can be performed using Informatica BES.

      1) Read BE Business Entity Service-

          It is used to read the data from the base object. the result does not include soft-deleted records.

      2) Write BE Business Entity Service- 

            It is used to create, update and delete parent and child business entity elements. we cannot perform trust override with this service.

      3) Search BE Business Entity Service-

            It is used to root record in the business entity. We can use match operation while searching records to achieve fuzzy searches.

C) Understanding Business Entity Services Architecture 

        Business Entity Services is an integral part of Informatica MDM and it is available as wed service. The consuming system can write web service consumer and consume BES, for reading and write operation. Once the operation is invoked when the data is searched against the base object or updated against the base object.



D) What are the REST Methods Supported 

      Here is the list of HTTP methods supported in Business Entity Services

           GET - used to retrieve details about record, task, or a file 

           POST - used to create task or record 

           PUT -  used to update root, child records, or a file.

          PATCH - used to update task partially 

          DELETE - used to delete records or file


Learn more about Business Entity Services here -  



Tuesday, June 22, 2021

When Tokenization job gets executed in Informatica MDM?

Are you looking for the details about when the tokenization job executed in Information? Would you be interested in knowing the relationship between a match job and a tokenization job? If so, then you reached the right place. In this article we will explore the instances during which tokenization job gets triggered:

A. Why we need a tokenization process in MDM?
          Tokenization is a process that generates the token-based fuzzy match key and other match column configuration. The tokens generated are stored in the table <BO> STRP. This STRP table is used as input to match the process. The match process matches the records based on these tokens.





B. When the tokenization process executed
           The tokenization process gets triggered at various points and These are:
          1)  Manual tokenization execution
          2)  During the match process
          3)  During the load process

1) Manual tokenization job execution
       We can execute the tokenization job in Informatica MDM manually whenever it is needed. It is recommended that tokenization need to executed separately instead of in conjunction with load or match job.





2)  Tokenization process as part of match job
       If it is required to update the match token, the match process in Informatica automatically starts the match job's tokenization process. This scenario may happen if new records have been added or updated existing records.


3) Tokenization process during load job.
         If we enable the `Generate Match Tokens on Load' property on the base object then, when records loaded in the base object at that time the tokenization job will be triggered. It is not recommended to enable this property as it will adversely impact load job.






Learn more about the tokenization process here-


Wednesday, May 19, 2021

Informatica MDM match job and Tokenization job

       Are you interested in knowing what is the relationship between the Informatica MDM match and tokenization job? Would you also like to know what all conditions needed in order to tokenization job as part of match job? If so, then you reached the right place. In this article on Informatica MDM, we will learn match and tokenization job relationship.

A] What is the tokenization job in Informatica MDM?

           Tokenization is a process in Informatica during which match columns are used to generate tokens. These are based on the type of fuzzy match key column, another match column, and type of data present in the system 

          An example of the token is X01Z530K. It is an alphanumeric value for a given record.





B] What is a match job in Informatica MDM?

      By using the match process the record is matched in order to consolidate the record from various source systems. The match process uses tokens that are generated as part of the tokenization job.

       Once records are matched then those are stored in <BO Table>_MTCH table. The merge or consolidation process uses the MTCH table to consolidate the records.

C] Dose the match job always generate Match tokens?

        The answer is No. The match job will not always generate match tokens automatically. There various factors as part of a match job.





D]  What are factors are needed in order to generate tokens as part of the match job automatically?

         The following conditions are required to be there in order to generate a match token when we execute the match job.

     1. The base object should be fuzzy match enabled.

     2. The important ' Fuzzy Match Key ' column must be defined in match columns.

     3. Match ruleset should have at least on Fuzzy match rule.

     4. If generate Match Tokens job already executed then match job will not generate tokens.

     5. If the value of Dirty Indicator column value is 0 ( zero) for all the records then match job will not generate tokens

More details about the tokenization job are explained in the video -







Friday, April 9, 2021

How to fix ORA-00604-Error occurred at recursive SQL level 1

 Are you working on a project where the oracle database is being used for implementation? Are you also facing an ORA-00604 and looking for fixing this error? If so, then you reached the right place. In this article, we are going to see how to fix ORA-00604-Error that occurred at recursive SQL level 1.





What is ORA-00604?

The error message ORA-00604-Error occurred at recursive SQL level 1 is commonly noticed in the application logs or identified by oracle user. This error message is a little complex and has its own challenges to fix. Let's understand what is the root cause of it.

The root cause of ORA-00604 

There are several causes of this error, however, the main cause is processing a recursive SQL statement. you might have a question what is the recursive statement? A recursive statement SQL statement that is applied to internal dictionary tables.

Example of ORA-00604

Assume that you are using Oracle 11g or 12c and you are getting 'ORA -00604-Error occurred at recursive SQL level 1' where table or view does not exist one of the causes for this is a trigger. The trigger might be 

1) to insert records into an audit log table or

2) to fire DDL statements or

3) to drop the audit log table Let's see possible option to fix this error





 How to fix ORA-00604 - 

Option 1:- Analyze and fix trigger error in order to determine if the error is related to DB trigger, execute statement below-

ALTER SYSTEM SET "_system_trig_enabled"=FALSE; 

Find the trigger which causes the issue & disables it.

Option 2:- syntax error in the SQL generated in the application. If an ORA error caused because of an error or bug in the  SQL code then reach out to developers of the code & fix it.

Option 3:- Oracle support If the error still persists then reach out to the oracle support team.


Learn more about oracle in detail here -












Sunday, March 21, 2021

Understanding Timeline in Informatica MDM

Timeline is one of the useful features in Informatica Master Data Management. It enables us to manage various versions of the business records. The timeline is totally different from the history of the records updates and inserts. In this article, we will focus on Timeline granularity and Timeline actions features. So let's dive in.


Timeline Granularity

In Informatica MDM, We use time measurement such as years, months, days, etc to define effective periods for versions of records. This time measurement is nothing but timeline granularity. We can define timeline granularity as a year, month, day, hour, minute, or seconds.




Timeline Action

The timeline action is nothing but an action to perform for entities for which you track data change events. We can perform add, edit actions on a record and edit effective period. 

a. Timeline action = 1 --> This value is set if we update the data
b. Timeline action = 2 --> This value is set if we update Effective Period
c. Timeline action = 4 --> This value is set if we Add Effective Period

Versioning Sequence :

The stage table column VERSION_SEQ  is set to 1 by default and gets changed based on the new update.


Learn more interesting facts about Informatica MDM Timeline



Monday, February 15, 2021

What are the different phases of Informatica MDM Implemantion


Are you planning to implement the Informatica Master Data Management project? Are you looking for what are the details you need to capture? If so, then you reached the right place. In this article, we will understand what are the phases of Informatica MDM implementation. We will also see what will be the outcome of each phase.


Introduction:

There are 5 phases in Informatica Master Data Management implementation.




1. Discovery Phase
2. Design Phase
3. Development Phase
4. Testing Phase
5. Deployment Phase



Let's understand each of these phases one by one.


A. Discovery Phase:

We need to perform the activities below as part of the discovery phase -
  • Perform data profiling
  • Conduct interviews and workshops 
  • Document solution requirements 
  • Define conceptual architecture 
  • Define the logical data model 

What will be output?
The output of this phase will be 1. Functional Requirements Specification 2. Testing Plan


B. Design Phase:

We need to perform the activities below as part of the Design phase -
  • Define detailed business rules 
  • Develop system design specifications for all technology components 
  • Develop Traceability Matrix 
  • Establish DEV Environment 

What will be output?
The output of this phase will be 1. System Design Specification 2. Traceability Matrix

C. Development  Phase:

We need to perform the activities below as part of the Development phase -
  • Configure and develop MDM components
  • Perform unit testing
  • Develop Test Scripts 
  • Develop Test Schedule 
  • Update Traceability Matrix 
  • Establish QA and PROD Environment 

What will be output?
The output of this phase will be 1. Configured and developed MDM components 2. Test Scripts 3. Traceability Matrix





D. Testing Phase:

We need to perform the activities below as part of the Testing phase -
  • Execute System Testing 
  • Execute Rules Testing 
  • Execute Integration Testing 
  • Execute User Acceptance Testing 
  • Develop Test Summary Report 

What will be output?
The output of this phase will be 1. Tested, production-ready MDM solution 2. Test Script Results 3. Test Summary Report 

E. Deployment Phase:

We need to perform the activities below as part of the Deployment phase -
Deploy the Solution into the Production Environment 
Execute the initial IDL Initiate 
Gray Area Reconciliation 

What will be output?
The output of this phase will be  1. Deployed solution Initial data loaded and ready for reconciliation





You can learn more about Informatica Master Data Management here -



Thursday, January 21, 2021

Informatica MDM Installation Checklist


 Are you looking for an article about how to prepare a checklist for Informatica MDM Installation? You might have gone through the Informatica MDM Installation guide and must be facing issues from where to start. The best way to start is nothing preparing checklist. In this article, we will see what are the main topics to consider for check and we will also see a sample checklist file.





Introduction

After going through more than 100 pages in the Installation guide, I realized that we need to prepare a checklist for each component of Informatica MDM. Informatica MDM comes with Hub Server, Process Server, Informatica Data Director Configuration Manager, Provisioning Tool, Business Process Management i.e. Active VOS, Elastic Search, Application Server (such as Jboss, WebSphere, Weblogic), Database (such as SQL Server, DB2, Oracle). Each of these components has a separate set of instructions. Sometimes we get lost or overwhelmed with all these instructions. So take a long breath and start documenting the main section as like mentioned in our next section

Components to consider for installation

The number of components that are needed for installation is based on the business needs. However, there few components are commonly required irrespective of business need e.g. Process Server, Hub Server. So first document all possible components which you are planning to install. Here is a sample list of components. 


Informatica MDM Installation checklist

The checklist contains details about each of the components which are captured in the previous section. e.g. Create the MDM Hub Master Database section will have details about database name, server name, port, and credentials. You can access the complete checklist here.


For reference, the checklist will look like as below





If you are looking for more details about Informatica MDM then here is a video for Informatica MDM.







Saturday, December 12, 2020

How to prepare for Informatica MDM Developer Certification?

 If you planning to get Informatica MDM Developer certified and wanted to know how to prepare for Informatica MDM Developer Certification. If so, then you reached the right place. In this article, we will discuss all the topics on which Informatica Master Data Management certification is based on. We will also discuss the Informatica MDM Certification cost. With that, I would like to congratulate very first steps towards your MDM Developer certification and good luck with your exam.






Introduction

In this article, we will discuss all the topics on which questions will be asked in the certification examination. We will provide the necessary video tutorial and other material. With help of this material and proper training, you will be ready for your certification. You do not have to look for Informatica MDM certification dumps and avoid online frauds. You can find Informatica MMD training online either on Udemy or other platforms. You can also take training from Informatica but these are a little expensive. The Informatica Certification exam includes questions that are multiple-choice, multiple responses, or true/false formats. The exam has 70 questions with a passing score of 70 percent. The duration of the exam is 90 minutes. The exam access will expire 90 days after purchase, with no refund.


Informatica MDM Certification cost

Informatica MDM Certification costs $340 USD, it includes a second free attempt. After that, if you would like to attempt again then the original cost will apply. This cost varies from time to time, so this cannot be a guaranteed price.

Informatica MDM Developer Certification

In this section, we will discuss the Informatica MDM syllabus. I would strongly recommend going through these topics before taking any Informatica MDM training online or attending any Informatica certification program. If you already working on the MDM project and have good experience in Informatica MDM real-time projects then it will be easy to get certified. 





A. Introduction to MDM Multidomain Edition

E. Overview of Match and Merge Processes

F. Fuzzy matching

G. Configure of queries and packages

H. User Objects

I. Introduction to Entity 360

J. MDM Entity 360 Architecture





Tuesday, December 1, 2020

Infomatica MDM - MDM Installation Topology

 Are you planning to install the Informatica MDM hub in the Development or Production environment? And looking for the details about the best possible way to make use of your infrastructure? If so, then you have reached the right place. In this article, we will explore different Informatica MDM installation topologies






Introduction

Basically, there are three types of topologies recommended by Informatica. We can use one of them while installing the Informatica MDM hub, based on project needs and benefits which we are looking for. Here is a list of recommended topologies

a. MDM Topology for Clusters

b. No Cluster - No High Availability

c. No Cluster - High Availability


A. MDM Topology for Clusters

In this type of topology Hub Server and Process server resides in a different machine and these are clustered together.


Characteristics:




B. No Cluster - No High Availability

In this topology, Hub server and Process servers are not clustered, hence we will not achieve high availability.



Characteristics:



C. No Cluster - High Availability

In this type of topology, Hub Server and Process are not clustered, however an external load balancer can be used to make the MDM system highly available.




Characteristics:







The detailed information types of MDM styles are provided here -















Friday, November 13, 2020

Informatica MDM - How to fix an error - ORA-01555: snapshot too old?

While working on Informatica MDM jobs, I came across one issue. The issue is an ORA-01555: snapshot too old. This error message was reported while running the tokenization job. If you are also noticing a similar issue then this article will help. This article provides details about an error message and a solution to fix it.






Error Message:

The detailed error message is as below -

java.sql.BatchUpdateException: ORA-01555: snapshot too old: rollback segment number 11 with name "_SYSSMU11_2399779032$" too small

SIP-16084: Error occurred while verifying the need to tokenize records. Return code 12801, 

Error SQLException During VerifyNeedToStrip :ORA-12801: error signaled in parallel query server P000,
ORA-01555: snapshot too old: rollback segment number 30 with name "_SYSSMU30_2998435469$" too small.
 at com.siperian.common.SipRuntimeException.createNotExternalized(SipRuntimeException.java:74)
 at com.delos.cmx.server.interact.caller.InteractCleanseClient.executeGenerateMatchTokens(InteractCleanseClient.java:460)


Solution:

To fix the issue perform the steps below -

A. Database Issue
First, analyze if there is any database issue going on. If the database looks good then perform the steps below

Step 1: Stop any job running as 'INCOMPLETE'
Step 2: Stop the Application server
Step 3: Verify undo_retention value by running the query below on db side
           
             show parameter undo_retention;

Step 4: If the value is lower then increase the value to 4000 by executing the command below

            ALTER SYSTEM SET UNDO_RETENTION = 4000;

Step 5: Increase Undo Tablespace to Auto Extended on.

Step 6: Restart the database servers with the clear cache

Step 7: Drop T$ table if present any in database

Step 8: Start the MDM servers with a clear cache.






This will fix the issue. I hope this is helpful. You can learn more about job tuning here -





What is Virtual Private Snowflake ?

                 Are you looking for details of Virtual Private Snowflake or VPS? Are you also interested in knowing how to use an account l...