In the latest times, business organizations run and improve the growth structure based on the data. They are utilizing some powerful tools to slice and dice, play with the data to reach out to potential customers for consistently driving business. Informatica IDQ is one of the latest web-based tools that is extensively combining customers and business organizations by allowing them to perform the data related functionalities effectively.
In this article, you can go through the set of Informatica IDQ interview questions most frequently asked in the interview panel. And these listed questions will help you to crack the interview as the topmost industry experts curate these at HKR Trainings.
Now let's have a look into the IDQ interview questions based on the levels.
Ans: Worklet refers to the group of workflow tasks that are accumulated in the set in a classified and organized format. The following are the different options available.
Ans: IDQ refers to Informatica Data Quality which is a tool used by the data analysts to deploy the data quality for all the different kinds of workloads. IDQ integrates with Informatica Power center to ensure that data quality is delivered. IDQ is specifically designed to perform functionalities like data matching, cleansing, and profiling.
IDQ includes two core components that are listed below:
Data quality workbench: The data quality workbench is the component that performs data analysis and creates the rules accordingly for data enrichment and standardization.
Data quality server: The Data Quality Server is used to run the programs, file sharing, and accomplish the plans in a networked environment. It also supports networking through different domains and communicates with workbench over TCP/IP.
Wish to make a career in the world of IDQ? Start with HKR'S IDQ Online Training !
Ans: The following are the differences between connected lookup and unconnected lookup.
Connected Lookup: The connected lookup is a lookup that is capable of participating in all the workflows and receives the inputs from other transformations directly. Connected lookup is used in both static and dynamic cache, supporting user-defined values.
Unconnected Lookup: Unconnected lookup is capable of receiving the inputs based on the last known position. It returns only one column value. Unconnected lookup is not capable of supporting user-defined values.
Ans: A predefined event states that the event is predefined. It is likely to be a file watch event. A predefined event waits for a particular file to arrive at a specific location.
A user-defined event is a collection or flow of the tasks to be performed in the workflow process. The user-defined events are created and utilized based on the requirement.
Ans: Address Doctor in IDQ is a transformation primarily used to compare the input address data with the reference data. It is designed to determine the accuracy of the input address and relatively fixes the errors in those addresses. It is the validator transformation available to rectify and make the necessary changes to the data as per the business requirements.
Ans: The standalone command task could be used to execute shell commands anywhere in the workflow process.
Ans: It is a transformation to verify I/P data with address reference data to ensure accuracy. It will resolve problems if it has been found.
Ans: An enterprise data warehouse is termed as a central and corporate repository that helps in storing and managing the historical data of an enterprise. The information would come from different sources like CRM’s, recordings, ERP’s, etc.
Ans: A surrogate key is the substitute/replacement of the primary key within the database. It is a unique identifier used to represent each row in the table. It is either represented in the form of an integer or a digit. The surrogate key has more significance because it makes the updation easier though the primary key is changed.
Ans: The command task is specifically used to perform the shell commands. It could be anywhere in the workflow. Command task can be called as pre-session or post-session command as it can run the commands accordingly within the workflow process.
Ans: The term transformation represents the nature of the activity. It is called the repository object that has the capability of generating, modifying, and passing the data.
There are different types of transformations that are available in Informatica. They are listed below:
Aggregator transformation: Aggregator transformation is an active transformation used to perform calculations on the data. The aggregate transformation stores the aggregate data in the aggregate cache.
Filter transformation: A filter transformation is a connected and active transformation used to filter the rows that do not meet the condition in the mapping.
Lookup transformation: Lookup transformation is an active, connected, or unconnected transformation used to look up the data in the relational table.
Rank transformation: Rank transformation is a connected and active transformation used to perform the selection of the top or bottom rank.
Expression transformation: The expression transformation allows the calculation of values in a single row. It also involves in performing non-aggregate calculations
Joiner transformation: Joiner transformation is a connected and active transformation used to join two sources of data. The source of the data can be either from one origin or two different origins.
Normalizer transformation: Normalizer transformation is an active transformation that helps in the conversion of a single row into multiple rows or multiple rows into a single row.
Router transformation: Router transformation is a connected and active transformation that has the capability of capturing the data that do not meet the condition.
Ans: A static cache refers to the cache that cannot be modified and will remain the same during the session. Informatica cache by default is a static cache.
A dynamic cache refers to the cache that allows modification during the session by performing functions like insertion or updating of the records.
Ans: Update transformation strategy allows the processing of the source data in a row format, row by row. Update strategy transformation is used when there is a need for insertion or updation of some logic in the row.
There are multiple options available in the update strategy, which are listed below:
Ans: There are three different types of dimensions available in Informatica IDQ. They are listed below:
STOP command: The STOP command helps in informing the integration service to stop reading the data from the source. It performs other functionalities like writing, committing, and processing the data.
ABORT command: The ABORT command can terminate the session only after a time frame of 60 seconds. This happens when the integration service is unable to complete the processing and committing of data within the given time.
Ans: Informatica includes two different types of loadings: Normal loading and bulk loading.
Normal Loading: This type of loading is the process that allows the loading of records in a sequential format: one by one. Normal loading takes a lot of time to the target source.
Bulk Loading: Bulk loading is the process that allows the loading of multiple records at once. The time taken for loading the data using the Bulk loading process is very less when compared to the normal loading.
Ans: The following are the different tools available in the workflow manager listed below:
Control M is one of the alternative tools that is specifically used for scheduling the processes other than the workflow manager.
Ans: A file that is created in a Wordpad or text editor is called as a parameter file. The following are the different data values that can be defined in a parameter file.
Ans: The following are the different components that are required to be installed during the Informatica power center installation.
Ans: Yes, there is an option to determine the through the option in Informatica. The throughput option is available in the Informatica workflow monitor. You will need to navigate to the workflow monitor, search for the session button, right-click on it followed by run properties. You will be able to view the throughput option under source/target statistics.
Ans: There is a different way where the record can be updated in the table without the usage of Update strategy. To perform this functionality, a key has to be defined in the Informatica level. The next step is to connect the key and update the fields that you would like to update in the mapping target. In the session level, the target property should be set up as “Update”, and the Update checkbox has to be checked.
Ans: The power center integration service, referred to as an application service specifically designed to run the sessions and workflows related to data management. The data center integration service referred to as an application service specifically designed to perform data integration tasks for different tools like Analyst tool, developer tool, and external clients. The Developer tool and the analyst tool will work together and send the data integration requests to the data integration service to perform some functionalities. These functionalities could either include SQL data services, running and previewing of data profiles, mappings, etc.
Active transformation: Active transformation in Informatica refers to the process that allows performing the changes to the number of rows that have gone through the mapping. The following are some of the active transformations:
Passive transformation: Passive transformation in Informatica refers to the process that does not allow performing changes to the number of rows that have gone through the mapping. The following are some of the passive transformations:
Ans: The performance of aggregator transformation can be improved if the records are sorted before performing the aggregator or passing the aggregator to the records. The sorted input option is checked, which is available under aggregator properties. The sorting should happen on those sets of columns that are used by the group by option.
Sessions: A session in Informatica refers to the set of instructions defined to transfer the data. These instructions are guided towards the power center server by letting it know when to perform the data transfer from the defined source to targets. In short, the conversion of data from source to data using defined instructions is called a session.
Batches: A batch is the set of tasks to be performed. The tasks include sessions, email, write, commands, etc. The execution of batches is used to combine different session executions. This can be done either in a serial or parallel manner. The pmcmd command is used to run the batch.
Ans: Stored process transformation is an essential method for populating and preserving databases.
If you have any doubts on IDQ, then get them clarified from IDQ Industry experts on our IDQ community!
Ans: Yes, there is an option to export an object from IDQ to the Power center tool. The following are the steps to be followed to export an object.
Powercenter repository service:
The PowerCenter repository service includes folder-based security and is utilized by both power center application services and clients.
Model repository service:
The model repository service includes project-based security which is the same as that of other application services like Analyst service, Developer tool, Data integration service, analyst tool.
Ans: Live data and staged data are two options available during profile creation for data drilling.
Live Data: For any type of profile definition, Live data is the default option set up under profile creation. Live data will help in accessing the row data on the source.
Staged Data: All the data represented in the rows is staged in the profiling warehouse.
Some of the sources that use mainframes and other BigData, drilling of data is not available on Live data. This is because of its higher expenses. Hence, the default mode of drill-down that uses such applications will be staged data.
Join Analysis profiling: Join Analysis profiling is the process that is used to analyze the degree of potential joins between two data columns. The results of the Join analysis will be represented in the form of a Venn diagram as a percentage or numerical value.
Multiple profiles: The multiple profile option helps in creating profiles on various tables at once. This option is available in the profile wizard window that allows prefix and suffix with some default values. The steps to be followed to create multiple profiles is: Select the objects ->right-click->Select the desired profile->Select multiple profiles.
Compare profile: Compare profile in Informatica IDQ provides the flexibility to compare the output from two transformations. It is beneficial as it is capable of providing the before and after pictures of the data. It is also compatible to be used with any source object, mapplet, a transformation that has outputs.
Ans: Target load order in Informatica refers to the list of activities where the priorities can be defined. The data loading happens in the Informatica server based on the priority. If there are a set of source qualifiers linked to multiple targets, then the order can be defined, or the order can be dictated so that the server will load the data accordingly to the targets.
Ans: Slowing changing dimensions are those dimensions that will be changing in a short time. They are called SCD in short. There are three different types of slowly changing dimensions in Informatica. They are:
Ans: The data transformation manager process is created once the load manager performs the validations for the sessions. The data transformation manager is the secondary process that is associated with the sessions in Informatica.
Ans: A parameter file is nothing more than a file generated in a text editor or a word pad. In a parameter file, the following different values can be defined:
Ans: Target load order has been nothing but a list of all operations where priority can be determined. Based on this priority, the data will be loaded to the Informatica server.
If you have a list of Source Qualifiers linked to multiple targets, you can specify an order or dictate an order to the Informatica server so that the data can be loaded into the targets.
Ans: Yeah, an object can be exported from the IDQ to the Powercenter tool.
Ans: The four output files are Session Log ,Workflow Log,Errors Log, Badfile.
Ans: Aggregator output improves significantly if the records are sorted before the aggregator is passed and the "sorted input" option under the aggregator properties is checked. The record set should be sorted on the columns used in Group By process.
Ans: The target table can be revised without the use of 'Update Strategy.' To do this we need to identify the key at the Informatica level in the target table, and then we need to link the key and the field that we want to update to the Target mapping. At the session stage, we can set the target property as "Update as Update" and check the "Update" checkbox.
Ans: Lookup transformation can be configured as a "Active" transformation from Informatica 9x.
Batch starts on 30th Oct 2021, Weekend batch
Batch starts on 3rd Nov 2021, Weekday batch
Batch starts on 7th Nov 2021, Weekend batch