ETL Testing Interview Questions

  1. How does an ETL Testing operational information store function?
  2. Define the term regression testing?
  3. Define the term ETL validator?
  4. What do you understand by the term data source view?
  5. Describe the various operation used in ETL systems.
  6. What do you understand by the term Hash partitioning?
  7. What is the procedure for processing a flat file with fixed length layout?
  8. What is purpose of impact analysis within ETL systems?
  9. What methodology can be utilized for extraction of SAP data with the help of Informatica?
  10. Define the term Round Robin partitioning.
  11. When is connect and unconnected looks–ups used?
  12. Describe data profiling with respect to ETL system.
  13. What do you understand by the terms initial load and full load?
  14. Explain the three-level hierarchy of a typical ETL cycle.
ETL Testing Interview Questions

In information warehousing engineering, ETL is a vital segment, which deals with the information for any business procedure. ETL represents Extract, Transform and Load. Extract does the procedure involved in perusing information from a database. Transform handles the changing over of information into a configuration that could be proper for announcing an investigation. While the load does the way toward composing the information into the objective database. Many organization is seeking for candidates well – versed in ETL Testing to help them grow. Thus, if you are seeking for a job in the automation test field, then the below ETL Testing Interview Questions can help you prepare.

The article also contains a list of features along with pros and cons, that will give you an overview of the topic, therefore, enhancing your knowledge. This article will help not only beginners but also advanced candidates who want to succeed in their professional career.

Download ETL Testing Interview Questions PDF

Totaled information is stacked into the EDW after it is populated in operational information store or ODS. Fundamentally ODS is additionally semi DWH helping the investigation of business information. Information industriousness period in ODS is as a rule in the scope of 30-45 days and not more.

Regression testing is the point at which we make changes to information change and accumulation principles to include another usefulness and help the analyzer to discover new errors. The bugs that show up in information which comes in Regression testing is also known as Regression.

ETL Validator is an information testing device that significantly improves the testing of Data Integration, Data Warehouse, and Data Migration ventures. It utilizes our licensed ELV engineering to Extract, Load and Validate information from information sources. For example, databases, level documents, BI frameworks, Hadoop and XML.

A data source view permits to characterize the social pattern, which will be utilized in the investigation of administrations databases. As opposed to straightforwardly from information source items, measurements, and 3D squares are made from information source sees.

The following operations are used in ETL systems concentrates the information from your transnational framework, which can be an Oracle, Microsoft, or some other social database, Transforms the information by performing information purifying tasks. A heap is a procedure of composing the information into the objective database.

In Hash partitioning, the Informatica server would apply for a hash work to segment keys to amass information among the segments. It is utilized to guarantee the preparing of gathering of columns with a similar dividing key in the same segment.

A layout of the file with fixed length ought to incorporate the record name, where the field starts; its length; and its information type (typically content or number). In some cases, the end position is provided, while for cases when that it is not provided, you need to figure the end position of each field dependent on its starting position and length in the event that it is required by your ETL instrument.

In most ETL apparatuses, you in all likelihood need to physically enter the document format of the level record once. After the design is entered, the apparatus recalls the format and expects that equivalent format each time it interfaces with the genuine level record. On the off chance that the document design changes or the information move off its appointed positions, the ETL procedure must be modified to fall flat.

When preparing fixed length level documents, endeavor to approve that the places of the information in the record are exact. A fast check to approve the positions is to test any date (or time) field to ensure it is a substantial date. On the off chance that the positions are moved, the date field in all likelihood contains alpha characters or unreasonable numbers. Different fields with unmistakable spaces can be tried similarly. XML offers increasingly solid approval capacities. On the off chance that information approval or consistency is an issue, endeavor to persuade the information supplier to convey the information in XML design.

Impact analysis looks at the metadata related with an article (for this situation a table or section) and figures out what is influenced by an adjustment in its structure or substance. Evolving information arranging articles can break forms that are essential to appropriately stacking the information distribution center. Permitting specially appointed changes to information arranging objects is adverse to the achievement of your task.

When a table is made in the arranging region, you should play out an effective investigation before any progressions are made to it. Numerous ETL apparatus merchants give sway investigation usefulness. Yet this usefulness is regularly neglected amid the ETL item verification of-idea since it is a back-room work and not by any stretch of the imagination vital until the information stockroom is ready for action and starts to advance.

  • With the power associate alternative, you separate SAP information utilizing Informatica
  • Introduce and design the PowerConnect apparatus
  • Import the source into the Source Analyzer. Among Informatica and SAP PowerConnect go about as a gateway. The subsequent stage is to produce the ABAP code for the mapping then no one but Informatica can pull information from SAP
  • To associate and import sources from outside frameworks PowerConnect is utilized.

During Round Robin partitioning, the information is equitably circulated by the Informatica among every one of the segments. It is utilized when the quantity of columns in the procedure in every one of the segments is about the equivalent.

In the event that arrival port just one, at that point we can go for detached. More than one return port is preposterous with Unconnected. In the event that more than one return port, at that point, go you should use Connected Looks – ups.

On the off chance that you require dynamic reserve that is where your information will change progressively, at that point you can go for associated query. On the off chance that your information is static where your information won't change when the session loads you can go for Unconnected Look – ups.

Data profiling is an efficient examination of the quality, degree, and setting of an information source to enable an ETL framework to be assembled. At one outrageous, an extremely spotless information source that has been all around kept up before it touches base at the information distribution center requires insignificant change and human mediation to stack legitimately into conclusive measurement tables and actuality tables.

In ETL, the initial load is the procedure for populating all information warehousing tables for the absolute first time. While on the other hand, full load implies that when the information is stacked for the very first time every set record are stacked at a stretch, contingent upon its volume. It would delete all substance in the table and would reload new information.

The typical three-level hierarchy of an ETL cycle is composed of the following:

  • Staging Layer − This layer is utilized to store the information extricated from various source information frameworks.
  • Data Integration Layer − This layer changes the information from the organizing layer and moves the information to a database, where the information is masterminded into various leveled gatherings, frequently called measurements, and into actualities and total realities. The blend of realities and measurements tables in a DW framework is known as an outline.
  • Access Layer − This layer is utilized by end-clients to recover the information for explanatory detailing.

The various testing procedures included in ETL include the following:

  • Confirm whether the information is changing accurately as per business prerequisites
  • Confirm that the anticipated information is stacked into the information stockroom with no truncation and information misfortune
  • Ensure that ETL application reports invalid information and replaces with default esteems
  • Ensure that information loads at expected time allotment to improve versatility and execution.