Datastage 8.1 interview questions and answers pdf
File Name: datastage 8.1 interview questions and answers .zip
- Top DataStage Interview Questions and Answers
- DataStage Interview Questions
- Datastage Interview Questions
Coursejet provides top 50 Docker Interview Questions and Answers for aspirants to crack their Docker interview. Here, we have compiled all the important topics of Docker like Docker container, Docker object, Docker architecture, Docker image, repository, hypervisor, virtualization, containerization, Docker registry, Docker file, Kubernetes and more. From beginners to experts, everyone can excel their career using these top 50 Docker interview questions and answers.
Top DataStage Interview Questions and Answers
Consists of fields with numeric values. Dimension table - Table with Unique Primary Key. Load - Data should be first loaded into dimensional table. Based on the primary key values in dimensional table, the data should be loaded into Fact table. What is the default cache size? How do you change the cache size if needed? Default cache size is MB. We can increase it by going into Datastage Administrator and selecting the Tunable Tab and specify the cache size over there. What are types of Hashed File?
A Hashed File is classified broadly into 2 types. Dynamic files do not perform as well as a well, designed static file, but do perform better than a badly designed one. When creating a dynamic file you can specify the following Although all of these have default values By Default Hashed file is "Dynamic - Type Random 30 D" 4. What does a Config File in parallel extender consist of?
A Config file consists of the following. In a Hashed File, the size of the file keeps changing randomly. If the size of the file increases it is called as "Modulus". If the size of the file decreases it is called as "Splitting". What are Stage Variables, Derivations and Constants?
Stage Variable - An intermediate processing variable that retains value during read and doesn? Derivation - Expression that specifies value to be passed on to the target column. Constant - Conditions that are either true or false that specifies flow of data with a link. Types of views in Datastage Director? Types of Parallel Processing? A Parallel Processing is broadly classified into 2 types. Orchestrate Vs Datastage Parallel Extender?
Importance of Surrogate Key in Data warehousing? Most importance of using it is it is independent of underlying database. Surrogate Key is not affected by the changes going on with a database.
How to run a Shell Script within the scope of a Data stage job? How to handle Date conversions in Datastage? A We use a "Iconv" function - Internal Conversion. A Using "dsjob" command as follows. Functionality of Link Partitioner and Link Collector? Link Partitioner: It actually splits data into various partitions or data flows using various partition methods.
Link Collector: It collects the data coming from partitions, merges it into a single data flow and loads to target. Types of Dimensional Modeling? A Dimensional modeling is again sub divided into 2 types. Denormalized form. More normalized form. Differentiate Primary Key and Partition Key?
Primary Key is a combination of unique and not null. It can be a collection of key values called as composite primary key. Partition Key is a just a part of Primary Key. There are several methods of partition like Hash, DB2, and Random etc.
While using Hash partition we specify the Partition Key. Differentiate Database data and Data warehouse data? Containers Usage and Types? Container is a collection of stages used for the purpose of Reusability. There are 2 types of Containers. Plug-In: a Good Performance. Only one database c Cannot handle Stored Procedures. Dimension Modelling types along with their significance Data Modelling is Broadly classified into 2 types. Server Component. Client Component:?
Data Stage Administrator. Data Stage Manager? Data Stage Designer? Data Stage Director Server Components:? Data Stage Engine? Meta Data Repository? Contains set of properties We can set the buffer size by default MB We can increase the buffer size. We can set the Environment Variables. In tunable we have in process and inter-process In-process? Data read in sequentially Inter-process? It reads the data as it comes. It just interfaces to metadata.
We can import table definitions. We can export the Data stage components in. We can compile the job. We can run the job. We can declare stage variable in transform, we can call routines, transform, macros, functions. We can write constraints. Data Stage Director: We can run the jobs. We can schedule the jobs. Schedule can be done daily, weekly, monthly, quarterly We can monitor the jobs.
We can release the jobs. Q 23 What is Meta Data Repository? Meta Data is a data about the data. It also contains? Query statistics? ETL statistics? Business subject area? Source Information? Target Information?
Source to Target mapping Information. Q 24 What is Data Stage Engine? It is a JAVA engine running at the background. Q 25 What is Dimensional Modeling? Dimensional Modeling is a logical design technique that seeks to present the data in a standard framework that is, intuitive and allows for high performance access. Q 26 What is Star Schema? Star Schema is a de-normalized multi-dimensional model. It contains centralized fact tables surrounded by dimensions table.
Dimension Table: It contains a primary key and description about the fact table. Fact Table: It contains foreign keys to the dimension tables, measures and aggregates. Q 27 What is surrogate Key? We can store up to 2 billion record.
Q 28 Why we need surrogate key? It is used for integrating the data may help better for primary key. Index maintenance, joins, table size, key updates, disconnected inserts and partitioning.
DataStage Interview Questions
A DWH is a RDBMS which is especially designed for analysis of the business and decisons making to achive the business goals but not for the business transformation. A single, complete and consistent store of data obtained from a variety of different sources made available to end users in a what they can understand and use in a business context. A Data ware house is a relational Database Management System which is specifically design for Business analysis and making decisions to achieve the business goal. A data ware house is design to support decision system dss. It is also called historical database. Because it will keep all the historical information.
Consists of fields with numeric values. Dimension table - Table with Unique Primary Key. Load - Data should be first loaded into dimensional table. Based on the primary key values in dimensional table, the data should be loaded into Fact table. What is the default cache size?
Define Data Stage? Ans: A data stage is basically a tool that is used to design, develop and execute various applications to fill multiple tables in data warehouse or data marts. It is a program for Windows servers that extracts data from databases and change them into data warehouses. Explain how a source file is populated? Name the command line functions to import and export the DS jobs? Ans: To import the DS jobs, dsimport. What is the difference between Datastage 7.
Datastage Interview Questions
A list of top frequently asked DataStage Interview Questions and answers are given below. DataStage is one of the most powerful ETL tools. It comes with the feature of graphical visualizations for data integration. It extracts, transforms, and loads data from source to the target.
Scenario: Objective: The interface aims to consolidate the round-trip ticket cost of passengers. Description: The source file is a. There are many repetitions in the source file Reverse Duplication: eg.
In DataStage interview questions, you will get all the information you need to clear your DataStage interviews with care. As part of these interview questions and answers, you will learn the difference between DataStage and Informatica, DataStage routine, join, merge, and lookup stages comparison, quality state, job control, performance tuning of jobs, repository table, data type conversions, and more. DataStage is a commonly used ETL tool in the current market. In this DataStage Interview Questions blog, we have shared an extremely useful set of questions and answers intended for cracking DataStage interviews. Here, we have provided in-depth solutions to the DataStage interview questions that are beneficial for freshers and experienced professionals alike.