Artificial Intelligence
Articles
eBooks
Interview Questions
Videos
Keras
Articles
Create Model using both Sequential and Functional API in Keras
Deep dive into Keras - Convolution Neural Network (CNN)
Deep Dive into Modules provided by Keras Library
Detail understanding of Keras Layers
Detailed understanding of Keras applications
Detailed understanding of the Keras Model compilation process
How Keras help in Deep Learning and Architecture of Keras Library
How to write a simple MPL based Artificial Neural Network to perform regression prediction.
Keras Backend Implementations and overview of Deep Learning
Overview of Deep Learning Library Keras and How to install Keras Library on your machine
Write a simple Long Short-Term Memory (LSTM) based RNN to do sequence analysis
eBooks
Interview Questions
Videos
Create Model using both Sequential and Functional API in Keras
Deep dive into Keras - Convolution Neural Network (CNN)
Deep Dive into Modules provided by Keras Library
Detail understanding of Keras Layers
Detailed understanding of Keras applications
Detailed understanding of the Keras Model compilation process
How Keras help in Deep Learning and Architecture of Keras Library
How to write a simple MPL based Artificial Neural Network to perform regression prediction.
Keras Backend Implementations and overview of Deep Learning
Overview of Deep Learning Library Keras and How to install Keras Library on your machine
Write a simple Long Short-Term Memory (LSTM) based RNN to do sequence analysis
Tensor Flow
Articles
Concept of Agents and Environments in AI
Hidden Layer Perceptron in TensorFlow
Multi-layer Perceptron in TensorFlow
Concept of Fuzzy Logic Systems
Deep Dive into TensorFlow Playground
Difference between TensorFlow and Keras
Difference between TensorFlow and PyTorch
Difference between TensorFlow and Theano
How to Install TensorFlow Through pip in Windows
Idea of Intelligence and components of Intelligence
Implementation of Neural Network in TensorFlow
Linear Regression in TensorFlow
Machine Learning and Deep Learning
How to Install TensorFlow through Anaconda
Introduction of Convolutional Neural Network in TensorFlow
Long short-term memory (LSTM) RNN in TensorFlow
What are Artificial Neural Networks?
Working of Convolutional Neural Network
Advantages and Disadvantages of TensorFlow
Architecture of TensorFlow explained
AI - Popular Search Algorithms
Artificial Intelligence - Research Areas
Artificial Neural Network in TensorFlow
CIFAR-10 and CIFAR-100 Dataset in TensorFlow
Classification of Neural Network in TensorFlow
TensorFlow Single and Multiple GPU
TensorFlow Security and TensorFlow Vs Caffe
Style Transferring in TensorFlow
Single Layer Perceptron in TensorFlow
Robotics in Artificial Intelligence
Recurrent Neural Network (RNN) in TensorFlow
eBooks
Interview Questions
Videos
Concept of Agents and Environments in AI
Hidden Layer Perceptron in TensorFlow
Multi-layer Perceptron in TensorFlow
Concept of Fuzzy Logic Systems
Deep Dive into TensorFlow Playground
Difference between TensorFlow and Keras
Difference between TensorFlow and PyTorch
Difference between TensorFlow and Theano
How to Install TensorFlow Through pip in Windows
Idea of Intelligence and components of Intelligence
Implementation of Neural Network in TensorFlow
Linear Regression in TensorFlow
Machine Learning and Deep Learning
How to Install TensorFlow through Anaconda
Introduction of Convolutional Neural Network in TensorFlow
Long short-term memory (LSTM) RNN in TensorFlow
What are Artificial Neural Networks?
Working of Convolutional Neural Network
Advantages and Disadvantages of TensorFlow
Architecture of TensorFlow explained
AI - Popular Search Algorithms
Artificial Intelligence - Research Areas
Artificial Neural Network in TensorFlow
CIFAR-10 and CIFAR-100 Dataset in TensorFlow
Classification of Neural Network in TensorFlow
TensorFlow Single and Multiple GPU
TensorFlow Security and TensorFlow Vs Caffe
Style Transferring in TensorFlow
Single Layer Perceptron in TensorFlow
Robotics in Artificial Intelligence
Recurrent Neural Network (RNN) in TensorFlow
Data Science Introduction and How to set up python
How to use Hibernate Query Language
Handling Arrays and Strings in PHP
Cookies and Sessions Handling in PHP
Cookies and Sessions Handling in PHP
Concept of Agents and Environments in AI
Hidden Layer Perceptron in TensorFlow
Multi-layer Perceptron in TensorFlow
Concept of Fuzzy Logic Systems
Deep Dive into TensorFlow Playground
Difference between TensorFlow and Keras
Difference between TensorFlow and PyTorch
Difference between TensorFlow and Theano
How to Install TensorFlow Through pip in Windows
Idea of Intelligence and components of Intelligence
Implementation of Neural Network in TensorFlow
Linear Regression in TensorFlow
Machine Learning and Deep Learning
How to Install TensorFlow through Anaconda
Introduction of Convolutional Neural Network in TensorFlow
Long short-term memory (LSTM) RNN in TensorFlow
What are Artificial Neural Networks?
Working of Convolutional Neural Network
Advantages and Disadvantages of TensorFlow
Architecture of TensorFlow explained
AI - Popular Search Algorithms
Artificial Intelligence - Research Areas
Artificial Neural Network in TensorFlow
CIFAR-10 and CIFAR-100 Dataset in TensorFlow
Classification of Neural Network in TensorFlow
Create Model using both Sequential and Functional API in Keras
Deep dive into Keras - Convolution Neural Network (CNN)
Deep Dive into Modules provided by Keras Library
Detail understanding of Keras Layers
Detailed understanding of Keras applications
Detailed understanding of the Keras Model compilation process
How Keras help in Deep Learning and Architecture of Keras Library
How to write a simple MPL based Artificial Neural Network to perform regression prediction.
Keras Backend Implementations and overview of Deep Learning
Overview of Deep Learning Library Keras and How to install Keras Library on your machine
Write a simple Long Short-Term Memory (LSTM) based RNN to do sequence analysis
Model Evaluation and Model Prediction in Keras
How to create our own Customized Layer in Keras Library
TensorFlow Single and Multiple GPU
TensorFlow Security and TensorFlow Vs Caffe
Style Transferring in TensorFlow
Single Layer Perceptron in TensorFlow
Robotics in Artificial Intelligence
Recurrent Neural Network (RNN) in TensorFlow
Overview of Artificial Intelligence and its Application
Natural Language Processing in AI
Test Article Main differences between Selenium RC and Selenium WebDriver - Dont Delete
Top Artificial Intelligence Interview Questions and Answers
Top Oracle DBA Interview Questions and Answers
Basics of Splunk and Installation of Splunk Environment
Microsoft Azure Solutions Architect Certification Exam Questions (AZ-300 & AZ-301)
Best Approach for Storing data to AWS DynamoDB and S3 – AWS Implementation
Maintain High Availability in AWS with anticipated Additional Load
BI and Visualization
Articles
eBooks
Videos
Cognos Analytics
Articles
Perform Report Operations in IBM Cognos
Introduction to IBM Cognos and its Components and Services
How to open, create, save, run and print report in Cognos
How to open, create and save Analysis in Analysis Studio in Cognos
How to create report in Report Studio
How to create List and CrossTab Report in Cognos
How to create a package using Cognos
Filters and Custom Calculations in Cognos
Data Warehouse Schemas, ETL and Reporting Tools
Cognos Studios and other capabilities
eBooks
Videos
Perform Report Operations in IBM Cognos
Introduction to IBM Cognos and its Components and Services
How to open, create, save, run and print report in Cognos
How to open, create and save Analysis in Analysis Studio in Cognos
How to create report in Report Studio
How to create List and CrossTab Report in Cognos
How to create a package using Cognos
Filters and Custom Calculations in Cognos
Data Warehouse Schemas, ETL and Reporting Tools
Cognos Studios and other capabilities
Cognos - Relationships in Metadata Model
Top Tableau Desktop Interview Questions and Answers
Top Tableau Server Interview Questions and Answers
Top Power BI Interview Questions and Answers
Top Cognos TM1 Interview Questions and Answers
Cognos TM1
eBooks
Interview Questions
Videos
Microsoft Excel
Articles
How to Merge & Wrap Cells, Borders and Shades and Apply Formatting in Excel
BackStage View and Explore Window in Excel
Creating Formulas, Copying Formulas in Excel
Data Sorting and Using Ranges in Excel
Data Tables and Pivot Tables in Excel
Excel Fill Handle and Excel If Function
Freeze Panes and Conditional Format in Excel
Header and Footer, Page Break and Set Background in Excel
How to add Graphics and Perform Cross-Referencing in Excel
How to Create and Copy Worksheet in Excel
How to Create Worksheets in Excel
How to enter values and move around in Excel
How to Insert Comments and Add Text Box in Excel
How to Open, Close, Delete and Hide Worksheet in Excel
How to Perform Copy & Paste, Find & Replace in Excel
Using Styles, Themes and Templates in Excel
Using Functions and Built-in Functions in Excel
Translate Worksheet and workbook Security in Excel
Simple Charts and Pivot Charts in Excel
Sheet Options, Adjust Margins and Page Orientation in Excel
Printing Worksheets and Email workbooks in Excel
Perform Spell Check, Zoom In-Out and Use Special Symbols in Excel
How to use COUNT, COUNTIF, and COUNTIFS Function and Advanced If in Excel
How to Undo Changes, Setting Cell and Fonts, Text Decoration in Excel
How to Select, Insert, Delete and Move Data in Excel
How to Rotate Cells, Setting Colors and Text Alignment in Excel
eBooks
Interview Questions
Videos
How to Merge & Wrap Cells, Borders and Shades and Apply Formatting in Excel
BackStage View and Explore Window in Excel
Creating Formulas, Copying Formulas in Excel
Data Sorting and Using Ranges in Excel
Data Tables and Pivot Tables in Excel
Excel Fill Handle and Excel If Function
Freeze Panes and Conditional Format in Excel
Header and Footer, Page Break and Set Background in Excel
How to add Graphics and Perform Cross-Referencing in Excel
How to Create and Copy Worksheet in Excel
How to Create Worksheets in Excel
How to enter values and move around in Excel
How to Insert Comments and Add Text Box in Excel
How to Open, Close, Delete and Hide Worksheet in Excel
How to Perform Copy & Paste, Find & Replace in Excel
Using Styles, Themes and Templates in Excel
Using Functions and Built-in Functions in Excel
Translate Worksheet and workbook Security in Excel
Simple Charts and Pivot Charts in Excel
Sheet Options, Adjust Margins and Page Orientation in Excel
Printing Worksheets and Email workbooks in Excel
Perform Spell Check, Zoom In-Out and Use Special Symbols in Excel
How to use COUNT, COUNTIF, and COUNTIFS Function and Advanced If in Excel
How to Undo Changes, Setting Cell and Fonts, Text Decoration in Excel
How to Select, Insert, Delete and Move Data in Excel
How to Rotate Cells, Setting Colors and Text Alignment in Excel
OBIEEE
Articles
Concept of Testing Repository in OBIEE
Understanding Schemas in OBIEE
Overview of Oracle Business Intelligence Edition (OBIEE)
Multiple Logical Table Sources, Calculation Measures and Dimension Hierarchies
Level-Based Measures and Aggregates in OBIEE
Deep Dive into Repositories in OBIEE
eBooks
Interview Questions
Videos
Concept of Testing Repository in OBIEE
Understanding Schemas in OBIEE
Overview of Oracle Business Intelligence Edition (OBIEE)
Multiple Logical Table Sources, Calculation Measures and Dimension Hierarchies
Level-Based Measures and Aggregates in OBIEE
Deep Dive into Repositories in OBIEE
Pentaho
Articles
User interfaces available in Pentaho and their navigation
Overview of Pentaho and How to install Pentaho on your system
How to use the Pentaho Reporting Designer
How to use Grouping in Pentaho
How to use Functions in Reports in Pentaho
How to create Chart Report in Pentaho
eBooks
Interview Questions
Videos
User interfaces available in Pentaho and their navigation
Overview of Pentaho and How to install Pentaho on your system
How to use the Pentaho Reporting Designer
How to use Grouping in Pentaho
How to use Functions in Reports in Pentaho
How to create Chart Report in Pentaho
Power BI
Articles
Visualization Options in Power BI
Power BI Data Sources and How to connect with them
Power BI - Supported Data Sources
Power BI - Comparison with Other BI Tools
Overview of Power BI Embedded, Power BI Gateway and Power BI Report Server
Overview of Business Intelligence (BI) and Power BI
How to use various DAX functions in Power BI
How to Share Power BI Dashboard
How to Integrate Excel in Power BI
How to Download and Install Power BI Desktop
eBooks
Interview Questions
Videos
Visualization Options in Power BI
Power BI Data Sources and How to connect with them
Power BI - Supported Data Sources
Power BI - Comparison with Other BI Tools
Overview of Power BI Embedded, Power BI Gateway and Power BI Report Server
Overview of Business Intelligence (BI) and Power BI
How to use various DAX functions in Power BI
How to Share Power BI Dashboard
How to Integrate Excel in Power BI
How to Download and Install Power BI Desktop
Qlik View
Articles
List Box and Multi Box in QlikView
Navigation Options in QlikView
Overview of Data files (QVD) in QlikView
Processing Web Files in QlikView
Resident Load, Preceding Load and Incremental Load in QlikView
How to create Cross Tables in QlikView
How to create Pie Chart in QlikView
Straight Tables and Pivot Tables in QlikView
Database Connection in QlikView
Dimensions and Measures in QlikView
Usage of Keep Command in QlikView
Using Peek and RangeSum Function in QlikView
Handling Delimited Files in QlikView
Handling Excel Files in QlikView
How to create Bar Chart in QlikView
Using Match and Rank Function in QlikView
Overview of QlikView and How to install QlikView on your machine
Inline Data and Scripting in QlikView
Data Transformation in QlikView
Creating Dashboard in QlikView
Concept of Star Schema and Synthetic Key in QlikView
Concatenation and Master Calendar in QlikView
Column Manipulation in QlikView
eBooks
Interview Questions
Videos
List Box and Multi Box in QlikView
Navigation Options in QlikView
Overview of Data files (QVD) in QlikView
Processing Web Files in QlikView
Resident Load, Preceding Load and Incremental Load in QlikView
How to create Cross Tables in QlikView
How to create Pie Chart in QlikView
Straight Tables and Pivot Tables in QlikView
Database Connection in QlikView
Dimensions and Measures in QlikView
Usage of Keep Command in QlikView
Using Peek and RangeSum Function in QlikView
Handling Delimited Files in QlikView
Handling Excel Files in QlikView
How to create Bar Chart in QlikView
Using Match and Rank Function in QlikView
Overview of QlikView and How to install QlikView on your machine
Inline Data and Scripting in QlikView
Data Transformation in QlikView
Creating Dashboard in QlikView
Concept of Star Schema and Synthetic Key in QlikView
Concatenation and Master Calendar in QlikView
Column Manipulation in QlikView
Circular Reference in QlikView
QLikSense
Articles
Navigating in Qlik Sense Selections
Qlik Sense Conditional Functions
Qlik Sense Counter and Exponential and Logarithmic Functions
Qlik Sense Developer: Roles and Responsibilities
Overview of Gauge Chart in Qlik Sense
Qlik Sense Advantages and Limitations
Qlik Sense Architecture Components
Qlik Sense Capabilities for people, Groups and Organizations
What is Qlik Sense Pivot Table?
Ways of Qlik Sense Collaboration
Qlik Sense Formatting Functions
Qlik Sense distribution and Trigonometric and HyperBolic Functions
Qlik Sense Mapping and Logical Functions
Qlik Sense Financial Functions
Types of Qlik Sense Aggregation Functions
Tableau vs Qlik Sense vs Power BI
Significance of Text and Image in Qlik Sense
Set Analysis and Set Expressions in Qlik Sense
QlikView Vs Qlik Sense: Overview
Using a Scatter Plot in Qlik Sense
Types of Operators in Qlik Sense
Qlik Sense System Requirements
Treemap Visualization in Qlik Sense
Qlik Sense Interpretation Functions
Modulo Functions in Qlik Sense
Key Performance Indicators (KPI) in Qlik Sense
Introduction to Qlik Sense Mashup
How to Manage Content and Resources in Qlik Management Console
How to Interact With Qlik Sense Visualizations?
How to Interact with Qlik Sense interface
How to create Qlik Sense Application
General Numeric Functions in Qlik Sense
Concept of Social Engineering Attacks and Cross-Site Scripting
Components of Qlik Sense Desktop
eBooks
Interview Questions
Videos
Navigating in Qlik Sense Selections
Qlik Sense Conditional Functions
Qlik Sense Counter and Exponential and Logarithmic Functions
Qlik Sense Developer: Roles and Responsibilities
Overview of Gauge Chart in Qlik Sense
Qlik Sense Advantages and Limitations
Qlik Sense Architecture Components
Qlik Sense Capabilities for people, Groups and Organizations
What is Qlik Sense Pivot Table?
Ways of Qlik Sense Collaboration
Qlik Sense Formatting Functions
Qlik Sense distribution and Trigonometric and HyperBolic Functions
Qlik Sense Mapping and Logical Functions
Qlik Sense Financial Functions
Types of Qlik Sense Aggregation Functions
Tableau vs Qlik Sense vs Power BI
Significance of Text and Image in Qlik Sense
Set Analysis and Set Expressions in Qlik Sense
QlikView Vs Qlik Sense: Overview
Using a Scatter Plot in Qlik Sense
Types of Operators in Qlik Sense
Qlik Sense System Requirements
Treemap Visualization in Qlik Sense
Qlik Sense Interpretation Functions
Modulo Functions in Qlik Sense
Key Performance Indicators (KPI) in Qlik Sense
Introduction to Qlik Sense Mashup
How to Manage Content and Resources in Qlik Management Console
How to Interact With Qlik Sense Visualizations?
How to Interact with Qlik Sense interface
How to create Qlik Sense Application
General Numeric Functions in Qlik Sense
Concept of Social Engineering Attacks and Cross-Site Scripting
Components of Qlik Sense Desktop
SSAS
eBooks
Interview Questions
Videos
SSIS
eBooks
Interview Questions
Videos
SSRS
eBooks
Interview Questions
Videos
Tableau Desktop
Articles
How to create Pareto Chart in Tableau
How to create Gantt Chart in Tableau
How to create Dual Axis Chart, Box Plot and Heat Map in Tableau
How to create Crosstab and Motion Chart in Tableau
How to create Bump and Bubble Chart in Tableau
How to create Bar, Line and Pie Chart in Tableau
How to Build Hierarchy and Groups in Tableau
Filter Operations and Extract Filters in Tableau
Different Tools of Tableau and Tableau Architecture
Understanding Tableau Navigation and Data Terminology
Understanding Tableau Desktop Workspace
Data Window, Data Types, Data Aggregation and File Types in Tableau
Top 10 Data Visualization Tools
Tableau Quick and Context Filters
Perform Table Calculations in Tableau
Perform Data Sorting in Tableau
Perform Calculation and Operators and Functions in Tableau
Overview of Tableau and Data Visualization
How to perform Numeric, String and Date Calculations in Tableau
How to Join Data in Tableau using multiple sources
Condition Filters, Data Source and Top Filters in Tableau
Comparison of Tableau and Power BI
How to install Tableau on your system
How to create Waterfall, Bullet and Area Chart in Tableau
eBooks
Interview Questions
Videos
How to create Pareto Chart in Tableau
How to create Gantt Chart in Tableau
How to create Dual Axis Chart, Box Plot and Heat Map in Tableau
How to create Crosstab and Motion Chart in Tableau
How to create Bump and Bubble Chart in Tableau
How to create Bar, Line and Pie Chart in Tableau
How to Build Hierarchy and Groups in Tableau
Filter Operations and Extract Filters in Tableau
Different Tools of Tableau and Tableau Architecture
Understanding Tableau Navigation and Data Terminology
Understanding Tableau Desktop Workspace
Data Window, Data Types, Data Aggregation and File Types in Tableau
Top 10 Data Visualization Tools
Tableau Quick and Context Filters
Perform Table Calculations in Tableau
Perform Data Sorting in Tableau
Perform Calculation and Operators and Functions in Tableau
Overview of Tableau and Data Visualization
How to perform Numeric, String and Date Calculations in Tableau
How to Join Data in Tableau using multiple sources
Condition Filters, Data Source and Top Filters in Tableau
Comparison of Tableau and Power BI
How to install Tableau on your system
How to create Waterfall, Bullet and Area Chart in Tableau
Tableau Server
TIBCO BW
eBooks
Videos
How to Merge & Wrap Cells, Borders and Shades and Apply Formatting in Excel
How to Clone Repository in Git
Navigating in Qlik Sense Selections
Qlik Sense Conditional Functions
Qlik Sense Counter and Exponential and Logarithmic Functions
Qlik Sense Developer: Roles and Responsibilities
Overview of Gauge Chart in Qlik Sense
Qlik Sense Advantages and Limitations
Qlik Sense Architecture Components
Qlik Sense Capabilities for people, Groups and Organizations
What is Qlik Sense Pivot Table?
Ways of Qlik Sense Collaboration
Qlik Sense Formatting Functions
Qlik Sense distribution and Trigonometric and HyperBolic Functions
Qlik Sense Mapping and Logical Functions
Qlik Sense Financial Functions
Types of Qlik Sense Aggregation Functions
Tableau vs Qlik Sense vs Power BI
Significance of Text and Image in Qlik Sense
Set Analysis and Set Expressions in Qlik Sense
QlikView Vs Qlik Sense: Overview
Using a Scatter Plot in Qlik Sense
Types of Operators in Qlik Sense
Qlik Sense System Requirements
List Box and Multi Box in QlikView
Navigation Options in QlikView
Overview of Data files (QVD) in QlikView
Processing Web Files in QlikView
Resident Load, Preceding Load and Incremental Load in QlikView
How to create Cross Tables in QlikView
How to create Pie Chart in QlikView
Straight Tables and Pivot Tables in QlikView
Database Connection in QlikView
Dimensions and Measures in QlikView
Usage of Keep Command in QlikView
Using Peek and RangeSum Function in QlikView
Handling Delimited Files in QlikView
Handling Excel Files in QlikView
How to create Bar Chart in QlikView
How to create Pareto Chart in Tableau
BackStage View and Explore Window in Excel
Creating Formulas, Copying Formulas in Excel
Treemap Visualization in Qlik Sense
Overview of SSRS and its Architecture
Overview of SSIS and why SSIS is required
Introduction to TIBCO Business Works (TIBCO BW)
Overview of Tableau Server and How to install it
Overview of SSAS and its Architecture
Using Match and Rank Function in QlikView
Data Sorting and Using Ranges in Excel
Data Tables and Pivot Tables in Excel
Excel Fill Handle and Excel If Function
Freeze Panes and Conditional Format in Excel
Overview of QlikView and How to install QlikView on your machine
Header and Footer, Page Break and Set Background in Excel
Inline Data and Scripting in QlikView
How to add Graphics and Perform Cross-Referencing in Excel
How to Create and Copy Worksheet in Excel
How to Create Worksheets in Excel
How to enter values and move around in Excel
How to Insert Comments and Add Text Box in Excel
How to Open, Close, Delete and Hide Worksheet in Excel
How to Perform Copy & Paste, Find & Replace in Excel
Qlik Sense Interpretation Functions
Setting Up Distributed Servers in Tableau Server
Concept of Testing Repository in OBIEE
Understanding Schemas in OBIEE
Modulo Functions in Qlik Sense
Key Performance Indicators (KPI) in Qlik Sense
Introduction to Qlik Sense Mashup
Data Transformation in QlikView
Creating Dashboard in QlikView
Concept of Star Schema and Synthetic Key in QlikView
How to Manage Content and Resources in Qlik Management Console
Concatenation and Master Calendar in QlikView
Column Manipulation in QlikView
How to Interact With Qlik Sense Visualizations?
How to Interact with Qlik Sense interface
Circular Reference in QlikView
Aggregate Functions in QlikView
Perform Report Operations in IBM Cognos
Overview of Oracle Business Intelligence Edition (OBIEE)
Introduction to IBM Cognos and its Components and Services
How to open, create, save, run and print report in Cognos
Multiple Logical Table Sources, Calculation Measures and Dimension Hierarchies
How to open, create and save Analysis in Analysis Studio in Cognos
Level-Based Measures and Aggregates in OBIEE
How to create report in Report Studio
Deep Dive into Repositories in OBIEE
Concept of Data Warehouse and Dimension Modelling
How to create List and CrossTab Report in Cognos
How to create a package using Cognos
How to create Qlik Sense Application
General Numeric Functions in Qlik Sense
Concept of Social Engineering Attacks and Cross-Site Scripting
Business and Presentation Layer of OBIEE explained
Filters and Custom Calculations in Cognos
Components of Qlik Sense Desktop
BI Tools for giant Data Visualization
Data Warehouse Schemas, ETL and Reporting Tools
Aggregation Functions in Qlik Sense
How to create Gantt Chart in Tableau
How to create Dual Axis Chart, Box Plot and Heat Map in Tableau
How to create Crosstab and Motion Chart in Tableau
How to create Bump and Bubble Chart in Tableau
How to create Bar, Line and Pie Chart in Tableau
Introduction to Cognos TM1 Perspective
How to Setup TM1 Application Server
How to Build Hierarchy and Groups in Tableau
How to Configure Security in TM1
Concept of Dimensions in Cognos TM1
Filter Operations and Extract Filters in Tableau
Cognos TM1 Installation and Configuration
Different Tools of Tableau and Tableau Architecture
Understanding Tableau Navigation and Data Terminology
Understanding Tableau Desktop Workspace
Data Window, Data Types, Data Aggregation and File Types in Tableau
Top 10 Data Visualization Tools
Tableau Quick and Context Filters
Perform Table Calculations in Tableau
Perform Data Sorting in Tableau
Perform Calculation and Operators and Functions in Tableau
Overview of Tableau and Data Visualization
How to perform Numeric, String and Date Calculations in Tableau
How to Join Data in Tableau using multiple sources
Condition Filters, Data Source and Top Filters in Tableau
Comparison of Tableau and Power BI
How to install Tableau on your system
How to create Waterfall, Bullet and Area Chart in Tableau
How to create Tree Maps and Heat Maps in Tableau
How to create Scatter Plot and Histogram Chart in Tableau
User interfaces available in Pentaho and their navigation
Overview of Pentaho and How to install Pentaho on your system
How to use the Pentaho Reporting Designer
How to use Grouping in Pentaho
How to use Functions in Reports in Pentaho
How to create Chart Report in Pentaho
How to add Page Footer Fields in Pentaho
Formatting Report Elements in Pentaho Reporting Designer
Using Styles, Themes and Templates in Excel
Using Functions and Built-in Functions in Excel
Translate Worksheet and workbook Security in Excel
Simple Charts and Pivot Charts in Excel
Sheet Options, Adjust Margins and Page Orientation in Excel
Printing Worksheets and Email workbooks in Excel
Perform Spell Check, Zoom In-Out and Use Special Symbols in Excel
How to use COUNT, COUNTIF, and COUNTIFS Function and Advanced If in Excel
How to Undo Changes, Setting Cell and Fonts, Text Decoration in Excel
How to Select, Insert, Delete and Move Data in Excel
How to Rotate Cells, Setting Colors and Text Alignment in Excel
How to Perform Data Validation and Data Filtering in Excel
Cognos Studios and other capabilities
Cognos - Relationships in Metadata Model
Visualization Options in Power BI
Power BI Data Sources and How to connect with them
Power BI - Supported Data Sources
Power BI - Comparison with Other BI Tools
Overview of Power BI Embedded, Power BI Gateway and Power BI Report Server
Overview of Business Intelligence (BI) and Power BI
How to use various DAX functions in Power BI
How to Share Power BI Dashboard
How to Integrate Excel in Power BI
How to Download and Install Power BI Desktop
How to create Power BI Dashboard and Reports
Top Qlik Sense Interview Questions and Answers
Top Microsoft BI Interview Questions and Answers
Top TIBCO Spotfire Interview Questions and Answers
Top OBIEE Interview Questions and Answers
Top Tableau Desktop Interview Questions and Answers
Top Tableau Server Interview Questions and Answers
Top Qlik View Interview Questions and Answers
Top TIBCO Business Works Interview Questions and Answers
Top Oracle Hyperion Interview Questions and Answers
Top Power BI Interview Questions and Answers
Top Pentaho Interview Questions and Answers
Top Cognos TM1 Interview Questions and Answers
Top IBM DataStage Interview Questions and Answers
Top IBM Cognos Analytics Interview Questions and Answers
Best Approach for Storing data to AWS DynamoDB and S3 – AWS Implementation
Maintain High Availability in AWS with anticipated Additional Load
Big Data
eBooks
Videos
Aapche Cassandra
Articles
Deep dive into Cassandra Query Language Collections and user defined data types.
Deep dive into Cassandra Shell Commands
How to Create and Alter Tables in Apache Cassandra
How to Create and Drop Indexes in Apache Cassandra
How to create, alter and drop Keyspaces in Cassandra
How to Drop and Truncate Tables in Apache Cassandra
How to set up Both cqlsh and Java environments to work with Cassandra
How to Perform CRUD ( Create , Read , Update and Delete ) Operations in Table in Apache Cassandra
Introduction to Apache Cassandra, History and Architecture
Overview of How Cassandra Stores its data
Overview of important class in Cassandra and introduction of Cassandra query shell language
eBooks
Interview Questions
Videos
Deep dive into Cassandra Query Language Collections and user defined data types.
Deep dive into Cassandra Shell Commands
How to Create and Alter Tables in Apache Cassandra
How to Create and Drop Indexes in Apache Cassandra
How to create, alter and drop Keyspaces in Cassandra
How to Drop and Truncate Tables in Apache Cassandra
How to set up Both cqlsh and Java environments to work with Cassandra
How to Perform CRUD ( Create , Read , Update and Delete ) Operations in Table in Apache Cassandra
Introduction to Apache Cassandra, History and Architecture
Overview of How Cassandra Stores its data
Overview of important class in Cassandra and introduction of Cassandra query shell language
Apache NiFi
Articles
How to Monitor System statistics using Apache NiFi
Concept of Logging in Apache NiFi
Basic Concepts of Apache NiFi and its Installation
Deep Dive into Apache Nifi – Flow Files, Queues, Process Groups and Labels
Deep dive into Apache NiFi-Processors
Detailed understanding of Apache NiFi -Templates
How to Administer Apache NiFi and Create Flows in Apache NiFi
Understanding Apache NiFi API’s with request and response example
Understanding Apache Nifi Processors Categorization and its relationship
Introduction to Apache NiFi, its History, Features and Architecture
eBooks
Interview Questions
Videos
How to Monitor System statistics using Apache NiFi
Concept of Logging in Apache NiFi
Basic Concepts of Apache NiFi and its Installation
Deep Dive into Apache Nifi – Flow Files, Queues, Process Groups and Labels
Deep dive into Apache NiFi-Processors
Detailed understanding of Apache NiFi -Templates
How to Administer Apache NiFi and Create Flows in Apache NiFi
Understanding Apache NiFi API’s with request and response example
Understanding Apache Nifi Processors Categorization and its relationship
Introduction to Apache NiFi, its History, Features and Architecture
Apache Oozie
eBooks
Interview Questions
Videos
Apache Pig
Articles
Explanation of Apache Pig Group and Cogroup Operators
Detailed Study of Architecture of Apache Pig
Deep Dive into Pig Latin Diagnostic Operators
Deep Dive into Apache Pig Functions: Load & Store, Bag & Tuple, String, Date-time, Math
Apache Pig Basics, Features and Comparison with MapReduce, Hive & SQL and History of Apache Pig
Explanation of Shell and Utility Commands provided by Apache Grunt Shell
How to Install Apache Pig and Configure Pig
How to Load data to Apache Pig from Hadoop File System
How to run Apache Pig Scripts in Batch Mode
How to Store data in Apache Pig using Store Operator
How to use Cross Operator and Union Operator in Pig Latin
How to use Split and Filter Operator in Apache Pig Latin
How to use the Join Operators in Pig Latin
How to use Distinct, For Each, Order By, Limit Operators and Eval Functions in Apache Pig
eBooks
Interview Questions
Videos
Explanation of Apache Pig Group and Cogroup Operators
Detailed Study of Architecture of Apache Pig
Deep Dive into Pig Latin Diagnostic Operators
Deep Dive into Apache Pig Functions: Load & Store, Bag & Tuple, String, Date-time, Math
Apache Pig Basics, Features and Comparison with MapReduce, Hive & SQL and History of Apache Pig
Explanation of Shell and Utility Commands provided by Apache Grunt Shell
How to Install Apache Pig and Configure Pig
How to Load data to Apache Pig from Hadoop File System
How to run Apache Pig Scripts in Batch Mode
How to Store data in Apache Pig using Store Operator
How to use Cross Operator and Union Operator in Pig Latin
How to use Split and Filter Operator in Apache Pig Latin
How to use the Join Operators in Pig Latin
How to use Distinct, For Each, Order By, Limit Operators and Eval Functions in Apache Pig
Apache Spark
Articles
Overview of Scala programming language and How to install Scala on your system
How to Install Apache Spark on your system
How to perform pattern matching in Scala and use of Regex expressions
How to use Functions in Scala programming Language
How to use Collections in Scala
How to use Arrays in Scala Programming Language
How to perform Exception Handling in Scala Language
How to Deploy Spark Application on Cluster
Extractor Object in Scala and how to perform pattern matching using extractors
Details of Data Types and Basic Literals in Scala
Detailed understanding of Operators in Scala Language
Deep dive into File Handling in Scala
Deep dive into Advanced programming in Spark
Basics of Scala Programming Language
Concept of String Manipulation in Scala
Conditional statements and Loop control structures in Scala
Concept of Resilient Distributed Datasets (RDD) in Apache Spark
How to use Classes and Objects in Scala programming
Overview of Apache Spark Framework
Spark Core and implementation of RDD transformations and actions in RDD programming
eBooks
Interview Questions
Videos
Overview of Scala programming language and How to install Scala on your system
How to Install Apache Spark on your system
How to perform pattern matching in Scala and use of Regex expressions
How to use Functions in Scala programming Language
How to use Collections in Scala
How to use Arrays in Scala Programming Language
How to perform Exception Handling in Scala Language
How to Deploy Spark Application on Cluster
Extractor Object in Scala and how to perform pattern matching using extractors
Details of Data Types and Basic Literals in Scala
Detailed understanding of Operators in Scala Language
Deep dive into File Handling in Scala
Deep dive into Advanced programming in Spark
Basics of Scala Programming Language
Concept of String Manipulation in Scala
Conditional statements and Loop control structures in Scala
Concept of Resilient Distributed Datasets (RDD) in Apache Spark
How to use Classes and Objects in Scala programming
Overview of Apache Spark Framework
Spark Core and implementation of RDD transformations and actions in RDD programming
Detailed understanding of Scala Access Modifiers
Apache Sqoop
Articles
How to use the Apache Sqoop Eval and Codegen tool
How to list out the databases and tables of a particular database using Sqoop
How to import data and tables from MySQL to Hadoop HDFS
How to export data back from Hadoop HDFS to RDBMS and Create and maintain the Sqoop jobs
Introduction, Installation and Configuration of Apache Sqoop
eBooks
Interview Questions
Videos
How to use the Apache Sqoop Eval and Codegen tool
How to list out the databases and tables of a particular database using Sqoop
How to import data and tables from MySQL to Hadoop HDFS
How to export data back from Hadoop HDFS to RDBMS and Create and maintain the Sqoop jobs
Introduction, Installation and Configuration of Apache Sqoop
Apache Storm
Articles
Application of Apache Storm Framework in Yahoo Finance
Concept of Cluster Architecture in Apache Storm
Introduction to Apache Storm and Core Concepts of Apache Storm
How to Install Apache Storm framework on your machine
How to implement Mobile Call log Analyzer using Apache Storm
How Apache Storm is used in Twitter
eBooks
Interview Questions
Videos
Application of Apache Storm Framework in Yahoo Finance
Concept of Cluster Architecture in Apache Storm
Introduction to Apache Storm and Core Concepts of Apache Storm
How to Install Apache Storm framework on your machine
How to implement Mobile Call log Analyzer using Apache Storm
How Apache Storm is used in Twitter
Detailed understanding of Workflow of Apache Storm
Hadoop and MapReduce
Articles
Concept of Combiners in Hadoop MapReduce
Concept of MapReduce in BigData
Detailed understanding of Hadoop Architecture and Hadoop Distributed File System (HDFS)
Concept of Partitioner in MapReduce and its implementation using example
Deep dive into Hadoop administration
Deep dive into the MapReduce API
Detailed understanding of Hadoop Distributed File System (HDFS)
Phases of MapReduce Data flow and detailed understanding of Mapreduce API
Overview of YARN and its components and benefits of YARN
Overview of Big Data and Hadoop, Big Data technologies
Implementation of Word Count program using Hadoop MapReduce
Operation of MapReduce in Hadoop framework using Java
Implementation of Character Count program using Hadoop MapReduce
How to set up Hadoop Multi-Node Cluster on a distributed environment
How to perform operations in Hadoop and commands used in Hadoop
How to install Hadoop on your system
eBooks
Videos
Concept of Combiners in Hadoop MapReduce
Concept of MapReduce in BigData
Detailed understanding of Hadoop Architecture and Hadoop Distributed File System (HDFS)
Concept of Partitioner in MapReduce and its implementation using example
Deep dive into Hadoop administration
Deep dive into the MapReduce API
Detailed understanding of Hadoop Distributed File System (HDFS)
Phases of MapReduce Data flow and detailed understanding of Mapreduce API
Overview of YARN and its components and benefits of YARN
Overview of Big Data and Hadoop, Big Data technologies
Implementation of Word Count program using Hadoop MapReduce
Operation of MapReduce in Hadoop framework using Java
Implementation of Character Count program using Hadoop MapReduce
How to set up Hadoop Multi-Node Cluster on a distributed environment
How to perform operations in Hadoop and commands used in Hadoop
How to install Hadoop on your system
How to install Hadoop Framework on your system
How the MapReduce Algorithm works using example
HBase
Articles
Deep dive into HBase architecture
Deep dive into Java Client API for HBase and its associated classes
How to create and List Table in HBase shell
How to create data in an HBase table
How to delete data in Table in HBase
How to enable and disable a Table using HBase shell
How to install HBase and configure on your system
How to make changes to an existing Table and describe it in HBase
How to read data from Table in HBase
How to start HBase interactive shell and how HBase general commands works
How to Stop HBase using Java API
How to update data in Table using HBase Shell
How to verify the existence of a Table and How to Drop a Table in HBase
Overview of HBase, its Advantages, Features and history
Deep dive into HBase Scan, Count and Truncate command and how to achieve security in HBase
Deep dive into HBase Scan, Count and Truncate command and how to achieve security in HBase
eBooks
Interview Questions
Videos
Deep dive into HBase architecture
Deep dive into Java Client API for HBase and its associated classes
How to create and List Table in HBase shell
How to create data in an HBase table
How to delete data in Table in HBase
How to enable and disable a Table using HBase shell
How to install HBase and configure on your system
How to make changes to an existing Table and describe it in HBase
How to read data from Table in HBase
How to start HBase interactive shell and how HBase general commands works
How to Stop HBase using Java API
How to update data in Table using HBase Shell
How to verify the existence of a Table and How to Drop a Table in HBase
Overview of HBase, its Advantages, Features and history
Top Apache HBase Interview Questions and Answers
Deep dive into HBase Scan, Count and Truncate command and how to achieve security in HBase
Deep dive into HBase Scan, Count and Truncate command and how to achieve security in HBase
Hive and Impala
Articles
Concept of Partitioning of table in Hive
Detailed understanding of built-in functions available in Hive
Different Data Types in Hive which are involved in creation of table.
How to Alter the attributes of a table and delete a Table in Hive
How to create a table in Hive and how to insert data into it
How to create and drop a database in Hive
How to create and manage Views and Create and Drop an index in Hive
How to install Hive on your system
How to perform Join operations in Hive Query Language (HQL)
How to use the select statement in Hive Query Language
Introduction to Impala, its features, advantages and disadvantages
How to start Impala Shell and the various options of the shell
How to select a database using Command and select database using Hue Browser in Impala
How to perform changes on a given table and how to delete table in Impala
How to fetch the data from one or more tables in a database and fetch description in Impala
How to download, install and set up Impala in your system
How to create a table in the required database in Impala
How to Create, Alter and Drop a View in Impala
Explanation of Union Clause, With Clause and Distinct Operator in Impala
Explanation of Limit Clause and Offset Clause in Impala
Data Types in Impala Query Language
Detailed understanding of Architecture of Impala
Explanation of Order by Clause, Group by Clause and Having Clause in Impala
How to add new records into an existing table in a database using INSERT in Impala
How to create a database in Impala
eBooks
Videos
Concept of Partitioning of table in Hive
Detailed understanding of built-in functions available in Hive
Different Data Types in Hive which are involved in creation of table.
How to Alter the attributes of a table and delete a Table in Hive
How to create a table in Hive and how to insert data into it
How to create and drop a database in Hive
How to create and manage Views and Create and Drop an index in Hive
How to install Hive on your system
How to perform Join operations in Hive Query Language (HQL)
How to use the select statement in Hive Query Language
Introduction to Impala, its features, advantages and disadvantages
How to start Impala Shell and the various options of the shell
How to select a database using Command and select database using Hue Browser in Impala
How to perform changes on a given table and how to delete table in Impala
How to fetch the data from one or more tables in a database and fetch description in Impala
How to download, install and set up Impala in your system
How to create a table in the required database in Impala
How to Create, Alter and Drop a View in Impala
Explanation of Union Clause, With Clause and Distinct Operator in Impala
Explanation of Limit Clause and Offset Clause in Impala
Data Types in Impala Query Language
Detailed understanding of Architecture of Impala
Explanation of Order by Clause, Group by Clause and Having Clause in Impala
How to add new records into an existing table in a database using INSERT in Impala
How to create a database in Impala
How to drop a database in Impala
Top Apache Impala Interview Questions and Answers
Top Apache Hive Interview Questions and Answers
MongoDB
Articles
Advanced Indexing in MongoDB and Limitation of Indexing in MongoDB
Concept of Capped Collections and Auto-Increment Sequence in MongoDB
Concept of Map Reduce in MongoDB
Concept of Relationships and Database References in MongoDB
Concept of Sharding process and How to create a backup in MongoDB
Data Modelling in MongoDB and How to create and Drop database in MongoDB
Deep dive into Covered Queries in MongoDB and Analyzing queries
Deep dive into Replication process in MongoDB
How to Create and Drop a collection using MongoDB
How to Insert, Update, Delete and Query Document in MongoDB Collection
How to Install MongoDB on your system
How to limit records using MongoDB ad use projection in MongoDB
How to Set up MongoDB JDBC driver
How to sort records in MongoDB and concept of Indexing and Aggregation in MongoDB
How to use Regex Expressions and Text Search in MongoDB
MongoDB Administration using RockMongo and concept of GridFS in MongoDB
Overview of MongoDB, its history and purpose of building MongoDB
Understand NoSQL Databases and MongoDB advantages over Relational DBMS
eBooks
Interview Questions
Videos
Advanced Indexing in MongoDB and Limitation of Indexing in MongoDB
Concept of Capped Collections and Auto-Increment Sequence in MongoDB
Concept of Map Reduce in MongoDB
Concept of Relationships and Database References in MongoDB
Concept of Sharding process and How to create a backup in MongoDB
Data Modelling in MongoDB and How to create and Drop database in MongoDB
Deep dive into Covered Queries in MongoDB and Analyzing queries
Deep dive into Replication process in MongoDB
How to Create and Drop a collection using MongoDB
How to Insert, Update, Delete and Query Document in MongoDB Collection
How to Install MongoDB on your system
How to limit records using MongoDB ad use projection in MongoDB
How to Set up MongoDB JDBC driver
How to sort records in MongoDB and concept of Indexing and Aggregation in MongoDB
How to use Regex Expressions and Text Search in MongoDB
MongoDB Administration using RockMongo and concept of GridFS in MongoDB
Overview of MongoDB, its history and purpose of building MongoDB
Understand NoSQL Databases and MongoDB advantages over Relational DBMS
Splunk
Articles
Deep dive into Splunk Search processing Language (SPL)
How to perform Basic Search in Splunk
How to perform searching using fields in Splunk
How to perform Time Range search in Splunk
How to share and export the search result in Splunk
A Deep Dive into Splunk Web Interface
eBooks
Videos
Deep dive into Splunk Search processing Language (SPL)
How to perform Basic Search in Splunk
How to perform searching using fields in Splunk
How to perform Time Range search in Splunk
How to share and export the search result in Splunk
Top Splunk SIEM Interview Questions and Answers
Top Splunk Interview Questions and Answers
A Deep Dive into Splunk Web Interface
Application of Apache Storm Framework in Yahoo Finance
Deep dive into Cassandra Query Language Collections and user defined data types.
Deep dive into Cassandra Shell Commands
How to Create and Alter Tables in Apache Cassandra
How to Create and Drop Indexes in Apache Cassandra
How to create, alter and drop Keyspaces in Cassandra
How to Drop and Truncate Tables in Apache Cassandra
Concept of Combiners in Hadoop MapReduce
Concept of MapReduce in BigData
Detailed understanding of Hadoop Architecture and Hadoop Distributed File System (HDFS)
Concept of Partitioner in MapReduce and its implementation using example
Deep dive into Hadoop administration
Deep dive into the MapReduce API
Detailed understanding of Hadoop Distributed File System (HDFS)
How to set up Both cqlsh and Java environments to work with Cassandra
How to Perform CRUD ( Create , Read , Update and Delete ) Operations in Table in Apache Cassandra
Explanation of Apache Pig Group and Cogroup Operators
Detailed Study of Architecture of Apache Pig
Deep Dive into Pig Latin Diagnostic Operators
Deep Dive into Apache Pig Functions: Load & Store, Bag & Tuple, String, Date-time, Math
Apache Pig Basics, Features and Comparison with MapReduce, Hive & SQL and History of Apache Pig
Phases of MapReduce Data flow and detailed understanding of Mapreduce API
Overview of YARN and its components and benefits of YARN
Overview of Big Data and Hadoop, Big Data technologies
Implementation of Word Count program using Hadoop MapReduce
Operation of MapReduce in Hadoop framework using Java
Implementation of Character Count program using Hadoop MapReduce
How to set up Hadoop Multi-Node Cluster on a distributed environment
How to perform operations in Hadoop and commands used in Hadoop
How to install Hadoop on your system
How to install Hadoop Framework on your system
How the MapReduce Algorithm works using example
Introduction to Apache Cassandra, History and Architecture
Overview of How Cassandra Stores its data
Overview of important class in Cassandra and introduction of Cassandra query shell language
Concept of Partitioning of table in Hive
Detailed understanding of built-in functions available in Hive
Different Data Types in Hive which are involved in creation of table.
How to Alter the attributes of a table and delete a Table in Hive
How to create a table in Hive and how to insert data into it
How to create and drop a database in Hive
How to create and manage Views and Create and Drop an index in Hive
How to install Hive on your system
How to perform Join operations in Hive Query Language (HQL)
How to use the select statement in Hive Query Language
Deep dive into Splunk Search processing Language (SPL)
Explanation of Shell and Utility Commands provided by Apache Grunt Shell
How to perform Basic Search in Splunk
How to Install Apache Pig and Configure Pig
How to perform searching using fields in Splunk
How to Load data to Apache Pig from Hadoop File System
How to perform Time Range search in Splunk
How to run Apache Pig Scripts in Batch Mode
How to Store data in Apache Pig using Store Operator
How to share and export the search result in Splunk
How to use Cross Operator and Union Operator in Pig Latin
How to use Split and Filter Operator in Apache Pig Latin
How to use the Join Operators in Pig Latin
How to use Distinct, For Each, Order By, Limit Operators and Eval Functions in Apache Pig
User Defined Functions in Apache Pig Latin
Advanced Indexing in MongoDB and Limitation of Indexing in MongoDB
Concept of Capped Collections and Auto-Increment Sequence in MongoDB
Concept of Map Reduce in MongoDB
Concept of Relationships and Database References in MongoDB
Concept of Sharding process and How to create a backup in MongoDB
Data Modelling in MongoDB and How to create and Drop database in MongoDB
Deep dive into Covered Queries in MongoDB and Analyzing queries
Deep dive into Replication process in MongoDB
How to Create and Drop a collection using MongoDB
How to Insert, Update, Delete and Query Document in MongoDB Collection
How to Install MongoDB on your system
How to limit records using MongoDB ad use projection in MongoDB
Deep dive into HBase architecture
Deep dive into Java Client API for HBase and its associated classes
How to create and List Table in HBase shell
How to create data in an HBase table
How to delete data in Table in HBase
How to enable and disable a Table using HBase shell
How to Set up MongoDB JDBC driver
How to install HBase and configure on your system
How to make changes to an existing Table and describe it in HBase
How to read data from Table in HBase
How to sort records in MongoDB and concept of Indexing and Aggregation in MongoDB
How to start HBase interactive shell and how HBase general commands works
How to Stop HBase using Java API
How to update data in Table using HBase Shell
How to verify the existence of a Table and How to Drop a Table in HBase
Overview of HBase, its Advantages, Features and history
How to use Regex Expressions and Text Search in MongoDB
Overview of Scala programming language and How to install Scala on your system
MongoDB Administration using RockMongo and concept of GridFS in MongoDB
Overview of MongoDB, its history and purpose of building MongoDB
Understand NoSQL Databases and MongoDB advantages over Relational DBMS
How to Monitor System statistics using Apache NiFi
Concept of Logging in Apache NiFi
How to use the Apache Sqoop Eval and Codegen tool
How to list out the databases and tables of a particular database using Sqoop
How to import data and tables from MySQL to Hadoop HDFS
How to export data back from Hadoop HDFS to RDBMS and Create and maintain the Sqoop jobs
How to Install Apache Spark on your system
How to perform pattern matching in Scala and use of Regex expressions
How to use Functions in Scala programming Language
How to use Collections in Scala
How to use Arrays in Scala Programming Language
How to perform Exception Handling in Scala Language
How to Deploy Spark Application on Cluster
Extractor Object in Scala and how to perform pattern matching using extractors
Details of Data Types and Basic Literals in Scala
Detailed understanding of Operators in Scala Language
Deep dive into File Handling in Scala
Deep dive into Advanced programming in Spark
Basics of Scala Programming Language
Concept of String Manipulation in Scala
Conditional statements and Loop control structures in Scala
Introduction to Impala, its features, advantages and disadvantages
How to start Impala Shell and the various options of the shell
How to select a database using Command and select database using Hue Browser in Impala
How to perform changes on a given table and how to delete table in Impala
How to fetch the data from one or more tables in a database and fetch description in Impala
How to download, install and set up Impala in your system
How to create a table in the required database in Impala
How to Create, Alter and Drop a View in Impala
Explanation of Union Clause, With Clause and Distinct Operator in Impala
Explanation of Limit Clause and Offset Clause in Impala
Data Types in Impala Query Language
Concept of Resilient Distributed Datasets (RDD) in Apache Spark
How to use Classes and Objects in Scala programming
Overview of Apache Spark Framework
Spark Core and implementation of RDD transformations and actions in RDD programming
Introduction, Installation and Configuration of Apache Sqoop
Basic Concepts of Apache NiFi and its Installation
Deep Dive into Apache Nifi – Flow Files, Queues, Process Groups and Labels
Deep dive into Apache NiFi-Processors
Detailed understanding of Apache NiFi -Templates
How to Administer Apache NiFi and Create Flows in Apache NiFi
Understanding Apache NiFi API’s with request and response example
Understanding Apache Nifi Processors Categorization and its relationship
Detailed understanding of Architecture of Impala
Explanation of Order by Clause, Group by Clause and Having Clause in Impala
How to add new records into an existing table in a database using INSERT in Impala
How to create a database in Impala
How to drop a database in Impala
Detailed understanding of Scala Access Modifiers
How to use Variables in Scala with the help of example
Introduction to Apache NiFi, its History, Features and Architecture
Concept of Cluster Architecture in Apache Storm
Introduction to Apache Storm and Core Concepts of Apache Storm
How to Install Apache Storm framework on your machine
How to implement Mobile Call log Analyzer using Apache Storm
How Apache Storm is used in Twitter
Detailed understanding of Workflow of Apache Storm
Deep Dive into Trident – an extension of Apache Storm
Top Splunk SIEM Interview Questions and Answers
Top Big Data Hadoop Interview Questions and Answers
Top MongoDB Interview Questions and Answers
Top Scala Interview Questions and Answers
Top Splunk Interview Questions and Answers
Top Hadoop Administration Interview Questions and Answers
Top Apache Sqoop Interview Questions and Answers
Top Apache NiFi Interview Questions and Answers
Top Apache Impala Interview Questions and Answers
Top Apache HBase Interview Questions and Answers
Top Apache Flume Interview Questions and Answers
Top Apache Spark Interview Questions and Answers
Top Apache Pig Interview Questions and Answers
Top Apache Cassandra Interview Questions and Answers
Top Apache Hive Interview Questions and Answers
Top Apache Oozie Interview Questions and Answers
Top Apache Storm Interview Questions and Answers
Deep dive into HBase Scan, Count and Truncate command and how to achieve security in HBase
Deep Dive into Apache NiFi User Interface
Deep dive into built-in operators of Hive
Concept of Atomic Operations in MongoDB
A Deep Dive into Splunk Web Interface
Deep Dive into Apache Oozie Workflow
How to Configure Oozie Workflow using Property File
Concept of Coordinators applications using Apache Oozie
Basics of Apache Oozie and Oozie Editors
Deep Dive into Oozie Bundle System and CLI & Extensions
Process of Data Ingestion in Splunk Environment
Deep Dive into Apache NiFi User Interface
Deep dive into HBase Scan, Count and Truncate command and how to achieve security in HBase
Basics of Splunk and Installation of Splunk Environment
Top Apache Oozie Interview Questions and Answers You must Prepare Gaurav
Blockchain
Articles
Introduction to Ethereum and Smart Contracts
Ethereum - Interacting with Deployed Contract
Ethereum – Attaching Wallet to Ganache Blockchain
Ethereum - Creating Contract Users
Concept of Blockchain Double Spending and Bitcoin Cash
Bitcoin Forks and SegWit and BlockChain Merkel Tree
Comparison between Blockchain and Database
Basic Components of Bitcoin and Blockchain Proof of Work
Ethereum - Solidity for Contract Writing
Ethereum - Ganache for Blockchain
eBooks
Interview Questions
Videos
BlockChain and Ethereum
Articles
eBooks
Interview Questions
Videos
Introduction to Ethereum and Smart Contracts
Ethereum - Interacting with Deployed Contract
Ethereum – Attaching Wallet to Ganache Blockchain
Ethereum - Creating Contract Users
Concept of Blockchain Double Spending and Bitcoin Cash
Bitcoin Forks and SegWit and BlockChain Merkel Tree
Comparison between Blockchain and Database
Basic Components of Bitcoin and Blockchain Proof of Work
Ethereum - Solidity for Contract Writing
Ethereum - Ganache for Blockchain
Overview and History of Blockchain
Overview of Bitcoin and Key Concepts of Bitcoin
Top BlockChain Interview Questions and Answers
Top Ethereum Interview Questions and Answers
Best Approach for Storing data to AWS DynamoDB and S3 – AWS Implementation
Maintain High Availability in AWS with anticipated Additional Load
Cloud Computing
Articles
eBooks
Interview Questions
Videos
AWS
Articles
How to Use Amazon Machine Learning
How to use Amazon KCL and set up Amazon EMR
How to Set Up Amazon RDS (Relational Database Service)
How to Configure AWS Direct Connect
How to Configure Amazon Simple Storage Service (S3)
How to Configure Amazon Route 53
How AWS CloudFront Delivers the Content
Amazon Elastic Block Storage (EBS) and Storage Gateway
How to use Simple Workflow Service (SWF) and Amazon WorkMail
Understanding of AWS Management Console
How to Set Up AWS Data Pipeline
eBooks
Videos
How to Use Amazon Machine Learning
How to use Amazon KCL and set up Amazon EMR
How to Set Up Amazon RDS (Relational Database Service)
How to Configure AWS Direct Connect
How to Configure Amazon Simple Storage Service (S3)
How to Configure Amazon Route 53
How AWS CloudFront Delivers the Content
Amazon Elastic Block Storage (EBS) and Storage Gateway
How to use Simple Workflow Service (SWF) and Amazon WorkMail
Understanding of AWS Management Console
How to Set Up AWS Data Pipeline
How to Create Amazon Workspaces
Top Azure Developer Interview Questions and Answers
Top Amazon Web Services (AWS) Interview Questions and Answers
Azure
Articles
How to configure Azure Cloud Service
How to configure Azure Load Balancer
How to Configure Azure Storage Security
How to create Azure Mobile App
Overview of Microsoft Azure and Cloud Computing
Creating App Service Plan in Azure Portal
Azure Virtual Machines and Compute Service
Azure Virtual Machine Scale Set and Auto Scaling
Azure Table, Queue and Disk Storage
Azure Storage Monitoring and Resource Tool
Azure Storage Building Blocks and Storage Account
Azure Storage account and Blob service configuration
Azure SQL Managed Instance and SQL Stretch Database
Azure SQL Database and its Configuration
Azure Network Service and Azure Virtual Network
Azure Media Service and Database Service
Azure Backup and Virtual Machine Security
Azure Availability Zones and Sets and VNet Connectivity
Azure App Service Monitoring and Azure CDN
eBooks
Interview Questions
Videos
How to configure Azure Cloud Service
How to configure Azure Load Balancer
How to Configure Azure Storage Security
How to create Azure Mobile App
Overview of Microsoft Azure and Cloud Computing
Creating App Service Plan in Azure Portal
Azure Virtual Machines and Compute Service
Azure Virtual Machine Scale Set and Auto Scaling
Azure Table, Queue and Disk Storage
Azure Storage Monitoring and Resource Tool
Azure Storage Building Blocks and Storage Account
Azure Storage account and Blob service configuration
Azure SQL Managed Instance and SQL Stretch Database
Azure SQL Database and its Configuration
Azure Network Service and Azure Virtual Network
Azure Media Service and Database Service
Azure Backup and Virtual Machine Security
Azure Availability Zones and Sets and VNet Connectivity
Azure App Service Monitoring and Azure CDN
Azure App Service Backup and Security
How to Use Amazon Machine Learning
How to use Amazon KCL and set up Amazon EMR
How to Set Up Amazon RDS (Relational Database Service)
How to Configure AWS Direct Connect
How to Configure Amazon Simple Storage Service (S3)
How to Configure Amazon Route 53
How AWS CloudFront Delivers the Content
Amazon Elastic Block Storage (EBS) and Storage Gateway
How to use Simple Workflow Service (SWF) and Amazon WorkMail
Understanding of AWS Management Console
How to configure Azure Cloud Service
How to configure Azure Load Balancer
How to Configure Azure Storage Security
How to create Azure Mobile App
Overview of Microsoft Azure and Cloud Computing
Creating App Service Plan in Azure Portal
Azure Virtual Machines and Compute Service
Azure Virtual Machine Scale Set and Auto Scaling
Azure Table, Queue and Disk Storage
Azure Storage Monitoring and Resource Tool
Azure Storage Building Blocks and Storage Account
Azure Storage account and Blob service configuration
How to Set Up AWS Data Pipeline
How to Create Amazon Workspaces
Azure SQL Managed Instance and SQL Stretch Database
Azure SQL Database and its Configuration
Azure Network Service and Azure Virtual Network
Azure Media Service and Database Service
Azure Backup and Virtual Machine Security
Azure Availability Zones and Sets and VNet Connectivity
Azure App Service Monitoring and Azure CDN
Azure App Service Backup and Security
Azure API Apps and API Management
Top Azure Developer Interview Questions and Answers
Top Azure Architect Interview Questions and Answers
Top Amazon Web Services (AWS) Interview Questions and Answers
Best Approach for Storing data to AWS DynamoDB and S3 – AWS Implementation
Migration of 3-tier e-commerce web application using Amazon web Services (AWS)
Cyber Security
Articles
eBooks
Interview Questions
Videos
Ethical Hacking
Articles
Concept of Enumeration in Ethical Hacking
Concept of Exploitation in Ethical Hacking
Concept of Social Engineering Attacks and Cross-Site Scripting
Concept of SQL Injection Attack
Concept of TCP/IP Hijacking and Trojan Attacks
DDOS Attacks in Ethical Hacking
Ethical Hacking - Fingerprinting
Ethical Hacking - Footprinting
Processes in Ethical Hacking and Reconnaissance
eBooks
Interview Questions
Videos
Concept of Enumeration in Ethical Hacking
Concept of Exploitation in Ethical Hacking
Concept of Social Engineering Attacks and Cross-Site Scripting
Concept of SQL Injection Attack
Concept of TCP/IP Hijacking and Trojan Attacks
DDOS Attacks in Ethical Hacking
Ethical Hacking - Fingerprinting
Ethical Hacking - Footprinting
Processes in Ethical Hacking and Reconnaissance
Concept of Enumeration in Ethical Hacking
Concept of Exploitation in Ethical Hacking
Concept of Social Engineering Attacks and Cross-Site Scripting
Concept of SQL Injection Attack
Concept of TCP/IP Hijacking and Trojan Attacks
DDOS Attacks in Ethical Hacking
Ethical Hacking - Fingerprinting
Ethical Hacking - Footprinting
Processes in Ethical Hacking and Reconnaissance
Data Science
Articles
Regression Analysis in Machine learning
Regression vs Classification in Machine Learning
Simple Linear Regression in Machine Learning
Naïve Bayes Classifier Algorithm
Support Vector Machine Algorithm
Logistic Regression in Machine Learning
Linear Regression in Machine Learning
K-Nearest Neighbor (KNN) Algorithm for Machine Learning
eBooks
Interview Questions
Videos
Machine Learning
Python with Data Science
Articles
Processing JSON Data in Python and Matplotlib
Processing Unstructured Data and rectilinear regression and Chi-Square Test in Python
P-Value and Correlation in Python
Python - Data Science Introduction
Relational Databases in Python
Perform Data Cleansing in Python
Performing Data Wrangling in Python
Introduction to Pandas, NumPy and SciPy Libraries
How to Read HTML Pages in Python
How to interact with MongoDB in Python
Box Plots and Scatter Plots and Heat Maps in Python
Bubble Charts and 3D Charts in Python
Data Aggregation and binomial distribution in Python
How to create Geographical Maps and Graphs in Python
Measuring Central Tendency and Variance in Python
eBooks
Interview Questions
Videos
Processing JSON Data in Python and Matplotlib
Processing Unstructured Data and rectilinear regression and Chi-Square Test in Python
P-Value and Correlation in Python
Python - Data Science Introduction
Relational Databases in Python
Perform Data Cleansing in Python
Performing Data Wrangling in Python
Introduction to Pandas, NumPy and SciPy Libraries
How to Read HTML Pages in Python
How to interact with MongoDB in Python
Box Plots and Scatter Plots and Heat Maps in Python
Bubble Charts and 3D Charts in Python
Data Aggregation and binomial distribution in Python
How to create Geographical Maps and Graphs in Python
Measuring Central Tendency and Variance in Python
Normal, Binomial and Poisson distribution in Python
R Language
Articles
Arrays and Factors in R Language
Binomial Distribution and Poisson Regression in R
Analysis of Covariance in R Language
Decision making and Loops in R Language
Handling Excel and Binary Files in R
Handling XML Files in R Language
How to create Line Graphs in R
How to create Scatterplots in R
How to create Histograms and Box Plots in R
Random Forest and Survival Analysis in R
Operators and Variables in R Language
Normal Distribution in R Language
Multiple and Logistic Regression in R
eBooks
Interview Questions
Videos
Arrays and Factors in R Language
Binomial Distribution and Poisson Regression in R
Analysis of Covariance in R Language
Decision making and Loops in R Language
Handling Excel and Binary Files in R
Handling XML Files in R Language
How to create Line Graphs in R
How to create Scatterplots in R
How to create Histograms and Box Plots in R
Random Forest and Survival Analysis in R
Operators and Variables in R Language
Normal Distribution in R Language
Multiple and Logistic Regression in R
SAS
Articles
One Way Anova and Hypothesis Testing
Overview of SAS and its Features
SAS - Basic Syntax and Program Structure
How to Perform Standard Deviation in SAS
How to perform Correlation Analysis in SAS
How to perform Bland Altman Analysis
SAS Applications and Loops and Decision Making
SAS Intelligence Platform Architecture
Strings Manipulation and Arrays in SAS
How to Format Data Sets in SAS
How to create Scatter Plots in SAS
How to create Pie Charts in SAS
How to create Histogram and Simulations in SAS
How to create Box Plots in SAS
How to Create Bar Charts in SAS
How to Concatenate Data Sets in SAS
How to calculate Arithmetic Mean and Handling Data and Time
Frequency Distributions and Cross Tabulations in SAS
eBooks
Interview Questions
Videos
One Way Anova and Hypothesis Testing
Overview of SAS and its Features
SAS - Basic Syntax and Program Structure
How to Perform Standard Deviation in SAS
How to perform Correlation Analysis in SAS
How to perform Bland Altman Analysis
SAS Applications and Loops and Decision Making
SAS Intelligence Platform Architecture
Strings Manipulation and Arrays in SAS
How to Format Data Sets in SAS
How to create Scatter Plots in SAS
How to create Pie Charts in SAS
How to create Histogram and Simulations in SAS
How to create Box Plots in SAS
How to Create Bar Charts in SAS
How to Concatenate Data Sets in SAS
How to calculate Arithmetic Mean and Handling Data and Time
Frequency Distributions and Cross Tabulations in SAS
Fishers Exact Tests and Repeated Measure Analysis in SAS
Regression Analysis in Machine learning
Regression vs Classification in Machine Learning
Simple Linear Regression in Machine Learning
Naïve Bayes Classifier Algorithm
Support Vector Machine Algorithm
Logistic Regression in Machine Learning
Linear Regression in Machine Learning
K-Nearest Neighbor (KNN) Algorithm for Machine Learning
Difference between Supervised and Unsupervised Learning
Classification Algorithm in Machine Learning
How to get datasets for Machine Learning
Processing JSON Data in Python and Matplotlib
Processing Unstructured Data and rectilinear regression and Chi-Square Test in Python
P-Value and Correlation in Python
Python - Data Science Introduction
Relational Databases in Python
One Way Anova and Hypothesis Testing
Overview of SAS and its Features
SAS - Basic Syntax and Program Structure
How to Perform Standard Deviation in SAS
How to perform Correlation Analysis in SAS
How to perform Bland Altman Analysis
SAS Applications and Loops and Decision Making
SAS Intelligence Platform Architecture
Strings Manipulation and Arrays in SAS
Perform Data Cleansing in Python
Performing Data Wrangling in Python
Introduction to Pandas, NumPy and SciPy Libraries
How to Read HTML Pages in Python
How to interact with MongoDB in Python
Box Plots and Scatter Plots and Heat Maps in Python
Arrays and Factors in R Language
Binomial Distribution and Poisson Regression in R
Bubble Charts and 3D Charts in Python
Analysis of Covariance in R Language
Difference between Artificial intelligence and Machine learning
Data Preprocessing in Machine learning
Data Aggregation and binomial distribution in Python
How to create Geographical Maps and Graphs in Python
Measuring Central Tendency and Variance in Python
Introduction to Machine Learning
Normal, Binomial and Poisson distribution in Python
Installing Anaconda and Python
Applications of Machine learning
Handle Date and Time in Python
Decision making and Loops in R Language
Handling Excel and Binary Files in R
Handling XML Files in R Language
How to create Line Graphs in R
How to create Scatterplots in R
How to create Histograms and Box Plots in R
How to Format Data Sets in SAS
How to create Scatter Plots in SAS
How to create Pie Charts in SAS
How to create Histogram and Simulations in SAS
How to create Box Plots in SAS
How to Create Bar Charts in SAS
How to Concatenate Data Sets in SAS
How to calculate Arithmetic Mean and Handling Data and Time
Frequency Distributions and Cross Tabulations in SAS
Fishers Exact Tests and Repeated Measure Analysis in SAS
Advantages and Disadvantages of SAS Programming Language
Random Forest and Survival Analysis in R
Operators and Variables in R Language
Normal Distribution in R Language
Multiple and Logistic Regression in R
Linear Regression in R Language
Top Data Science Interview Questions and Answers
Top Machine Learning Interview Questions and Answers
Top SAS Interview Questions and Answers
Top Python Interview Questions and Answers
Data Warehousing and ETL
Articles
eBooks
Interview Questions
Videos
ETL Testing
eBooks
Interview Questions
Videos
Informatica
Articles
Aggregator Transformation in Informatica
Concept of Informatica (Big Data Management) BDM
Informatica Master Data Management (MDM) Process
Lookup and Normalizer Transformation in Informatica
Performance Tuning and Partitioning in Informatica
Rank Transformation in Informatica
Router and Joiner Transformation in Informatica
Source Qualifier Transformation in Informatica
Transaction Control Transformation in Informatica
Sequence Generator Transformation in Informatica
eBooks
Interview Questions
Videos
Aggregator Transformation in Informatica
Concept of Informatica (Big Data Management) BDM
Informatica Master Data Management (MDM) Process
Lookup and Normalizer Transformation in Informatica
Performance Tuning and Partitioning in Informatica
Rank Transformation in Informatica
Router and Joiner Transformation in Informatica
Source Qualifier Transformation in Informatica
Transaction Control Transformation in Informatica
Sequence Generator Transformation in Informatica
Concept of Informatica IDQ (Informatica Data Quality)
Concept of ETL Pipeline and Files
Overview of ELT Testing and its Architecture
Aggregator Transformation in Informatica
Concept of Informatica (Big Data Management) BDM
Informatica Master Data Management (MDM) Process
Lookup and Normalizer Transformation in Informatica
Performance Tuning and Partitioning in Informatica
Rank Transformation in Informatica
Router and Joiner Transformation in Informatica
Source Qualifier Transformation in Informatica
Transaction Control Transformation in Informatica
Sequence Generator Transformation in Informatica
Concept of Informatica IDQ (Informatica Data Quality)
Installation of Informatica PowerCenter
Comparison between ETL and ELT
Detailed understanding of ETL (Extraction, Transformation and Loading) Testing
Databases
Articles
eBooks
Interview Questions
Videos
MS-SQL Server
Articles
Backup and restore a database in SQL Server
Concept of Primary Key in SQL Server
CRUD Operations of Data in MS SQL Server
How to Enable, Disable and Drop a Foreign Key
Popular Functions in MS SQL Server
SQL Server BETWEEN Condition (Operator)
SQL Server Comparison Operator
Create and Delete Table in MS SQL Server
SQL Server DISTINCT and GROUP BY Clause
eBooks
Interview Questions
Videos
Backup and restore a database in SQL Server
Concept of Primary Key in SQL Server
CRUD Operations of Data in MS SQL Server
How to Enable, Disable and Drop a Foreign Key
Popular Functions in MS SQL Server
SQL Server BETWEEN Condition (Operator)
SQL Server Comparison Operator
Create and Delete Table in MS SQL Server
SQL Server DISTINCT and GROUP BY Clause
Oracle DBA
Overview of Oracle Tablespace Group
Overview of Oracle Database and its Architecture
Oracle ALTER USER and DROP USER
Introduction to Oracle Data Pump Import and Export tool
Introduction to Oracle CREATE USER statement
How to use the Oracle STARTUP command to start out an Oracle Database instance
How to shut down the Oracle Database
How to Manage Tablespaces in Oracle
How To List Users within the Oracle Database
How to Grant SELECT Object Privilege on One or More Tables to a User and Unlock a User in Oracle
How to Grant All Privileges to a User in Oracle
How to Grant All Privileges to a User in Oracle
How to Create User Profiles in Oracle
How to Create Oracle Database Links
How to Alter and Drop Roles in Oracle
Oracle PL-SQL
eBooks
Interview Questions
Videos
Date and Time Handling in PL-SQL
Constants and Literals and Operators in PL-SQL
Conditions and Loops in PL-SQL
Backup and restore a database in SQL Server
Concept of Primary Key in SQL Server
CRUD Operations of Data in MS SQL Server
How to Enable, Disable and Drop a Foreign Key
Popular Functions in MS SQL Server
SQL Server BETWEEN Condition (Operator)
SQL Server Comparison Operator
Create and Delete Table in MS SQL Server
SQL Server DISTINCT and GROUP BY Clause
SQL Server NOT Condition (Operator)
Overview of Oracle Tablespace Group
Overview of Oracle Database and its Architecture
Oracle ALTER USER and DROP USER
Introduction to Oracle Data Pump Import and Export tool
Introduction to Oracle CREATE USER statement
How to use the Oracle STARTUP command to start out an Oracle Database instance
How to shut down the Oracle Database
How to Manage Tablespaces in Oracle
How To List Users within the Oracle Database
How to Grant SELECT Object Privilege on One or More Tables to a User and Unlock a User in Oracle
How to Grant All Privileges to a User in Oracle
How to Grant All Privileges to a User in Oracle
How to Create User Profiles in Oracle
How to Create Oracle Database Links
How to Alter and Drop Roles in Oracle
How to Alter and Drop Oracle Database Link
Introduction to PL-SQL and Environment setup
Top Oracle PL-SQL Interview Questions and Answers
Top MS-SQL Server Interview Questions and Answers
Best Approach for Storing data to AWS DynamoDB and S3 – AWS Implementation
Maintain High Availability in AWS with anticipated Additional Load
DevOps
eBooks
Interview Questions
Videos
Ansible
Articles
Overview of YAML and Ad-hoc commands in Ansible
A Detailed comparison of Ansible and Puppet
A Detailed comparison of Ansible Vs Chef
Detailed understanding of concept of Playbooks in Ansible
Deep dive into Pip module in Ansible
How to perform troubleshooting in Ansible
How to use variables in playbooks in Ansible and concept of exception handling
Overview of Ansible, its History and How to set-up Ansible on your machine
eBooks
Interview Questions
Videos
Overview of YAML and Ad-hoc commands in Ansible
A Detailed comparison of Ansible and Puppet
A Detailed comparison of Ansible Vs Chef
Detailed understanding of concept of Playbooks in Ansible
Deep dive into Pip module in Ansible
How to perform troubleshooting in Ansible
How to use variables in playbooks in Ansible and concept of exception handling
Overview of Ansible, its History and How to set-up Ansible on your machine
Chef
Articles
Chef-Client as Daemon and Chef-Shell
Concept of Libraries , Definition and setting environment variable
Concept of Lightweight Resource Provider and Blueprints in Chef
Concept of Templates and Dynamically Configuring Recipes
Dealing with Files and Software packages and Community Cookbooks
Execute Cookbook on Node and run Chef-Client
Detailed understanding of Resources in Chef
How to Set up Chef on your system
How to set up Test Kitchen Workflow
How to write Cross-Platform Cookbooks
Overview of Chef and its Architecture
Plain Ruby with Chef DSL and Ruby Gems with Recipes
Testing Cookbook with Test Kitchen
Roles in Chef and perform environment specific configuration
eBooks
Interview Questions
Videos
Chef-Client as Daemon and Chef-Shell
Concept of Libraries , Definition and setting environment variable
Concept of Lightweight Resource Provider and Blueprints in Chef
Concept of Templates and Dynamically Configuring Recipes
Dealing with Files and Software packages and Community Cookbooks
Execute Cookbook on Node and run Chef-Client
Detailed understanding of Resources in Chef
How to Set up Chef on your system
How to set up Test Kitchen Workflow
How to write Cross-Platform Cookbooks
Overview of Chef and its Architecture
Plain Ruby with Chef DSL and Ruby Gems with Recipes
Testing Cookbook with Test Kitchen
Roles in Chef and perform environment specific configuration
Docker
Articles
Concept of Docker Cloud Service
Deep dive into Docker Architecture
Concept of public repositories in Docker
Concept of Container Linking and Storage in Docker
Building a web Server Docker File
Working with Docker Toolbox and how to use the Jenkins Docker image from Docker Hub
Overview of Docker and its features
Managing ports and private registries in Docker
Instruction commands in Docker
How to work with Containers in Docker
How to Set-up MongoDB in Docker
How to set up Node.js in Docker
How to set up Kubernetes in Docker
How to set up ASP.net in Docker
How to perform Continuous integration using Jenkins in Docker
How to install Docker on Windows
eBooks
Interview Questions
Videos
Concept of Docker Cloud Service
Deep dive into Docker Architecture
Concept of public repositories in Docker
Concept of Container Linking and Storage in Docker
Building a web Server Docker File
Working with Docker Toolbox and how to use the Jenkins Docker image from Docker Hub
Overview of Docker and its features
Managing ports and private registries in Docker
Instruction commands in Docker
How to work with Containers in Docker
How to Set-up MongoDB in Docker
How to set up Node.js in Docker
How to set up Kubernetes in Docker
How to set up ASP.net in Docker
How to perform Continuous integration using Jenkins in Docker
How to install Docker on Windows
How to install docker on Linux
Git and GitHub
eBooks
Interview Questions
Videos
Concept of Git Index and Git Head
Comparison of Git with SVN and Mercurial
Deep dive into Git Branching Model
Git Repository and How to Fork Repository
Git Terminology and General Tools
How to Clone Repository in Git
Working with Remote Repository
Version Control System and its Types
Overview of GitHub and Comparison of Git and GitHub
Overview of Git and its features
Merging Branches and Resolve conflicts in Git
How to use Git via the command line
How to switch branches without committing the current branch in Git
How to perform Rebasing in Git
How to Install Git on Linux (Ubuntu) and Mac
Jenkins
Articles
Deep dive into Metrics and Trends for builds
Server maintenance and Plugins Management in Jenkins
Perform Continuous Deployment using Jenkins
Overview of Jenkins, its History and Architecture
How to take Back-up in Jenkins using Backup plugin
How to set up Git and Maven Plugin in Jenkins
How to set up Distributed build and Automated deployment in Jenkins
How to set up Build jobs in Jenkins
How to run Remote tests using Jenkins
How to perform Notification, Reporting and Code Analysis
How to perform Junit Testing in Jenkins
How to perform Automation Testing in Jenkins
How to install Jenkins on your system
Comparison of Jenkins with Ansible and Hudson Frameworks
Comparison of Jenkins with Bamboo and TeamCity
eBooks
Interview Questions
Videos
Deep dive into Metrics and Trends for builds
Server maintenance and Plugins Management in Jenkins
Perform Continuous Deployment using Jenkins
Overview of Jenkins, its History and Architecture
How to take Back-up in Jenkins using Backup plugin
How to set up Git and Maven Plugin in Jenkins
How to set up Distributed build and Automated deployment in Jenkins
How to set up Build jobs in Jenkins
How to run Remote tests using Jenkins
How to perform Notification, Reporting and Code Analysis
How to perform Junit Testing in Jenkins
How to perform Automation Testing in Jenkins
How to install Jenkins on your system
Comparison of Jenkins with Ansible and Hudson Frameworks
Comparison of Jenkins with Bamboo and TeamCity
Comparison of Jenkins with GoCD and Maven Tools
Kubernetes
Articles
eBooks
Interview Questions
Videos
How to setup Kubernetes on your machine
How to Set up Kubernetes Dashboard
How to manage Deployments and Concept of Kubernetes Volume
How to achieve Autoscaling in Kubernetes cluster
Deep dive into Kubectl command line utility
Create an Application for Kubernetes deployment
Concept of Secrets, Network Policy and Kubernetes API
Concept of Replication Controller and Replica Sets
Concept of Node, Service and Pod in Kubernetes
Concept of Images and creating a Job in Kubernetes
Namespace, Labels and Selectors in Kubernetes
Overview of Kubernetes and its Architecture and components
Maven
Articles
Introduction to Maven and How to Set up Maven Environment
How to manage Maven Project in NetBeans and IntelliJ IDEA
How to manage a web-based project using Maven
How to import Maven Project in Eclipse IDE
How to create documentation of Application in Maven
How to automate the Deployment process in Maven
Deep dive into Build Automation
Creating Java Project in Maven
Concept of Project Object Model (POM) in Maven
Concept of Maven Repositories and Plugins in Maven
eBooks
Interview Questions
Videos
Introduction to Maven and How to Set up Maven Environment
How to manage Maven Project in NetBeans and IntelliJ IDEA
How to manage a web-based project using Maven
How to import Maven Project in Eclipse IDE
How to create documentation of Application in Maven
How to automate the Deployment process in Maven
Deep dive into Build Automation
Creating Java Project in Maven
Concept of Project Object Model (POM) in Maven
Concept of Maven Repositories and Plugins in Maven
Nagios
Articles
Look into Nagios Features, applications, Hosts and services and Commands
Overview of Nagios, its architecture and Nagios products
Ports and protocols and Add-ons and plugins in Nagios
Detailed understanding of Checks and States in Nagios
How to run Nagios plugins on other machines remotely using NRPE
eBooks
Interview Questions
Videos
Look into Nagios Features, applications, Hosts and services and Commands
Overview of Nagios, its architecture and Nagios products
Ports and protocols and Add-ons and plugins in Nagios
Detailed understanding of Checks and States in Nagios
How to run Nagios plugins on other machines remotely using NRPE
Puppet
Articles
Implementation of Live working demo project in Puppet
How to Set-up and configure Puppet Master
How to install and configure r10k tool and validate puppet setup
How to install and configure puppet on your machine
How to define Functions and Custom functions in Puppet
Concept of Templating in Puppet
Concept of Type and Provider in Puppet
How to create custom environment in Puppet
Detailed understanding of architecture of puppet and its components and application of puppet
Detail understanding of environment conf file in puppet
Deep Dive into Resources in Puppet
Concept of Resource Abstraction Layer (RAL) in Puppet
Concept of File Server in Puppet
Concept of Facter and Facts in Puppet
Understanding Puppet Manifest files and How to write a manifest file in Puppet
Overview of Puppet and its components and concept of configuration management
How to Set-up Puppet agent and How to sign and check for SSL Ceritficate
eBooks
Interview Questions
Videos
How to use RESTful APIs in Puppet
Implementation of Live working demo project in Puppet
How to Set-up and configure Puppet Master
How to install and configure r10k tool and validate puppet setup
How to install and configure puppet on your machine
How to define Functions and Custom functions in Puppet
Concept of Templating in Puppet
Concept of Type and Provider in Puppet
How to create custom environment in Puppet
Detailed understanding of architecture of puppet and its components and application of puppet
Detail understanding of environment conf file in puppet
Deep Dive into Resources in Puppet
Concept of Resource Abstraction Layer (RAL) in Puppet
Concept of File Server in Puppet
Concept of Facter and Facts in Puppet
Understanding Puppet Manifest files and How to write a manifest file in Puppet
Overview of Puppet and its components and concept of configuration management
How to Set-up Puppet agent and How to sign and check for SSL Ceritficate
Concept of Git Index and Git Head
Comparison of Git with SVN and Mercurial
Deep dive into Git Branching Model
Git Repository and How to Fork Repository
Git Terminology and General Tools
How to Clone Repository in Git
Deep dive into Metrics and Trends for builds
Concept of Docker Cloud Service
Server maintenance and Plugins Management in Jenkins
How to use RESTful APIs in Puppet
Implementation of Live working demo project in Puppet
Perform Continuous Deployment using Jenkins
Overview of Jenkins, its History and Architecture
How to take Back-up in Jenkins using Backup plugin
How to set up Git and Maven Plugin in Jenkins
How to set up Distributed build and Automated deployment in Jenkins
How to set up Build jobs in Jenkins
How to run Remote tests using Jenkins
How to perform Notification, Reporting and Code Analysis
How to perform Junit Testing in Jenkins
How to perform Automation Testing in Jenkins
How to install Jenkins on your system
Deep dive into Docker Architecture
Concept of public repositories in Docker
Concept of Container Linking and Storage in Docker
Building a web Server Docker File
Comparison of Jenkins with Ansible and Hudson Frameworks
How to setup Kubernetes on your machine
How to Set up Kubernetes Dashboard
How to manage Deployments and Concept of Kubernetes Volume
How to achieve Autoscaling in Kubernetes cluster
Deep dive into Kubectl command line utility
Create an Application for Kubernetes deployment
Concept of Secrets, Network Policy and Kubernetes API
Concept of Replication Controller and Replica Sets
Concept of Node, Service and Pod in Kubernetes
Concept of Images and creating a Job in Kubernetes
How to Set-up and configure Puppet Master
How to install and configure r10k tool and validate puppet setup
How to install and configure puppet on your machine
How to define Functions and Custom functions in Puppet
Concept of Templating in Puppet
Concept of Type and Provider in Puppet
How to create custom environment in Puppet
Detailed understanding of architecture of puppet and its components and application of puppet
Detail understanding of environment conf file in puppet
Deep Dive into Resources in Puppet
Concept of Resource Abstraction Layer (RAL) in Puppet
Concept of File Server in Puppet
Concept of Facter and Facts in Puppet
Working with Docker Toolbox and how to use the Jenkins Docker image from Docker Hub
Overview of Docker and its features
Managing ports and private registries in Docker
Instruction commands in Docker
How to work with Containers in Docker
How to Set-up MongoDB in Docker
How to set up Node.js in Docker
How to set up Kubernetes in Docker
How to set up ASP.net in Docker
How to perform Continuous integration using Jenkins in Docker
How to install Docker on Windows
How to install docker on Linux
Namespace, Labels and Selectors in Kubernetes
Comparison of Jenkins with Bamboo and TeamCity
Comparison of Jenkins with GoCD and Maven Tools
Comparison of Jenkins with Travis CI and Circle CI
Working with Remote Repository
Version Control System and its Types
Overview of GitHub and Comparison of Git and GitHub
Overview of Git and its features
Merging Branches and Resolve conflicts in Git
How to use Git via the command line
How to switch branches without committing the current branch in Git
How to perform Rebasing in Git
How to Install Git on Linux (Ubuntu) and Mac
How to create a new Blank Repository and commit code in it
Overview of Kubernetes and its Architecture and components
Monitor processes in Kubernetes
Introduction to Maven and How to Set up Maven Environment
How to manage Maven Project in NetBeans and IntelliJ IDEA
How to manage a web-based project using Maven
How to import Maven Project in Eclipse IDE
How to create documentation of Application in Maven
How to automate the Deployment process in Maven
Deep dive into Build Automation
Creating Java Project in Maven
Concept of Project Object Model (POM) in Maven
Concept of Maven Repositories and Plugins in Maven
Concept of Dependency Management in Maven
Understanding Puppet Manifest files and How to write a manifest file in Puppet
Overview of Puppet and its components and concept of configuration management
Look into Nagios Features, applications, Hosts and services and Commands
Overview of Nagios, its architecture and Nagios products
Ports and protocols and Add-ons and plugins in Nagios
Detailed understanding of Checks and States in Nagios
How to run Nagios plugins on other machines remotely using NRPE
Chef-Client as Daemon and Chef-Shell
Concept of Libraries , Definition and setting environment variable
Concept of Lightweight Resource Provider and Blueprints in Chef
Concept of Templates and Dynamically Configuring Recipes
Dealing with Files and Software packages and Community Cookbooks
Execute Cookbook on Node and run Chef-Client
Detailed understanding of Resources in Chef
How to Set up Chef on your system
How to set up Test Kitchen Workflow
How to write Cross-Platform Cookbooks
Overview of Chef and its Architecture
Plain Ruby with Chef DSL and Ruby Gems with Recipes
Testing Cookbook with Test Kitchen
How to Set-up Puppet agent and How to sign and check for SSL Ceritficate
Roles in Chef and perform environment specific configuration
Overview of YAML and Ad-hoc commands in Ansible
A Detailed comparison of Ansible and Puppet
A Detailed comparison of Ansible Vs Chef
Detailed understanding of concept of Playbooks in Ansible
Deep dive into Pip module in Ansible
How to perform troubleshooting in Ansible
How to use variables in playbooks in Ansible and concept of exception handling
Overview of Ansible, its History and How to set-up Ansible on your machine
Concept of Advanced Execution with Ansible
Popular DevOps and DevOps Automation Tools
Comparison between DevOps and Agile methodologies
Concept of DevOps Pipeline and Who are DevOps Engineers
Overview of DevOps and its Architecture
DevOps Training Certification and Azure and AWS DevOps
How to set-up Nagios on Ubuntu
Top Docker Interview Questions and Answers
Top Ansible Interview Questions and Answers
Top Chef Interview Questions and Answers
Top Git and GitHub Interview Questions and Answers
Top DevOps Interview Questions and Answers
Top Puppet Interview Questions and Answers
Top Nagios Interview Questions and Answers
Top Kubernetes Interview Questions and Answers
Digital Marketing
Articles
Understanding Mobile marketing
Understanding Google Analytics
Online Marketing - Web Analytics
Why can we need an SEO Friendly Website?
Concept of Pay Per Click (PPC) and Conversion Rate Optimization (CRO) explained
Online Marketing - Impact, Pros & Cons
Online Marketing - Blogs, banners and forums
Introduction to Online Marketing
Digital Marketing using Twitter and LinkedIn
Digital Marketing using Social Media and YouTube
Digital Marketing using Facebook and Pinterest
Digital Marketing using Content marketing and Email Marketing
eBooks
Interview Questions
Videos
SEO and SMM
Articles
Social Media Marketing using Blogs
Social Media Marketing using Facebook
Social Media Marketing using Google Plus
Social Media Marketing using Linkedin
Social Media Marketing using Pinterest
Social Media Marketing using Twitter
Social Media Marketing using Video
Social Media Analysis and Monitoring Social Media Accounts
SMM - Image Optimization and Social Bookmarking
eBooks
Interview Questions
Videos
Social Media Marketing using Blogs
Social Media Marketing using Facebook
Social Media Marketing using Google Plus
Social Media Marketing using Linkedin
Social Media Marketing using Pinterest
Social Media Marketing using Twitter
Social Media Marketing using Video
Social Media Analysis and Monitoring Social Media Accounts
SMM - Image Optimization and Social Bookmarking
Social Media Marketing using Blogs
Social Media Marketing using Facebook
Social Media Marketing using Google Plus
Social Media Marketing using Linkedin
Social Media Marketing using Pinterest
Social Media Marketing using Twitter
Social Media Marketing using Video
Understanding Mobile marketing
Understanding Google Analytics
Online Marketing - Web Analytics
Why can we need an SEO Friendly Website?
Concept of Pay Per Click (PPC) and Conversion Rate Optimization (CRO) explained
Online Marketing - Impact, Pros & Cons
Online Marketing - Blogs, banners and forums
Introduction to Online Marketing
Digital Marketing using Twitter and LinkedIn
Digital Marketing using Social Media and YouTube
Digital Marketing using Facebook and Pinterest
Digital Marketing using Content marketing and Email Marketing
Overview of Digital Marketing and SEO
Social Media Analysis and Monitoring Social Media Accounts
SMM - Image Optimization and Social Bookmarking
SEO Strategy to Optimize Keywords and Metatags
Affiliate Marketing and Email Marketing
Frontend Development
Articles
eBooks
Interview Questions
Videos
Angular JS
Articles
Create Angular Application and Angular MVC Architecture
Custom Directives in Angular JS
Dependency Injection in Angular JS
Directives and Filters in Angular JS
Embedding Html Pages within HTML page
Expressions and Controllers in Angular JS
How to create Forms in Angular JS
How to create Single Page Application via multiple views
Internationalization in Angular JS
Services Architecture in Angular JS
Spring Angular CRUD Application
Spring Angular Login & Logout Application
Spring Angular Search Field Application
Tables and HTML DOM in Angular JS
Using Directives and Expressions in Angular JS
eBooks
Interview Questions
Videos
Create Angular Application and Angular MVC Architecture
Custom Directives in Angular JS
Dependency Injection in Angular JS
Directives and Filters in Angular JS
Embedding Html Pages within HTML page
Expressions and Controllers in Angular JS
How to create Forms in Angular JS
How to create Single Page Application via multiple views
Internationalization in Angular JS
Services Architecture in Angular JS
Spring Angular CRUD Application
Spring Angular Login & Logout Application
Spring Angular Search Field Application
Tables and HTML DOM in Angular JS
Using Directives and Expressions in Angular JS
React JS
Articles
Comparison Between AngularJS and ReactJS
How to implement flux pattern in React Applications
How to Animate elements using React
Error Handling using Error Boundaries
Environment Setup for React JS
Component Life Cycle Methods in React JS
Comparison between ReactJS and React Native
Overview of ReactJS and its Features
Overview of React Redux with an example
How to set up Router for an app
Using Refs and Keys in React JS
eBooks
Interview Questions
Videos
Comparison Between AngularJS and ReactJS
How to implement flux pattern in React Applications
How to Animate elements using React
Error Handling using Error Boundaries
Environment Setup for React JS
Component Life Cycle Methods in React JS
Comparison between ReactJS and React Native
Overview of ReactJS and its Features
Overview of React Redux with an example
How to set up Router for an app
Using Refs and Keys in React JS
Create Angular Application and Angular MVC Architecture
Custom Directives in Angular JS
Dependency Injection in Angular JS
Directives and Filters in Angular JS
Embedding Html Pages within HTML page
Expressions and Controllers in Angular JS
How to create Forms in Angular JS
How to create Single Page Application via multiple views
Internationalization in Angular JS
Services Architecture in Angular JS
Spring Angular CRUD Application
Spring Angular Login & Logout Application
Spring Angular Search Field Application
Tables and HTML DOM in Angular JS
Using Directives and Expressions in Angular JS
How to Setup AngularJS Environment
Comparison Between AngularJS and ReactJS
How to implement flux pattern in React Applications
How to Animate elements using React
Error Handling using Error Boundaries
Environment Setup for React JS
Component Life Cycle Methods in React JS
Comparison between ReactJS and React Native
Overview of ReactJS and its Features
Overview of React Redux with an example
How to set up Router for an app
Using Refs and Keys in React JS
Understanding ReactJS Components
Top React JS Interview Questions and Answers
IOT
Articles
IoT project of controlling home light using WiFi Node MCU, and Relay module
IoT project of Sonar system using Ultrasonic Sensor HC-SR04 and Arduino device
IoT project of Temperature and Pressure measurement using Pressure sensor BMP180 and Arduino device
IoT (Internet of Things) Project: Google Firebase controlling LED with NodeMCU
IoT link Communication Protocol
IoT Decision Framework and Architecture
IoT in Energy and Biometrics Domain
IoT in Security Camera and Smart Home
IoT in Smart Agriculture and Healthcare Domain
IoT Network Layer and Session Layer Protocols
IoT – Platform and Thing Worx in IoT
IoT Project Google Firebase controlling LED using Android App
IoT Project: Google Firebase using NodeMCU ESP8266
IoT project of controlling home light using WiFi Node MCU, and Relay module
Overview of Internet of Things (IoT)
CISCO Virtualized Packet Zone and Salesforce in IoT
Embedded Devices (System) in (IoT) and IoT Ecosystem
GE Predix Platform and Eclipse IoT
How is IoT transforming businesses and IoT in transportation
eBooks
Interview Questions
Videos
IoT project of controlling home light using WiFi Node MCU, and Relay module
IoT project of Sonar system using Ultrasonic Sensor HC-SR04 and Arduino device
IoT project of Temperature and Pressure measurement using Pressure sensor BMP180 and Arduino device
IoT (Internet of Things) Project: Google Firebase controlling LED with NodeMCU
IoT link Communication Protocol
IoT Decision Framework and Architecture
IoT in Energy and Biometrics Domain
IoT in Security Camera and Smart Home
IoT in Smart Agriculture and Healthcare Domain
IoT Network Layer and Session Layer Protocols
IoT – Platform and Thing Worx in IoT
IoT Project Google Firebase controlling LED using Android App
IoT Project: Google Firebase using NodeMCU ESP8266
IoT project of controlling home light using WiFi Node MCU, and Relay module
Overview of Internet of Things (IoT)
CISCO Virtualized Packet Zone and Salesforce in IoT
Embedded Devices (System) in (IoT) and IoT Ecosystem
GE Predix Platform and Eclipse IoT
How is IoT transforming businesses and IoT in transportation
Internet of Things – Contiki and Security Flaws
Internet of Things – Security and Identity Protection
Top Internet of Things (IoT) Interview Questions and Answers
Mobile Development
Articles
eBooks
Interview Questions
Videos
Operating Systems
Articles
eBooks
Interview Questions
Videos
Programming and Frameworks
Articles
Cookies in Laravel based web applications
Encryption and Hashing in Laravel
How to create Blade Templates Layout
How to Create Façade in Laravel
How to perform Redirections and connect to Database
Installation Process of Laravel
Introduction to Laravel and its History
Laravel vs CodeIgniter and Laravel Vs Symphony
Laravel vs Django and Laravel vs WordPress
Middleware Mechanism in Laravel
Process of Authentication and Authorization in Laravel
Responses in Laravel web applications
Understanding Release Process in Laravel
How to setup Check/Money Order payment method in Magento 2
Dynamic Content Handling in PHP
eBooks
Interview Questions
Videos
Hibernate and Spring
Articles
How to use Node Package Manager and REPL Terminal
Handling GET and POST Request in NodeJS
Using Sessions and POJO Classes in Hibernate
Transaction Management in Spring
Overview and Architecture of Spring Framework
ORM Overview and Overview of Hibernate
IoC Containers, AOP and JDBC Framework in Spring
Injecting Inner Beans and Collections in Spring
How to use Criteria Queries in Hibernate
How to perform Java Based Configuration in Spring
How to Install Hibernate and its Configuration
eBooks
Interview Questions
Videos
How to use Node Package Manager and REPL Terminal
Handling GET and POST Request in NodeJS
Using Sessions and POJO Classes in Hibernate
Transaction Management in Spring
Overview and Architecture of Spring Framework
ORM Overview and Overview of Hibernate
IoC Containers, AOP and JDBC Framework in Spring
Injecting Inner Beans and Collections in Spring
How to use Criteria Queries in Hibernate
How to perform Java Based Configuration in Spring
How to Install Hibernate and its Configuration
Java
Articles
Variables and Keywords in Java
Transaction Management and Batch Processing in JDBC
StringBuffer and StringBuilder Class in Java
String Vs StringBuffer Vs StringBuilder
Stream API Improvement in Java 9
Static Binding and Dynamic Binding and Final Keyword
Serialization and Reflection in Java
Properties class and Generics in Java
Method Parameter Reflection in Java
Java StringJoiner and ArrayList Vs Vector
Java Queue and Deque Interface
Java Parallel Array Sorting and Type Inference
Java Networking and Socket Programming
Java Nested Interface and Method Overloading and Overriding
Java Method References and Functional Interfaces
Java Garbage Collection and Java Runtime Class
Java forEach loop and Collectors
Java Comments and Naming Conventions
Java 9 Process API Improvement
Java 9 Module System and Control Panel
Java 9 Anonymous Inner Classes Improvement and SafeVarargs Annotation
Introduction to Java and History of Java
Inter-thread communication and Deadlock in Java
How to write the Hello World Java program
How to create Immutable class in Java
Features of Java and C++ Vs Java
ExceptionHandling with MethodOverriding in Java
Difference between JDK, JRE, and JVM
Deep Dive into Threads in Java
Deep Dive into LinkedList in Java
Deep dive into LinkedHashMap and TreeMap
Deep dive into HashSet , LinkedHashSet and TreeSet
Deep Dive into HashMap in Java
Deep Dive into ArrayList in Java
Conditional Statements in Java
Concept of Method Overloading and Method Overriding in Java
Concept of Inheritance and Aggregation in Java
Comparable and Comparator interface in Java
eBooks
Interview Questions
Videos
Variables and Keywords in Java
Transaction Management and Batch Processing in JDBC
StringBuffer and StringBuilder Class in Java
String Vs StringBuffer Vs StringBuilder
Stream API Improvement in Java 9
Static Binding and Dynamic Binding and Final Keyword
Serialization and Reflection in Java
Properties class and Generics in Java
Method Parameter Reflection in Java
Java StringJoiner and ArrayList Vs Vector
Java Queue and Deque Interface
Java Parallel Array Sorting and Type Inference
Java Networking and Socket Programming
Java Nested Interface and Method Overloading and Overriding
Java Method References and Functional Interfaces
Java Garbage Collection and Java Runtime Class
Java forEach loop and Collectors
Java Comments and Naming Conventions
Java 9 Process API Improvement
Java 9 Module System and Control Panel
Java 9 Anonymous Inner Classes Improvement and SafeVarargs Annotation
Introduction to Java and History of Java
Inter-thread communication and Deadlock in Java
How to write the Hello World Java program
How to create Immutable class in Java
Features of Java and C++ Vs Java
ExceptionHandling with MethodOverriding in Java
Difference between JDK, JRE, and JVM
Deep Dive into Threads in Java
Deep Dive into LinkedList in Java
Deep dive into LinkedHashMap and TreeMap
Deep dive into HashSet , LinkedHashSet and TreeSet
Deep Dive into HashMap in Java
Deep Dive into ArrayList in Java
Conditional Statements in Java
Concept of Method Overloading and Method Overriding in Java
Concept of Inheritance and Aggregation in Java
Comparable and Comparator interface in Java
JSP
eBooks
Interview Questions
Videos
Laravel
Articles
Understanding Release Process in Laravel
Responses in Laravel web applications
Process of Authentication and Authorization in Laravel
Middleware Mechanism in Laravel
Laravel vs Django and Laravel vs WordPress
Laravel vs CodeIgniter and Laravel Vs Symphony
Introduction to Laravel and its History
Installation Process of Laravel
How to perform Redirections and connect to Database
How to Create Façade in Laravel
How to create Blade Templates Layout
Encryption and Hashing in Laravel
Cookies in Laravel based web applications
Contracts and CSRF Protection in Laravel
Available Validation Rules of Laravel
eBooks
Interview Questions
Videos
Understanding Release Process in Laravel
Responses in Laravel web applications
Process of Authentication and Authorization in Laravel
Middleware Mechanism in Laravel
Laravel vs Django and Laravel vs WordPress
Laravel vs CodeIgniter and Laravel Vs Symphony
Introduction to Laravel and its History
Installation Process of Laravel
How to perform Redirections and connect to Database
How to Create Façade in Laravel
How to create Blade Templates Layout
Encryption and Hashing in Laravel
Cookies in Laravel based web applications
Contracts and CSRF Protection in Laravel
Available Validation Rules of Laravel
Magento
Articles
Architecture of Magento 2 and Product Overview
How to use the multi language feature of Magento
How to Setup System Theme, Page Title, Layout and New Pages in Magento
How to Setup Shipping Rates and Payment Plans in Magento
How to setup shipping methods in Magento 2
How to Setup Paypal Payment and Google checkout in Magento
How to Setup Newsletter in Magento
How to Setup Google Analytics Youtube Videos and Facebook Likes in Magento
How to setup Check/Money Order payment method in Magento 2
How to set up Zero Subtotal Checkout payment method in Magento 2
How to set up the tax rules, tax rates, and tax zones in Magento 2
How to set up Purchase Order (PO) payment method in Magento 2
How to set up Order Emails in Magento 2
How to set up multiple websites, stores, and store views in Magento 2
How to Set up Contact, Categories, Products and Inventory in Magento
How to set up Cash on Delivery (COD) payment method in Magento 2
How to set up Bank Transfer payment method in Magento 2
How to set up Authorize.net method in Magento 2
How to Manage Tax Classes in Magento
How to Install Magento on your system
How to install Magento 2 using Composer
How to install Magento 2 on windows
Ways for Site Optimization in Magento
Store Configuration in Magento 2
Search Engine Optimization in Magento 2
Products and their Types in Magento 2
Overview of Magento and its Features
Orders Life Cycle in Magento 2
Ways for Site Optimization in Magento
Basic Configuration in Magento 2
Create and Manage CMS (Content Management System) in Magento 2
How to add the product on Home page in Magento 2
How to configure and Manage the Inventory in Magento 2
How to create Attribute Sets in Magento 2
How to create Product Attributes in Magento 2
How to create Product Category in Magento 2
eBooks
Interview Questions
Videos
Architecture of Magento 2 and Product Overview
How to use the multi language feature of Magento
How to Setup System Theme, Page Title, Layout and New Pages in Magento
How to Setup Shipping Rates and Payment Plans in Magento
How to setup shipping methods in Magento 2
How to Setup Paypal Payment and Google checkout in Magento
How to Setup Newsletter in Magento
How to Setup Google Analytics Youtube Videos and Facebook Likes in Magento
How to setup Check/Money Order payment method in Magento 2
How to set up Zero Subtotal Checkout payment method in Magento 2
How to set up the tax rules, tax rates, and tax zones in Magento 2
How to set up Purchase Order (PO) payment method in Magento 2
How to set up Order Emails in Magento 2
How to set up multiple websites, stores, and store views in Magento 2
How to Set up Contact, Categories, Products and Inventory in Magento
How to set up Cash on Delivery (COD) payment method in Magento 2
How to set up Bank Transfer payment method in Magento 2
How to set up Authorize.net method in Magento 2
How to Manage Tax Classes in Magento
How to Install Magento on your system
How to install Magento 2 using Composer
How to install Magento 2 on windows
Ways for Site Optimization in Magento
Store Configuration in Magento 2
Search Engine Optimization in Magento 2
Products and their Types in Magento 2
Overview of Magento and its Features
Orders Life Cycle in Magento 2
Ways for Site Optimization in Magento
Basic Configuration in Magento 2
Create and Manage CMS (Content Management System) in Magento 2
How to add the product on Home page in Magento 2
How to configure and Manage the Inventory in Magento 2
How to create Attribute Sets in Magento 2
How to create Product Attributes in Magento 2
How to create Product Category in Magento 2
NodeJS
Articles
Scaffolding and Middleware in ExpressJS
Overview of expressJS, installation and Request-response model
NodeJS environment setup and Creating First Application
How to scale application in NodeJS and concept of packaging
Event Driven Programming in NodeJS
Cookies Management, Routing and Template Engine in ExpressJS
eBooks
Interview Questions
Videos
Scaffolding and Middleware in ExpressJS
Overview of expressJS, installation and Request-response model
NodeJS environment setup and Creating First Application
How to scale application in NodeJS and concept of packaging
Event Driven Programming in NodeJS
Cookies Management, Routing and Template Engine in ExpressJS
PHP
Articles
Variable Types and Constant Types in PHP
Operations in MySQL DB using PHP
Object Oriented Programming in PHP
Login with Facebook and Paypal Integration in PHP
Dynamic Content Handling in PHP
How to Install PHP on your system
How to access information from DB using PHP and AJAX
Error and Exception Handling in PHP
CRUD operations in MySQL DB using PHP
eBooks
Interview Questions
Videos
Variable Types and Constant Types in PHP
Operations in MySQL DB using PHP
Object Oriented Programming in PHP
Login with Facebook and Paypal Integration in PHP
Dynamic Content Handling in PHP
How to Install PHP on your system
How to access information from DB using PHP and AJAX
Error and Exception Handling in PHP
CRUD operations in MySQL DB using PHP
Python
Articles
Variable Types and Basic Operators in Python
Time Series, Geographical and Graph Data in Python
Sending Email using SMTP in Python
Processing CSV, JSON and XLS Data in Python
MySQL Database Access in Python
Multithreaded Programming in Python
Introduction to Python and Installing Python
How to draw different Charts in Python
Handling Relational and NoSQL Databases in Python
Handling Date and Time in Python
Extension Programming with C in Python
Data Wrangling and Data Aggregations in Python
Data Science Libraries in Python
eBooks
Interview Questions
Videos
Variable Types and Basic Operators in Python
Time Series, Geographical and Graph Data in Python
Sending Email using SMTP in Python
Processing CSV, JSON and XLS Data in Python
MySQL Database Access in Python
Multithreaded Programming in Python
Introduction to Python and Installing Python
How to draw different Charts in Python
Handling Relational and NoSQL Databases in Python
Handling Date and Time in Python
Extension Programming with C in Python
Data Wrangling and Data Aggregations in Python
Data Science Libraries in Python
Servlet
eBooks
Interview Questions
Videos
Spring Boot
Articles
How to write a Scheduler on the Spring applications and CORS Support
Service Components in Spring Boot
Tracing Micro Service Logs in Spring Boot
How to perform Bootstrapping on a Spring Boot application
How to use Spring Boot JDBC driver connection to connect the database
How to write a unit test case by using Mockito and Web Controller
Spring Boot - Code Structure and Build Systems
Spring Boot - Enabling Swagger2
Spring Boot - Google Cloud Platform
Spring Boot - Rest Controller Unit Test
Spring Boot - Securing Web Applications
Spring Boot - Tomcat Deployment
Spring Boot Architecture and Why Spring Boot is used
Spring Boot Security mechanisms and OAuth2 with JWT
Spring Vs Spring Boot Vs Spring MVC
Application Properties in Spring Boot
How to implement the SMS sending and making voice calls by using Spring Boot with Twilio
Building RESTful Web Services using Spring Boot
Consuming RESTful Web Services by using jQuery AJAX
Create a Web Application in Spring Boot using Thymeleaf
Creating Servlet Filter using Spring Boot
Exception Handling in Spring Boot
File Handling using Spring Boot
How to add the Google OAuth2 Sign-In by using Spring Boot application with Gradle build
How to build a Eureka Server using Spring Boot
How to build an interactive web application by using Spring Boot with Web sockets
How to Build Spring Boot Admin Server and Client
How to Create Applications that consume Restful Web Services
How to Create Spring Cloud Configuration Server
How to Configure Flyway Database in your Spring Boot application
How to create a Docker Image using Maven and Gradle
How to create a Spring Boot Application using Maven and Gradle
How to create Zuul Proxy Server application in Spring Boot
How to implement the Apache Kafka in Spring Boot application
eBooks
Interview Questions
Videos
How to write a Scheduler on the Spring applications and CORS Support
Service Components in Spring Boot
Tracing Micro Service Logs in Spring Boot
How to perform Bootstrapping on a Spring Boot application
How to use Spring Boot JDBC driver connection to connect the database
How to write a unit test case by using Mockito and Web Controller
Spring Boot - Code Structure and Build Systems
Spring Boot - Enabling Swagger2
Spring Boot - Google Cloud Platform
Spring Boot - Rest Controller Unit Test
Spring Boot - Securing Web Applications
Spring Boot - Tomcat Deployment
Spring Boot Architecture and Why Spring Boot is used
Spring Boot Security mechanisms and OAuth2 with JWT
Spring Vs Spring Boot Vs Spring MVC
Application Properties in Spring Boot
How to implement the SMS sending and making voice calls by using Spring Boot with Twilio
Building RESTful Web Services using Spring Boot
Consuming RESTful Web Services by using jQuery AJAX
Create a Web Application in Spring Boot using Thymeleaf
Creating Servlet Filter using Spring Boot
Exception Handling in Spring Boot
File Handling using Spring Boot
How to add the Google OAuth2 Sign-In by using Spring Boot application with Gradle build
How to build a Eureka Server using Spring Boot
How to build an interactive web application by using Spring Boot with Web sockets
How to Build Spring Boot Admin Server and Client
How to Create Applications that consume Restful Web Services
How to Create Spring Cloud Configuration Server
How to Configure Flyway Database in your Spring Boot application
How to create a Docker Image using Maven and Gradle
How to create a Spring Boot Application using Maven and Gradle
How to create Zuul Proxy Server application in Spring Boot
How to implement the Apache Kafka in Spring Boot application
Variable Types and Basic Operators in Python
Time Series, Geographical and Graph Data in Python
Sending Email using SMTP in Python
Processing CSV, JSON and XLS Data in Python
MySQL Database Access in Python
Multithreaded Programming in Python
Introduction to Python and Installing Python
How to draw different Charts in Python
Handling Relational and NoSQL Databases in Python
Handling Date and Time in Python
Extension Programming with C in Python
Data Wrangling and Data Aggregations in Python
Data Science Libraries in Python
Calendar and Date and Time in Python
Scaffolding and Middleware in ExpressJS
Overview of expressJS, installation and Request-response model
NodeJS environment setup and Creating First Application
How to use Node Package Manager and REPL Terminal
How to scale application in NodeJS and concept of packaging
Handling GET and POST Request in NodeJS
Event Driven Programming in NodeJS
Cookies Management, Routing and Template Engine in ExpressJS
Concept of Callbacks and Streams in NodeJS
Comparison of NodeJS with other programming languages
Using Sessions and POJO Classes in Hibernate
Transaction Management in Spring
Overview and Architecture of Spring Framework
ORM Overview and Overview of Hibernate
IoC Containers, AOP and JDBC Framework in Spring
Architecture of Magento 2 and Product Overview
Injecting Inner Beans and Collections in Spring
How to use Criteria Queries in Hibernate
How to perform Java Based Configuration in Spring
How to Install Hibernate and its Configuration
Environment Setup for Spring Framework
Variables and Keywords in Java
Transaction Management and Batch Processing in JDBC
StringBuffer and StringBuilder Class in Java
String Vs StringBuffer Vs StringBuilder
Stream API Improvement in Java 9
Static Binding and Dynamic Binding and Final Keyword
Serialization and Reflection in Java
Properties class and Generics in Java
Method Parameter Reflection in Java
Java StringJoiner and ArrayList Vs Vector
Java Queue and Deque Interface
Java Parallel Array Sorting and Type Inference
Java Networking and Socket Programming
Java Nested Interface and Method Overloading and Overriding
Java Method References and Functional Interfaces
Java Garbage Collection and Java Runtime Class
Java forEach loop and Collectors
Java Comments and Naming Conventions
Java 9 Process API Improvement
Java 9 Module System and Control Panel
Java 9 Anonymous Inner Classes Improvement and SafeVarargs Annotation
Introduction to Java and History of Java
Inter-thread communication and Deadlock in Java
How to write the Hello World Java program
How to create Immutable class in Java
Features of Java and C++ Vs Java
ExceptionHandling with MethodOverriding in Java
Difference between JDK, JRE, and JVM
Deep Dive into Threads in Java
Deep Dive into LinkedList in Java
Deep dive into LinkedHashMap and TreeMap
Deep dive into HashSet , LinkedHashSet and TreeSet
Deep Dive into HashMap in Java
Deep Dive into ArrayList in Java
Conditional Statements in Java
Concept of Method Overloading and Method Overriding in Java
Concept of Inheritance and Aggregation in Java
Comparable and Comparator interface in Java
Call by Value and Call by Reference in Java
Cookies in Laravel based web applications
Encryption and Hashing in Laravel
How to create Blade Templates Layout
How to Create Façade in Laravel
How to perform Redirections and connect to Database
Installation Process of Laravel
Introduction to Laravel and its History
Laravel vs CodeIgniter and Laravel Vs Symphony
Laravel vs Django and Laravel vs WordPress
Middleware Mechanism in Laravel
Process of Authentication and Authorization in Laravel
Responses in Laravel web applications
Understanding Release Process in Laravel
How to setup Check/Money Order payment method in Magento 2
Dynamic Content Handling in PHP
Object Oriented Programming in PHP
How to use the multi language feature of Magento
How to Setup System Theme, Page Title, Layout and New Pages in Magento
How to Setup Shipping Rates and Payment Plans in Magento
How to setup shipping methods in Magento 2
How to Setup Paypal Payment and Google checkout in Magento
How to Setup Newsletter in Magento
How to Setup Google Analytics Youtube Videos and Facebook Likes in Magento
How to setup Check/Money Order payment method in Magento 2
How to set up Zero Subtotal Checkout payment method in Magento 2
How to set up the tax rules, tax rates, and tax zones in Magento 2
How to set up Purchase Order (PO) payment method in Magento 2
How to set up Order Emails in Magento 2
How to set up multiple websites, stores, and store views in Magento 2
How to Set up Contact, Categories, Products and Inventory in Magento
How to set up Cash on Delivery (COD) payment method in Magento 2
How to set up Bank Transfer payment method in Magento 2
How to set up Authorize.net method in Magento 2
How to Manage Tax Classes in Magento
How to Install Magento on your system
How to install Magento 2 using Composer
How to install Magento 2 on windows
Ways for Site Optimization in Magento
Store Configuration in Magento 2
Search Engine Optimization in Magento 2
Products and their Types in Magento 2
Overview of Magento and its Features
Orders Life Cycle in Magento 2
Ways for Site Optimization in Magento
Basic Configuration in Magento 2
Create and Manage CMS (Content Management System) in Magento 2
How to add the product on Home page in Magento 2
How to configure and Manage the Inventory in Magento 2
How to create Attribute Sets in Magento 2
How to create Product Attributes in Magento 2
How to create Product Category in Magento 2
How to generate Order Report in Magento 2
How to create Product in Magento 2
Variable Types and Constant Types in PHP
Operations in MySQL DB using PHP
Object Oriented Programming in PHP
Login with Facebook and Paypal Integration in PHP
Dynamic Content Handling in PHP
How to Install PHP on your system
How to access information from DB using PHP and AJAX
Error and Exception Handling in PHP
CRUD operations in MySQL DB using PHP
Handling Arrays and Strings in PHP
Standard Tag Library (JSTL) in JSP
Page Redirecting and Hits Counter and Auto Refresh
Overview of Java Server Pages and its Architecture
How to Access Database with JSP
Servlets - Server HTTP Response
Servlets - Page Redirection and Auto Refresh
Internationalization in Servlets
Handling Date and Time using Servlets
Exception Handling in Servlets
Overview of Servlets and setup of Environment
How to write a Scheduler on the Spring applications and CORS Support
Service Components in Spring Boot
Tracing Micro Service Logs in Spring Boot
How to perform Bootstrapping on a Spring Boot application
How to use Spring Boot JDBC driver connection to connect the database
How to write a unit test case by using Mockito and Web Controller
Spring Boot - Code Structure and Build Systems
Spring Boot - Enabling Swagger2
Spring Boot - Google Cloud Platform
Spring Boot - Rest Controller Unit Test
Spring Boot - Securing Web Applications
Spring Boot - Tomcat Deployment
Spring Boot Architecture and Why Spring Boot is used
Spring Boot Security mechanisms and OAuth2 with JWT
Spring Vs Spring Boot Vs Spring MVC
Application Properties in Spring Boot
How to implement the SMS sending and making voice calls by using Spring Boot with Twilio
Building RESTful Web Services using Spring Boot
Consuming RESTful Web Services by using jQuery AJAX
Create a Web Application in Spring Boot using Thymeleaf
Creating Servlet Filter using Spring Boot
Exception Handling in Spring Boot
File Handling using Spring Boot
How to add the Google OAuth2 Sign-In by using Spring Boot application with Gradle build
How to build a Eureka Server using Spring Boot
How to build an interactive web application by using Spring Boot with Web sockets
How to Build Spring Boot Admin Server and Client
How to Create Applications that consume Restful Web Services
How to Create Spring Cloud Configuration Server
How to Configure Flyway Database in your Spring Boot application
How to create a Docker Image using Maven and Gradle
How to create a Spring Boot Application using Maven and Gradle
How to create Zuul Proxy Server application in Spring Boot
How to implement the Apache Kafka in Spring Boot application
How to implement the Internationalization in Spring Boot
How to implement the Hystrix in a Spring Boot application
Login with Facebook and Paypal Integration in PHP
Understanding Release Process in Laravel
Responses in Laravel web applications
Process of Authentication and Authorization in Laravel
Middleware Mechanism in Laravel
Laravel vs Django and Laravel vs WordPress
Laravel vs CodeIgniter and Laravel Vs Symphony
Introduction to Laravel and its History
Installation Process of Laravel
How to perform Redirections and connect to Database
How to Create Façade in Laravel
How to create Blade Templates Layout
Encryption and Hashing in Laravel
Cookies in Laravel based web applications
Contracts and CSRF Protection in Laravel
Available Validation Rules of Laravel
Artisan Console for interaction in Laravel
Application Structure of Laravel
Expression Language (EL) in JSP
Expression Language (EL) in JSP
Project Management and Methodologies
Articles
eBooks
Interview Questions
Videos
Robotic Process Automation
eBooks
Interview Questions
Videos
RPA-UiPath
Articles
eBooks
Interview Questions
Videos
Working of RPA and its Services
Understanding User Interface Components
UiPath Studio - Workflow Design
RPA Use Cases and Applications
RPA Life Cycle and Implementation
Recording using UiPath in Detail
Keyboard Shortcuts and Customization in UiPath Studio
Key Basics of UiPath and the related concepts
Installation of UiPath on your local system
How to work with Automation Projects in UiPath and their Debugging methods
How to deal and work with variables and arguments in UiPath
Data Scraping and Screen Scraping in UiPath
Comparison of RPA and AI, Test Automation and Traditional Automation
Architecture and Components of RPA
Advantages and drawbacks of RPA
Top Robotic Process Automation (RPA) with UiPath Interview Questions and Answers
Salesforce
Articles
Different Levels of Data Access in Salesforce
Variables & Formulas in Salesforce
Using Records, Fields and Tables in Salesforce
Using Forms and List Controllers in Salesforce
Creating Static Resources in Salesforce
Standard and Custom Objects in Salesforce platform
Overview of Salesforce and its architecture
Master Detail Relationship in Salesforce
Lookup Relationship in Salesforce
How to Import Data in Salesforce
How to Export Data from Salesforce
How to Define Sharing Rules in Salesforce
How to create Visual force Pages in Salesforce
How to create Reports and Dashboards in Salesforce
Get Started with Salesforce - Environment
How to Create a Role Hierarchy in Salesforce
eBooks
Interview Questions
Videos
Apex Programming
Articles
Classes and Methods in Apex programming language
Concept of Objects and Interfaces in Apex programming language
Database Methods and process of executing the Apex class in Salesforce
Deployment in Salesforce using Sandbox
Enterprise Application Development Example
How to Perform Debugging in Apex
How to perform the various Database Modification Functionalities in Salesforce
How to perform Unit Testing in Apex
Overview of Apex Programming and its environment
Search Functionality using SOSL and SOQL
Understand Batch Processing in Salesforce Apex
Understanding deciding, Loops and Collections in Apex
Understanding Governor Limits in Salesforce Apex
Understanding the info Types and variables in Apex programming language
Understanding the environment for Salesforce Apex development
Understanding the String Manipulation, Arrays and Constants in Apex programming language
eBooks
Interview Questions
Videos
Classes and Methods in Apex programming language
Concept of Objects and Interfaces in Apex programming language
Database Methods and process of executing the Apex class in Salesforce
Deployment in Salesforce using Sandbox
Enterprise Application Development Example
How to Perform Debugging in Apex
How to perform the various Database Modification Functionalities in Salesforce
How to perform Unit Testing in Apex
Overview of Apex Programming and its environment
Search Functionality using SOSL and SOQL
Understand Batch Processing in Salesforce Apex
Understanding deciding, Loops and Collections in Apex
Understanding Governor Limits in Salesforce Apex
Understanding the info Types and variables in Apex programming language
Understanding the environment for Salesforce Apex development
Understanding the String Manipulation, Arrays and Constants in Apex programming language
Different Levels of Data Access in Salesforce
Variables & Formulas in Salesforce
Using Records, Fields and Tables in Salesforce
Using Forms and List Controllers in Salesforce
Creating Static Resources in Salesforce
Standard and Custom Objects in Salesforce platform
Overview of Salesforce and its architecture
Master Detail Relationship in Salesforce
Lookup Relationship in Salesforce
How to Import Data in Salesforce
How to Export Data from Salesforce
How to Define Sharing Rules in Salesforce
How to create Visual force Pages in Salesforce
How to create Reports and Dashboards in Salesforce
Get Started with Salesforce - Environment
How to Create a Role Hierarchy in Salesforce
Classes and Methods in Apex programming language
Concept of Objects and Interfaces in Apex programming language
Database Methods and process of executing the Apex class in Salesforce
Deployment in Salesforce using Sandbox
Enterprise Application Development Example
How to Perform Debugging in Apex
How to perform the various Database Modification Functionalities in Salesforce
How to perform Unit Testing in Apex
Overview of Apex Programming and its environment
Search Functionality using SOSL and SOQL
Understand Batch Processing in Salesforce Apex
Understanding deciding, Loops and Collections in Apex
Understanding Governor Limits in Salesforce Apex
Understanding the info Types and variables in Apex programming language
Understanding the environment for Salesforce Apex development
Understanding the String Manipulation, Arrays and Constants in Apex programming language
Using Formula Fields in Salesforce
SAP
Articles
unv Universe in SAP Business Object
Using Formula Bar and Universe Operations in SAP Universe Designer
Using LOVs and Create, Edit and Save a Universe
How to Display Financial Tables in SAP Simple Finance
Concept of Period Lock Transaction in SAP Simple Finance
Concept of Asset Scrapping in SAP Simple Finance
Create Default Account Assignment in SAP Simple Finance
How to Create a Primary Cost in G-L Account
Asset Accounting in SAP Simple Finance
Concept of Integrated Business Planning and Integration of Simple Finance with other Modules
eBooks
Interview Questions
Videos
SAP Business Object
Articles
Using Filters in SAP BO Analysis
Sheets and Sharing Workspaces in SAP BO Analysis
Perform Conditional Formatting in SAP BO Analysis
Overview of SAP Business Object Analysis
How to create a Workspace in SAP Business Objects
How to Connect to SAP BW in SAP Business Objects
Export Options in SAP BO Analysis
Concept of Sub Analysis in SAP BO
eBooks
Interview Questions
Videos
Using Filters in SAP BO Analysis
Sheets and Sharing Workspaces in SAP BO Analysis
Perform Conditional Formatting in SAP BO Analysis
Overview of SAP Business Object Analysis
How to create a Workspace in SAP Business Objects
How to Connect to SAP BW in SAP Business Objects
Export Options in SAP BO Analysis
Concept of Sub Analysis in SAP BO
Calculations in SAP BO Analysis
SAP Hana
Articles
Alert Monitoring and Logging in SAP Hana
Authentications and Authorization Methods in SAP HANA
DXC Replication Method and CTL Method and MDX provider in SAP Hana
Excel Integration with SAP Hana and Bi 4.0 Connectivity to Hana Views
User Administration & Role Management and Security Overview in SAP Hana
Usage of SQL Script in SAP Hana
SQL Triggers, Synonym and Data Profiling in SAP Hana
SQL Overview and Data Types in SAP Hana
SQL Functions and Operators in SAP Hana
SQL Expressions, Stored Procedures and Sequences in SAP Hana
Packages and Attribute and Analytic View in SAP Hana
Modeling and Schemas in SAP HANA
Log Based and ETL Based Replication in SAP Hana
License Management and Auditing in SAP Hana
Information Modeler and System Monitor in SAP HANA
High Availability and Backup and Recovery in SAP Hana
Export and Import Options in Sap Hana
eBooks
Videos
Alert Monitoring and Logging in SAP Hana
Authentications and Authorization Methods in SAP HANA
DXC Replication Method and CTL Method and MDX provider in SAP Hana
Excel Integration with SAP Hana and Bi 4.0 Connectivity to Hana Views
User Administration & Role Management and Security Overview in SAP Hana
Usage of SQL Script in SAP Hana
SQL Triggers, Synonym and Data Profiling in SAP Hana
SQL Overview and Data Types in SAP Hana
SQL Functions and Operators in SAP Hana
SQL Expressions, Stored Procedures and Sequences in SAP Hana
Packages and Attribute and Analytic View in SAP Hana
Modeling and Schemas in SAP HANA
Log Based and ETL Based Replication in SAP Hana
License Management and Auditing in SAP Hana
Information Modeler and System Monitor in SAP HANA
High Availability and Backup and Recovery in SAP Hana
Export and Import Options in Sap Hana
Data Replication Overview in SAP Hana
Analytic Privileges and Information Composer in SAP Hana
SAP Hana Adminstration
Articles
SAP HANA Admin Studio and System Management
Overview of SAP HANA Administration
SAP HANA License Management and Multitenant DB Container Management
Smart Data Access and Integration with Hadoop
How to Start, Stop and Monitor a HANA System
HANA XS Application Service and Data Provisioning in SAP Hana
Data Compression and Solman Integration in SAP Hana
eBooks
Interview Questions
Videos
SAP HANA Admin Studio and System Management
Overview of SAP HANA Administration
SAP HANA License Management and Multitenant DB Container Management
Smart Data Access and Integration with Hadoop
How to Start, Stop and Monitor a HANA System
HANA XS Application Service and Data Provisioning in SAP Hana
Data Compression and Solman Integration in SAP Hana
SAP Hana Finance
Articles
Profitability Analysis and Management Accounting in SAP Simple Finance
Overview of SAP Hana and SAP Hana Finance
Migration and Manual Reposting of Costs in SAP Simple Finance
How to Display Financial Tables in SAP Simple Finance
Concept of Period Lock Transaction in SAP Simple Finance
Concept of Asset Scrapping in SAP Simple Finance
Create Default Account Assignment in SAP Simple Finance
How to Create a Primary Cost in G-L Account
Ledger Management in SAP Simple Finance
Reporting Options and G/L Accounting in SAP Simple Finance
Universal Journal and Document Number in SAP Simple Finance
SAP Simple Finance Architecture and Deployment Options
Asset Accounting in SAP Simple Finance
Concept of Integrated Business Planning and Integration of Simple Finance with other Modules
eBooks
Interview Questions
Videos
Profitability Analysis and Management Accounting in SAP Simple Finance
Overview of SAP Hana and SAP Hana Finance
Migration and Manual Reposting of Costs in SAP Simple Finance
How to Display Financial Tables in SAP Simple Finance
Concept of Period Lock Transaction in SAP Simple Finance
Concept of Asset Scrapping in SAP Simple Finance
Create Default Account Assignment in SAP Simple Finance
How to Create a Primary Cost in G-L Account
Ledger Management in SAP Simple Finance
Reporting Options and G/L Accounting in SAP Simple Finance
Universal Journal and Document Number in SAP Simple Finance
SAP Simple Finance Architecture and Deployment Options
Asset Accounting in SAP Simple Finance
Concept of Integrated Business Planning and Integration of Simple Finance with other Modules
SAP Hana Logistics
Articles
Supply Chain Planning and Integrated Business Planning in SAP Hana Logistics
Overview of SAP Hana Simple Logistics
MRP Procedures and Key Features in SAP Simple Logistics
MIGO Transactions in SAP Simple Logistics
Manufacturing Process in SAP Simple Logistics
Invoice Management and Operational Procurement in SAP Simple Logistics
How to Manage Business Partner in SAP Simple Logistics
How to Execute MRP Live planning
How to Create Business Partner in SAP HANA Logistics
Fiori UX and Deployment and Procurement Types in SAP Hana Logistics
Execute Discrete Production in SAP Hana Logistics
Contract Management and Perform Procurement Transfer Stock in SAP Hana Logistics
eBooks
Interview Questions
Videos
Supply Chain Planning and Integrated Business Planning in SAP Hana Logistics
Overview of SAP Hana Simple Logistics
MRP Procedures and Key Features in SAP Simple Logistics
MIGO Transactions in SAP Simple Logistics
Manufacturing Process in SAP Simple Logistics
Invoice Management and Operational Procurement in SAP Simple Logistics
How to Manage Business Partner in SAP Simple Logistics
How to Execute MRP Live planning
How to Create Business Partner in SAP HANA Logistics
Fiori UX and Deployment and Procurement Types in SAP Hana Logistics
Execute Discrete Production in SAP Hana Logistics
Contract Management and Perform Procurement Transfer Stock in SAP Hana Logistics
Concept of Simplification Item in SAP Simple Logistics
SAP UDT & IDT
Articles
Building Data Foundation in SAP IDT
Building Query in Query Panel, Publishing in SAP IDT
Business Layer Properties in SAP IDT
Dealing with Published Universes in SAP IDT
Deploying Universe in SAP Universe Designer
Format Editor Overview in SAP IDT
How to create universe in SAP IDT
How to use Table Browser and Derived Tables in SAP Universal Designer
Joins In Data Foundation in SAP IDT
Managing Connections in SAP IDT
Managing Resources in Repository, Qualifiers and Owners
OLAP Data Sources in SAP Universe Designer
Overview of SAP Universe Designer
unv Universe in SAP Business Object
Using Formula Bar and Universe Operations in SAP Universe Designer
Using LOVs and Create, Edit and Save a Universe
Concept of Calculated Measures and Aggregate Awareness
Business Layer View in SAP IDT
eBooks
Interview Questions
Videos
Building Data Foundation in SAP IDT
Building Query in Query Panel, Publishing in SAP IDT
Business Layer Properties in SAP IDT
Dealing with Published Universes in SAP IDT
Deploying Universe in SAP Universe Designer
Format Editor Overview in SAP IDT
How to create universe in SAP IDT
How to use Table Browser and Derived Tables in SAP Universal Designer
Joins In Data Foundation in SAP IDT
Managing Connections in SAP IDT
Managing Resources in Repository, Qualifiers and Owners
OLAP Data Sources in SAP Universe Designer
Overview of SAP Universe Designer
unv Universe in SAP Business Object
Using Formula Bar and Universe Operations in SAP Universe Designer
Using LOVs and Create, Edit and Save a Universe
Concept of Calculated Measures and Aggregate Awareness
Business Layer View in SAP IDT
Sap Webi
Articles
Working with Reports in SAP Webi
Sending Documents in SAP Web Intelligence
Query Filters and Filters Type in SAP Webi
Queries using Bex and Analysis View in SAP Webi
How to use Formulas and Variables in SAP Webi
How to use Breaks, Sorts and Ranking Data in SAP Webi
How to Create SAP Webi documents
How to achieve Conditional Formatting in SAP Webi
eBooks
Interview Questions
Videos
Working with Reports in SAP Webi
Sending Documents in SAP Web Intelligence
Query Filters and Filters Type in SAP Webi
Queries using Bex and Analysis View in SAP Webi
How to use Formulas and Variables in SAP Webi
How to use Breaks, Sorts and Ranking Data in SAP Webi
How to Create SAP Webi documents
How to achieve Conditional Formatting in SAP Webi
SAP HANA Admin Studio and System Management
Overview of SAP HANA Administration
SAP HANA License Management and Multitenant DB Container Management
Smart Data Access and Integration with Hadoop
Building Data Foundation in SAP IDT
Building Query in Query Panel, Publishing in SAP IDT
Business Layer Properties in SAP IDT
Dealing with Published Universes in SAP IDT
Deploying Universe in SAP Universe Designer
Format Editor Overview in SAP IDT
How to create universe in SAP IDT
How to use Table Browser and Derived Tables in SAP Universal Designer
Joins In Data Foundation in SAP IDT
Managing Connections in SAP IDT
Managing Resources in Repository, Qualifiers and Owners
OLAP Data Sources in SAP Universe Designer
Overview of SAP Universe Designer
unv Universe in SAP Business Object
Using Formula Bar and Universe Operations in SAP Universe Designer
Using LOVs and Create, Edit and Save a Universe
Concept of Calculated Measures and Aggregate Awareness
Business Layer View in SAP IDT
Profitability Analysis and Management Accounting in SAP Simple Finance
Overview of SAP Hana and SAP Hana Finance
Migration and Manual Reposting of Costs in SAP Simple Finance
How to Display Financial Tables in SAP Simple Finance
Concept of Period Lock Transaction in SAP Simple Finance
Concept of Asset Scrapping in SAP Simple Finance
Create Default Account Assignment in SAP Simple Finance
How to Create a Primary Cost in G-L Account
Ledger Management in SAP Simple Finance
Reporting Options and G/L Accounting in SAP Simple Finance
Universal Journal and Document Number in SAP Simple Finance
SAP Simple Finance Architecture and Deployment Options
Alert Monitoring and Logging in SAP Hana
Authentications and Authorization Methods in SAP HANA
DXC Replication Method and CTL Method and MDX provider in SAP Hana
Excel Integration with SAP Hana and Bi 4.0 Connectivity to Hana Views
Working with Reports in SAP Webi
Sending Documents in SAP Web Intelligence
Query Filters and Filters Type in SAP Webi
Queries using Bex and Analysis View in SAP Webi
How to use Formulas and Variables in SAP Webi
How to use Breaks, Sorts and Ranking Data in SAP Webi
How to Create SAP Webi documents
How to achieve Conditional Formatting in SAP Webi
Filtering Report Data in SAP Webi
Drill Options in Reports and Sharing Reports in SAP Webi
Supply Chain Planning and Integrated Business Planning in SAP Hana Logistics
Overview of SAP Hana Simple Logistics
MRP Procedures and Key Features in SAP Simple Logistics
MIGO Transactions in SAP Simple Logistics
Manufacturing Process in SAP Simple Logistics
Invoice Management and Operational Procurement in SAP Simple Logistics
How to Manage Business Partner in SAP Simple Logistics
How to Execute MRP Live planning
How to Create Business Partner in SAP HANA Logistics
Fiori UX and Deployment and Procurement Types in SAP Hana Logistics
Execute Discrete Production in SAP Hana Logistics
Contract Management and Perform Procurement Transfer Stock in SAP Hana Logistics
Concept of Simplification Item in SAP Simple Logistics
Analyze Sales Orders in SAP Simple Logistics
User Administration & Role Management and Security Overview in SAP Hana
Usage of SQL Script in SAP Hana
SQL Triggers, Synonym and Data Profiling in SAP Hana
SQL Overview and Data Types in SAP Hana
SQL Functions and Operators in SAP Hana
SQL Expressions, Stored Procedures and Sequences in SAP Hana
Packages and Attribute and Analytic View in SAP Hana
Modeling and Schemas in SAP HANA
Log Based and ETL Based Replication in SAP Hana
License Management and Auditing in SAP Hana
Information Modeler and System Monitor in SAP HANA
High Availability and Backup and Recovery in SAP Hana
Export and Import Options in Sap Hana
Data Replication Overview in SAP Hana
Analytic Privileges and Information Composer in SAP Hana
Using Filters in SAP BO Analysis
Sheets and Sharing Workspaces in SAP BO Analysis
Perform Conditional Formatting in SAP BO Analysis
Overview of SAP Business Object Analysis
How to create a Workspace in SAP Business Objects
Asset Accounting in SAP Simple Finance
Concept of Integrated Business Planning and Integration of Simple Finance with other Modules
How to Connect to SAP BW in SAP Business Objects
Export Options in SAP BO Analysis
Concept of Sub Analysis in SAP BO
Calculations in SAP BO Analysis
Aggregations and Hierarchies in SAP BO Analysis
SAP IDT - Overview and User Interface
Creating Parameters and Schemas in SAP Universe Designer
How to Start, Stop and Monitor a HANA System
HANA XS Application Service and Data Provisioning in SAP Hana
Data Compression and Solman Integration in SAP Hana
Authentication Methods supported by SAP HANA
Auditing Activities in SAP Hana
Top SAP S4 HANA Logistics Interview Questions and Answers
Top SAP S4 HANA Finance Interview Questions and Answers
Top SAP HANA Interview Questions and Answers
Software Testing
Articles
eBooks
Interview Questions
Videos
Selenium WebDriver
Articles
How to run your Selenium Test Scripts on IE Browser
How to run your Selenium Test Scripts on Firefox Browser
Comparison of Selenium vs QTP and Selenium Tool Suite
How to run your Selenium Test Scripts on Safari Browser
Overview of Selenium WebDriver
Overview of Selenium, its features and limitations
Scrolling an internet page in Selenium WebDriver
Selenium IDE- Locating Strategies by Identifier and By Id
Selenium IDE- Locating Strategies by Name, XPath , CSS and DOM
How to run your Selenium Test Scripts on Chrome Browser
How to Handle Alerts in Selenium WebDriver
Selenium WebDriver - Navigation and Web Element Commands
How to handle radio buttons and checkbox in selenium web driver
Selenium WebDriver - Browser Commands
Selenium WebDriver- Locating Strategies and Handling Drop-downs
Comparison between Selenium WebDriver and Selenium RC
Creating Test Cases Manually in Selenium IDE
How to create Login test suit in Selenium IDE
How to create your First Selenium Automation Test Script
Selenium IDE- Commands (Selenese)
Using Assertions in Selenium WebDriver
Overview of Selenium Integrated Development Environment (IDE)
eBooks
Interview Questions
Videos
How to run your Selenium Test Scripts on IE Browser
How to run your Selenium Test Scripts on Firefox Browser
Comparison of Selenium vs QTP and Selenium Tool Suite
How to run your Selenium Test Scripts on Safari Browser
Overview of Selenium WebDriver
Overview of Selenium, its features and limitations
Scrolling an internet page in Selenium WebDriver
Selenium IDE- Locating Strategies by Identifier and By Id
Selenium IDE- Locating Strategies by Name, XPath , CSS and DOM
How to run your Selenium Test Scripts on Chrome Browser
How to Handle Alerts in Selenium WebDriver
Selenium WebDriver - Navigation and Web Element Commands
How to handle radio buttons and checkbox in selenium web driver
Selenium WebDriver - Browser Commands
Selenium WebDriver- Locating Strategies and Handling Drop-downs
Comparison between Selenium WebDriver and Selenium RC
Creating Test Cases Manually in Selenium IDE
How to create Login test suit in Selenium IDE
How to create your First Selenium Automation Test Script
Selenium IDE- Commands (Selenese)
Using Assertions in Selenium WebDriver
Overview of Selenium Integrated Development Environment (IDE)
Selenium with Maven
Articles
Execute Selenium code through Maven and TestNG
How to Configure Selenium using NUnit in Visual Studio
How to Configure Selenium with Visual Studio in C#
How to handle or download dependency Jar using Maven
Write a Selenium test script using C#
Selenium Test Script using NUnit
How to write a Selenium test script using C#
eBooks
Interview Questions
Videos
Execute Selenium code through Maven and TestNG
How to Configure Selenium using NUnit in Visual Studio
How to Configure Selenium with Visual Studio in C#
How to handle or download dependency Jar using Maven
Write a Selenium test script using C#
Selenium Test Script using NUnit
How to write a Selenium test script using C#
Test NG
Articles
How to Run test cases in TestNG without java compiler
Overview of TestNG and its Features
Importance of XML file in TestNG Configuration
How to use TestNG Annotation Attributes
How to Run test cases with Regex in TestNG
How to install TestNG Framework and Configuration in Eclipse
How to enable and disable test cases in TestNG
eBooks
Interview Questions
Videos
How to Run test cases in TestNG without java compiler
Overview of TestNG and its Features
Importance of XML file in TestNG Configuration
How to use TestNG Annotation Attributes
How to Run test cases with Regex in TestNG
How to install TestNG Framework and Configuration in Eclipse
How to enable and disable test cases in TestNG
How to run your Selenium Test Scripts on IE Browser
How to run your Selenium Test Scripts on Firefox Browser
Comparison of Selenium vs QTP and Selenium Tool Suite
How to run your Selenium Test Scripts on Safari Browser
Overview of Selenium WebDriver
Overview of Selenium, its features and limitations
Scrolling an internet page in Selenium WebDriver
Selenium IDE- Locating Strategies by Identifier and By Id
Selenium IDE- Locating Strategies by Name, XPath , CSS and DOM
How to run your Selenium Test Scripts on Chrome Browser
How to Handle Alerts in Selenium WebDriver
Selenium WebDriver - Navigation and Web Element Commands
How to handle radio buttons and checkbox in selenium web driver
Selenium WebDriver - Browser Commands
Selenium WebDriver- Locating Strategies and Handling Drop-downs
Comparison between Selenium WebDriver and Selenium RC
Creating Test Cases Manually in Selenium IDE
How to create Login test suit in Selenium IDE
How to create your First Selenium Automation Test Script
Execute Selenium code through Maven and TestNG
How to Configure Selenium using NUnit in Visual Studio
How to Configure Selenium with Visual Studio in C#
How to handle or download dependency Jar using Maven
Write a Selenium test script using C#
Selenium Test Script using NUnit
How to write a Selenium test script using C#
Write and Execute the Selenium test script
Using Maven with Selenium TestNG
Selenium IDE- Commands (Selenese)
Using Assertions in Selenium WebDriver
Overview of Selenium Integrated Development Environment (IDE)
How to Run test cases in TestNG without java compiler
Overview of TestNG and its Features
Importance of XML file in TestNG Configuration
How to use TestNG Annotation Attributes
How to Run test cases with Regex in TestNG
How to install TestNG Framework and Configuration in Eclipse
How to enable and disable test cases in TestNG
How to create TestNG Listeners
Top Apache Impala Interview Questions and Answers
Last updated on Feb 18 2022Table of Contents
What is Impala?
Basically, for processing huge volumes of data Impala is an MPP (Massive Parallel Processing) SQL query engine which is stored in Hadoop cluster. Moreover, this is an advantage that it is an open-source software which is written in C++ and Java. Also, it offers high performance and low latency compared to other SQL engines for Hadoop.
To be more specific, it is a highest performing SQL engine that offers the fastest way to access data that is stored in Hadoop Distributed File System HDFS.
Why we need Impala Hadoop?
Along with the scalability and flexibility of Apache Hadoop, Impala combines the SQL support and multi-user performance of a traditional analytic database, by utilizing standard components. Like HDFS, HBase, Metastore, YARN, and Sentry.
Also, users can communicate with HDFS or HBase using SQL queries With Impala, even in a faster way compared to other SQL engines like Hive.
It can read almost all the file formats used by Hadoop. Like Parquet, Avro, RCFile.
Moreover, it uses the same metadata, SQL syntax (Hive SQL), ODBC driver, and user interface (Hue Beeswax) as Apache Hive. Also, offers a familiar and unified platform for batch-oriented or real-time queries.
Impala is not based on MapReduce algorithms, unlike Apache Hive.
Hence, Impala faster than Apache Hive, since it reduces the latency of utilizing MapReduce.
What are Impala Architecture Components?
Basically, the Impala engine consists of different daemon processes that run on specific hosts within your CDH cluster.
i. The Impala Daemon
While it comes to Impala Daemon, it is one of the core components of the Hadoop Impala. Basically, it runs on every node in the CDH cluster. It generally identified by the Impalad process.
Moreover, we use it to read and write the data files. In addition, it accepts the queries transmitted from impala-shell command, ODBC, JDBC or Hue.
ii. The Impala Statestore
To check the health of all Impala Daemons on all the data nodes in the Hadoop cluster we use The Impala Statestore. Also, we call it a process statestored.
However, only in the Hadoop cluster one such process we need on one host.
The major advantage of this Daemon is it informs all the Impala Daemons if an Impala Daemon goes down. Hence, they can avoid the failed node while distributing future queries.
iii. The Impala Catalog Service
The Catalog Service tells metadata changes from Impala SQL statements to all the Datanodes in Hadoop cluster. Basically, by Daemon process catalogd it is physically represented. Also, we only need one such process on one host in the Hadoop cluster.
Generally, as catalog services are passed through statestored, statestored and catalogd process will be running on the same host.
Moreover, it also avoids the need to issue REFRESH and INVALIDATE METADATA statements. Even when the metadata changes are performed by statements issued through Impala.
How to call Impala Built-in Functions.
In order to call any of these Impala functions by using the SELECT statement. Basically, for any required arguments we can omit the FROM clause and supply literal values, for the most function:
select abs(-1);
select concat(‘The rain ‘, ‘in Spain’);
select po
What is Impala Data Types?
There is a huge set of data types available in Impala. Basically, those Impala Data Types we use for table columns, expression values, and function arguments and return values. Each Impala Data Types serves a specific purpose. Types are:
- BIGINT
- BOOLEAN
- CHAR
- DECIMAL
- DOUBLE
- FLOAT
- INT
- SMALLINT
- STRING
- TIMESTAMP
- TINYINT
- VARCHAR
- ARRAY
- Map
- Struct
State some advantages of Impala:
There are several advantages of Cloudera Impala. So, here is a list of those advantages.
- Fast Speed
Basically, we can process data that is stored in HDFS at lightning-fast speed with traditional SQL knowledge, by using Impala.
- No need to move data
However, while working with Impala, we don’t need data transformation and data movement for data stored on Hadoop. Even if the data processing is carried where the data resides (on Hadoop cluster),
- Easy Access
Also, we can access the data that is stored in HDFS, HBase, and Amazon s3 without the knowledge of Java (MapReduce jobs), by using Impala. That implies we can access them with a basic idea of SQL queries.
- Short Procedure
Basically, while we write queries in business tools, the data has to be gone through a complicated extract-transform-load (ETL) cycle. However, this procedure is shortened with Impala. Moreover, with the new techniques, time-consuming stages of loading & reorganizing is resolved. Like, exploratory data analysis & data discovery making the process faster.
- File Format
However, for large-scale queries typical in data warehouse scenarios, Impala is pioneering the use of the Parquet file format, a columnar storage layout. Basically, that is very optimized for it.
State some disadvantages of Impala.
Some of the drawbacks of using Impala are as follows −
i. No support SerDe
There is no support for Serialization and Deserialization in Impala.
ii. No custom binary files
Basically, we cannot read custom binary files in Impala. It only read text files.
iii. Need to refresh
However, we need to refresh the tables always, when we add new records/ files to the data directory in HDFS.
iv. No support for triggers
Also, it does not provide any support for triggers.
v. No Updation
In Impala, We cannot update or delete individual records.
How to control Access to Data in Impala?
Basically, through Authorization, Authentication, and Auditing we can control data access in Cloudera Impala. Also, for user authorization, we can use the Sentry open source project. Sentry includes a detailed authorization framework for Hadoop. Also, associates various privileges with each user of the computer. In addition, by using authorization techni we can control access to Impala data.
How Apache Impala Works with CDH
This below graphic illustrates how Impala is positioned in the broader Cloudera environment:
So, above Architecture diagram, implies how Impala relates to other Hadoop components. Like HDFS, the Hive Metastore database, client programs [ JDBC and ODBC applications] and the Hue web UI.
There are following components the Impala solution is composed of. Such as:
i. Clients
To issue queries or complete administrative tasks such as connecting to Impala we can use these interfaces.
ii. Hive Metastore
to store information about the data available to Impala, we use it.
iii. Impala
Basically, a process, which runs on DataNodes, coordinates and executes queries. By using Impala clients, each instance of Impala can receive, plan, and coordinate queries. However, all queries are distributed among Impala nodes. So, these nodes then act as workers, executing parallel query fragments.
iv. HBase and HDFS
It is generally a Storage for data to be queried.
Relational Databases and Impala
Here are some of the key differences between SQL and Impala Query language
- Impala
It uses an SQL like query language that is similar to HiveQL.
- Relational databases
It use SQL language.
- Impala
In Impala, you cannot update or delete individual records.
- Relational Databases
Here, it is possible to update or delete individual records.
- Impala
It does not support transactions.
- Relational databases
It supports transactions.
- Impala
It does not support indexing.
- Relational Databases
It supports indexing.
Hive, HBase, and Impala.
Here is a comparative analysis of HBase, Hive, and Impala.
– HBase
HBase is wide-column store database based on Apache Hadoop. It uses the concepts of BigTable.
– Hive
Hive is a data warehouse software. Using this, we can access and manage large distributed datasets, built on Hadoop.
-Impala
Impala is a tool to manage, analyze data that is stored on Hadoop.
-HBase
The data model of HBase is wide column store.
– Hive
Hive follows the Relational model.
-Impala
Impala follows the Relational model.
– HBase
HBase is developed using Java language.
– Hive
Hive is developed using Java language.
-Impala
Impala is developed using C++.
– HBase
The data model of HBase is schema-free.
– Hive
Here, the data model of Hive is Schema-based.
-Impala
The data model of Impala is Schema-based.
– HBase
HBase provides Java, RESTful and, Thrift API’s.
–Hive
Hive provides JDBC, ODBC, Thrift API’s.
-Impala
Impala provides JDBC and ODBC API’s.
– HBase
Supports programming languages like C, C#, C++, Groovy, Java PHP, Python, and Scala.
–Hive
Supports programming languages like C++, Java, PHP, and Python.
-Impala
Impala supports all languages supporting JDBC/ODBC.
– HBase
It offers support for triggers.
–Hive
Hive does not provide any support for triggers.
-Impala
It does not provide any support for triggers.
How Do I Configure Hadoop High Availability (ha) For Impala?
To relay rets back and forth to the Impala servers we can set up a proxy server, for load balancing and high availability.
What Is the Maximum Number of Rows In A Table?
We cannot say any maximum number. Because some customers have used Impala to query a table with over a trillion rows.
On Which Hosts Does Impala Run?
However, for good performance, Cloudera strongly recommends running the impala daemon on each Data Node. But it is not a hard requirement. Since the data must be transmitted from one host to another for processing by “remote reads” if there are data blocks with no Impala daemons running on any of the hosts containing replicas of those blocks, queries involving that data could be very inefficient. Although, it is a condition Impala normally tries to avoid.
How Are Joins Performed in Impala?
Using a cost-based method, Impala automatically determines the most efficient order in which to join tables, on the basis of their overall size and number of rows. As per new feature, for efficient join performance, the COMPUTE STATS statement gathers information about each table that is crucial. For join queries, Impala chooses between two techniques, known as “broadcast joins” and “partitioned joins”.
How Is Impala Metadata Managed?
There are two pieces of metadata, Impala uses. Such as the catalog information from the Hive metastore and the file metadata from the NameNode. Currently, this metadata is lazily populated and cached when an impala needs it to plan a query.
What Load Do Concurrent Queries Produce on The Namenode?
The load Impala generates is very similar to MapReduce. Impala contacts the NameNode during the planning phase to get the file metadata. Every impala will read files as part of normal processing of the query.
What size is recommended for each node?
Generally, in each node, 128 GB RAM is recommended.
Is It possible to share data files between different components?
By using Impala it is possible to share data files between different components with no copy or export/import step.
Does if offer scaling?
It provides distributed queries for convenient scaling in a cluster environment. Also, offers to use of cost-effective commodity hardware.
Is There A Dual Table?
To running queries against a single-row table named DUAL to try out expressions, built-in functions, and UDFs. It does not have a DUAL table. Also, we can issue a SELECTstatement without any table name, to achieve the same result,
select 2+2;
select substr(‘hello’,2,1);
select pow(10,6);
How Do I Load A Big Csv File into A Partitioned Table?
In order to load a data file into a partitioned table, use a two-stage process. Especially, when the data file includes fields like year, month, and so on that correspond to the partition key columns. to bring the data into an unpartitioned text table, use the LOAD DATA or CREATE EXTERNAL TABLE statement.
Further, use an INSERT … SELECT statement to copy the data from the unpartitioned table to a partitioned one. Also, include a PARTITION clause in the INSERTstatement to specify the partition key columns.
How Do I Try Impala Out?
To look at the core features and functionality on Impala, the easiest way to try out Impala is to download the Cloudera QuickStart VM and start the Impala service through Cloudera Manager, then use impala-shell in a terminal window or the Impala Query UI in the Hue web interface.
To do performance testing and try out the management features for Impala on a cluster, you need to move beyond the QuickStart VM with its virtualized single-node environment. Ideally, download the Cloudera Manager software to set up the cluster, then install the Impala software through Cloudera Manager.
Does Cloudera Offer a VM For Demonstrating Impala?
Cloudera offers a demonstration VM called the QuickStart VM, available in VMWare, VirtualBox, and KVM formats. For more information, see the Cloudera QuickStart VM. After booting the QuickStart VM, many services are turned off by default; in the Cloudera Manager UI that appears automatically, turn on Impala and any other components that you want to try out.
Where Can I Find Impala Documentation?
Starting with Impala 1.3.0, Impala documentation is integrated with the CDH 5 documentation, in addition to the standalone Impala documentation for use with CDH 4. For CDH 5, the core Impala developer and administrator information remains in the associated Impala documentation portion. Information about Impala release notes, installation, configuration, startup, and security is embedded in the corresponding CDH 5 guides.
- New features
- Known and fixed issues
- Incompatible changes
- Installing Impala
- Upgrading Impala
- Configuring Impala
- Starting Impala
- Security for Impala
- CDH Version and Packaging Information
How Much Memory Is Required?
Although Impala is not an in-memory database, when dealing with large tables and large result sets, you should expect to dedicate a substantial portion of physical memory for the impala daemon. Recommended physical memory for an Impala node is 128 GB or higher. If practical, devote approximately 80% of physical memory to Impala.
The amount of memory required for an Impala operation depends on several factors:
The file format of the table. Different file formats represent the same data in more or fewer data files. The compression and encoding for each file format might require a different amount of temporary memory to decompress the data for analysis.
Whether the operation is a SELECT or an INSERT. For example, Parquet tables require relatively little memory to query, because Impala reads and decompresses data in 8 MB chunks. Inserting into a Parquet table is a more memory-intensive operation because the data for each data file (potentially hundreds of megabytes, depending on the value of the PARQUET_FILE_SIZE query option) is stored in memory until encoded, compressed, and written to disk.
Whether the table is partitioned or not, and whether a query against a partitioned table can take advantage of partition pruning.
Whether the final result set is sorted by the ORDER BY clause. Each Impala node scans and filters a portion of the total data, and applies the LIMIT to its own portion of the result set. In Impala 1.4.0 and higher, if the sort operation requires more memory than is available on any particular host, Impala uses a temporary disk work area to perform the sort. The intermediate result sets are all sent back to the coordinator node, which does the final sorting and then applies the LIMIT clause to the final result set.
For example, if you execute the query:
select * from giant_table order by some_column limit 1000;
and your cluster has 50 nodes, then each of those 50 nodes will transmit a maximum of 1000 rows back to the coordinator node. The coordinator node needs enough memory to sort (LIMIT * cluster_size) rows, although in the end the final result set is at most LIMIT rows, 1000 in this case.
Likewise, if you execute the query:
select * from giant_table where test_val > 100 order by some_column;
then each node filters out a set of rows matching the WHERE conditions, sorts the results (with no size limit), and sends the sorted intermediate rows back to the coordinator node. The coordinator node might need substantial memory to sort the final result set, and so might use a temporary disk work area for that final phase of the query.
Whether the query contains any join clauses, GROUP BY clauses, analytic functions, or DISTINCT operators. These operations all require some in-memory work areas that vary depending on the volume and distribution of data. In Impala 2.0 and later, these kinds of operations utilize temporary disk work areas if memory usage grows too large to handle.
The size of the result set. When intermediate results are being passed around between nodes, the amount of data depends on the number of columns returned by the query. For example, it is more memory-efficient to query only the columns that are actually needed in the result set rather than always issuing SELECT *.
The mechanism by which work is divided for a join query. You use the COMPUTE STATS statement, and query hints in the most difficult cases, to help Impala pick the most efficient execution plan.
What Features from Relational Databases or Hive Are Not Available In Impala?
Querying streaming data.
Deleting individual rows. You delete data in bulk by overwriting an entire table or partition, or by dropping a table.
Indexing (not currently). LZO-compressed text files can be indexed outside of Impala, as described in Using LZO-Compressed Text Files.
Full text search on text fields. The Cloudera Search product is appropriate for this use case.
Custom Hive Serializer/Deserializer classes (SerDes). Impala supports a set of common native file formats that have built-in SerDes in CDH.
Checkpointing within a query. That is, Impala does not save intermediate results to disk during long-running queries. Currently, Impala cancels a running query if any host on which that query is executing fails. When one or more hosts are down, Impala reroutes future queries to only use the available hosts, and Impalad detects when the hosts come back up and begins using them again. Because a query can be submitted through any Impala node, there is no single point of failure. In the future, we will consider adding additional work allocation features to Impala, so that a running query would complete even in the presence of host failures.
Encryption of data transmitted between Impala daemons.
Hive indexes.
Non-Hadoop data stores, such as relational databases.
How Do I Know How Many Impala Nodes Are In My Cluster?
The Impala statestore keeps track of how many impalad nodes are currently available. You can see this information through the statestore web interface. For example, at the URL http://statestore_host:25010/metrics you might see lines like the following:
statestore.live-backends:3
statestore.live-backends.list:[host1:22000, host1:26000, host2:22000]
The number of impalad nodes is the number of list items referring to port 22000, in this case two. (Typically, this number is one less than the number reported by the statestore.live-backends line.) If an impalad node became unavailable or came back after an outage, the information reported on this page would change appropriately.
Are Results Returned as They Become Available, Or All at Once When A Query Completes?
Impala streams results whenever they are available, when possible. Certain SQL operations (aggregation or ORDER BY) require all of the input to be ready before Impala can return results.
Why Does My Select Statement Fail?
When a SELECT statement fails, the cause usually falls into one of the following categories:
A timeout because of a performance, capacity, or network issue affecting one particular node.
Excessive memory use for a join query, resulting in automatic cancellation of the query.
A low-level issue affecting how native code is generated on each node to handle particular WHERE clauses in the query. For example, a machine instruction could be generated that is not supported by the processor of a certain node. If the error message in the log suggests the cause was an illegal instruction, consider turning off native code generation temporarily, and trying the query again.
Malformed input data, such as a text data file with an enormously long line, or with a delimiter that does not match the character specified in the FIELDS TERMINATED BY clause of the CREATE TABLE statement.
Does Impala Performance Improve as It Is Deployed to More Hosts in A Cluster in Much the Same Way That Hadoop Performance Does?
Yes. Impala scales with the number of hosts. It is important to install Impala on all the DataNodes in the cluster, because otherwise some of the nodes must do remote reads to retrieve data not available for local reads. Data locality is an important architectural aspect for Impala performance.
Is the Hdfs Block Size Reduced to Achieve Faster Query Results?
No. Impala does not make any changes to the HDFS or HBase data sets.
The default Parquet block size is relatively large (256 MB in Impala 2.0 and later; 1 GB in earlier releases). You can control the block size when creating Parquet files using the PARQUET_FILE_SIZE query option.
Does Impala Use Caching?
Impala does not cache table data. It does cache some table and file metadata. Although queries might run faster on subsequent iterations because the data set was cached in the OS buffer cache, Impala does not explicitly control this.
Impala takes advantage of the HDFS caching feature in CDH 5. You can designate which tables or partitions are cached through the CACHED and UNCACHED clauses of the CREATE TABLE and ALTER TABLE statements. Impala can also take advantage of data that is pinned in the HDFS cache through the hdfscacheadmin command.
Is Impala Intended To Handle Real Time Queries In Low-latency Applications Or Is It For Ad Hoc Queries For The Purpose Of Data Exploration?
Ad-hoc queries are the primary use case for Impala. We anticipate it being used in many other situations where low-latency is required. Whether Impala is appropriate for any particular use-case depends on the workload, data size and query volume.
How Does Impala Compare to Hive and Pig?
Impala is different from Hive and Pig because it uses its own daemons that are spread across the cluster for queries. Because Impala does not rely on MapReduce, it avoids the startup overhead of MapReduce jobs, allowing Impala to return results in real time.
Can I Do Transforms or Add New Functionality?
Impala adds support for UDFs in Impala 1.2. You can write your own functions in C++, or reuse existing Java-based Hive UDFs. The UDF support includes scalar functions and user-defined aggregate functions (UDAs). User-defined table functions (UDTFs) are not currently supported.
Impala does not currently support an extensible serialization-deserialization framework (SerDes), and so adding extra functionality to Impala is not as straightforward as for Hive or Pig.
Can Any Impala Query Also Be Executed in Hive?
Yes. There are some minor differences in how some queries are handled, but Impala queries can also be completed in Hive. Impala SQL is a subset of HiveQL, with some functional limitations such as transforms.
Is Hive an Impala Requirement?
The Hive metastore service is a requirement. Impala shares the same metastore database as Hive, allowing Impala and Hive to access the same tables transparently.
Hive itself is optional, and does not need to be installed on the same nodes as Impala. Currently, Impala supports a wider variety of read (query) operations than write (insert) operations; you use Hive to insert data into tables that use certain file formats.
Is Impala Production Ready?
Impala has finished its beta release cycle, and the 1.0, 1.1, and 1.2 GA releases are production ready. The 1.1.x series includes additional security features for authorization, an important requirement for production use in many organizations. The 1.2.x series includes important performance features, particularly for large join queries. Some Cloudera customers are already using Impala for large workloads.
The Impala 1.3.0 and higher releases are bundled with corresponding levels of CDH 5. The number of new features grows with each release.
How Do I Configure Hadoop High Availability (ha) For Impala?
You can set up a proxy server to relay rets back and forth to the Impala servers, for load balancing and high availability.
What Is the Maximum Number of Rows in A Table?
There is no defined maximum. Some customers have used Impala to query a table with over a trillion rows.
On Which Hosts Does Impala Run?
Cloudera strongly recommends running the impala daemon on each DataNode for good performance. Although this topology is not a hard requirement, if there are data blocks with no Impala daemons running on any of the hosts containing replicas of those blocks, queries involving that data could be very inefficient. In that case, the data must be transmitted from one host to another for processing by “remote reads”, a condition Impala normally tries to avoid.
How Does Impala Achieve Its Performance Improvements?
These are the main factors in the performance of Impala versus that of other Hadoop components and related technologies.
Impala avoids MapReduce. While MapReduce is a great general parallel processing model with many benefits, it is not designed to execute SQL. Impala avoids the inefficiencies of MapReduce in these ways:
Impala does not materialize intermediate results to disk. SQL queries often map to multiple MapReduce jobs with all intermediate data sets written to disk.
Impala avoids MapReduce start-up time. For interactive queries, the MapReduce start-up time becomes very noticeable. Impala runs as a service and essentially has no start-up time.
Impala can more naturally disperse query plans instead of having to fit them into a pipeline of map and reduce jobs. This enables Impala to parallelize multiple stages of a query and avoid overheads such as sort and shuffle when unnecessary.
Impala uses a more efficient execution engine by taking advantage of modern hardware and technologies:
Impala generates runtime code. Impala uses LLVM to generate assembly code for the query that is being run. Individual queries do not have to pay the overhead of running on a system that needs to be able to execute arbitrary queries.
Impala uses available hardware instructions when possible. Impala uses the supplemental SSE3 (SSSE3) instructions which can offer tremendous speedups in some cases. (Impala 2.0 and 2.1 required the SSE4.1 instruction set; Impala 2.2 and higher relax the restriction again so only SSSE3 is required.)
Impala uses better I/O scheduling. Impala is aware of the disk location of blocks and is able to schedule the order to process blocks to keep all disks busy.
Impala is designed for performance. A lot of time has been spent in designing Impala with sound performance-oriented fundamentals, such as tight inner loops, inline function calls, minimal branching, better use of cache, and minimal memory usage.
What Happens When the Data Set Exceeds Available Memory?
Currently, if the memory required to process intermediate results on a node exceeds the amount available to Impala on that node, the query is cancelled. You can adjust the memory available to Impala on each node, and you can fine-tune the join strategy to reduce the memory required for the biggest queries. We do plan on supporting external joins and sorting in the future.
Keep in mind though that the memory usage is not directly based on the input data set size. For aggregations, the memory usage is the number of rows after grouping. For joins, the memory usage is the combined size of the tables excluding the biggest table, and Impala can use join strategies that divide up large joined tables among the various nodes rather than transmitting the entire table to each node.
Is There an Update Statement?
Impala does not currently have an UPDATE statement, which would typically be used to change a single row, a small group of rows, or a specific column. The HDFS-based files used by typical Impala queries are optimized for bulk operations across many megabytes of data at a time, making traditional UPDATE operations inefficient or impractical.
You can use the following techniques to achieve the same goals as the familiar UPDATE statement, in a way that preserves efficient file layouts for subsequent queries:
Replace the entire contents of a table or partition with updated data that you have already staged in a different location, either using INSERT OVERWRITE, LOAD DATA, or manual HDFS file operations followed by a REFRESH statement for the table. Optionally, you can use built-in functions and expressions in the INSERT statement to transform the copied data in the same way you would normally do in an UPDATE statement, for example to turn a mixed-case string into all uppercase or all lowercase.
To update a single row, use an HBase table, and issue an INSERT … VALUES statement using the same key as the original row. Because HBase handles duplicate keys by only returning the latest row with a particular key value, the newly inserted row effectively hides the previous one.
Why Do I Have to Use Refresh and Invalidate Metadata, What Do They Do?
In Impala 1.2 and higher, there is much less need to use the REFRESH and INVALIDATE METADATA statements:
The new impala-catalog service, represented by the catalog daemon, broadcasts the results of Impala DDL statements to all Impala nodes. Thus, if you do a CREATE TABLE statement in Impala while connected to one node, you do not need to do INVALIDATE METADATA before issuing queries through a different node.
The catalog service only recognizes changes made through Impala, so you must still issue a REFRESH statement if you load data through Hive or by manipulating files in HDFS, and you must issue an INVALIDATE METADATA statement if you create a table, alter a table, add or drop partitions, or do other DDL statements in Hive.
Because the catalog service broadcasts the results of REFRESH and INVALIDATE METADATA statements to all nodes, in the cases where you do still need to issue those statements, you can do that on a single node rather than on every node, and the changes will be automatically recognized across the cluster, making it more convenient to load balance by issuing queries through arbitrary Impala nodes rather than always using the same coordinator node.
How Do I Load a Big Csv File into A Partitioned Table?
To load a data file into a partitioned table, when the data file includes fields like year, month, and so on that correspond to the partition key columns, use a two-stage process. First, use the LOAD DATA or CREATE EXTERNAL TABLE statement to bring the data into an unpartitioned text table. Then use an INSERT … SELECT statement to copy the data from the unpartitioned table to a partitioned one. Include a PARTITION clause in the INSERT statement to specify the partition key columns.
Can I Do Insert … Select * Into A Partitioned Table?
When you use the INSERT … SELECT * syntax to copy data into a partitioned table, the columns corresponding to the partition key columns must appear last in the columns returned by the SELECT *. You can create the table with the partition key columns defined last. Or, you can use the CREATE VIEW statement to create a view that reorders the columns: put the partition key columns last, then do the INSERT … SELECT * from the view.
What Kinds of Impala Queries Or Data Are Best Suited For HBase?
HBase tables are ideal for queries where normally you would use a key-value store. That is, where you retrieve a single row or a few rows, by testing a special unique key column using the = or IN operators.
HBase tables are not suitable for queries that produce large result sets with thousands of rows. HBase tables are also not suitable for queries that perform full table scans because the WHERE clause does not ret specific values from the unique key column.
Use HBase tables for data that is inserted one row or a few rows at a time, such as by the INSERT … VALUES syntax. Loading data piecemeal like this into an HDFS-backed table produces many tiny files, which is a very inefficient layout for HDFS data files.
If the lack of an UPDATE statement in Impala is a problem for you, you can simulate single-row updates by doing an INSERT … VALUES statement using an existing value for the key column. The old row value is hidden; only the new row value is seen by queries.
HBase tables are often wide (containing many columns) and sparse (with most column values NULL). For example, you might record hundreds of different data points for each user of an online service, such as whether the user had registered for an online game or enabled particular account features. With Impala and HBase, you could look up all the information for a specific customer efficiently in a single query. For any given customer, most of these columns might be NULL, because a typical customer might not make use of most features of an online service.
Why we need Impala Hadoop?
Along gone the scalability and malleability of Apache Hadoop, Impala combines the SQL avow and multi-adherent doing of a traditional diagnostic database, by utilizing satisfactory components. Like HDFS, HBase, Metastore, YARN, and Sentry.
Also, users can communicate as soon as HDFS or HBase using SQL queries With Impala, even in a faster way compared to connection SQL engines when Hive.
It can right to use as regards all the file formats used by Hadoop. Like Parquet, Avro, RCFile.
Moreover, it uses the same metadata, SQL syntax (Hive SQL), ODBC driver, and adherent interface (Hue Beeswax) as Apache Hive. Also, offers a familiar and unified platform for batch-oriented or authentic-epoch queries.
State some Impala Hadoop Benefits.
Some of the further are:
- Impala is extremely occurring to date SQL interface. Especially data scientists and analysts already know.
- It in addition to offers the conduct yourself to query high volumes of data (Big Data) in Apache Hadoop.
- Also, it provides distributed queries for convenient scaling in cluster vibes. It offers to use of cost-active commodity hardware.
- By using Impala it is reachable to portion data files surrounded by rotate components together along as well as no copy or export/import step.
How to call Impala Built-in Functions.
In order to call any of these Impala functions by using the SELECT avowal. Basically, for any required arguments we can omit the FROM clause and supply literal values, for the most go to come:
choose abs(-1);
choose concat(The rain, in Spain);
choose po
What are the best features of Impala?
There are several best features of Impala. They are:
Open Source
Basically, sedated the Apache license, Cloudera Impala is manageable freely as a retrieve of the source.
In-memory Processing
While it’s come to meting out, Cloudera Impala supports in-memory data giving out. That implies without any data vigor it accesses/analyzes data that is stored a propos Hadoop data nodes.
Easy Data Access
However, using SQL-in imitation of queries, we can easily access data using Impala. Moreover, Impala offers Common data admission interfaces. That includes:
- JDBC driver.
- ODBC driver.
Faster Access
While we compare Impala to subsidiary SQL engines, Impala offers faster access to the data in HDFS.
Storage Systems
We can easily accrual data in storage systems. Such as HDFS, Apache HBase, and Amazon s3.
- HDFS file formats: delimited text files, Parquet, Avro, Sequence File, and RCFile.
- Compression codec’s: Snappy, GZIP, Deflate, BZIP.
Easy Integration
It is attainable to merge Impala subsequent to issue penetration tools. Such as; Tableau, Pentaho, Micro strategy, and Zoom data.
Joins and Functions
Including SELECT, joins, and aggregate functions, Impala offers the most common SQL-92 features of Hive Query Language (HiveQL).
What are Impala Architecture Components?
Basically, the Impala engine consists of swap daemon processes that run approaching specific hosts within your CDH cluster.
i. The Impala Daemon
While it comes to Impala Daemon, it is one of the core components of the Hadoop Impala. Basically, it runs re all nodes in the CDH cluster. It generally identified by the Impalad process.
Moreover, we use it to the gate and write the data files. In supporter, it accepts the queries transmitted from impala-shell command, ODBC, JDBC, or Hue.
ii. The Impala state store
To check the health of all Impala Daemons re all the data nodes in the Hadoop cluster we use The Impala state store. Also, we call it a process state stored.
However, only in the Hadoop cluster one such process we compulsion approximately one host.
The major advantage of this Daemon is it informs all the Impala Daemons if an Impala Daemon goes beside. Hence, they can avoid the fruitless node even if distributing well ahead queries.
iii. The Impala Catalog Service
The Catalog Service tells metadata changes from Impala SQL statements to each and everyone one of the Datanodes in the Hadoop cluster. Basically, by the Daemon process catalog, it is physically represented. Also, we without help habit one such process in a report to one host in the Hadoop cluster.
Generally, as catalog facilities are passed through state stored, state stored and catalog processes will be giving out on the subject of the same host.
Moreover, it next avoids the dependence on business REFRESH and invalidates METADATA statements. Even in addition to the metadata changes are performed by statements issued through Impala.
State some advantages of Impala
There are several advantages of Cloudera Impala. So, here is a list of those advantages.
Fast Speed
Basically, we can process data that is stored in HDFS at lightning-unexpected animatronics when avowed SQL knowledge, by using Impala.
No dependence to touch data
However, even if life in the tune of Impala, we don’t obsession data transformation and data pursuit for data stored a propos the order of Hadoop. Even if the data paperwork is carried where the data resides (happening for Hadoop cluster),
Easy Access
Also, we can entry the data that is stored in HDFS, HBase, and Amazon s3 without the knowledge of Java (MapReduce jobs), by using Impala. That implies we can access them as soon as a basic idea of SQL queries.
Short Procedure
Basically, even though we write queries in business tools, the data has to be as soon as through a complicated extract-transform-load (ETL) cycle. However, this procedure is edited behind Impala. Moreover, in the ventilate of the new techni, period-absorbing stages of loading & reorganizing are tote going on. Like, exploratory data discovery and data analysis making the process faster.
File Format
However, for large-scale queries typically in data warehouse scenarios, Impala is pioneering the use of the Parquet file format, a columnar storage layout. Basically, that is no study optimized for it.
However, there are many more advantages to Impala. Follow partner; advantages of Impala
State some disadvantages of Impala.
i. No retain Service.
There is no maintenance for Serialization and Deserialization in Impala.
ii. No custom binary files
Basically, we cannot entre custom binary files in Impala. It single-handedly pretentiousness in text files.
iii. Need to refresh
However, we showing off to refresh the tables always, surrounded by we ensure subsidiary records/ files to the data manual in HDFS.
iv. No, retain for triggers
Also, it does not pay for any end for triggers.
v. No Updation
In Impala, We can’t update or delete individual records.
However, there are many more disadvantages to Impala. Follow partner; disadvantages of Impala
How to control Access to Data in Impala?
Basically, through Authorization, Authentication, and Auditing we can control data entry in Cloudera Impala. Also, for adherent endorsement, we can use the Sentry admission source project. Sentry includes a detailed endorsement framework for Hadoop. Also, partners various privileges once each fanatic of the computer. In adding together, by using official approval techniques we can control the entrance to Impala data.
What are the names of Daemons in Impala?
They are:
i. ImpalaD (impala Daemon)
ii. StatestoreD
iii. CatalogD
How Do I Try Impala Out?
To see at the core features and functionality taking place for Impala, the easiest showing off to attempt out Impala is to download the Cloudera Quick Start VM and begin the Impala help through Cloudera Manager, later use impala-shell in a terminal window or the Impala Query UI in the Hue web interface.
To benefit vibrancy psychotherapy and attempt out the dealing out features for Impala concerning a cluster, you habit to shake up subsequently more the QuickStart VM gone its virtualized single-node setting. Ideally, download the Cloudera Manager software to set happening the cluster, also install the Impala software through Cloudera Manager.
Is Avro supported?
Yes, Avro is supported. Impala has always been practiced to query Avro tables. You can use the Impala LOAD DATA statement to load existing Avro data files into a table. Starting behind Impala 1.4, you can make Avro tables subsequently Impala. Currently, you yet use the INSERT avowal in Hive to copy data from the abnormal table into an Avro table.
Are Results Returned as they Become Available, Or All at Once When A Query Completes?
Impala streams result whenever they are easy to realize too, gone than possible. Certain SQL operations (aggregation or ORDER BY) require every single one of the input to be ready back Impala can compensation results.
Does Impala Performance Improve As It Is Deployed To More Hosts In A Cluster In Much The Same Way That Hadoop Performance Does?
Yes. Impala scales once the number of hosts. It is important to install Impala upon every the Data Nodes in the cluster because on the other hand some of the nodes must realize cold reads to relationships data not understandable for local reads. Data locality is an important architectural aspect for Impala appears in.
Is the HDFS Block Size Reduced To Achieve Faster Query Results?
No. Impala does not make any changes to the HDFS or HBase data sets. The default Parquet block size is relatively large (256 MB in Impala 2.0 and far-off along; 1 GB in earlier releases). You can recommend the block size gone creating Parquet files using the PARQUET_FILE_SIZE query inconsistent.
Can Impala Be Used for Complex Event Processing?
For example, in an industrial setting, many agents may generate large amounts of data. Can Impala be used to analyze this data, checking for notable changes in the setting?
Complex Event Processing (CEP) is usually performed by dedicated stream-supervision systems. Impala is not a stream-supervision system, as it most closely resembles a relational database.
Is Impala Intended To Handle Real-Time Queries in Low-latency Applications or Is It For Ad Hoc Queries For The Purpose Of Data Exploration?
Ad-hoc queries are the primary use encounter for Impala. We anticipate it mammal used in many subsidiary situations where low-latency is required. Whether Impala is taken control of any particular use-encounter depends upon the workload, data size, and query volume.
How Does Impala Compare to Hive and Pig?
Impala is alternating from Hive and Pig because it uses its own daemons that are overdue across the cluster for queries. Because Impala does not rely upon MapReduce, it avoids the startup overhead of MapReduce jobs, allowing Impala to reward results in genuine grow obsolete.
Is Hive an Impala Requirement?
The Hive metastore facilitate is a requirement. Impala shares the associated metastore database as Hive, allowing Impala and Hive to entrance the same tables transparently.
The hive itself is optional and does not obsession to be installed upon the same nodes as Impala. Currently, Impala supports a wider variety of dealings (query) operations than write (embellish) operations; you use Hive to sum data into tables that use sure file formats.
How Do I Try Impala Out?
To examine the middle capabilities and capability on Impala, the easiest manner to try out Impala is to down load the Cloudera QuickStart VM and begin the Impala provider thru Cloudera Manager, then use impala-shell in a terminal window or the Impala Query UI in the Hue web interface.
To do overall performance checking out and strive out the management features for Impala on a cluster, you want to transport beyond the QuickStart VM with its virtualized unmarried-node environment. Ideally, download the Cloudera Manager software to installation the cluster, then installation the Impala software thru Cloudera Manager.
Does Cloudera Offer A Vm For Demonstrating Impala?
Cloudera gives an illustration VM called the QuickStart VM, available in VMWare, VirtualBox, and KVM formats. For extra facts, see the Cloudera QuickStart VM. After booting the QuickStart VM, many services are turned off by means of default; in the Cloudera Manager UI that appears automatically, turn on Impala and some other components that you need to strive out.
What Are the Main Features of Impala?
A large set of SQL statements, together with SELECT and INSERT, with joins, Subqueries in Impala SELECT Statements, and Impala Analytic Functions. Highly well matched with HiveQL, and also including a few dealer extensions. For more information.
Distributed, excessive-overall performance queries.
Using Cloudera Manager, you can set up and manage your Impala offerings. Cloudera Manager is the first-class way to get began with Impala to your cluster.
Using Hue for queries.
Appending and putting statistics into tables through the INSERT announcement.
ODBC: Impala is certified to run in opposition to MicroStrategy and Tableau, with restrictions. For more facts, see Configuring Impala to Work with ODBC.
Querying facts saved in HDFS and HBase in a single .
In Impala 2.2.0 and higher, querying statistics saved within the Amazon Simple Storage Service (S3).
Concurrent purchaser rets. Each Impala daemon can take care of multiple concurrent purchaser rets. The results on overall performance depend on your unique hardware and workload.
Kerberos authentication. For extra statistics.
Partitions. With Impala SQL, you may create partitioned tables with the CREATE TABLE statement, and add and drop partitions with the ALTER TABLE assertion. Impala additionally takes advantage of the partitioning present in Hive tables.
Does Impala Support Generic Jdbc?
Impala supports the HiveServer2 JDBC motive force.
Is Avro Supported?
Yes, Avro is supported. Impala has constantly been able to Avro tables. You can use the Impala LOAD DATAstatement to load present Avro data files into a table. Starting with Impala 1.Four, you could create Avro tables with Impala. Currently, you continue to use the INSERT statement in Hive to replicate records from any other table into an Avro table.
How Do I Know How Many Impala Nodes Are In My Cluster?
The Impala statestore maintains music of how many impalad nodes are currently to be had. You can see this information through the statestore internet interface. For instance, on the URL http://statestore_host:25010/metrics you would possibly see traces just like the following:
statestore.Live-backends:3
statestore.Live-backends.List:[host1:22000, host1:26000, host2:22000]
The wide variety of impalad nodes is the quantity of list gadgets regarding port 22000, in this example two. (Typically, this wide variety is one much less than the variety stated via the statestore.Live-backends line.) If an impalad node have become unavailable or got here lower back after an outage, the facts reported on this page could exchange appropriately.
Are Results Returned As They Become Available, Or All At Once When A Query Completes?
Impala streams consequences whenever they’re available, whilst feasible. Certain SQL operations (aggregation or ORDER BY) require all of the input to be ready before Impala can go back effects.
Where Can I Find Impala Documentation?
Starting with Impala 1. Three.0, Impala documentation is incorporated with the CDH 5 documentation, similarly to the standalone Impala documentation to be used with CDH four. For CDH 5, the center Impala developer and administrator information remains in the associated Impala documentation portion. Information about Impala launch notes, set up, configuration, startup, and security are embedded in the corresponding CDH 5 publications.
New capabilities
Known and fixed issues
Incompatible modifications
Installing Impala
Upgrading Impala
Configuring Impala
Starting Impala
Security for Impala
CDH Version and Packaging Information
Why Does My Insert Statement Fail?
When an INSERT announcement fails, it is also the result of exceeding some restrict inside a Hadoop element, normally HDFS.
An INSERT right into a partitioned table may be a strenuous operation because of the opportunity of commencing many files and associated threads simultaneously in HDFS. Impala 1.1.1 includes some improvements to distribute the work more correctly, so that the values for every partition are written with the aid of a single node, in place of as a separate statistics record from every node.
Certain expressions within the SELECT part of the INSERT assertion can complicate the execution planning and result in an inefficient INSERT operation. Try to make the column information sorts of the supply and destination tables healthy up, as an example with the aid of doing ALTER TABLE … REPLACE COLUMNS at the supply table if important. Try to avoid CASE expressions in the SELECT element, because they make the result values more difficult to are expecting than shifting a column unchanged or passing the column through a integrated characteristic.
Be organized to elevate some limits in the HDFS configuration settings, either briefly at some point of the INSERT or permanently in case you regularly run such INSERT statements as part of your ETL pipeline.
The aid utilization of an INSERT declaration can range relying at the document format of the destination desk. Inserting right into a Parquet desk is reminiscence-in depth, due to the fact the statistics for each partition is buffered in memory until it reaches 1 gigabyte, at which point the facts record is written to disk. Impala can distribute the paintings for an INSERT extra effectively while records are to be had for the source desk that is queried in the course of the INSERT statement.
Is Mapreduce Required For Impala? Will Impala Continue To Work As Expected If Mapreduce Is Stopped?
Impala does not use MapReduce at all.
Can Impala Be Used For Complex Event Processing?
For instance, in an business surroundings, many dealers might also generate huge quantities of statistics. Can Impala be used to analyze this information, checking for remarkable modifications within the surroundings?
Complex Event Processing (CEP) is commonly executed through dedicated movement-processing structures. Impala is not a movement-processing machine, as it maximum carefully resembles a relational database.
Is Impala Intended to Handle Real Time Queries In Low-latency Applications Or Is It For Ad Hoc Queries For The Purpose Of Data Exploration?
Ad-hoc queries are the primary use case for Impala. We assume it being used in many different situations wherein low-latency is required. Whether Impala is suitable for any unique use-case depends at the workload, records length and query volume.
Can I Use Impala to Query Data Already Loaded Into Hive And HBase?
There are no extra steps to allow Impala to tables managed via Hive, whether they’re saved in HDFS or HBase. Make sure that Impala is configured to get admission to the Hive metastore effectively and you need to be ready to head. Keep in mind that impalad, by default, runs because the impala consumer, so that you may want to adjust a few report permissions depending on how strict your permissions are currently.
Is Hive an Impala Requirement?
The Hive metastore carrier is a requirement. Impala shares the equal metastore database as Hive, permitting Impala and Hive to get right of entry to the same tables transparently.
Hive itself is non-compulsory, and does now not want to be installed on the identical nodes as Impala. Currently, Impala supports a greater variety of study (query) operations than write (insert) operations; you operate Hive to insert information into tables that use certain record codecs.
What Happens If There Is An Error In Impala?
There isn’t always a unmarried factor of failure in Impala. All Impala daemons are completely able to cope with incoming queries. If a machine fails however, all queries with fragments going for walks on that machine will fail. Because queries are anticipated to return quickly, you can just rerun the query if there may be a failure.
The longer answer: Impala need to be able to hook up with the Hive metastore. Impala aggressively caches metadata so the metastore host have to have minimal load. Impala is predicated at the HDFS NameNode, and, in CDH4, you may configure HA for HDFS. Impala also has centralized services, called the statestore andcatalog offerings, that run on one host most effective. Impala continues to execute queries if the statestore host is down, however it’s going to not get country updates. For example, if a host is introduced to the cluster even as the statestore host is down, the existing instances of impalad walking on the opposite hosts will not find out about this new host. Once the statestore system is restarted, all of the statistics it serves is robotically reconstructed from all strolling Impala daemons.
What Is the Maximum Number of Rows in A Table?
There isn’t any defined most. Some clients have used Impala to a desk with over one thousand billion rows.
On Which Hosts Does Impala Run?
Cloudera strongly recommends going for walks the impalad daemon on every DataNode for desirable performance. Although this topology isn’t a difficult requirement, if there are data blocks with no Impala daemons jogging on any of the hosts containing replicas of these blocks, queries related to that data could be very inefficient. In that case, the information must be transmitted from one host to another for processing by “remote reads”, a situation Impala usually tries to keep away from.
What Load Do Concurrent Queries Produce On The Namenode?
The load Impala generates may be very much like MapReduce. Impala contacts the NameNode during the making plans segment to get the report metadata (this is best run at the host the query turned into sent to). Every impalad will read documents as part of normal processing
How Does Impala Achieve Its Performance Improvements?
These are the main factors in the overall performance of Impala as opposed to that of other Hadoop components and associated technologies.
Impala avoids MapReduce. While MapReduce is a first-rate standard parallel processing model with many blessings, it isn’t designed to execute SQL. Impala avoids the inefficiencies of MapReduce in these methods:
Impala does now not materialize intermediate consequences to disk. SQL queries often map to more than one MapReduce jobs with all intermediate data units written to disk.
Impala avoids MapReduce begin-up time. For interactive queries, the MapReduce start-up time will become very sizeable. Impala runs as a carrier and basically has no start-up time.
Impala can greater evidently disperse plans as a substitute of getting too healthy them into a pipeline of map and reduce jobs. This enables Impala to parallelize multiple ranges of a and keep away from overheads which includes sort and shuffle when needless.
Impala uses a greener execution engine by taking benefit of modern hardware and technology:
Impala generates runtime code. Impala uses LLVM to generate meeting code for the query that is being run. Individual queries do now not ought to pay the overhead of running on a system that needs so that it will execute arbitrary queries.
Impala makes use of to be had hardware commands when feasible. Impala makes use of the supplemental SSE3 (SSSE3) instructions that can provide terrific speedups in a few instances. (Impala 2.Zero and a couple of.1 required the SSE4.1 practise set; Impala 2.2 and higher loosen up the limit again so most effective SSSE3 is required.)
Impala makes use of better I/O scheduling. Impala is privy to the disk vicinity of blocks and is capable of agenda the order to procedure blocks to maintain all disks busy.
Impala is designed for performance. A lot of time has been spent in designing Impala with sound performance-oriented basics, consisting of tight internal loops, inlined function calls, minimum branching, better use of cache, and minimum reminiscence usage.
What Happens When the Data Set Exceeds Available Memory?
Currently, if the memory required to process intermediate results on a node exceeds the amount available to Impala on that node, the is cancelled. You can adjust the reminiscence to be had to Impala on every node, and you may great-song the be part of method to reduce the memory required for the largest queries. We do plan on helping external joins and sorting in the destiny.
Keep in thoughts although that the reminiscence utilization isn’t always immediately primarily based on the input information set size. For aggregations, the memory utilization is the wide variety of rows after grouping. For joins, the memory usage is the mixed length of the tables apart from the largest table, and Impala can use be a part of strategies that divide up large joined tables the various nodes in preference to transmitting the entire table to every node.
What Are the Most Memory-extensive Operations?
If a fails with an error indicating “memory restriction passed”, you would possibly suspect a memory leak. The problem ought to clearly be a query this is structured in a manner that reasons Impala to allocate extra memory than you count on, surpassed the memory allocated for Impala on a specific node. Some examples of query or table structures which are specially reminiscence-in depth are:
INSERT statements the usage of dynamic partitioning, into a desk with many different walls. (Particularly for tables the use of Parquet layout, in which the records for every partition is held in reminiscence till it reaches the overall block length in size before it is written to disk.) Consider breaking apart such operations into numerous exclusive INSERT statements, as an instance to load facts 365 days at a time rather than for all years right away.
GROUP BY on a unique or excessive-cardinality column. Impala allocates a few handler structures for each unique value in a GROUP BY . Having thousands and thousands of different GROUP BY values could exceed the reminiscence limit.
Queries concerning very wide tables, with lots of columns, specifically with many STRING columns. Because Impala permits a STRING value to be as much as 32 KB, the intermediate outcomes during such queries should require giant memory allocation.
Can Impala Do User-defined Functions (udfs)?
Impala 1.2 and better does help UDFs and UDAs. You can either write native Impala UDFs and UDAs in C++, or reuse UDFs (but not UDAs) at the beginning written in Java to be used with Hive.
Why Do I Have to Use Refresh and Invalidate Metadata, What Do They Do?
In Impala 1.2 and higher, there may be plenty much less need to apply the REFRESH and INVALIDATE METADATA statements:
The new impala-catalog service, represented with the aid of the catalog daemon, declares the results of Impala DDL statements to all Impala nodes. Thus, if you do a CREATE TABLE declaration in Impala while related to 1 node, you do now not want to do INVALIDATE METADATA before issuing queries thru a exclusive node.
The catalog provider only recognizes changes made via Impala, so that you ought to nonetheless issue a REFRESH statement if you load data thru Hive or by means of manipulating documents in HDFS, and you have to problem an INVALIDATE METADATA announcement if you create a table, modify a desk, upload or drop walls, or do different DDL statements in Hive.
Because the catalog service announces the effects of REFRESH and INVALIDATE METADATA statements to all nodes, inside the cases wherein you do nonetheless need to trouble those statements, you could try this on a single node instead of on every node, and the changes may be mechanically recognized across the cluster, making it more handy to load stability with the aid of issuing queries through arbitrary Impala nodes instead of constantly using the same coordinator node.
Why Is Space Not Freed Up When I Issue Drop Table?
Impala deletes statistics files when you trouble a DROP TABLE on an internal desk, however now not an outside one. By default, the CREATE TABLE assertion creates internal tables, wherein the documents are managed with the aid of Impala. An outside table is created with a CREATE EXTERNAL TABLE declaration, where the documents reside in a place outside the control of Impala. Issue a DESCRIBE FORMATTED statement to test whether a table is inner or external. The keyword MANAGED_TABLE indicates an internal desk, from which Impala can delete the data files. The keyword EXTERNAL_TABLE shows an external desk, wherein Impala will leave the information documents untouched while you drop the table.
Even when you drop an internal table and the documents are eliminated from their authentic region, you may not get the hard drive space lower back straight away. By default, documents which can be deleted in HDFS cross into a special garbage can listing, from which they’re purged after a time frame (via default, 6 hours). For history statistics on the garbage can mechanism.
What Kinds of Impala Queries Or Data Are Best Suited For HBase?
HBase tables are perfect for queries in which generally you would use a key-price shop. That is, in which you retrieve a unmarried row or a few rows, by using checking out a unique key column the use of the = or IN operators.
HBase tables are not appropriate for queries that produce big end result units with hundreds of rows. HBase tables are also now not suitable for queries that perform full desk scans because the WHERE clause does no longer ret specific values from the precise key column.
Use HBase tables for information that is inserted one row or a few rows at a time, which include by means of the INSERT … VALUES syntax. Loading facts piecemeal like this into an HDFS-subsidized desk produces many tiny documents, that’s a completely inefficient format for HDFS records files.
If the dearth of an UPDATE assertion in Impala is a trouble for you, you may simulate single-row updates by doing an INSERT … VALUES declaration using an existing value for the key column. The antique row fee is hidden; most effective the brand-new row cost is visible by way of queries.
HBase tables are regularly extensive (containing many columns) and sparse (with most column values NULL). For example, you might document loads of different statistics factors for each user of a web service, which includes whether or not the person had registered for a web sport or enabled specific account capabilities. With Impala and HBase, you may look up all of the records for a particular patron efficiently in an unmarried. For any given consumer, most of those columns is probably NULL, due to the fact an ordinary patron might not employ maximum features of a web carrier.
What Load Do Concurrent Queries Produce on The Namenode?
The load Impala generates is very similar to MapReduce. Impala contacts the NameNode during the planning phase to get the file metadata (this is only run on the host the query was sent to). Every impalad will read files as part of normal processing of the query.
State some Impala Hadoop Benefits.
Some of the benefits are:
- Impala is very familiar SQL interface. Especially data scientists and analysts already know.
- It also offers the ability to query high volumes of data (“Big Data“) in Apache Hadoop.
- Also, it provides distributed queries for convenient scaling in a cluster environment. It offers to use of cost-effective commodity hardware.
- By using Impala it is possible to share data files between different components with no copy or export/import step.
What are the best features of Impala?
There are several best features of Impala. They are:
- Open Source
Basically, under the Apache license, Cloudera Impala is available freely as open source.
- In-memory Processing
While it’s come to processing, Cloudera Impala supports in-memory data processing. That implies without any data movement it accesses/analyzes data that is stored on Hadoop data nodes.
- Easy Data Access
However, using SQL-like queries, we can easily access data using Impala. Moreover, Impala offers Common data access interfaces. That includes:
i. JDBC driver.
ii. ODBC driver.
- Faster Access
While we compare Impala to another SQL engines, Impala offers faster access to the data in HDFS.
- Storage Systems
We can easily store data in storage systems. Such as HDFS, Apache HBase, and Amazon s3.
i. HDFS file formats: delimited text files, Parquet, Avro, SequenceFile, and RCFile.
ii. Compression codecs: Snappy, GZIP, Deflate, BZIP.
- Easy Integration
It is possible to integrate Impala with business intelligence tools. Such as; Tableau, Pentaho, Micro strategy, and Zoom data.
- Joins and Functions
Including SELECT, joins, and aggregate functions, Impala offers most common SQL-92 features of Hive Query Language (HiveQL).
What are Impala Built-in Functions?
In order to perform several functions like mathematical calculations, string manipulation, date calculations, and other kinds of data transformations directly in SELECT statements we can use Impala Built-in Functions. we can get results with all formatting, calculating, and type conversions applied, with the built-in functions SQL query in Impala. Despite performing time-consuming postprocessing in another application we can use the Impala Built-in Functions.
Impala support following categories of built-in functions. Such as:
- Mathematical Functions
- Type Conversion Functions
- Date and Time Functions
- Conditional Functions
- String Functions
- Aggregation functions
Describe Impala Shell (impala-shell Command).
Basically, to set up databases and tables, insert data, and issue queries, we can use the Impala shell tool (impala-shell). Moreover, we can submit SQL statements in an interactive session for ad hoc queries and exploration. Also, to process a single statement or a script file or to process a single statement or a script file we can specify command-line options.
In addition, it supports all the same SQL statements listed in Impala SQL Statements along with some shell-only commands. Hence, that we can use for tuning performance and diagnosing problems.
Does Impala Use Caching?
No. There is no provision of caching table data in Impala. However, it does cache some table and file metadata. But queries might run faster on subsequent iterations because the data set was cached in the OS buffer cache, Impala does not explicitly control this.
Although, in CDH 5, Impala takes advantage of the HDFS caching feature. Hence, we can designate which tables or partitions are cached through the CACHED and UNCACHED clauses of the CREATE TABLE and ALTER TABLE statements. Also, through the hdfscacheadmin command, Impala can take advantage of data that is pinned in the HDFS cache.
What are the names of Daemons in Impala?
They are:
i. ImpalaD (impala Daemon)
ii. StatestoreD
iii. CatalogD
What are distinct Operators in Impala?
While we want to filter the results or to remove duplicates, we use The DISTINCT operator in a SELECT statement:
— Returns the unique values from one column.
— NULL is included in the set of values if any rows have a NULL in this column.
select distinct c_birth_country from Employees;
— Returns the unique combinations of values from multiple columns.
select distinct c_salutation, c_last_name from Employees;
Moreover, to find how many different values a column contains, we can use DISTINCT in combination with an aggregation function.Typically COUNT():
— Counts the unique values from one column.
— NULL is not included as a distinct value in the count.
select count(distinct c_birth_country) from Employees;
— Counts the unique combinations of values from multiple columns.
select count(distinct c_salutation, c_last_name) from Employees;
However, make sure that using DISTINCT in more than one aggregation function in the same query is not supported by Impala SQL.
To understand more, we could not have a single query with both COUNT(DISTINCT c_first_name) and COUNT(DISTINCT c_last_name) in the SELECT list.
What is Troubleshooting for Impala?
Basically, being able to diagnose and debug problems in Impala, is what we call Impala Troubleshooting/performance tuning. It includes performance, network connectivity, out-of-memory conditions, disk space usage, and crash or hangs conditions in any of the Impala-related daemons. However, there are several ways, we can follow for diagnosing and debugging of above-mentioned problems. Such as:
- Impala performance tuning
- Impala Troubleshooting Quick Reference.
- Troubleshooting Impala SQL Syntax Issues
- Impala Web User Interface for Debugging
However, to learn them in detail, follow the link: Steps to Impala Troubleshooting.
Does Impala Support Generic Jdbc?
It supports the HiveServer2 JDBC driver.
Is Avro Supported?
Yes, it supports Avro. Impala has always been able to query Avro tables. To load existing Avro data files into a table, we can use the Impala LOAD DATAstatement.
How Do I Know How Many Impala Nodes Are In My Cluster?
Basically, how many impalad nodes are currently available, The Impala statestore keeps track. Through the statestore web interface, we can see this information.
Can Any Impala Query Also Be Executed In Hive?
Yes. Impala queries can also be completed in Hive. However, there are some minor differences in how some queries are handled. Also, with some functional limitations, Impala SQL is a subset of HiveQL, such as transforms.
What Are Good Use Cases For Impala As Opposed To Hive Or MapReduce?
For interactive exploratory analytics on large data sets, Impala is well-suited to executing SQL queries. Also, for very long-running, batch-oriented tasks, Hive and MapReduce are appropriate. Likes ETL.
Is Mapreduce Required for Impala? Will Impala Continue to Work as Expected If Mapreduce Is Stopped?
No. Impala does not use MapReduce at all.
Can Impala Be Used for Complex Event Processing?
By dedicated stream-processing systems, Complex Event Processing (CEP) is usually performed. Impala most closely resembles a relational database. Hence, it is not a stream-processing system.
What Happens If There Is an Error In Impala?
However, there is not a single point of failure in Impala. To handle incoming queries all Impala daemons are fully able. All queries with fragments running on that machine will fail if a machine fails, however. We can just rerun the query if there is a failure because queries are expected to return quickly.
How Does Impala Process Join Queries for Large Tables?
To allow joins between tables and result sets of various sizes, Impala utilizes multiple strategies. While, joining a large table with a small one, the data from the small table is transmitted to each node for intermediate processing. The data from one of the tables are divided into pieces, and each node processes only selected pieces, when joining two large tables.
What Is Impala’s Aggregation Strategy?
It only supports in-memory hash aggregation. If the memory requirements for a join or aggregation operation exceed the memory limit for a particular host, In Impala 2.0 and higher, It uses a temporary work area on disk to help the query complete successfully.
What is the no. of threads created by ImpalaD?
Here, no. of threads created by impalaD = 2 or 3x no of cores.
What does Impala do for fast access?
For fast access, ImpalaD’s caches the metadata.
What Impala use for Authentication?
It supports Kerberos authentication.
What is used to store data generally?
In order to store information about the data available to Impala, we use it. Let’s understand this with the example. Here, the Metastore lets Impala know what databases are available. Also, it informs about what the structure of those databases is.
Can Impala Do User-defined Functions (udfs)?
Impala 1.2 and higher does support UDFs and UDAs. we can either write native Impala UDFs and UDAs in C++ or reuse UDFs (but not UDAs) originally written in Java for use with Hive.
Can I Do Insert … Select * Into A Partitioned Table?
The columns corresponding to the partition key columns must appear last in the columns returned by the SELECT * when you use the INSERT … SELECT * syntax to copy data into a partitioned table. We can create the table with the partition key columns defined last.
Also, we can use the CREATE VIEW statement to create a view that reorders the columns: put the partition key columns last, then do the INSERT … SELECT * from the view.
How can it help for avoiding costly modeling?
It is a single system for Big Data processing and analytics. Hence, through this customer can avoid costly modeling and ETL just for analytics.
Does Impala Support Generic Jdbc?
Impala supports the HiveServer2 JDBC driver.
Is the Hdfs Block Size Reduced to Achieve Faster Query Results?
No. Impala does not make any changes to the HDFS or HBase data sets.
Basically, the default Parquet block size is relatively large (256 MB in Impala 2.0 and later; 1 GB in earlier releases). Also, we can control the block size when creating Parquet files using the PARQUET_FILE_SIZE query option.
State Use cases of Impala.
Impala Use Cases and Applications are:
- Do BI-style Queries on Hadoop
While it comes to BI/analytic queries on Hadoop especially those which are not delivered by batch frameworks such as Apache Hive, Impala offers low latency and high concurrency for them. Moreover, it scales linearly, even in multi-tenant environments.
- Unify Your Infrastructure
In Impala, there is no redundant infrastructure or data conversion/duplication is possible. Hence, that implies we need to utilize the same file and data formats and metadata, security, and resource management frameworks as your Hadoop deployment.
- Implement Quickly
Basically, Impala utilizes the same metadata and ODBC driver for Apache Hive users. Such as Hive, Impala supports SQL. Hence, we do not require to think about re-inventing the implementation wheel.
- Count on Enterprise-class Security
However, there is a beautiful feature of Authentication. So, for that Impala is integrated with native Hadoop security and Kerberos. Moreover, we can also ensure that the right users and applications are authorized for the right data by using the Sentry module.
- Retain Freedom from Lock-in
Also, it is available easily, which mean it is an Open source (Apache License).
Where Can I Get Sample Data to Try?
You can get scripts that produce data files and set up an environment for TPC-DS style benchmark tests from this GitHub repository. In addition to being useful for experimenting with performance, the tables are suited to experimenting with many aspects of SQL on Impala: they contain a good mixture of data types, data distributions, partitioning, and relational data suitable for join queries.
What Are the Main Features of Impala?
A large set of SQL statements, including SELECT and INSERT, with joins, Subqueries in Impala SELECT Statements, and Impala Analytic Functions. Highly compatible with HiveQL, and also including some vendor extensions. For more information.
Distributed, high-performance queries.
Using Cloudera Manager, you can deploy and manage your Impala services. Cloudera Manager is the best way to get started with Impala on your cluster.
Using Hue for queries.
Appending and inserting data into tables through the INSERT statement.
ODBC: Impala is certified to run against MicroStrategy and Tableau, with restrictions. For more information, see Configuring Impala to Work with ODBC.
Querying data stored in HDFS and HBase in a single query.
In Impala 2.2.0 and higher, querying data stored in the Amazon Simple Storage Service (S3).
Concurrent client rets. Each Impala daemon can handle multiple concurrent clients rets. The effects on performance depend on your particular hardware and workload.
Kerberos authentication. For more information.
Partitions. With Impala SQL, you can create partitioned tables with the CREATE TABLE statement, and add and drop partitions with the ALTER TABLE statement. Impala also takes advantage of the partitioning present in Hive tables.
Does Impala Support Generic Jdbc?
Impala supports the HiveServer2 JDBC driver.
Is Avro Supported?
Yes, Avro is supported. Impala has always been able to query Avro tables. You can use the Impala LOAD DATAstatement to load existing Avro data files into a table. Starting with Impala 1.4, you can create Avro tables with Impala. Currently, you still use the INSERT statement in Hive to copy data from another table into an Avro table.
Why Does My Insert Statement Fail?
When an INSERT statement fails, it is usually the result of exceeding some limit within a Hadoop component, typically HDFS.
An INSERT into a partitioned table can be a strenuous operation due to the possibility of opening many files and associated threads simultaneously in HDFS. Impala 1.1.1 includes some improvements to distribute the work more efficiently, so that the values for each partition are written by a single node, rather than as a separate data file from each node.
Certain expressions in the SELECT part of the INSERT statement can complicate the execution planning and result in an inefficient INSERT operation. Try to make the column data types of the source and destination tables match up, for example by doing ALTER TABLE … REPLACE COLUMNS on the source table if necessary. Try to avoid CASE expressions in the SELECT portion, because they make the result values harder to predict than transferring a column unchanged or passing the column through a built-in function.
Be prepared to raise some limits in the HDFS configuration settings, either temporarily during the INSERT or permanently if you frequently run such INSERT statements as part of your ETL pipeline.
The resource usage of an INSERT statement can vary depending on the file format of the destination table. Inserting into a Parquet table is memory-intensive, because the data for each partition is buffered in memory until it reaches 1 gigabyte, at which point the data file is written to disk. Impala can distribute the work for an INSERT more efficiently when statistics are available for the source table that is queried during the INSERT statement.
What Are Good Use Cases for Impala As Opposed To Hive Or Mapreduce?
Impala is well-suited to executing SQL queries for interactive exploratory analytics on large data sets. Hive and MapReduce are appropriate for very long running, batch-oriented tasks such as ETL.
Is Mapreduce Required for Impala? Will Impala Continue to Work As Expected If Mapreduce Is Stopped?
Impala does not use MapReduce at all.
Can Impala Be Used for Complex Event Processing?
For example, in an industrial environment, many agents may generate large amounts of data. Can Impala be used to analyze this data, checking for notable changes in the environment?
Complex Event Processing (CEP) is usually performed by dedicated stream-processing systems. Impala is not a stream-processing system, as it most closely resembles a relational database.
Can I Use Impala to Query Data Already Loaded into Hive and HBase?
There are no additional steps to allow Impala to query tables managed by Hive, whether they are stored in HDFS or HBase. Make sure that Impala is configured to access the Hive metastore correctly and you should be ready to go. Keep in mind that impala, by default, runs as the impala user, so you might need to adjust some file permissions depending on how strict your permissions are currently.
What Happens If There Is an Error in Impala?
There is not a single point of failure in Impala. All Impala daemons are fully able to handle incoming queries. If a machine fails however, all queries with fragments running on that machine will fail. Because queries are expected to return quickly, you can just rerun the query if there is a failure.
The longer answer: Impala must be able to connect to the Hive metastore. Impala aggressively caches metadata so the metastore host should have minimal load. Impala relies on the HDFS NameNode, and, in CDH4, you can configure HA for HDFS. Impala also has centralized services, known as the statestore andcatalog services, that run on one host only. Impala continues to execute queries if the statestore host is down, but it will not get state updates. For example, if a host is added to the cluster while the statestore host is down, the existing instances of impalad running on the other hosts will not find out about this new host. Once the statestore process is restarted, all the information it serves is automatically reconstructed from all running Impala daemons.
How Are Joins Performed in Impala?
By default, Impala automatically determines the most efficient order in which to join tables using a cost-based method, based on their overall size and number of rows. (This is a new feature in Impala 1.2.2 and higher.) The COMPUTE STATS statement gathers information about each table that is crucial for efficient join performance. Impala chooses between two techni for join queries, known as “broadcast joins” and “partitioned joins”.
How Does Impala Process Join Queries for Large Tables?
Impala utilizes multiple strategies to allow joins between tables and result sets of various sizes. When joining a large table with a small one, the data from the small table is transmitted to each node for intermediate processing. When joining two large tables, the data from one of the tables is divided into pieces, and each node processes only selected pieces.
What Is Impala’s Aggregation Strategy?
Impala currently only supports in-memory hash aggregation. In Impala 2.0 and higher, if the memory requirements for a join or aggregation operation exceed the memory limit for a particular host, Impala uses a temporary work area on disk to help the query complete successfully.
How Is Impala Metadata Managed?
Impala uses two pieces of metadata: the catalog information from the Hive metastore and the file metadata from the NameNode. Currently, this metadata is lazily populated and cached when an impala needs it to plan a query.
The REFRESH statement updates the metadata for a particular table after loading new data through Hive. The INVALIDATE METADATA Statement statement refreshes all metadata, so that Impala recognizes new tables or other DDL and DML changes performed through Hive.
In Impala 1.2 and higher, a dedicated catalogue daemon broadcasts metadata changes due to Impala DDL or DML statements to all nodes, reducing or eliminating the need to use the REFRESH and INVALIDATE METADATAstatements.
What Are the Most Memory-intensive Operations?
If a query fails with an error indicating “memory limit exceeded”, you might suspect a memory leak. The problem could actually be a query that is structured in a way that causes Impala to allocate more memory than you expect, exceeded the memory allocated for Impala on a particular node. Some examples of query or table structures that are especially memory-intensive are:
INSERT statements using dynamic partitioning, into a table with many different partitions. (Particularly for tables using Parquet format, where the data for each partition is held in memory until it reaches the full block size in size before it is written to disk.) Consider breaking up such operations into several different INSERT statements, for example to load data one year at a time rather than for all years at once.
GROUP BY on a unique or high-cardinality column. Impala allocates some handler structures for each different value in a GROUP BY query. Having millions of different GROUP BY values could exceed the memory limit.
Queries involving very wide tables, with thousands of columns, particularly with many STRING columns. Because Impala allows a STRING value to be up to 32 KB, the intermediate results during such queries could require substantial memory allocation.
When Does Impala Hold on to or Return Memory?
Impala allocates memory using tcmalloc, a memory allocator that is optimized for high concurrency. Once Impala allocates memory, it keeps that memory reserved to use for future queries. Thus, it is normal for Impala to show high memory usage when idle. If Impala detects that it is about to exceed its memory limit (defined by the -mem_limit startup option or the MEM_LIMIT query option), it deallocates memory not needed by the current queries.
When issuing queries through the JDBC or ODBC interfaces, make sure to call the appropriate close method afterwards. Otherwise, some memory associated with the query is not freed.
Can Impala Do User-defined Functions (udfs)?
Impala 1.2 and higher does support UDFs and UDAs. You can either write native Impala UDFs and UDAs in C++, or reuse UDFs (but not UDAs) originally written in Java for use with Hive.
Why Is Space Not Freed Up When I Issue Drop Table?
Impala deletes data files when you issue a DROP TABLE on an internal table, but not an external one. By default, the CREATE TABLE statement creates internal tables, where the files are managed by Impala. An external table is created with a CREATE EXTERNAL TABLE statement, where the files reside in a location outside the control of Impala. Issue a DESCRIBE FORMATTED statement to check whether a table is internal or external. The keyword MANAGED_TABLE indicates an internal table, from which Impala can delete the data files. The keyword EXTERNAL_TABLE indicates an external table, where Impala will leave the data files untouched when you drop the table.
Even when you drop an internal table and the files are removed from their original location, you might not get the hard drive space back immediately. By default, files that are deleted in HDFS go into a special trashcan directory, from which they are purged after a period of time (by default, 6 hours). For background information on the trashcan mechanism.
Is There a Dual Table?
You might be used to running queries against a single-row table named DUAL to try out expressions, built-in functions, and UDFs. Impala does not have a DUAL table. To achieve the same result, you can issue a SELECT statement without any table name:
select 2+2;
select substr(‘hello’,2,1);
select pow(10,6);
What is Impala?
Basically, for running huge volumes of data Impala is an MPP (Massive Parallel Processing) SQL query engine that is stored in the Hadoop cluster. Moreover, this is an advantage that it is admission-source software which is written in C++ and Java. Also, it offers high take steps and low latency compared to appendage SQL engines for Hadoop.
To be more specific, it is the highest-the stage SQL engine that offers the fastest habit to admission data that is stored in Hadoop Distributed File System HDFS.
What is Impala Data Types?
There is a loud set of data types attainable in Impala. Basically, those Impala Data Types we use for table columns, aeration values, and skirmish arguments and recompense values. Each Impala Data Types serves a specific strive for. Types are:
- BIGINT
- BOOLEAN
- CHAR
- DECIMAL
- DOUBLE
- FLOAT
- INT
- SMALLINT
- STRING
- TIMESTAMP
- TINYINT
- VARCHAR
- ARRAY
- Map
- Struct
Describe Impala Shell (impala-shell Command).
Basically, to set taking place databases and tables, put in data, and matter queries, we can use the Impala shell tool (impala-shell). Moreover, we can agree with SQL statements in an interactive session for ad hoc queries and exploration. Also, to process a single confirmation or a script file or to process a single publication or a script file we can specify command-descent options.
In addition, together with going on, it supports every single one of the related SQL statements listed in Impala SQL Statements along with bearing in mind some shell-unaided commands. Hence, we can use tuning take steps and diagnosing problems.
Does Impala Use Caching?
No. There is no provision of caching data in Impala. However, it does cache some tables and file metadata. But queries might control faster harshly speaking subsequent iterations because the data set was cached in the OS buffer cache, Impala does not explicitly run this.
Although, in CDH 5, it takes advantage of the HDFS caching feature. Hence, we can apportion which tables or partitions are cached through the CACHED and UNCACHED clauses of the CREATE TABLE and ALTER TABLE statements. Also, through the hdfscacheadmin command, Impala can pronounce-calling data that is pinned in the HDFS cache.
Does Cloudera Offer A VM For Demonstrating Impala?
Cloudera offers a demonstration VM called the QuickStart VM, to hand in VMWare, VirtualBox, and KVM formats. For more recommendations, see the Cloudera QuickStart VM. After booting the QuickStart VM, many services are turned off by default; in the Cloudera Manager UI that appears automatically, perspective upon Impala and any new components that you lack to a goal out.
Where Can I Find Impala Documentation?
Starting when Impala 1.3.0, Impala documentation is integrated behind the CDH 5 documentation, in add together to the standalone Impala documentation for use as soon as CDH 4. For CDH 5, the core Impala developer and administrator information remain in the associated Impala documentation allocation. Information nearly Impala handy explanation, installation, configuration, startup, and security is embedded in the corresponding CDH 5 guides.
- New features
- Known and unmodified issues
- Incompatible changes
- Installing Impala
- Upgrading Impala
- configuring Impala
- starting Impala
- security for Impala
- DH Version and Packaging Information
Where Can I Get Sample Data to Try?
you can get sticking together of scripts that fabricate data files and settings in the works a mood for TPC-DS style benchmark tests from this Github repository. In appendage to being useful for experimenting later than war, the tables are suited to experimenting later with many aspects of SQL upon Impala: they contain a courteous merged of data types, data distributions, partitioning, and relational data occurring to within enough limits for colleague queries.
Does Impala Use Caching?
Impala does not cache table data. It does cache some tables and file metadata. Although queries might manage faster upon subsequent iterations because the data set was cached in the OS buffer cache, Impala does not explicitly control this.
Impala takes advantage of the HDFS caching feature in CDH 5. You can designate which tables or partitions are cached through the CACHED and UNCACHED clauses of the CREATE TABLE and ALTER TABLE statements. Impala can as well as infuriate data that is pinned in the HDFS cache through the hdfscacheadmin command.
What Are Good Use Cases for Impala as opposed to Hive Or Mapreduce?
Impala is competently-suited to executing SQL queries for interactive exploratory analytics upon large data sets. Hive and MapReduce are invading for certainly long-running, batch-oriented tasks such as ETL.
Can I Do Transforms or Add New Functionality?
Impala adds desist for UDFs in Impala 1.2. You can write your own functions in C++, or reuse existing Java-based Hive UDFs. The UDF desist includes scalar functions and devotee-defined aggregate functions (UDAs). User-defined table functions (UDTFs) are not currently supported.
Impala does not currently verify an extensible serialization-deserialization framework (SerDes), and correspondingly accumulation added functionality to Impala is not as understandable as for Hive or Pig.
Can Any Impala Query Also Be Executed In Hive?
Yes. There are some teenage differences in how some queries are handled, but Impala queries can plus be completed in Hive. Impala SQL is a subset of HiveQL, as soon as some lithe limitations such as transforms.
Can I Use Impala to Query Data Already Loaded into Hive and HBase?
There are no tallying steps to divulge Impala to query tables managed by Hive, whether they are stored in HDFS or HBase. Make sure that Impala is configured to admission the Hive metastore correctly and you should be ready to go. Keep in mind that impaled, by default, runs as the impala devotee, therefore you might crave to change some file permissions depending upon how strict your permissions are currently.
Where Can I Get Sample Data To Try?
You can get scripts that produce statistics files and set up an environment for TPC-DS style benchmark checks from this Github repository. In addition to being useful for experimenting with performance, the tables are ideal to experimenting with many aspects of SQL on Impala: they include an excellent mixture of statistics sorts, information distributions, partitioning, and relational data suitable for be a part of queries.
How Much Memory Is Required?
Although Impala isn’t an in-reminiscence database, when handling large tables and big result sets, you need to assume to devote a giant portion of physical memory for the impalad daemon. Recommended physical memory for an Impala node is 128 GB or better. If practical, dedicate about eighty% of physical reminiscence to Impala.
The quantity of memory required for an Impala operation relies upon on numerous elements:
The file layout of the table. Different file formats constitute the equal information in more or fewer records documents. The compression and encoding for every report layout may require a extraordinary quantity of temporary memory to decompress the records for evaluation.
Whether the operation is a SELECT or an INSERT. For example, Parquet tables require notably little reminiscence to query, due to the fact Impala reads and decompresses information in 8 MB chunks. Inserting into a Parquet desk is a extra memory-intensive operation due to the fact the data for every facts record (potentially masses of megabytes, relying at the fee of the PARQUET_FILE_SIZE query choice) is saved in reminiscence until encoded, compressed, and written to disk.
Whether the table is partitioned or not, and whether a against a partitioned desk can take advantage of partition pruning.
Whether the final end result set is sorted through the ORDER BY clause. Each Impala node scans and filters a part of the total data, and applies the LIMIT to its very own part of the end result set. In Impala 1.4.0 and higher, if the kind operation calls for more reminiscence than is to be had on any unique host, Impala makes use of a transient disk work region to perform the type. The intermediate end result units are all despatched back to the coordinator node, which does the very last sorting and then applies the LIMIT clause to the final result set.
For example, if you execute the :
choose * from giant_table order by some_column limit a thousand;
and your cluster has 50 nodes, then every of these 50 nodes will transmit a maximum of 1000 rows again to the coordinator node. The coordinator node wishes enough reminiscence to sort (LIMIT * cluster_size) rows, despite the fact that in the long run the final end result set is at maximum LIMIT rows, 1000 in this case.
Likewise, if you execute the :
choose * from giant_table where test_val > a hundred order by some_column;
then each node filters out a set of rows matching the WHERE situations, types the effects (with out a size restrict), and sends the looked after intermediate rows lower back to the coordinator node. The coordinator node would possibly need considerable reminiscence to sort the very last end result set, and so might use a transient disk work region for that final phase of the.
Whether the query carries any be a part of clauses, GROUP BY clauses, analytic functions, or DISTINCT operators. These operations all require a few in-reminiscences works areas that modify depending on the extent and distribution of statistics. In Impala 2. Zero and later, these types of operations utilize brief disk work regions if memory usage grows too massive to handle.
The length of the result set. When intermediate consequences are being exceeded around between nodes, the amount of data depends on the quantity of columns lower back via the. For instance, it’s miles greater memory-efficient to simplest the columns which are clearly wanted within the end result set instead of constantly issuing SELECT *.
The mechanism by using which work is divided for a be part of query. You use the COMPUTE STATS announcement, and query pointers in the most difficult instances, to assist Impala pick out the most efficient execution plan.
What Features from Relational Databases or Hive Are Not Available in Impala?
Querying streaming statistics.
Deleting character rows. You delete facts in bulk by means of overwriting an entire desk or partition, or by means of losing a desk.
Indexing (now not currently). LZO-compressed textual content documents may be listed outdoor of Impala, as defined in Using LZO-Compressed Text Files.
Full text search on textual content fields. The Cloudera Search product is appropriate for this use case.
Custom Hive Serializer/Deserializer instructions (SerDes). Impala supports a set of common native report codecs that have integrated SerDes in CDH.
Checkpointing within a query. That is, Impala does now not keep intermediate results to disk throughout lengthy-strolling queries. Currently, Impala cancels a jogging if any host on which that is executing fails. When one or more hosts are down, Impala reroutes destiny queries to handiest use the available hosts, and Impala detects while the hosts come lower back up and starts off evolved using them once more. Because a query can be submitted through any Impala node, there’s no single point of failure. In the future, we can recall including extra work allocation functions to Impala, so that a strolling query would complete even in the presence of host screw ups.
Encryption of information transmitted among Impala daemons.
Hive indexes.
Non-Hadoop information stores, including relational databases.
Why Does My Select Statement Fail?
When a SELECT statement fails, the reason normally falls into one of the following classes:
A timeout due to a overall performance, potential, or network trouble affecting one precise node.
Excessive memory use for a be part of , resulting in automatic cancellation of the .
A low-stage trouble affecting how local code is generated on every node to deal with particular WHERE clauses within the . For instance, a device training can be generated that is not supported by means of the processor of a positive node. If the mistake message within the log indicates the cause was an illegal practise, remember turning off native code era quickly, and attempting the again.
Malformed enter statistics, such as a textual content facts file with an relatively lengthy line, or with a delimiter that doesn’t in shape the man or woman precise within the FIELDS TERMINATED BY clause of the CREATE TABLE announcement.
Does Impala Performance Improve as It Is Deployed to More Hosts In A Cluster In Much The Same Way That Hadoop Performance Does?
Yes. Impala scales with the range of hosts. It is important to put in Impala on all the DataNodes within the cluster, because in any other case some of the nodes should do remote reads to retrieve records now not to be had for local reads. Data locality is a crucial architectural aspect for Impala performance.
Is the Hdfs Block Size Reduced To Achieve Faster Query Results?
No. Impala does now not make any changes to the HDFS or HBase records units.
The default Parquet block length is surprisingly massive (256 MB in Impala 2.Zero and later; 1 GB in earlier releases). You can manage the block size when developing Parquet documents the use of the PARQUET_FILE_SIZE choice.
Does Impala Use Caching?
Impala does now not cache table facts. It does cache some table and record metadata. Although queries may run quicker on next iterations because the data set changed into cached within the OS buffer cache, Impala does now not explicitly control this.
Impala takes gain of the HDFS caching function in CDH 5. You can designate which tables or walls are cached via the CACHED and UNCACHED clauses of the CREATE TABLE and ALTER TABLE statements. Impala also can take gain of facts this is pinned within the HDFS cache via the hdfscacheadmin command.
What Are Good Use Cases For Impala As Opposed To Hive Or Mapreduce?
Impala is properly-suitable to executing SQL queries for interactive exploratory analytics on massive statistics sets. Hive and MapReduce are suitable for terribly lengthy running, batch-oriented obligations such as ETL.
How Does Impala Compare To Hive And Pig?
Impala isn’t the same as Hive and Pig because it uses its own daemons which can be unfold throughout the cluster for queries. Because Impala does no longer rely upon MapReduce, it avoids the startup overhead of MapReduce jobs, allowing Impala to return consequences in real time.
Can I Do Transforms or Add New Functionality?
Impala provides support for UDFs in Impala 1.2. You can write your own functions in C++, or reuse present Java-primarily based Hive UDFs. The UDF help includes scalar features and user-described combination features (UDAs). User-defined desk features (UDTFs) aren’t presently supported.
Impala does no longer presently assist an extensible serialization-deserialization framework (SerDes), and so including more functionality to Impala isn’t as honest as for Hive or Pig.
Can Any Impala Query Also Be Executed In Hive?
Yes. There are a few minor variations in how a few queries are handled, however Impala queries can also be completed in Hive. Impala SQL is a subset of HiveQL, with some purposeful boundaries along with transforms.
Is Impala Production Ready?
Impala has completed its beta launch cycle, and the 1.0, 1.1, and 1.2 GA releases are manufacturing ready. The 1.1.X series consists of extra security functions for authorization, an crucial requirement for production use in lots of agencies. The 1.2.X series includes crucial overall performance functions, specifically for huge be part of queries. Some Cloudera clients are already the use of Impala for large workloads.
The Impala 1.Three.Zero and better releases are bundled with corresponding tiers of CDH five. The quantity of latest features grows with each release.
How Do I Configure Hadoop High Availability (ha) For Impala?
You can set up a proxy server to relay rets to and fro to the Impala servers, for load balancing and excessive availability.
How Are Joins Performed In Impala?
By default, Impala automatically determines the maximum efficient order wherein to sign up for tables using a price-primarily based approach, primarily based on their usual size and range of rows. (This is a new characteristic in Impala 1.2.2 and better.) The COMPUTE STATS declaration gathers data about each table that is vital for efficient join performance. Impala chooses among strategies for be part of queries, known as “broadcast joins” and “partitioned joins”.
How Does Impala Process Join Queries for Large Tables?
Impala utilizes multiple strategies to permit joins between tables and end result units of various sizes. When joining a large table with a small one, the information from the small desk is transmitted to each node for intermediate processing. When becoming a member of huge tables, the facts from one of the tables is split into pieces, and each node techniques best decided on pieces.
What Is Impala’s Aggregation Strategy?
Impala presently only helps in-memory hash aggregation. In Impala 2. Zero and higher, if the reminiscence necessities for a be part of or aggregation operation exceed the memory limit for a particular host, Impala uses a temporary paintings vicinity on disk to assist the complete successfully.
How Is Impala Metadata Managed?
Impala uses two pieces of metadata: the catalog records from the Hive metastore and the file metadata from the NameNode. Currently, this metadata is lazily populated and cached whilst an impalad needs it to plot a .
The REFRESH assertion updates the metadata for a particular table after loading new statistics thru Hive. The INVALIDATE METADATA Statement statement refreshes all metadata, in order that Impala acknowledges new tables or other DDL and DML changes executed through Hive.
In Impala 1.2 and better, a committed catalogd daemon declares metadata modifications due to Impala DDL or DML statements to all nodes, lowering or removing the need to use the REFRESH and INVALIDATE METADATAstatements.
When Does Impala Hold on To or Return Memory?
Impala allocates reminiscence the usage of tcmalloc, a memory allocator that is optimized for excessive concurrency. Once Impala allocates reminiscence, it keeps that reminiscence reserved to use for destiny queries. Thus, it’s far everyday for Impala to expose excessive memory usage while idle. If Impala detects that it’s far approximately to exceed its reminiscence restriction (described via the -mem_limit startup option or the MEM_LIMIT query alternative), it deallocates memory not wanted with the aid of the modern queries.
When issuing queries via the JDBC or ODBC interfaces, ensure to call the suitable near method afterwards. Otherwise, some reminiscence associated with the isn’t always freed.
Is There an Update Statement?
Impala does now not presently have an UPDATE assertion, which could generally be used to change a unmarried row, a small organization of rows, or a specific column. The HDFS-primarily based documents used by traditional Impala queries are optimized for bulk operations across many megabytes of facts at a time, making traditional UPDATE operations inefficient or impractical.
You can use the subsequent techniques to gain the same goals because the familiar UPDATE declaration, in a way that preserves efficient report layouts for next queries:
Replace the entire contents of a table or partition with updated facts which you have already staged in a one-of-a-kind place, both the usage of INSERT OVERWRITE, LOAD DATA, or manual HDFS document operations followed via a REFRESH statement for the table. Optionally, you can use built-in functions and expressions inside the INSERT statement to convert the copied statistics in the same manner you’ll typically do in an UPDATE assertion, for instance to turn a combined-case string into all uppercase or all lowercase.
To replace a unmarried row, use an HBase table, and difficulty an INSERT … VALUES assertion using the identical key because the authentic row. Because HBase handles reproduction keys by using simplest returning the modern-day row with a particular key price, the newly inserted row effectively hides the previous one.
Is There A Dual Table?
You are probably used to walking queries against a single-row desk named DUAL to try out expressions, integrated functions, and UDFs. Impala does no longer have a DUAL desk. To achieve the identical result, you may issue a SELECT statement with none table call:
choose 2+2;
pick substr(‘howdy’,2,1);
select pow(10,6);
How Do I Load A Big Csv File into A Partitioned Table?
To load a statistics file right into a partitioned desk, whilst the statistics document consists of fields like 12 months, month, and so forth that correspond to the partition key columns, use a two-degree procedure. First, use the LOAD DATA or CREATE EXTERNAL TABLE statement to deliver the statistics into an unpartitioned text desk. Then use an INSERT … SELECT statement to duplicate the facts from the unpartitioned table to a partitioned one. Include a PARTITION clause within the INSERT statement to specify the partition key columns.
Can I Do Insert … Select * Into A Partitioned Table?
When you operate the INSERT … SELECT * syntax to duplicate facts into a partitioned desk, the columns corresponding to the partition key columns must appear final within the columns lower back through the SELECT *. You can create the table with the partition key columns described ultimate. Or, you can use the CREATE VIEW statement to create a view that reorders the columns: positioned the partition key columns final, then do the INSERT … SELECT * from the view.
So, this brings us to the end of the Apache Impala Interview Questions blog.This Tecklearn ‘Top Apache Impala Interview Questions and Answers’ helps you with commonly asked questions if you are looking out for a job in Apache Impala or Big Data Domain. If you wish to learn Apache Impala and build a career in Big Data domain, then check out our interactive, Big Data Hadoop Architect Training, that comes with 24*7 support to guide you throughout your learning period.
https://www.tecklearn.com/course/bigdata-hadoop-architect-all-in-1-combo-course/
BigData Hadoop-Architect (All in 1) Combo Training
About the Course
Tecklearn’s BigData Hadoop-Architect (All in 1) combo includes the following Courses:
- BigData Hadoop Analyst
- BigData Hadoop Developer
- BigData Hadoop Administrator
- BigData Hadoop Tester
- Big Data Security with Kerberos
Why Should you take BigData Hadoop Combo Training?
- Average salary for a Hadoop Administrator ranges from approximately $104,528 to $141,391 per annum – Indeed.com
- Average salary for a Spark and Hadoop Developer ranges from approximately $106,366 to $127,619 per annum – Indeed.com
- Average salary for a Big Data Hadoop Analyst is $115,819– ZipRecruiter.com
What you will Learn in this Course?
Introduction
- The Case for Apache Hadoop
- Why Hadoop?
- Core Hadoop Components
- Fundamental Concepts
HDFS
- HDFS Features
- Writing and Reading Files
- NameNode Memory Considerations
- Overview of HDFS Security
- Using the Namenode Web UI
- Using the Hadoop File Shell
Getting Data into HDFS
- Ingesting Data from External Sources with Flume
- Ingesting Data from Relational Databases with Sqoop
- REST Interfaces
- Best Practices for Importing Data
YARN and MapReduce
- What Is MapReduce?
- Basic MapReduce Concepts
- YARN Cluster Architecture
- Resource Allocation
- Failure Recovery
- Using the YARN Web UI
- MapReduce Version 1
Planning Your Hadoop Cluster
- General Planning Considerations
- Choosing the Right Hardware
- Network Considerations
- Configuring Nodes
- Planning for Cluster Management
Hadoop Installation and Initial Configuration
- Deployment Types
- Installing Hadoop
- Specifying the Hadoop Configuration
- Performing Initial HDFS Configuration
- Performing Initial YARN and MapReduce Configuration
- Hadoop Logging
Installing and Configuring Hive, Impala, and Pig
- Hive
- Impala
- Pig
Hadoop Clients
- What is a Hadoop Client?
- Installing and Configuring Hadoop Clients
- Installing and Configuring Hue
- Hue Authentication and Authorization
Cloudera Manager
- The Motivation for Cloudera Manager
- Cloudera Manager Features
- Express and Enterprise Versions
- Cloudera Manager Topology
- Installing Cloudera Manager
- Installing Hadoop Using Cloudera Manager
- Performing Basic Administration Tasks Using Cloudera Manager
Advanced Cluster Configuration
- Advanced Configuration Parameters
- Configuring Hadoop Ports
- Explicitly Including and Excluding Hosts
- Configuring HDFS for Rack Awareness
- Configuring HDFS High Availability
Hadoop Security
- Why Hadoop Security Is Important
- Hadoop’s Security System Concepts
- What Kerberos Is and How it Works
- Securing a Hadoop Cluster with Kerberos
Managing and Scheduling Jobs
- Managing Running Jobs
- Scheduling Hadoop Jobs
- Configuring the Fair Scheduler
- Impala Query Scheduling
Cluster Maintenance
- Checking HDFS Status
- Copying Data Between Clusters
- Adding and Removing Cluster Nodes
- Rebalancing the Cluster
- Cluster Upgrading
Cluster Monitoring and Troubleshooting
- General System Monitoring
- Monitoring Hadoop Clusters
- Common Troubleshooting Hadoop Clusters
- Common Misconfigurations
Introduction to Pig
- What Is Pig?
- Pig’s Features
- Pig Use Cases
- Interacting with Pig
Basic Data Analysis with Pig
- Pig Latin Syntax
- Loading Data
- Simple Data Types
- Field Definitions
- Data Output
- Viewing the Schema
- Filtering and Sorting Data
- Commonly-Used Functions
Processing Complex Data with Pig
- Storage Formats
- Complex/Nested Data Types
- Grouping
- Built-In Functions for Complex Data
- Iterating Grouped Data
Multi-Dataset Operations with Pig
- Techniques for Combining Data Sets
- Joining Data Sets in Pig
- Set Operations
- Splitting Data Sets
Pig Troubleshooting and Optimization
- Troubleshooting Pig
- Logging
- Using Hadoop’s Web UI
- Data Sampling and Debugging
- Performance Overview
- Understanding the Execution Plan
- Tips for Improving the Performance of Your Pig Jobs
Introduction to Hive and Impala
- What Is Hive?
- What Is Impala?
- Schema and Data Storage
- Comparing Hive to Traditional Databases
- Hive Use Cases
Querying with Hive and Impala
- Databases and Tables
- Basic Hive and Impala Query Language Syntax
- Data Types
- Differences Between Hive and Impala Query Syntax
- Using Hue to Execute Queries
- Using the Impala Shell
Data Management
- Data Storage
- Creating Databases and Tables
- Loading Data
- Altering Databases and Tables
- Simplifying Queries with Views
- Storing Query Results
Data Storage and Performance
- Partitioning Tables
- Choosing a File Format
- Managing Metadata
- Controlling Access to Data
Relational Data Analysis with Hive and Impala
- Joining Datasets
- Common Built-In Functions
- Aggregation and Windowing
Working with Impala
- How Impala Executes Queries
- Extending Impala with User-Defined Functions
- Improving Impala Performance
Analyzing Text and Complex Data with Hive
- Complex Values in Hive
- Using Regular Expressions in Hive
- Sentiment Analysis and N-Grams
- Conclusion
Hive Optimization
- Understanding Query Performance
- Controlling Job Execution Plan
- Bucketing
- Indexing Data
Extending Hive
- SerDes
- Data Transformation with Custom Scripts
- User-Defined Functions
- Parameterized Queries
Importing Relational Data with Apache Sqoop
- Sqoop Overview
- Basic Imports and Exports
- Limiting Results
- Improving Sqoop’s Performance
- Sqoop 2
Introduction to Impala and Hive
- Introduction to Impala and Hive
- Why Use Impala and Hive?
- Comparing Hive to Traditional Databases
- Hive Use Cases
Modelling and Managing Data with Impala and Hive
- Data Storage Overview
- Creating Databases and Tables
- Loading Data into Tables
- HCatalog
- Impala Metadata Caching
Data Formats
- Selecting a File Format
- Hadoop Tool Support for File Formats
- Avro Schemas
- Using Avro with Hive and Sqoop
- Avro Schema Evolution
- Compression
Data Partitioning
- Partitioning Overview
- Partitioning in Impala and Hive
Capturing Data with Apache Flume
- What is Apache Flume?
- Basic Flume Architecture
- Flume Sources
- Flume Sinks
- Flume Channels
- Flume Configuration
Spark Basics
- What is Apache Spark?
- Using the Spark Shell
- RDDs (Resilient Distributed Datasets)
- Functional Programming in Spark
Working with RDDs in Spark
- A Closer Look at RDDs
- Key-Value Pair RDDs
- MapReduce
- Other Pair RDD Operations
Writing and Deploying Spark Applications
- Spark Applications vs. Spark Shell
- Creating the SparkContext
- Building a Spark Application (Scala and Java)
- Running a Spark Application
- The Spark Application Web UI
- Configuring Spark Properties
- Logging
Parallel Programming with Spark
- Review: Spark on a Cluster
- RDD Partitions
- Partitioning of File-based RDDs
- HDFS and Data Locality
- Executing Parallel Operations
- Stages and Tasks
Spark Caching and Persistence
- RDD Lineage
- Caching Overview
- Distributed Persistence
Common Patterns in Spark Data Processing
- Common Spark Use Cases
- Iterative Algorithms in Spark
- Graph Processing and Analysis
- Machine Learning
- Example: k-means
Preview: Spark SQL
- Spark SQL and the SQL Context
- Creating DataFrames
- Transforming and Querying DataFrames
- Saving DataFrames
- Comparing Spark SQL with Impala
Hadoop Testing
- Hadoop Application Testing
- Roles and Responsibilities of Hadoop Testing Professional
- Framework MRUnit for Testing of MapReduce Programs
- Unit Testing
- Test Execution
- Test Plan Strategy and Writing Test Cases for Testing Hadoop Application
Big Data Testing
- BigData Testing
- Unit Testing
- Integration Testing
- Functional Testing
- Non-Functional Testing
- Golden Data Set
System Testing
- Building and Set up
- Testing SetUp
- Solary Server
- Non-Functional Testing
- Longevity Testing
- Volumetric Testing
Security Testing
- Security Testing
- Non-Functional Testing
- Hadoop Cluster
- Security-Authorization RBA
- IBM Project
Automation Testing
- Query Surge Tool
Oozie
- Why Oozie
- Installation Engine
- Oozie Workflow Engine
- Oozie security
- Oozie Job Process
- Oozie terminology
- Oozie bundle
Got a question for us? Please mention it in the comments section and we will get back to you.
0 responses on "Top Apache Impala Interview Questions and Answers"