+44 1923526021 +91 989 345 2420
Stay Connected:

Data Science with BI

We are a Training, Research and Development Organisation based in UK.

Business Intelligence is a technology in which you visualize the data in the concise manner to get the company insights. These insights will help much in improving the business. You will learn about creating new ideas to make the decisions more précised and optimised.

We Offer Comprehensive set of Skills from initial concept down to the Final Outcome.

Data science

Course Content

Unit 1 : R Programing

The Introduction to R chapter will cover everything basic, starting from installing Rstudio to writing basic syntax and code in Rstudio. The history of R, introduction to R workspace and working environment and the various operators used in R are briefly discussed in this introduction part.
This chapter briefly discuss the types of data structure that R accepts as for the data input and to perform various functionalities on the data. The various operations that we can perform using these data structures is also covered in this chapter.
Here, you will be learning about how to load and read different types of data in R working environment. Also the chapter will provide a brief about how we can perform different functions and analysis on different types of data.
In this chapter the various functions that are very frequently used and which are essential for analysis of data in R, are discussed briefly. Also the utility of these functions with practical examples is covered in this chapter.
The working with text and date chapter includes the utility of various date functions and usage of packages that supports functionality for Date such as lubridate package in R. Working with text includes usage of text functions and understanding their functionality with practical examples.
The Descriptive Statistics chapter will cover the main elements of descriptive statistics, i.e., working with mean, median, variance, quantiles etc. in R with their practical implementation in Rstudio.
This Chapter will cover an overview of the statistical tests and methods on the various datasets that are available in Rstudio. The statistical tests that will be covered under this section includes application of Z-test, T-test, utility of normal distributed data etc.
In this chapter we will be covering the basics and some of the advanced methods that R supports and which helps the user in cleaning the data efficiently. The chapter also covers the importance of data generated after the application of these cleaning method on the raw data. The cleaning methods will cover data replacement and data elimination methods in R.
The data visualization with R will cover the practical implementation and visuals of the plots and charts created through ggplot2 package and also it covers an overview of the esiquisse package that supports the drag & drop functionality on data in R.
The chapter covers an overview on the various statistical models that can be applied on the given dataset in R. The statistical models covers an introduction on classification and regression models and their practical implementation using a given dataset in Rstudio.
The supervised learning with R will include the classification methods on data with a detailed implementation and explanation of the regression models such as Random Forest, Logistic Regression, Rpart etc. in Rstudio.
The unsupervised methods in R covers the application and utility of the unsupervised methods of classification and regression such as k-means, PCA etc. with examples performed on the dataset.
The Recommender system includes working on a practical example that builds a rating wise recommender system. The chapter will cover the algorithm of the Recommender system and its model building in R.
This chapter briefly describes the Machine Learning algorithm that will work on a real life data. The applied algorithm will work with dynamic data and the results showing thereof in R.
The chapter on deploying the script to production will cover the practical demonstration on how the codes and syntax written in script can be deployed and made work on the other datasets that are provided as input to the R script function.

Unit 2 : Big data

Under this topic we will learn and explore about Big Data and Hadoop with the importance of Hadoop. Also the challenges that you face in Big Data. The Fundamental design principles and the technologies used in Hadoop will also be covered here. You will learn about the architecture, RDBMS and use cases of Hadoop.
In this section you will learn about the Hadoop Cluster and the Architecture of Hadoop cluster. Here you will get to learn about the workflow in cluster. The way of reading files from HDFS and writing files to HDFS will be covered under this section.
In this section we will get the basic idea of Map Reduce and its necessity. Also you will get the idea about the terms mapper , reducer and shuffling. The concept of usage of map reduce and its flow will be covered under this section.
In this section you will learn about the terminology like Combiner, Partitioner and Counter. You will get the basic idea about the Input formats and Output formats. Also the concept of mapping and reducing join using MR. Also learn about the configuration of Hadoop.
In this section you will get an idea about the challenges or drawback of using Hadoop 1.0. Then you will learn about the new features added in Hadoop 2.0. Also its Cluster Architecture and Federation. The concept of Yarn, Hadoop Ecosystem and the Yarn MR Application flow will be covered in this section.
In this section you will learn about basic concept of Pig, its features, its use cases. Also you will learn how to interact with Pig. It will also cover the idea of Basic data analysis and the Latin syntax of Pig. Also you will how data is loaded , its data types ,field definitions and the output of data.
In this section you will learn about view of schema, how to filter and sorting data. The commonly used functions, processing complex data . It will cover the concept of grouping data and also techniques of combining data sets. Also you will get idea about how to join data sets and split data sets in Pig.
In this section you will explore about the fundamentals, architecture of Hive and how to load data and apply query on data in Hive. It will also cover the data types, operators and functions of HiveQL. You will also learn about hive tables, managed tables and external table. Also about Storage formats, Importing data , Altering tables and dropping table. It will also cover about data query, sorting and aggregating data and scripts of Map Reduce.
Under this section you will learn about CAP Theorem. Also you will learn about the concept and architecture of HBase. Also you will learn about Client API’s and its features. The section will also cover the data models and operations over it.
In this section you will learn about the basic concept of Sqoop. You will also explore about connecting relational database using Sqoop. You will also learn about importing and exporting data from and to MySQL respectively. Then also how to further connect MySQL data to hive and hbase. You will also learn about using queries in Sqoop.
In this section you will learn the basic concept of Flume, its importance, architecture and configuration. Then you will learn about the concept of Oozie, its architecture and its configuration. It will also cover about the properties of Oozie and Job Submission in Oozie.
In this section you will learn about you will learn about the basics of Apache Spark, its importance and benefits. It will also cover the overview of Batch Analytics in Hadoop Ecosystem and also real time analytics options. It will give an idea about how Spark is beneficial to professionals. Under this section you will also learn about Spark components and Executive architecture of Spark.
In this section you will learn about features of Scala. Also you will explore about Basic data types, Objects, Classes, List, and Maps in Scala. You will also learn about usage of function as object and usage of Anonymous Functions, along with higher order functions in Scala. You will also learn about Pattern matching, Traits and Collections in Scala.
In this section you will learn about Spark, its Application and its deployment. You will also get idea about Distributed Systems. You will also get to know about Spark for scalable system and its execution context. You will also get to learn about Parallelism and Caching in Spark. The section will also cover basic idea about RDD, its deep dive, dependencies and lineage.
In this section you will learn about Transformations, Actions, Clusters also about the Data Frames in Spark. Also you will learn about the basic introduction of SQL. You will also explore about Spark SQL with CSV, JSON and Database.
In this section you will learn about Features of Spark Streaming and its use case. Also you will learn about Dstreams and about Transformations on it. Also you will explore about Hadoop and its importance. In this you will learn about Distributed Computation and Functional Programming. Also you will get to know about MapReduce Framework and its job run.

Unit 3 : Business Intelligence

Under this section we will cover the basics of Tableau. Also it will give the insights of the importance and beneficial measures about working with Tableau. With this you will be able to clarify the quest about learning Tableau.
Under this we will learn how to connect or import data from different data sources. Also you will learn how different format of data can be imported to tableau.
Under this section you will cover about how you can clean data with the help of tableau. In general data cleaning is not done in tableau in vast area but in this we can clean data in minor level. In this you will learn the usage and working of data interpreter tool.
Under this you will learn about the actual workspace where you we will be working on the data set. You will be learning about how the different features will be useful for showing the different insights of data which will make data in much interactive manner.
Under this you will learn how to make data much more concise. This will make the data much more optimized which will help in depicting data in much more meaningful way. Mostly the data is prepared by removing the unwanted and noisy values from the data.
In this section you will learn about how to show data through different appropriate graph according to the data. Different formatting methods can be applied in graph to make the data in graphs appropriately descriptive. Different formatting filters will help in visualizing data in a understanding way.
In this section you will learn about the different filters in detail. By applying these filters as and when required on the graph will make the graphs much more informative. Also you will learn about how to edit the names of the data present on the row. Also we can change the diameters by ranging the values rather than the single value.
Under this section you will learn about creating additional fields. This is done when you need to analyse something different that is not given directly in the data source. You will learn in detail by creating new field with the help of existing fields simply or by applying conditions on them.
Under this particular section you will explore more filter tools which will make data more exploratory.
In this topic you will get to learn about framing the dashboard, which is basically concatenation of different graphs. In dashboard you bring different sheets depicting relative information on one sheet. Also you can add filter to it so if you change anywhere on dashboard it will reflect accordingly and vice versa.
Under this topic you will explore about how you can share your tableau project to client. This is one of the vital step in tableau course. With the help of publishing function you can create a simple link which can be then shared with the remote users to make them access the project that you made.
In this section you will learn about how you can create a full explanatory storyline. This is basically the sequence of different graph. Addition to this along with every sheet you can write caption or a short story guide with every sheet which will give the guideline narrating a sort of story. This will help you in explaining the data more concisely and optimally.

Unit 4 : Power BI

Power BI is a very powerful tool developed by Microsoft. As the name suggest it has been used for business intelligence purpose. In Power BI we will develop various reports and dashboards and do some visualizations.
Power BI allow us to create datasets from loaded data models. We can load data models from various sources like Excel, SQL Server etc. As I have already created a dataset in SQL Server. So we will directly load dataset from SQL Server.
Power BI and Power BI desktop are two different tools. Power BI is a windows application which can be downloaded in local pc where as Power BI is a cloud based service. To get started with Power BI then download Power BI desktop from https://powerbi.microsoft.com/en/downloads/.
Quick Insights is a feature of Power BI use to to generate interesting interactive visualizations based on data. Quick insights can be run on a specific dashboard tile.
To use Quick Insights we have load dataset in Power BI cloud. It is a feature of Power BI cloud used for analyses of data sets and finds patterns, trends and outliers.
In Power BI desktop or in Power BI cloud we will build various reports to make various visuals. The reports can be seen in report tab in Power BI desktop.
Dashboard is single page formed by combination of various reports. Power BI allows us to make multiple dashboards through different reports.
Power BI desktop is a free windows application which runs on windows computers. After downloading Power BI desktop, it will ask for sign up, we can use it without signing up but we will not able to publish our visualisation.
Data Modelling is a feature of Power BI. In Power BI we can do data modelling be adding or removing columns, splitting columns. Power BI allow us to write or edit queries. But as I already made dataset in SQL Server so we don’t need to do data modelling.
Power BI allow us to add or remove columns from our data table. We can add more data to the data table.
Power BI is a cloud-based business analytics service from Microsoft that enables anyone to visualize and analyze data, with better speed and efficiency. It is a powerful as well as a flexible tool for connecting with and analyzing a wide variety of data. Many businesses even consider it indispensable for data-science-related work. We will do visualization after publishing the data.

Unit 5 : SQL

• Introduction to T-SQL
• What is SQL Server
• What is SSMS
• What is Database
• Datatypes in SQL
• DDL Commands (Create, Alter, Drop)
• DML Commands (Select, Insert, Update, Delete, Truncate)
• Clauses ( Order by, Where By, Group By, Having)
• Aggregate Function
• Distinct, Null Values
• Operators (Arithmetic, logical, Conditional)
 o Arithmetic (+,-,/,*,++,--)
 o Logical Operator (<,>,<=,>=,!=,==)
 o Conditional (and, or)
 o In, not in, is null, between

• Alias Name for column and table name
• Introduction to Views
• Inner Join
• Left Join
• Right Join
• Full Outer Join
• Cross Join
• Self Join
• Introduction to SP
• Create Parameterized SP
• Execution SP
• Introduction to Trigger
• Creating trigger
• Creating cursor
• Use of trigger and cursor
• Introduction to Functions
• Types of funtions
o User Define function
 Inline table value function  Multiline table value function  Scalar funtion
o Built in function
 Date Funtion  String Function  Row Number  Rank, Dense Rank  Colease, is null
• Case when
• Union All


Coming soon


Coming soon


Coming soon

Fee structure for

“Kindly send your queries to info@academyofdatascience.com”

Fee structure for

“Kindly send your queries to info@academyofdatascience.com”

Fee structure for

“Kindly send your queries to info@academyofdatascience.com”



Batch Available


Batches Runninng


of Live Projects

Other Training Programmes

From The Blog

Sachin Raizada

Data Visualization on Power BI

Sachin Raizada

Visualization On Predictive Data