Bigdata Analytics with Hive,Spark,Sqoop
0.0 (0 ratings)
Instead of using a simple lifetime average, Udemy calculates a course's star rating by considering a number of different factors such as the number of ratings, the age of ratings, and the likelihood of fraudulent ratings.
3 students enrolled
Wishlisted Wishlist

Please confirm that you want to add Bigdata Analytics with Hive,Spark,Sqoop to your Wishlist.

Add to Wishlist

Bigdata Analytics with Hive,Spark,Sqoop

If you’re an R developer looking to harness the power of big data analytics with Hadoop, then this course is for you
0.0 (0 ratings)
Instead of using a simple lifetime average, Udemy calculates a course's star rating by considering a number of different factors such as the number of ratings, the age of ratings, and the likelihood of fraudulent ratings.
3 students enrolled
Created by Ashok M
Last updated 3/2017
English
Curiosity Sale
Current price: $10 Original price: $20 Discount: 50% off
30-Day Money-Back Guarantee
Includes:
  • 33 mins on-demand video
  • Full lifetime access
  • Access on mobile and TV
  • Certificate of Completion
What Will I Learn?
  • Understand how to work with SparkR
  • Will understand how to create dataframes in SparkR
  • understand how to create Linear regression
View Curriculum
Requirements
  • You should have basic understanding of Spark
  • You should have basic understanding of R
Description

SparkR is an R package that provides a light-weight frontend to use Apache Spark from R. In Spark 2.0.2, SparkR provides a distributed data frame implementation that supports operations like selection, filtering, aggregation etc. (similar to R data frames) but on large datasets. SparkR also supports distributed machine learning using MLlib.

You will learn how to create spark cluster in Databricks.

You will learn how to create dataframes and grouping data and aggregating data.

Hadoop is an Apache open source framework written in java that allows distributed processing of large datasets across clusters of computers using simple programming models. A Hadoop frame-worked application works in an environment that provides distributed storage and computation across clusters of computers. Hadoop is designed to scale up from single server to thousands of machines, each offering local computation and storage


Prerequisites:

You should have basic knowledge of Spark and R


  • Who have  some R experience that wants to learn about big data solutions
  • Who are interested in SparkR and Hadoop
  • Who are interested in Spark and cluster computing


Who is the target audience?
  • This is for all students
  • Who are interested in SparkR
  • Good for all Bigdata developers and Datascientists
Students Who Viewed This Course Also Viewed
Curriculum For This Course
+
Introduction
1 Lecture 01:20
+
Sqoop
4 Lectures 26:04


sqoop incremental loading
06:00

sqoop incremental loading --DEMO
11:05
+
Hive
2 Lectures 05:57
When to Use hive and Pig
01:55

Grouping and Aggregating Data
04:02
About the Instructor
Ashok M
2.4 Average rating
61 Reviews
340 Students
29 Courses
Architect

I am  Reddy having 10 years of IT experience.For the last 4 years I have been working on Bigdata.
From Bigdata perspective,I had working experience on Kafka,Spark,and Hbase,cassandra,hive technologies.
And also I had working experience with AWS and Java technologies.

I have the experience in desigining and implemeting lambda architecture solutions in bigdata

Has experience in Working with Rest API and worked in various domains like financial ,insurance,manufacuring.

I am so passinate about  new technologies.


BigDataTechnologies  is a online training provider and has many experienced lecturers who will proivde excellent training.

BigDataTechnologies has extensive experience in providing training for Java,AWS,iphone,Mapredue,hive,pig,hbase,cassandra,Mongodb,spark,storm and Kafka.

From skills that will help you to develop and future proof your career to immediate solutions to every day tech challenges.

Main objective is to provide high quality content to all students