Apache AVRO is a very popular data serialization format in the Hadoop technology stack.
It is used widely in Hadoop stack i.e in hive,pig,mapreduce componenets.
It stores metadata also along with Actual data.
It is a rowbased oriented data storage format.
Provides schema evaluation and block compression.
Metadata will be represented in JSON file
Avro depends heavily on its schema. It allows every data to be written with no prior knowledge of the schema. It serializes fast and the resulting serialized data is lesser in size. Schema is stored along with the Avro data in a file for any further processing.
In RPC, the client and the server exchange schemas during the connection. This exchange helps in the communication between same named fields, missing fields, extra fields, etc.
Avro schemas are defined with JSON that simplifies its implementation in languages with JSON libraries.
Like Avro, there are other serialization mechanisms in Hadoop such as Sequence Files, Protocol Buffers, and Thrift.
I am Reddy having 10 years of IT experience.For the last 4 years I have been working on Bigdata.
From Bigdata perspective,I had working experience on Kafka,Spark,and Hbase,cassandra,hive technologies.
And also I had working experience with AWS and Java technologies.
I have the experience in desigining and implemeting lambda architecture solutions in bigdata
Has experience in Working with Rest API and worked in various domains like financial ,insurance,manufacuring.
I am so passinate about new technologies.
BigDataTechnologies is a online training provider and has many experienced lecturers who will proivde excellent training.
BigDataTechnologies has extensive experience in providing training for Java,AWS,iphone,Mapredue,hive,pig,hbase,cassandra,Mongodb,spark,storm and Kafka.
From skills that will help you to develop and future proof your career to immediate solutions to every day tech challenges.
Main objective is to provide high quality content to all students