Get Even More Visitors To Your Blog, Upgrade To A Business Listing >>

HDFS Interview Question Answer Preparation

HDFS the Hadoop distributed file system is the file system project that supports bigdata in Hadoop ecosystem. Some big data companies like MapR do have their proprietary filesystem instead of HDFS.Many organizations do make use of HDFS and questions on HDFS file system is definitely going to be part of bigdata interview. Here are some HDFS interview questions to crack bigdata jobs 1) What is HDFS? HDFS stands for Hadoop distributed filesystem. It is a distributed file system that is used to manage distributed data across clusters. 2) What are the two major components of HDFS? Namenode, data node 3) What is the core concept behind design of HDFS filesystem? HDFS filesystem is designed with the following concepts in mind: Size of file - HDFS filesystem is meant to store files that are of gigabyte, terabyte in size. Real-time hadoop clusters store petabytes worth data in filesystem Performance with streaming data - IF we look at typical bigdata projects information from many different sources gets stored in HDFS, additional information is appended onto data in HDFS. Hadoop is write once read many implementation that should consider performance of data access as well. HDFS is designed to support streaming data HDFS failover capabilities - Hadoop clusters are a good alternate to expensive hardware as they are designed to be run in commodity hardware. The term commodity hardware is more generic. They are mid range servers from many different vendors that are reasonably stable. In case of failure there should be node failover provision. HDFS does support this via replication factor 4) Is HDFS a good fit for all the applications? Nope. HDFS may not be the right choice under following situations Low latency applications Lots of small files HDFS files can be written to in append-only fashion. Many applications can't write to it at same time 5) What is the purpose of distributed filesystem? Bigdata refers to the data whose volume does mandate storage of data across more than one commodity servers. The information is stored across network and set of machines across network behave as single physical entity. This mandates a filesystem that should take care of filesystem management across network. Thats wherein distributed filesystem comes into picture 6) What is the use of heartbeat in HDFS? Datanodes sends signals to namenode in HDFS environment. This is an indication that namenodes are functioning properly. The default interval of heartbeat is 3 seconds. The heartbeat interval is configurable and can be changed by setting dfs.heartbeat.interval value in hdfs-site.xml file 7) What command is used to check the status of daemon on HDFS? The jps command is a java command that needs jdk for running. This comamnd will produce list of all hadoop daemons that are currently running like Namenode, TestTracker,JobTracker. This can be thought of linux equivalent ps command 8) Which file should I make use of for changing the block size of HDFS files? hdfs-site.xml file has default size parameter. This value can be changed to change the block size of HDFS. This should be done during downtime as this process demands cluster restart 9) I want to modify the files present in HDFS. What should I do? HDFS works on the concept of write once, read many. All we can do is append data to an existing file. It is not possible to modify the files already present in HDFS 10) What is the use of hadoop archives? HDFS utilized hadoop archives concept to minimize the metadata information stored in namenode. This inturn conserves memory 11) Is HDP different than Hadoop? HDP the horton data platform is a distribution from hortonworks that is involved in apache hadoop enterprise distributions. It is popularly called HDP. HDP is at 2.3.2 now and supports big data for enterprises. So, is HDP different than apache hadoop? Nope. If so why do I need HDP. Instead I can download directly from apache software foundation Apache hadoop ecosystem is a set of tools that solve the big data challenges. Essential components of hadoop include hdfs, flume, spark, sqoop, hbase, hive, mapreduce, yarn a resource scheduler enhanced from map reduce jobs to name a few. Each and every component in itself is a separate project and has many different versions that get released at different point in time. This lack of synchronization can cause compatibility issue among many different components, can cause one or more components in hadoop cluster to break while upgrade, can affect performance functionality etc. As such there is a need to bundle these hadoop components, make sure they are properly functional, test them adn release them to field as standard enterprise distributions that are stable, relied upon. That's wherein vendors like hortonworks, cloudera come into picture. Hence, HDP is the hadoop flavor bundled, shipped to enterprise by hortonworks. Another popular hadoop distribution is CDH from clouderaStart learning hadoop to enter bigdata spaceHadoop the open source Apache foundation Project originally written in JAva code that makes use of Google filesystem as its base forms the framework to support big data All of us say big data. So far we have been processing TB's of data using existing relational dataabse management system. So, what exactly is Big Data? Let us first take a look at what are the three major things that are being addressed by Hadoop - It is popularly called 3 V's - Velocity, Volume, Variety Yep - These 3 acronyms form the basics of Big data Big data as the basic properties define ; 1) Grow at a spectacular rate. Good examples inlude data collected from sensors in offices, RFId, mobile phones etc 2) Are voluminous in nature 3) Can be - structure, unstructure, semi - structured - Come in different forms and variety To handle this kind of data relational database may not be sufficient. To handle this hadoop framework comes into picture Hadoop Framework A Quick Overview:- To kick start the big data arena it becomes mandate to know the ABCD "Hadoop". Apache hadoop is the framework built by team of Yahoo engineers in 2005. Originally built in c++ this project eventually has become stable in java. This is an open source project supported by Apache. Anyone can download, practise this binary for free. As with any popular frameworks, Apache hadoop is available in popular commercial flavours from Hortonworks, Cloudera etc Lets take a quick look at the pieces that will make the Hadoop framework work big 1) Apache hadoop - This is the framework on which big data is supported. This is considered hadoop data management tool 2) Hadoop Pig Latin - This is the scripting language used to process big data. As this is associated with data this becomes big data management tool 3) Apache Hadoop HBase - This is the NoSQL database from hadoop. This is database for big data 4) Apache Hadoop HDFS - The Hadoop distributed file system that hosts big data and is a data management tool 5) Apache Hadoop Ambari- The monitoring and management tool classified as Hadoop operational tool 6) Apache Hadoop zookeeper - This bigdata operational tool is used for configuration of hadoop framework 7) HAdoop squoop - This is used to migrate data from relational databases onto HAdoop HDFS 8) Hadoop flume - This forms big data aggregation. This includes aggregation of logs onto central repository Bigdata promising career for java developersAs an aspiring java programmer if you are exhausted and looking for a career change that travels on top of your prior java development experience but gives you best compensation, big data is the way to go Many vendors are in process of developing and implementing tools to support big data. One of the most popular vendor technology that supports big data is Cloudera Now, lets take a look at how it is possible to handle the big data challenges thrown at a java programmer. As a java programmer you can choose to start learning the following skills to grow big and make more money in your career 1) Start learning hadoop mapreduce. Scripting using pig or java is an essential skill. In some organizations mapreduce functions inherent in nosql databases like mongodb might come handy 2) Hadoop hive, spark and such distributed data processing platform experience is an essential skill to take up job as big data engineer 3) Experience with tools like cloudera manager would be a plus. Some employers prefer certification from cloudera 4) Experience working with nosql databases like hbase, mongodb would be a plus 5) Development experience with java is much preferred. However experience with python, R might come handy 6) This is a technology trend. Lots of attitude, ambition, self-learning is an essential skill 7) Must be very much comfortable working in linux environment, shell scripting, perl scripting, python scripting , ruby on rails scripting etc comes handy 8) Some employers prefer cloud knowledge and experience like AWS, essentially components of AWS including EC2, S3, EMR etc 9) Big data development is an agile environment and hence SDLC life cycle knowledge is a must Cloudera Distribution and apache hadoop quick overview : Data has grown form paper files to digital CD's, floppy disks, hard disk, storage SAN/NAS and now hadoop cluster is the trend. Apache software foundation called ASF runs set of projects to support data that are generated at faster pace, come in structured, unstructured form and that needs to be stored and processed to mine valuable business insights. This is wherein hadoop project has come into existence. In simple terms apache hadoop is the framework needed to store, process massive amounts of data. Set of machines presented to end user as cluster. In real world this is considered a cost saving measure as commodity hardware can be made use of for implementing apache hadoop cluster. Here are some interesting facts and features of apache hadoop 1) Fault tolerant - The basic unit of datastore is HDFS the hadoop distributed file system that is used to store bigdata that can come in many different forms 2) Scalable - possible to add more machines to cluster to meet the growing demand 3) Open source - Hadoop is not owned by any firm. Anyone can download source code. They can modify and run the code. Instead of downloading directly from apache website, look for distributions like CDH from cloudera, HDP from hortonworks that are apache ahdoop flavors bundled with appropriate components, tested and released for use by enterprises Projects are built around hadoop comprise hadoop ecosystem. Some components include: 1) Spark 2) Scala 3) Kafka 4) Ranger 5) Storm 6) Flume These tools that form part of apache hadoop ecosystem make hadoop easier to use Give details on cloudera, CDH and how is this related to hadoop? Cloudera offers enterprise solutions to solve the bigdata problem of enterprises. Just like Ubuntu, RHEL, Federacor or any other Linux distribution CDH is a licensed version of apache hadoop for enterprises. The service offerings of cloudera doesn't stop there. Cloudera Manager is the graphical user interface that can be used to manage hadoop cluster from UI. This cna be treated similar to oracle enterprise manager the GUI from oracle Career of cloudera bigdata analystWant to take up career as a Cloudera bigdata data analyst? Interested in learning the requisites? Here is the outline of requirements to emerge as a cloudera big data data analyst 1) The primary skill necessary to find career as adata analyst is SQL 2) In hadoop specific environment it becomes mandatory to learn tools like pig scripting, Hive, impala to analyse big data 3) Deals with high level analysis is involved. Dont need to be a developer 4) Learn how to get data other systems like datawarehouse, databases 5) Learn how to analyse big data set 6) Knowledge of basic relational databases comes handy 7) Basic unix commands definitely helps. Some interesting commands include mkdir,rm,cp,move etc 8) High level programming language knowledge definitely helps. Most preferred languages include java, python, perl 9) Knowledge of ETL, hadoop framework is a plus 10) You get to learn hadoop data ingestion tools, analysis using pig scripting, hive commands, impala to name a few Building hadoop cluster know howAs a administrator in infrastructure team say your manager wants you to come up with a plan for prospective hadoop cluster to support an upcoming big data project in your organization. Are you wondering from where to start? Here are some basic things that come handy as a first step in building hadoop cluster Before looking at installation options lets see what are all the ways to build the cluster. This can be done in one of the following ways Choosing hardware understanding hadoop architectureChoose set of commodity hardware in your datacenter. If not have a meeting with capacity planner to determine if you have commodity hardware in place. Use one or more of them. With the combined resource availability make sure you can start building hadoop cluster on your own. Typically when it comes to HDFS files the recommendation is to have atleast three nodes for redundancy purpose that does guarantee high availability. This is a totally different topic that we can discuss in detail. Essentially 3 commodity hardware machines might be needed. Here are the many different ways to build hadoop cluster 1) Utilize the commodity hardware in your organization 2) Rent hardware 3) Make use of cloud services like amazon web services, microsoft azure that make the hadoop cluster creation and hosting a piece of cake. All you need is to buy the appropriate virtual machines form these vendors, create and launch hadoop cluster in less timeframe. This comes with unique advantage of paying as and when your resource consumption increases. These Infrastructure as Service makes the job easy and simple Now, lets look at hadoop cluster installation options: Say, you choose to build hadoop cluster on your own, here are the installation options to be considered 1.1) Apache tarballs - This is one of the most time consuming task as you need to download appropriate binary tar balls from apache hadoop and related projects. You may have to decide on location of installation files, configuration files, log files in file system. Need to make sure file permissions are set correctly and so on. Also, unique thing to be noted is that you make sure the version of hadoop you download is compatible with Hive. The component compatability has not been tested and certified once you do all by yourself 1.2) Apache Packages - Starting with apache hadoop bigtop projects, to vendor packages form hortonworks, cloudera, MApR to name a few enterprise hadoop clusters mostly rely on RPM and debian packages from certified vendors. This ensures the component compatability like proper functioning of hadoop with hive, puppet to name a few that eases most of the work 1.3) Hadoop Cluster Management tools - Starting with Apache ambari to cloudera manager many GUI tools make this a piece of cake. Also, this comes with unique advantage of rolling upgrade that helps cluster upgrade with zero downtime. When there is a need to add more resources to cluster, the job becomes easy utilizing these tools. The tools come with heuritstics, best recommendation that comes handy while working with many different components of hadoop 12) What is Hortonworks dataflow and dataplatform real difference ? This month Hortonworks has released the latest version of hortonworks dataflow version 1.1. Hoetonworks DAta Platform popularly called HDP is the major project and product of hortonworks that is built on top of open-source hadoop ecosystem. Now, do hortonwoks data flow and data paltform represent the same? No. HDP the hortonworks data platform is the bundled version of open-source hadoop in a packaged format. Using a installer all the components that form part of hadoop project are chosen and bundled correctly. As the many different components in hadoop ecosystem have different version releases at different point-in-time and compatability is not always guaranteed, HDP is a stable solution for enterprises looking to have hadoop implemented as a customized, stable, tested package that is installed using installer Hadoop Dataflow on other hand is Apache nifi. This is the GUI tool used to design the dataflows using processors which are data extracting engines designed to work with many different datasources. Hadoop is meant for its data enrinchment. As such there are around 90 processors in HDP that can getfiles from local file system, extract information from twitter etc. This information can be put into HDFS the hadoop distributed filesystem and dataflow is designed using relationships. Once the drag and drop of the processors is done in GUI, appropriate properties are configured, relation ship is established and built appropriately dataflow gets initiated. As such HDF is for designing dataflow, HDP is the apache hadoop platform supporting enterprise big data projects starting with its HDFS the hadoop distributed file system



This post first appeared on FOOD PREPARATION TIPS AND USEFUL INFORMATION, please read the originial post: here

Share the post

HDFS Interview Question Answer Preparation

×

Subscribe to Food Preparation Tips And Useful Information

Get updates delivered right to your inbox!

Thank you for your subscription

×