SlideShare ist ein Scribd-Unternehmen logo
1 von 32
Apache Hadoop and Hive

              Dhruba Borthakur
          Apache Hadoop Developer
         Facebook Data Infrastructure
   dhruba@apache.org, dhruba@facebook.com
         Condor Week, April 22, 2009
Outline
• Architecture of Hadoop Distributed File System
• Hadoop usage at Facebook
• Ideas for Hadoop related research
Who Am I?
• Hadoop Developer
   – Core contributor since Hadoop’s infancy
   – Project Lead for Hadoop Distributed File System
• Facebook (Hadoop, Hive, Scribe)
• Yahoo! (Hadoop in Yahoo Search)
• Veritas (San Point Direct, Veritas File System)
• IBM Transarc (Andrew File System)
• UW Computer Science Alumni (Condor Project)
Hadoop, Why?
• Need to process Multi Petabyte Datasets
• Expensive to build reliability in each application.
• Nodes fail every day
  – Failure is expected, rather than exceptional.
  – The number of nodes in a cluster is not constant.
• Need common infrastructure
  – Efficient, reliable, Open Source Apache License
• The above goals are same as Condor, but
   – Workloads are IO bound and not CPU bound
Hive, Why?
• Need a Multi Petabyte Warehouse
• Files are insufficient data abstractions
   – Need tables, schemas, partitions, indices
• SQL is highly popular
• Need for an open data format
  – RDBMS have a closed data format
  – flexible schema
• Hive is a Hadoop subproject!
Hadoop & Hive History
• Dec 2004    – Google GFS paper published
•   July 2005 – Nutch uses MapReduce
•   Feb 2006 – Becomes Lucene subproject
•   Apr 2007 – Yahoo! on 1000-node cluster
•   Jan 2008 – An Apache Top Level Project
•   Jul 2008 – A 4000 node test cluster
• Sept 2008 – Hive becomes a Hadoop subproject
Who uses Hadoop?
•   Amazon/A9
•   Facebook
•   Google
•   IBM
•   Joost
•   Last.fm
•   New York Times
•   PowerSet
•   Veoh
•   Yahoo!
Commodity Hardware



Typically in 2 level architecture
– Nodes are commodity PCs
– 30-40 nodes/rack
– Uplink from rack is 3-4 gigabit
– Rack-internal is 1 gigabit
Goals of HDFS
• Very Large Distributed File System
  – 10K nodes, 100 million files, 10 PB
• Assumes Commodity Hardware
  – Files are replicated to handle hardware failure
  – Detect failures and recovers from them
• Optimized for Batch Processing
  – Data locations exposed so that computations can
  move to where data resides
  – Provides very high aggregate bandwidth
• User Space, runs on heterogeneous OS
HDFS Architecture
                                                                                                           Cluster Membership



                                            e                                     NameNode
                                          am
                                      ilen
                                  1. f




                                                                       es
                                                                    Nod
                                                          Id,   Data
                                                2.   Blck                          Secondary
                                                                                   NameNode
                                                 o

          Client




                            3.Read d
                                     ata



                                                                                                                                Cluster Membership




NameNode : Maps a file to a file-id and list of MapNodes
                                                                                               DataNodes
DataNode : Maps a block-id to a physical location on disk
SecondaryNameNode: Periodic merge of Transaction log
Distributed File System
• Single Namespace for entire cluster
• Data Coherency
  – Write-once-read-many access model
  – Client can only append to existing files
• Files are broken up into blocks
  – Typically 128 MB block size
  – Each block replicated on multiple DataNodes
• Intelligent Client
  – Client can find location of blocks
  – Client accesses data directly from DataNode
NameNode Metadata
• Meta-data in Memory
  – The entire metadata is in main memory
  – No demand paging of meta-data
• Types of Metadata
  – List of files
  – List of Blocks for each file
  – List of DataNodes for each block
  – File attributes, e.g creation time, replication factor
• A Transaction Log
  – Records file creations, file deletions. etc
DataNode
• A Block Server
  – Stores data in the local file system (e.g. ext3)
  – Stores meta-data of a block (e.g. CRC)
  – Serves data and meta-data to Clients
• Block Report
  – Periodically sends a report of all existing blocks to
  the NameNode
• Facilitates Pipelining of Data
  – Forwards data to other specified DataNodes
Block Placement
• Current Strategy
  -- One replica on local node
  -- Second replica on a remote rack
  -- Third replica on same remote rack
  -- Additional replicas are randomly placed
• Clients read from nearest replica
• Would like to make this policy pluggable
Data Correctness
• Use Checksums to validate data
  – Use CRC32
• File Creation
  – Client computes checksum per 512 byte
  – DataNode stores the checksum
• File access
  – Client retrieves the data and checksum from
  DataNode
  – If Validation fails, Client tries other replicas
NameNode Failure
• A single point of failure
• Transaction Log stored in multiple directories
  – A directory on the local file system
  – A directory on a remote file system (NFS/CIFS)
• Need to develop a real HA solution
Data Pipelining
• Client retrieves a list of DataNodes on which to place
  replicas of a block
• Client writes block to the first DataNode
• The first DataNode forwards the data to the next
  DataNode in the Pipeline
• When all replicas are written, the Client moves on to
  write the next block in file
Rebalancer
• Goal: % disk full on DataNodes should be similar
   –   Usually run when new DataNodes are added
   –   Cluster is online when Rebalancer is active
   –   Rebalancer is throttled to avoid network congestion
   –   Command line tool
Hadoop Map/Reduce
• The Map-Reduce programming model
  – Framework for distributed processing of large data
  sets
  – Pluggable user code runs in generic framework
• Common design pattern in data processing
  cat * | grep | sort    | unique -c | cat > file
   input | map | shuffle | reduce | output
• Natural for:
  – Log processing
   – Web search indexing
   – Ad-hoc queries
Hadoop at Facebook
• Production cluster
   –   4800 cores, 600 machines, 16GB per machine – April 2009
   –   8000 cores, 1000 machines, 32 GB per machine – July 2009
   –   4 SATA disks of 1 TB each per machine
   –   2 level network hierarchy, 40 machines per rack
   –   Total cluster size is 2 PB, projected to be 12 PB in Q3 2009


• Test cluster
   • 800 cores, 16GB each
Data Flow



Web Servers          Scribe Servers


                                              Network
                                              Storage




 Oracle RAC   Hadoop Cluster          MySQL
Hadoop and Hive Usage
• Statistics :
   –   15 TB uncompressed data ingested per day
   –   55TB of compressed data scanned per day
   –   3200+ jobs on production cluster per day
   –   80M compute minutes per day
• Barrier to entry is reduced:
   – 80+ engineers have run jobs on Hadoop platform
   – Analysts (non-engineers) starting to use Hadoop through
     Hive
Ideas for Collaboration
Condor and HDFS

• Run Condor jobs on Hadoop File System
   – Create HDFS using local disk on condor nodes
   – Use HDFS API to find data location
   – Place computation close to data location

• Support map-reduce data abstraction model
Power Management
• Power Management
   – Major operating expense
   – Power down CPU’s when idle
   – Block placement based on access pattern
      • Move cold data to disks that need less power
• Condor Green
Benchmarks
• Design Quantitative Benchmarks
   – Measure Hadoop’s fault tolerance
   – Measure Hive’s schema flexibility
• Compare above benchmark results
   – with RDBMS
   – with other grid computing engines
Job Sheduling
• Current state of affairs
   – FIFO and Fair Share scheduler
   – Checkpointing and parallelism tied together
• Topics for Research
   – Cycle scavenging scheduler
   – Separate checkpointing and parallelism
   – Use resource matchmaking to support
     heterogeneous Hadoop compute clusters
   – Scheduler and API for MPI workload
Commodity Networks
• Machines and software are commodity
• Networking components are not
   – High-end costly switches needed
   – Hadoop assumes hierarchical topology
• Design new topology based on commodity
  hardware
More Ideas for Research
• Hadoop Log Analysis
   – Failure prediction and root cause analysis
• Hadoop Data Rebalancing
   – Based on access patterns and load
• Best use of flash memory?
Summary
• Lots of synergy between Hadoop and Condor
• Let’s get the best of both worlds
Useful Links
• HDFS Design:
  – http://hadoop.apache.org/core/docs/current/hdfs_design.html
• Hadoop API:
  – http://hadoop.apache.org/core/docs/current/api/
• Hive:
  – http://hadoop.apache.org/hive/

Weitere ähnliche Inhalte

Was ist angesagt?

Hadoop Distributed File System
Hadoop Distributed File SystemHadoop Distributed File System
Hadoop Distributed File System
Vaibhav Jain
 
Dynamic Namespace Partitioning with Giraffa File System
Dynamic Namespace Partitioning with Giraffa File SystemDynamic Namespace Partitioning with Giraffa File System
Dynamic Namespace Partitioning with Giraffa File System
DataWorks Summit
 
Hadoop architecture meetup
Hadoop architecture meetupHadoop architecture meetup
Hadoop architecture meetup
vmoorthy
 
Hadoop Ecosystem | Hadoop Ecosystem Tutorial | Hadoop Tutorial For Beginners ...
Hadoop Ecosystem | Hadoop Ecosystem Tutorial | Hadoop Tutorial For Beginners ...Hadoop Ecosystem | Hadoop Ecosystem Tutorial | Hadoop Tutorial For Beginners ...
Hadoop Ecosystem | Hadoop Ecosystem Tutorial | Hadoop Tutorial For Beginners ...
Simplilearn
 
Hadoop Distributed File System
Hadoop Distributed File SystemHadoop Distributed File System
Hadoop Distributed File System
elliando dias
 
Hadoop 2.0 Architecture | HDFS Federation | NameNode High Availability |
Hadoop 2.0 Architecture | HDFS Federation | NameNode High Availability | Hadoop 2.0 Architecture | HDFS Federation | NameNode High Availability |
Hadoop 2.0 Architecture | HDFS Federation | NameNode High Availability |
Edureka!
 
Hadoop Training | Hadoop Training For Beginners | Hadoop Architecture | Hadoo...
Hadoop Training | Hadoop Training For Beginners | Hadoop Architecture | Hadoo...Hadoop Training | Hadoop Training For Beginners | Hadoop Architecture | Hadoo...
Hadoop Training | Hadoop Training For Beginners | Hadoop Architecture | Hadoo...
Simplilearn
 

Was ist angesagt? (20)

Hadoop Distributed File System
Hadoop Distributed File SystemHadoop Distributed File System
Hadoop Distributed File System
 
2013 feb 20_thug_h_catalog
2013 feb 20_thug_h_catalog2013 feb 20_thug_h_catalog
2013 feb 20_thug_h_catalog
 
Hadoop distributed file system
Hadoop distributed file systemHadoop distributed file system
Hadoop distributed file system
 
Dynamic Namespace Partitioning with Giraffa File System
Dynamic Namespace Partitioning with Giraffa File SystemDynamic Namespace Partitioning with Giraffa File System
Dynamic Namespace Partitioning with Giraffa File System
 
Hadoop 3.0 - Revolution or evolution?
Hadoop 3.0 - Revolution or evolution?Hadoop 3.0 - Revolution or evolution?
Hadoop 3.0 - Revolution or evolution?
 
Hadoop training in hyderabad-kellytechnologies
Hadoop training in hyderabad-kellytechnologiesHadoop training in hyderabad-kellytechnologies
Hadoop training in hyderabad-kellytechnologies
 
HDFS: Hadoop Distributed Filesystem
HDFS: Hadoop Distributed FilesystemHDFS: Hadoop Distributed Filesystem
HDFS: Hadoop Distributed Filesystem
 
Tutorial Haddop 2.3
Tutorial Haddop 2.3Tutorial Haddop 2.3
Tutorial Haddop 2.3
 
Hadoop architecture meetup
Hadoop architecture meetupHadoop architecture meetup
Hadoop architecture meetup
 
Introduction to Big Data & Hadoop
Introduction to Big Data & HadoopIntroduction to Big Data & Hadoop
Introduction to Big Data & Hadoop
 
Hadoop Ecosystem | Hadoop Ecosystem Tutorial | Hadoop Tutorial For Beginners ...
Hadoop Ecosystem | Hadoop Ecosystem Tutorial | Hadoop Tutorial For Beginners ...Hadoop Ecosystem | Hadoop Ecosystem Tutorial | Hadoop Tutorial For Beginners ...
Hadoop Ecosystem | Hadoop Ecosystem Tutorial | Hadoop Tutorial For Beginners ...
 
Hadoop Distributed File System
Hadoop Distributed File SystemHadoop Distributed File System
Hadoop Distributed File System
 
Hadoop Distributed File System
Hadoop Distributed File SystemHadoop Distributed File System
Hadoop Distributed File System
 
Hadoop Operations - Best practices from the field
Hadoop Operations - Best practices from the fieldHadoop Operations - Best practices from the field
Hadoop Operations - Best practices from the field
 
Big Data Meets HPC - Exploiting HPC Technologies for Accelerating Big Data Pr...
Big Data Meets HPC - Exploiting HPC Technologies for Accelerating Big Data Pr...Big Data Meets HPC - Exploiting HPC Technologies for Accelerating Big Data Pr...
Big Data Meets HPC - Exploiting HPC Technologies for Accelerating Big Data Pr...
 
Hadoop 2.0 Architecture | HDFS Federation | NameNode High Availability |
Hadoop 2.0 Architecture | HDFS Federation | NameNode High Availability | Hadoop 2.0 Architecture | HDFS Federation | NameNode High Availability |
Hadoop 2.0 Architecture | HDFS Federation | NameNode High Availability |
 
Lecture 2 part 1
Lecture 2 part 1Lecture 2 part 1
Lecture 2 part 1
 
Hadoop HDFS Architeture and Design
Hadoop HDFS Architeture and DesignHadoop HDFS Architeture and Design
Hadoop HDFS Architeture and Design
 
Hadoop Training | Hadoop Training For Beginners | Hadoop Architecture | Hadoo...
Hadoop Training | Hadoop Training For Beginners | Hadoop Architecture | Hadoo...Hadoop Training | Hadoop Training For Beginners | Hadoop Architecture | Hadoo...
Hadoop Training | Hadoop Training For Beginners | Hadoop Architecture | Hadoo...
 
What's new in hadoop 3.0
What's new in hadoop 3.0What's new in hadoop 3.0
What's new in hadoop 3.0
 

Ähnlich wie Borthakur hadoop univ-research

Константин Швачко, Yahoo!, - Scaling Storage and Computation with Hadoop
Константин Швачко, Yahoo!, - Scaling Storage and Computation with HadoopКонстантин Швачко, Yahoo!, - Scaling Storage and Computation with Hadoop
Константин Швачко, Yahoo!, - Scaling Storage and Computation with Hadoop
Media Gorod
 

Ähnlich wie Borthakur hadoop univ-research (20)

Hadoop ppt1
Hadoop ppt1Hadoop ppt1
Hadoop ppt1
 
Apache hadoop and hive
Apache hadoop and hiveApache hadoop and hive
Apache hadoop and hive
 
Hadoop introduction
Hadoop introductionHadoop introduction
Hadoop introduction
 
HDFS_architecture.ppt
HDFS_architecture.pptHDFS_architecture.ppt
HDFS_architecture.ppt
 
List of Engineering Colleges in Uttarakhand
List of Engineering Colleges in UttarakhandList of Engineering Colleges in Uttarakhand
List of Engineering Colleges in Uttarakhand
 
Hadoop.pptx
Hadoop.pptxHadoop.pptx
Hadoop.pptx
 
Hadoop.pptx
Hadoop.pptxHadoop.pptx
Hadoop.pptx
 
Hadoop
HadoopHadoop
Hadoop
 
Bigdata workshop february 2015
Bigdata workshop  february 2015 Bigdata workshop  february 2015
Bigdata workshop february 2015
 
Introduction to Hadoop Administration
Introduction to Hadoop AdministrationIntroduction to Hadoop Administration
Introduction to Hadoop Administration
 
Introduction to Hadoop Administration
Introduction to Hadoop AdministrationIntroduction to Hadoop Administration
Introduction to Hadoop Administration
 
Chapter2.pdf
Chapter2.pdfChapter2.pdf
Chapter2.pdf
 
Hadoop-Quick introduction
Hadoop-Quick introductionHadoop-Quick introduction
Hadoop-Quick introduction
 
Константин Швачко, Yahoo!, - Scaling Storage and Computation with Hadoop
Константин Швачко, Yahoo!, - Scaling Storage and Computation with HadoopКонстантин Швачко, Yahoo!, - Scaling Storage and Computation with Hadoop
Константин Швачко, Yahoo!, - Scaling Storage and Computation with Hadoop
 
Hadoop - Just the Basics for Big Data Rookies (SpringOne2GX 2013)
Hadoop - Just the Basics for Big Data Rookies (SpringOne2GX 2013)Hadoop - Just the Basics for Big Data Rookies (SpringOne2GX 2013)
Hadoop - Just the Basics for Big Data Rookies (SpringOne2GX 2013)
 
Big data Hadoop
Big data  Hadoop   Big data  Hadoop
Big data Hadoop
 
Cloud computing UNIT 2.1 presentation in
Cloud computing UNIT 2.1 presentation inCloud computing UNIT 2.1 presentation in
Cloud computing UNIT 2.1 presentation in
 
Hadoop introduction
Hadoop introductionHadoop introduction
Hadoop introduction
 
Hadoop on Azure, Blue elephants
Hadoop on Azure,  Blue elephantsHadoop on Azure,  Blue elephants
Hadoop on Azure, Blue elephants
 
Introduction to Hadoop Administration
Introduction to Hadoop AdministrationIntroduction to Hadoop Administration
Introduction to Hadoop Administration
 

Kürzlich hochgeladen

+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
?#DUbAI#??##{{(☎️+971_581248768%)**%*]'#abortion pills for sale in dubai@
 

Kürzlich hochgeladen (20)

GenAI Risks & Security Meetup 01052024.pdf
GenAI Risks & Security Meetup 01052024.pdfGenAI Risks & Security Meetup 01052024.pdf
GenAI Risks & Security Meetup 01052024.pdf
 
Apidays New York 2024 - The value of a flexible API Management solution for O...
Apidays New York 2024 - The value of a flexible API Management solution for O...Apidays New York 2024 - The value of a flexible API Management solution for O...
Apidays New York 2024 - The value of a flexible API Management solution for O...
 
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot TakeoffStrategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
 
Boost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdfBoost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdf
 
🐬 The future of MySQL is Postgres 🐘
🐬  The future of MySQL is Postgres   🐘🐬  The future of MySQL is Postgres   🐘
🐬 The future of MySQL is Postgres 🐘
 
Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organization
 
Manulife - Insurer Innovation Award 2024
Manulife - Insurer Innovation Award 2024Manulife - Insurer Innovation Award 2024
Manulife - Insurer Innovation Award 2024
 
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemkeProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
 
Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024
 
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost SavingRepurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
 
Top 10 Most Downloaded Games on Play Store in 2024
Top 10 Most Downloaded Games on Play Store in 2024Top 10 Most Downloaded Games on Play Store in 2024
Top 10 Most Downloaded Games on Play Store in 2024
 
Deploy with confidence: VMware Cloud Foundation 5.1 on next gen Dell PowerEdg...
Deploy with confidence: VMware Cloud Foundation 5.1 on next gen Dell PowerEdg...Deploy with confidence: VMware Cloud Foundation 5.1 on next gen Dell PowerEdg...
Deploy with confidence: VMware Cloud Foundation 5.1 on next gen Dell PowerEdg...
 
Strategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a FresherStrategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a Fresher
 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
 
TrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
TrustArc Webinar - Stay Ahead of US State Data Privacy Law DevelopmentsTrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
TrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)
 
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
 
Real Time Object Detection Using Open CV
Real Time Object Detection Using Open CVReal Time Object Detection Using Open CV
Real Time Object Detection Using Open CV
 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
 
MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024
 

Borthakur hadoop univ-research

  • 1. Apache Hadoop and Hive Dhruba Borthakur Apache Hadoop Developer Facebook Data Infrastructure dhruba@apache.org, dhruba@facebook.com Condor Week, April 22, 2009
  • 2. Outline • Architecture of Hadoop Distributed File System • Hadoop usage at Facebook • Ideas for Hadoop related research
  • 3. Who Am I? • Hadoop Developer – Core contributor since Hadoop’s infancy – Project Lead for Hadoop Distributed File System • Facebook (Hadoop, Hive, Scribe) • Yahoo! (Hadoop in Yahoo Search) • Veritas (San Point Direct, Veritas File System) • IBM Transarc (Andrew File System) • UW Computer Science Alumni (Condor Project)
  • 4. Hadoop, Why? • Need to process Multi Petabyte Datasets • Expensive to build reliability in each application. • Nodes fail every day – Failure is expected, rather than exceptional. – The number of nodes in a cluster is not constant. • Need common infrastructure – Efficient, reliable, Open Source Apache License • The above goals are same as Condor, but – Workloads are IO bound and not CPU bound
  • 5. Hive, Why? • Need a Multi Petabyte Warehouse • Files are insufficient data abstractions – Need tables, schemas, partitions, indices • SQL is highly popular • Need for an open data format – RDBMS have a closed data format – flexible schema • Hive is a Hadoop subproject!
  • 6. Hadoop & Hive History • Dec 2004 – Google GFS paper published • July 2005 – Nutch uses MapReduce • Feb 2006 – Becomes Lucene subproject • Apr 2007 – Yahoo! on 1000-node cluster • Jan 2008 – An Apache Top Level Project • Jul 2008 – A 4000 node test cluster • Sept 2008 – Hive becomes a Hadoop subproject
  • 7. Who uses Hadoop? • Amazon/A9 • Facebook • Google • IBM • Joost • Last.fm • New York Times • PowerSet • Veoh • Yahoo!
  • 8. Commodity Hardware Typically in 2 level architecture – Nodes are commodity PCs – 30-40 nodes/rack – Uplink from rack is 3-4 gigabit – Rack-internal is 1 gigabit
  • 9. Goals of HDFS • Very Large Distributed File System – 10K nodes, 100 million files, 10 PB • Assumes Commodity Hardware – Files are replicated to handle hardware failure – Detect failures and recovers from them • Optimized for Batch Processing – Data locations exposed so that computations can move to where data resides – Provides very high aggregate bandwidth • User Space, runs on heterogeneous OS
  • 10. HDFS Architecture Cluster Membership e NameNode am ilen 1. f es Nod Id, Data 2. Blck Secondary NameNode o Client 3.Read d ata Cluster Membership NameNode : Maps a file to a file-id and list of MapNodes DataNodes DataNode : Maps a block-id to a physical location on disk SecondaryNameNode: Periodic merge of Transaction log
  • 11. Distributed File System • Single Namespace for entire cluster • Data Coherency – Write-once-read-many access model – Client can only append to existing files • Files are broken up into blocks – Typically 128 MB block size – Each block replicated on multiple DataNodes • Intelligent Client – Client can find location of blocks – Client accesses data directly from DataNode
  • 12.
  • 13. NameNode Metadata • Meta-data in Memory – The entire metadata is in main memory – No demand paging of meta-data • Types of Metadata – List of files – List of Blocks for each file – List of DataNodes for each block – File attributes, e.g creation time, replication factor • A Transaction Log – Records file creations, file deletions. etc
  • 14. DataNode • A Block Server – Stores data in the local file system (e.g. ext3) – Stores meta-data of a block (e.g. CRC) – Serves data and meta-data to Clients • Block Report – Periodically sends a report of all existing blocks to the NameNode • Facilitates Pipelining of Data – Forwards data to other specified DataNodes
  • 15. Block Placement • Current Strategy -- One replica on local node -- Second replica on a remote rack -- Third replica on same remote rack -- Additional replicas are randomly placed • Clients read from nearest replica • Would like to make this policy pluggable
  • 16. Data Correctness • Use Checksums to validate data – Use CRC32 • File Creation – Client computes checksum per 512 byte – DataNode stores the checksum • File access – Client retrieves the data and checksum from DataNode – If Validation fails, Client tries other replicas
  • 17. NameNode Failure • A single point of failure • Transaction Log stored in multiple directories – A directory on the local file system – A directory on a remote file system (NFS/CIFS) • Need to develop a real HA solution
  • 18. Data Pipelining • Client retrieves a list of DataNodes on which to place replicas of a block • Client writes block to the first DataNode • The first DataNode forwards the data to the next DataNode in the Pipeline • When all replicas are written, the Client moves on to write the next block in file
  • 19. Rebalancer • Goal: % disk full on DataNodes should be similar – Usually run when new DataNodes are added – Cluster is online when Rebalancer is active – Rebalancer is throttled to avoid network congestion – Command line tool
  • 20. Hadoop Map/Reduce • The Map-Reduce programming model – Framework for distributed processing of large data sets – Pluggable user code runs in generic framework • Common design pattern in data processing cat * | grep | sort | unique -c | cat > file input | map | shuffle | reduce | output • Natural for: – Log processing – Web search indexing – Ad-hoc queries
  • 21. Hadoop at Facebook • Production cluster – 4800 cores, 600 machines, 16GB per machine – April 2009 – 8000 cores, 1000 machines, 32 GB per machine – July 2009 – 4 SATA disks of 1 TB each per machine – 2 level network hierarchy, 40 machines per rack – Total cluster size is 2 PB, projected to be 12 PB in Q3 2009 • Test cluster • 800 cores, 16GB each
  • 22. Data Flow Web Servers Scribe Servers Network Storage Oracle RAC Hadoop Cluster MySQL
  • 23. Hadoop and Hive Usage • Statistics : – 15 TB uncompressed data ingested per day – 55TB of compressed data scanned per day – 3200+ jobs on production cluster per day – 80M compute minutes per day • Barrier to entry is reduced: – 80+ engineers have run jobs on Hadoop platform – Analysts (non-engineers) starting to use Hadoop through Hive
  • 25. Condor and HDFS • Run Condor jobs on Hadoop File System – Create HDFS using local disk on condor nodes – Use HDFS API to find data location – Place computation close to data location • Support map-reduce data abstraction model
  • 26. Power Management • Power Management – Major operating expense – Power down CPU’s when idle – Block placement based on access pattern • Move cold data to disks that need less power • Condor Green
  • 27. Benchmarks • Design Quantitative Benchmarks – Measure Hadoop’s fault tolerance – Measure Hive’s schema flexibility • Compare above benchmark results – with RDBMS – with other grid computing engines
  • 28. Job Sheduling • Current state of affairs – FIFO and Fair Share scheduler – Checkpointing and parallelism tied together • Topics for Research – Cycle scavenging scheduler – Separate checkpointing and parallelism – Use resource matchmaking to support heterogeneous Hadoop compute clusters – Scheduler and API for MPI workload
  • 29. Commodity Networks • Machines and software are commodity • Networking components are not – High-end costly switches needed – Hadoop assumes hierarchical topology • Design new topology based on commodity hardware
  • 30. More Ideas for Research • Hadoop Log Analysis – Failure prediction and root cause analysis • Hadoop Data Rebalancing – Based on access patterns and load • Best use of flash memory?
  • 31. Summary • Lots of synergy between Hadoop and Condor • Let’s get the best of both worlds
  • 32. Useful Links • HDFS Design: – http://hadoop.apache.org/core/docs/current/hdfs_design.html • Hadoop API: – http://hadoop.apache.org/core/docs/current/api/ • Hive: – http://hadoop.apache.org/hive/

Hinweis der Redaktion

  1. This is the architecture of our backend data warehouing system. This system provides important information on the usage of our website, including but not limited to the number page views of each page, the number of active users in each country, etc. We generate 3TB of compressed log data every day. All these data are stored and processed by the hadoop cluster which consists of over 600 machines. The summary of the log data is then copied to Oracle and MySQL databases, to make sure it is easy for people to access.