zoukankan      html  css  js  c++  java
  • Hadoop ecosystem

    1. How did it all start- huge data on the web!
    2. Nutch built to crawl this web data
    3. Huge data had to saved- HDFS was born!
    4. How to use this data?
    5. Map reduce framework built for coding and running analytics – java, any language-streaming/pipes
    6. How to get in unstructured data – Web logs, Click streams, Apache logs, Server logs  – fuse,webdav, chukwa, flume, Scribe
    7. Hiho and sqoop for loading data into HDFS – RDBMS can join the Hadoop band wagon!
    8. High level interfaces required over low level map reduce programming– Pig, Hive, Jaql
    9. BI tools with advanced UI reporting- drilldown etc- Intellicus 
    10. Workflow tools over Map-Reduce processes and High level languages
    11. Monitor and manage hadoop, run jobs/hive, view HDFS – high level view- Hue, karmasphere, eclipse plugin, cacti, ganglia
    12. Support frameworks- Avro (Serialization), Zookeeper (Coordination)
    13. More High level interfaces/uses- Mahout, Elastic map Reduce
    14. OLTP- also possible – Hbase
  • 相关阅读:
    go语言goroutine
    go语言接口
    go语言的map
    go语言切片
    go语言数组
    django的信号机制
    python redis 实现简单的消息订阅
    scrapy中使用selenium来爬取页面
    尝试用tornado部署django
    控制台输出太多导致项目启动过慢
  • 原文地址:https://www.cnblogs.com/AloneSword/p/3789835.html
Copyright © 2011-2022 走看看