欧美三区_成人在线免费观看视频_欧美极品少妇xxxxⅹ免费视频_a级毛片免费播放_鲁一鲁中文字幕久久_亚洲一级特黄

MongoDB: Hadoop Integerateion 1

系統 2680 0

Hadoop and MongoDB Use Cases

The following are some example deployments with MongoDB and Hadoop. The goal is to provide a high-level description of how MongoDB and Hadoop can fit together in a typical Big Data stack. In each of the following examples MongoDB is used as the “operational” real-time data store and Hadoop is used for offline batch data processing and analysis.

Batch Aggregation

In several scenarios the built-in aggregation functionality provided by MongoDB is sufficient for analyzing your data. However in certain cases, significantly more complex data aggregation may be necessary. This is where Hadoop can provide a powerful framework for complex analytics.

In this scenario data is pulled from MongoDB and processed within Hadoop via one or more MapReduce jobs. Data may also be brought in from additional sources within these MapReduce jobs to develop a multi-datasource solution. Output from these MapReduce jobs can then be written back to MongoDB for later querying and ad-hoc analysis. Applications built on top of MongoDB can now use the information from the batch analytics to present to the end user or to drive other downstream features.


MongoDB: Hadoop Integerateion 1

Data Warehouse

In a typical production scenario, your application’s data may live in multiple datastores, each with their own query language and functionality. To reduce complexity in these scenarios, Hadoop can be used as a data warehouse and act as a centralized repository for data from the various sources.

In this situation, you could have periodic MapReduce jobs that load data from MongoDB into Hadoop. This could be in the form of “daily” or “weekly” data loads pulled from MongoDB via MapReduce. Once the data from MongoDB is available from within Hadoop, and data from other sources are also available, the larger dataset data can be queried against. Data analysts now have the option of using either MapReduce or Pig to create jobs that query the larger datasets that incorporate data from MongoDB.


MongoDB: Hadoop Integerateion 1

ETL Data

MongoDB may be the operational datastore for your application but there may also be other datastores that are holding your organization’s data. In this scenario it is useful to be able to move data from one datastore to another, either from your application’s data to another database or vice versa. Moving the data is much more complex than simply piping it from one mechanism to another, which is where Hadoop can be used.

In this scenario, Map-Reduce jobs are used to extract, transform and load data from one store to another. Hadoop can act as a complex ETL mechanism to migrate data in various forms via one or more MapReduce jobs that pull the data from one store, apply multiple transformations (applying new data layouts or other aggregation) and loading the data to another store. This approach can be used to move data from or to MongoDB, depending on the desired result.


MongoDB: Hadoop Integerateion 1

MongoDB: Hadoop Integerateion 1

MongoDB Connector for Hadoop

The MongoDB Connector for Hadoop is a plugin for Hadoop that provides the ability to use MongoDB as an input source and/or an output destination.

The source code is available on github where you can find a more comprehensive readme .

If you have questions please email the mongodb-user Mailing List . For any issues please file a ticket in Jira .

Installation

The MongoDB Connector for Hadoop uses Gradle tool for compilation. To build, simply invoke the jar task as seen with the following command:

            ./gradlew jar

          

The MongoDB Connector for Hadoop supports a number of Hadoop releases. You can change the Hadoop version supported by passing the hadoop_version parameter to gradle. For instance, to build against Apache Hadoop 2.2 use the following command:

            ./gradlew jar -Phadoop_version
            
              =
            
            2.2

          

After building, you will need to place the “core” jar and the mongo-java-driver in the lib directory of each Hadoop server.

For more complete install instructions please see the install instructions in the readme

?? ?

?

References

http://docs.mongodb.org/ecosystem/tools/hadoop/

http://docs.mongodb.org/ecosystem/use-cases/hadoop/

http://www.mongodb.com/press/integration-hadoop-and-mongodb-big-data%E2%80%99s-two-most-popular-technologies-gets-significant

?

MongoDB: Hadoop Integerateion 1


更多文章、技術交流、商務合作、聯系博主

微信掃碼或搜索:z360901061

微信掃一掃加我為好友

QQ號聯系: 360901061

您的支持是博主寫作最大的動力,如果您喜歡我的文章,感覺我的文章對您有幫助,請用微信掃描下面二維碼支持博主2元、5元、10元、20元等您想捐的金額吧,狠狠點擊下面給點支持吧,站長非常感激您!手機微信長按不能支付解決辦法:請將微信支付二維碼保存到相冊,切換到微信,然后點擊微信右上角掃一掃功能,選擇支付二維碼完成支付。

【本文對您有幫助就好】

您的支持是博主寫作最大的動力,如果您喜歡我的文章,感覺我的文章對您有幫助,請用微信掃描上面二維碼支持博主2元、5元、10元、自定義金額等您想捐的金額吧,站長會非常 感謝您的哦!!!

發表我的評論
最新評論 總共0條評論
主站蜘蛛池模板: 欧美成年性h版影视中文字幕 | 日韩福利视频 | 欧美日韩一区二区中文字幕 | 中文字幕在线电影观看 | 日本高清免费zzzzzzzz | 久久久久国产精品 | 亚洲黑人在线观看 | 久久亚洲精品视频 | 亚洲网视频 | 成人免费毛片高清视频 | 嘿咻嘿咻免费区在线观看吃奶 | 免费一级视频在线观看 | 中文字幕日韩欧美一区二区三区 | 国产成人精品一区二区三区四区 | 精品亚洲永久免费精品 | 欧美精品在线观看 | 中文字幕亚洲欧美 | 国产精品字幕 | 99精品国产福利在线观看 | 日韩伦理一区二区三区 | 婷婷久久五月天 | www.色黄 | 欧美国产二区 | 一级片性| 久久精品视频5 | 91看片在线看 | 涩涩色中文综合亚洲 | 夫妻性生活交换 | 亚洲日本一区二区三区 | 日韩欧美色综合 | 亚洲国产精品人人爽夜夜爽 | 亚洲高清一区二区三区 | 欧美vide| 色狠狠色狠狠综合天天 | 亚洲国产精品视频一区 | 二级黄绝大片中国免费视频 | 草草国产成人免费视频 | 亚洲人成在线精品 | 四虎影视在线影院在线观看观看 | 日韩欧美一区二区三区久久 | 一区二区三区四区在线视频 |