分享

有人用过 mongodb hadoop connector 吗?能指点一下吗

pig2 发表于 2013-11-16 03:38:08 [显示全部楼层] 只看大图 回帖奖励 阅读模式 关闭右栏 1 5636

已有(1)人评论

跳转到指定楼层
admin 发表于 2013-11-16 03:40:16
MongoDB+Hadoop Connector
CURRENT RELEASE: 1.0.0

The Mongo+Hadoop Connector (for brevitys sake, we’ll refer to it as mongo-hadoop in this documentation) is a series of plugins for the Apache Hadoop Platform to allow connectivity to MongoDB. This connectivity takes the form of allowing both reading MongoDB data into Hadoop (for use in MapReduce jobs as well as other components of the Hadoop ecosystem), as well as writing the results of Hadoop jobs out to MongoDB. A forthcoming release will also allow for reading and writing static BSON files (ala mongodump / mongorestore) to allow offline batching; commonly, users find this to be a beneficial feature to run analytics against backup data.

At this time, we support the “core” Hadoop APIs (now known as Hadoop Common), in the form of mongo-hadoop-core. There is additionally support for other pieces of the Hadoop Ecosystem, including Pig for ETL and Streaming for running Mongo+Hadoop jobs with Python (future releases will support additional scripting languages such as Ruby). Although it is not dependent upon Hadoop, we also provide a connector for the Flume distributed logging system.
具体可以参考下面的官网地址:

http://api.mongodb.org/hadoop/MongoDB%2BHadoop+Connector.html
https://github.com/mongodb/mongo-hadoop
回复

使用道具 举报

您需要登录后才可以回帖 登录 | 立即注册

本版积分规则

关闭

推荐上一条 /2 下一条