Storage and Retrieval of Large RDF Graph Using Hadoop and MapReduce

CLOUD COMPUTING, PROCEEDINGS(2009)

引用 172|浏览0
暂无评分
摘要
Handling huge amount of data scalably is a matter of concern for a long time. Same is true for semantic web data. Current semantic web frameworks lack this ability. In this paper, we describe a framework that we built using Hadoop to store and retrieve large number of RDF triples. We describe our schema to store RDF data in Hadoop Distribute File System. We also present our algorithms to answer a SPARQL query. We make use of Hadoop's MapReduce framework to actually answer the queries. Our results reveal that we can store huge amount of semantic web data in Hadoop clusters built mostly by cheap commodity class hardware and still can answer queries fast enough. We conclude that ours is a scalable framework, able to handle large amount of RDF data efficiently.
更多
查看译文
关键词
large rdf graph,rdf data,rdf triple,mapreduce framework,hadoop distribute file system,huge amount,large amount,current semantic web framework,semantic web data,data scalably,hadoop cluster,distributed file system,semantic web
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要