Hadoop and Map Reduce
Rationale
The main objective of this journal is to build a parallelized Vertical Search Engine on Apache Hadoop cluster by taking seed URLs of computer domain mining of Wikipedia. The extracted web pages are a creep and parsed using Apache Nutch crawler and it is stored in Apache HBase. The stored web pages are used for linguistic processing to remove the stop words and stemming the performance. It also uses the ranking algorithm to rank the results. It focuses on obtaining the most relevant results in a short time using distributed processing.
Methods
In search engines queries are mostly too short or indistinguishable and produce too many results which include irrelevant data. The obtained results are not browsed by the individual since it has a large number of results. The use of mobile phones and other ubiquitous devices has small screens has limited use. To produce relevant search results in less time a vertical search engine by parallelizing the vertical search engine using Hadoop and map-reduce a comparison is made with a non-parallelized vertical search engine. The method used is the distributed database over the Hadoop cluster and perform the search using the map-reduce. Map-reduce has been divided into three categories based on the existing deficiencies (Khezr and Navimipour, 2017) based on the improvement goals as a native variant, apache variant and third-party extensions. Most of the investigation is based on the Hadoop platform to enhance the database process and schedules in Hadoop map and reduce. The applications of Map-reduce is a promising technology like telecom, manufacturing, government organization, medical, etc. (Hashem IAT et al, 2016).
Tools
It combines the different advantages of vertical search engines like relevant findings with a computing capacity of parallel processing along with the Hadoop cluster and Map-reduce programming. The components used are cluster, database, URL dB, Crawler, Linguistic processing, Inverted Index and Ranking algorithm. Cluster formation is done using Apache Hadoop to provide the advantage of the distributed file system (HDFS). The database used in HBase is distributed non-relational database to facilitate the fault-tolerant. (Joshi and Mulay, 2018). Wikipedia is used as an URL dB as a seed to the crawl to bring the Robust and Scalable. Linguistic processing is used to remove the stop words like pronouns, conjunctions, prepositions, punctuations, symbols etc. and filter them. Porter-stemmer algorithm is used for map-reduce programming. URL indexing is done and the pages are ranked using the ranking algorithm.
Findings
The performance of the vertical search is reduced when the numbers of nodes in the cluster are increased due to the communication latency. The number of nodes and URLs are kept constant and the keywords are changed and the results produced are exponentially increased in time to fetch the result and the performance was decreased.
Conclusion
The aim of this article is to provide an alternative for search engines that provides relevant data and reduce the overall time spend in the search process. Parallelization provides a good performance benefit but has many limits in future it will be used to replace the search engine by increasing efficiency.
References
Hashem, I.A.T., Anuar, N.B., Gani, A., Yaqoob, I., Xia, F. and Khan, S.U. ,2016. MapReduce: Review and open challenges. Scientometrics, 109(1), pp.389–422.
Joshi, R. and Mulay, P. 2018. Mind Map based Survey of Conventional and Recent Clustering Algorithms: Learning’s for Development of Parallel and Distributed Clustering Algorithms. International Journal of Computer Applications, 181(4), pp.14–21.
Khezr, S.N. and Navimipour, N.J. 2017. MapReduce and Its Applications, Challenges, and Architecture: a Comprehensive Review and Directions for Future Research. Journal of Grid Computing, 15(3), pp.295–321.