ORIGINAL RESEARCH Efficiency measures for ranked pages by Markov Chain Principle Swati Jain 1 • Mukesh Rawat 1 Received: 28 May 2019 / Accepted: 11 October 2020 Ó Bharati Vidyapeeth’s Institute of Computer Applications and Management 2020 Abstract More often, the user likes to visit the web doc- uments that appear in few top excellent responses to the list of links provided by the search engine and these results are the most likely accurate results to the search query. The Information Retrieval by Search Engine helps in retrieving the most relevant pages for query. In this paper, we propose an ideal technique for link analysis by taking web graph structure and we focus around the ranking of such links. The relevancy of the links is evaluated by using Markov Chain Principle and also query keyword occurrence is given a weight-age to the overall ranking of the links. The term proximity and discounted cumulative gain are used to simulate results and the scores show that the proposed methodology efficiently enhances the ranking of the web pages. Keywords Markov chain Ranking Search engine Search query Transition probability Webgraph 1 Introduction A Search Engine is an online apparatus which alludes to a gigantic database of web assets that empowers clients to find data on the World Wide Web. To search for information, clients can scan data by sending catchphrases or keywords as input query, and after that search engine gives response within few sub-second times to extensive amount of user queries presented each day all around the globe. The results of searching provided by search engine consist of the sequence of responses which is termed as Search Engine Result Pages (SERPs). The results might be a mix of site pages, images and various types of records. 1.1 Components of Search Engine Mainly, a search engine has three fundamental components which are explained as: 1.1.1 Web Crawler Theprimary responsibility of a crawler is to identify and acquire documents to be processed by search engine. A web crawler is intended to pursue the connections on website pages to find new pages and download those pages. It is basically a software segment that visits web to accumulate information. Another synonym for it is spider or bots. 1.1.2 Database A database stores whole of the web information. It is a repository which consists of huge web resources such as web documents, audios, pictures, videos and various other files. 1.1.3 Search Interfaces Search Interface acts as a bridge between the end-user and database repository. It assists end-user to look within the database to find the relevant information. & Swati Jain vce.swatijain@gmail.com Mukesh Rawat mukesh.rawat@miet.ac.in 1 Department of Computer Science and Engineering, Meerut Institute of Engineering and Technology, Dr. A P J Abdul Kalam Technical University, Meerut 250005, Uttar Pradesh, India 123 Int. j. inf. tecnol. https://doi.org/10.1007/s41870-020-00549-y