There are currently more than a billion pages of information on every topic imaginable on the Internet. The question is: how can you find what you want? Computer algorithms can be written to search the Internet, but most are impractical because they must sacrifice precision for coverage. However, some engines have found interesting ways to quickly provide high-quality information. Page value ranking, topic-specific searches, and Meta search engines are three of the most popular search engines because they work smarter, not harder. While no commercial search engine will make its algorithm public, the basic structure can be deduced by testing the results. The reason for this is because there would be a thousand imitated sites, which means little to no profit for the developers. The most primitive of the searches is the sequential search, which examines each item in the list one at a time. Yet the vastness of the web immediately excludes this possibility. While sequential may return the best results, you will most likely never see any results due to the web's inflammatory growth rate. Even the fastest computers would take a long time, and all kinds of new pages will be created in that time. Some of the older "spiders" like Alta Vista are designed to literally randomly wander the web using links to other pages. This is achieved with high-speed servers with 300 connections open at the same time. These web “spiders” are content-based, meaning they actually read and classify the HTML on each page. A flaw of this is the verbal disagreement problem where you have a particular word that can describe two different concepts. Type a few words into the query and you'll be lucky if you can find anything related to what you're looking for. Query words can occur anywhere on a page and are likely to be taken out of context. Content-based searches can also be easily manipulated. Some tactics are very deceptive, for example "...some automotive websites have been reduced to writing 'Buy this car' dozens of times in hidden characters...a subliminal version of the AAAA car listing in the Yellow Pages"(1 ). The truth is, you would never know if a site was doing this unless you looked at the code, and most consumers don't look at the code. A less subtle tactic is to pay your way to the top. For example, the GoTo engine accepts payments from those who want to b...... half paper...... the metasearch engine can achieve several benefits:1 It will present users with a more sophisticated interface...2 Make the most accurate translation3 Obtain more complete and precise results4 Improve the selection of sources and the execution of priority decisions” (3).Once again the idea of optimizing the Internet through intelligent software emerges. It's just a matter of designing a certain algorithm that doesn't forget what it has learned. Most people didn't predict the enormous growth of the Internet in the 1990s. Computer algorithms have moved from small government programs to every personal computer in the world. You start with solving the most basic problems and end with solving the most complex problems. This obviously happens through a database that grows almost exponentially. Plain and simple, the Internet contains a lot of information. A crawler works 24 hours a day digging everywhere. The search engine extracts the parts that people want and delivers them to the Meta search engine. The Meta search engine discriminates further until you get exactly what you are looking for.
tags