Today we see the pattern of Baidu Aladdin. In my opinion, it is a useful supplement to the previous crawling, analysis, and search modes.
1. Crawling is time-consuming and labor-consuming. It is nothing more than getting the original webpage file, but the search engine's goal is to get an inverted index table. This is a helpless irony.
2. The number of existing web pages and the update speed cannot be achieved through diligence. Such as onlinedown websites, stock quotations, weather forecasts, etc. It makes no sense to capture information with a short update period. Just like the stock-market dashboard, it mainly involves real-time display. for search engines, that is, real-time search is required.
3. Partners, websites with high credibility, such as Sina and People's Network. Some exceptions of search engines can be ignored, such as cheating and keyword stacking.
4. The most important thing is whether there are several "dark networks", that is, the web pages read from the DB and dynamically displayed, which cannot be retrieved by the search engine. I think the more meaningful Aladdin's meaning is here. If it is not a website with a very high access volume, of course we do not need to do some DB-static processes. Direct submission allows you to search for all of our contents in depth and completely. That is undoubtedly a good thing.
5. New Question 1: Undoubtedly, such a submission must be reviewed. It may be said that many common websites may not be able to pass. There is also a technical comparison between the webpage display page and the submitted index. Determine consistency. Technically speaking, it is quite difficult. But then again, since it is review, if you find that it is seriously invalid, you should have no opinion if you have not discussed it.
6. New Question 2: Are the summaries of images, videos, or other files manually true? It is very difficult to verify this information.