Introducing the ONLY search engine optimization software product that has been 100% PROVEN to dramatically increase your rankings in Google, Yahoo, and Bing. Get it now!
This is a very general question but a very nice topic! Definitely upvoted :) However I am not satisfied with the answers provided so far, so I decided to write a rather lengthy answer on this.
This is a very general question but a very nice topic! Definitely upvoted :) However I am not satisfied with the answers provided so far, so I decided to write a rather lengthy answer on this. The reason I am not satisfied is that the answers are basically all true (I especially like the answer of kovshenin (+1), which is very graph theory related...), but the all are either too specific on certain factors or too general.It's like asking how to bake a cake and you get the following answers: You make a cake and you put it in the oven.
You definitely need sugar in it! What is a cake? The cake is a lie!
You won't be satisfied because you wan't to know what makes a good cake. And of course there are a lot or recipies. Of course Google is the most important player, but, depending on the use case, a search engine might include very different factors or weight them differently.
For example a search engine for discovering new independent music artists may put a malus on artists websites with a lots of external links in. A mainstream search engine will probably do the exact opposite to provide you with "relevant results". There are (as already said) over 200 factors that are published by Google.
So webmasters know how to optimize their websites. There are very likely many many more that the public is not aware of (in Google's case). But in the very borad and abstract term SEO optimazation you can generally break the important ones apart into two groups: How well does the answer match the question?
Or: How well does the pages content match the search terms? How popular/good is the answer? Or: What's the pagerank?
In both cases the important thing is that I am not talking about whole websites or domains, I am talking about single pages with a unique URL. It's also important that pagerank doesn't represent all factors, only the ones that Google categorizes as Popularity. And by good I mean other factors that just have nothing to do with popularity.
In case of Google the official statement is that they want to give relevant results to the user. Meaning that all algorithms will be optimized towards what the user wants.So after this long introduction (glad you are still with me...) I will give you a list of factors that I consider to be very important (at the moment): Category 1 (how good does the answer match the question? You will notice that a lot comes down to the structure of the document!
The page primarily deals with the exact question. Meaning: the question words appear in the pages title text or in heading paragraphs paragraphs. The same goes for the position of theese keywords.
The earlier in the page the better. Repeated often as well (if not too much which goes under the name of keywords stuffing). The whole website deals with the topic (keywords appear in the domain/subdomain) The words are an important topic in this page (internal links anchor texts jump to positions of the keyword or anchor texts / link texts contain the keyword).
The same goes if external links use the keywords in link text to link to this page Category 2 (how important/popular is the page?) You will notice that not all factors point towards this exact goal. Some are included (especially by Google) just to give pages a boost, that... well... that just deserved/earned it. Content is king!
The existence of unique content that can't be found or only very little in the rest of the web gives a boost. This is mostly measured by unordered combinations of words on a website that are generally used very little (important words). But there are much more sophisticated methods as well.
Recency - newer is better storical change (how often the page has updated in the past. Changing is good. ) External link popularity (how many links in?) If a page links another page the link is worth more if the page itself has a high pagerank.
External link diversity basically links from different root domains, but other factors play a role too. Factors like even how seperated are the webservers of linking sites geographically (according to their ip address). Trust Rank For example if big, trusted, established sites with redactional content link to you, you get a trust rank.
That's why a link from The New York Times is worth much more than some strange new website, even if it's PageRank is higher! Domain trust Your whole website gives a boost to your content if your domain is trusted. Well different factors count here.
Of course links from trusted sties to your domain, but it will even do good if you are in the same datacenter as important websites. Topic specific links in. If websites that can be resolved to a topic link to you and the query can be resolved to this topic as well, it's good.
Distribution of links in over time. If you earned a lot of links in in a short period of time, this will do you good at this time and the near future afterwards. But not so good later in time.
If you slow and steady earn links it will do you good for content that is "timeless". Links from restrited domains A link from a . Gov domain is worth a lot.
User click behaviour Whats the clickrate of your search result? Time spent on site Google analytics tracking, etc.It's also tracked if the user clicks back or clicks another result after opening yours. Collected user data Votes, rating, etc. , references in Gmail, etc.Now I will introduce a third category, and one or two points from above would go into this category, but I haven't thought of that... The category is: ** How important/good is your website in general ** All your pages will be ranked up a bit depending on the quality of your websites Factors include: Good site architecture (easy to navgite, structured.
Sitemaps, etc...) How established (long existing domains are worth more). Hoster information (what other websites are hosted near you? Search frequency of your exact name.
Last, but not least, I want to say that a lot of these theese factors can be enriched by semantic technology and new ones can be introduced. For example someone may search for Titanic and you have a website about icebergs ... that can be set into correlation which may be reflected. Newly introduced semantic identifiers.
For example OWL tags may have a huge impact in the future. For example a blog about the movie Titanic could put a sign on this page that it's the same content as on the Wikipedia article about the same movie. This kind of linking is currently under heavy development and establishment and nobody knows how it will be used.
Maybe duplicate content is filtered, and only the most important of same content is displayed? Or maybe the other way round? That you get presented a lot of pages that match your query.
Even if they don't contain your keywords? Google even applies factors in different relevance depending on the topic of your search query!
Thank you for taking the time to create such a wealthy answer but please read my question thoroughly, as im not talking about SEO, this has nothing to do with search engines apart from I would like to adopt a some techniques for a different usage – RobertPitt Oct 27 '10 at 22:21 @RoberPitt - This is to find relevant content. Google implements all of this to find relevant content (as you ask in the very first line). I think he answered your question quite well.
Just more detail than you may have wanted. Besides SEO, is nothing more than a web designer presenting relevant content to Google to increase page rank. Pretty links, h1 tags, page linking, etc.Are all methods Google searches to find relevant content.
– Jason Oct 28 '10 at 3:02.
Tricky, but I'll take a stab: An image (If applicable) The first image on the page the image with a name that includes the letters "logo" the image that renders closest to the top-left (or top-right) the image that appears most often on other pages of the site an image smaller than some maximum dimensions A.
1 +1 for actually giving answers that are somewhat relative to my question, What are the reason for the first h1 and first p tag. – RobertPitt Oct 22 '10 at 11:24 3 thx. The first 'h1' should be the biggest and most important heading of the page; if it contains the search term then the page is more likely to be relevant.
Similarly for the 'p'; the first paragraph on the page is more likely to contain words reflective of the rest of the page, like an introduction, or summary of what follows; so if it mentions the search term once or twice then the whole page is perhaps relevant. – John Mee Oct 22 '10 at 12:00 I think you should edit your answer and include what you wrote in your comment. :) – musicfreak Oct 25 '10 at 4:39 Great answers, g'luck for the bounty ;) – kovshenin Oct 28 '10 at 7:09.
I don't work at Google but around a year ago I read they had over 200 factors for ranking their search results. Of course the top ranking would be relevance, so your question is quite interesting in that sense. What is relevance and how do you calculate it?
There are several algorithms and I bet Google have their own, but ones I'm aware of are Pearson Correlation and Euclidean Distance. A good book I'd suggest on this topic (not necessarily search engines) is Programming Collective Intelligence by Toby Segaran (O'Reilly). A few samples from the book show how to fetch data from third-party websites via APIs or screen-scraping, and finding similar entries, which is quite nice.
Anyways, back to Google. Other relevance techniques are of course full-text searching and you may want to get a good book on MySQL or Sphinx for that matter. Suggested by @Chaoley was TSEP which is also quite interesting.
But really, I know people from a Russian search engine called Yandex here, and everything they do is under NDA, so I guess you can get close, but you cannot get perfect, unless you work at Google ;) Cheers.
In laymen's terms: I'm not talking about ranking, I am talking about scraping a page and finding the best bits – RobertPitt Oct 21 '10 at 11:16 In that case it's information extraction and data mining I guess, not relevancy – kovshenin Oct 21 '10 at 11:25 What I want to know how relevant paragraph A is to paragraph B by using keywords found in links,meta,title, and a title provided when the link is submitted to me, +1 for the book, very nice, title looks very promising – RobertPitt Oct 21 '10 at 11:57 3 So first step is to find the most important piece of information and extract it from both websites, next step is to calculate their relevance. Once again, Toby's book has a good sample of fetching data from a bunch of RSS feeds and grouping relevant sources, which is nice, but simpler because RSS is short and supports tags and categories. But you will have to look somewhere else for extraction techniques.
I suggest starting from "Mining the Social Web" by Matthew Russell. (yeah, I'm crazy about books) – kovshenin Oct 21 '10 at 15:12.
Actually answering your question (and not just generally about search engines): I believe going bit like Instapaper does would be the best option. Logic behind instapaper (I didn't create it so I certainly don't know inner-workings, but it's pretty easy to predict how it works): Find biggest bunch of text in text-like elements (relying on paragraph tags, while very elegant, won't work with those crappy sites that use div's instead of p's). Basically, you need to find good balance between block elements (divs, ps, etc.) and amount of text.
Come up with some threshold: if X number of words stays undivided by markup, that text belongs to main body text. Then expand to siblings keeping the text / markup threshold of some sort. Once you do the most difficult part — find what text belongs to actual article — it becomes pretty easy.
You can find first image around that text and use it as you thumbnail. This way you will avoid ads, because they will not be that close to body text markup-wise. Finally, coming up with keywords is the fun part.
You can do tons of things: order words by frequency, remove noise (ands, ors and so on) and you have something nice. Mix that with "prominent short text element above detected body text area" (i.e. Your article's heading), page title, meta and you have something pretty tasty.
All these ideas, if implemented properly, will be very bullet-proof, because they do not rely on semantic markup — by making your code complex you ensure even very sloppy-coded websites will be detected properly. Of course, it comes with downside of poor performance, but I guess it shouldn't be that poor. Tip: for large-scale websites, to which people link very often, you can set HTML element that contains the body text (that I was describing on point #1) manually.
This will ensure correctness and speed things up. Hope this helps a bit.
Most search engines look for the title and meta description in the head of the document, then heading one and text content in the body. Image alt tags and link titles are also considered. Last I read Yahoo was using the meta keyword tag but most don't.
You might want to download the open source files from The Search Engine Project (TSEP) on Sourceforge https://sourceforge. Net/projects/tsep/ and have a look at how they do it.
– RobertPitt Oct 19 '10 at 14:56 No way to turns CSS into a DOM representation that I know of, you'd need to use the file functions for that. If you're searching a single site with consistent code structure the entire exercise is easy, if you want to search multiple sites it's a lot harder. Another link, check out webmasterworld.
Com/perl/3460556. Htm for some more ideas. – Chaoley Oct 20 '10 at 7:58 5 -1 The big search engines (i.e.
Google and no one else :D) do not use meta description and meta keywords for site ranking... – NikiC Oct 21 '10 at 10:17 1 To be fair, he didn't say that they use meta data to rank the page, but to extract the content which is almost correct! – Mouhannad Oct 21 '10 at 10:30 2 I don't think we're talking about ranking here, the question was about finding relevant content. – kovshenin Oct 21 '10 at 11:03.
I'd just grab the first 'paragraph' of text. The way most people write stories/problems/whatever is that they first state the most important thing, and then elaborate. If you look at any random text and you can see it makes sense most of the time.
For example, you do it yourself in your original question. If you take the first three sentences of your original question, you have a pretty good summary of what you are trying to do. And, I just did it myself too: the gist of my comment is summarized in the first paragraph.
The rest is just examples and elaborations. If you're not convinced, take a look at a few recent articles I semi-randomly picked from Google News. Ok, that last one was not semi-random, I admit ;) Anyway, I think that this is a really simple approach that works most of the time.
You can always look at meta-descriptions, titles and keywords, but if they aren't there, this might be an option. Hope this helps.
There are lots of highly sophisticated algorithms for extracting the relevant content from a tag soup. If you're looking to build something usable your self, you could take a look at the source code for readability and port it over to php. I did something similar recently (Can't share the code, unfortunately).
The basic logic of readability is to find all block level tags and count the length of text in them, not counting children. Then each parent node is awarded a fragment (half) of the weight of each of its children. This is used to fund the largest block level tag that has the largest amount of plain text.
From here, the content is further cleaned up. It's not bullet proof by any means, but it works well in the majority of cases.
I would consider these building the code Check for synonyms and acronyms applying OCR on images to search as text(Abby Fine Reader and Recostar are nice, Tesseract is free and fine(no so fine as fine reader :) ) weight Fonts as well(size, boldness, underline, color) weight content depending on its place on page(like contents on upper side of page is more relevant) Also: An optinal text asked from the webmaster to define the page You can also check if you can find anything useful at Google search API: code.google.com/intl/tr/apis/ajaxsearch.
Google for 'web crawlers, robots, Spiders, and Intelligent Agents', might try them separately as well to get individual results. Web Crawler User-Agents Bots Data/Screen Scraping What I think you're looking for is Screen Scraping (with DOM) which Stack has a ton of Q&A on.
I fully understand what the entities there are to a search engine, I am specifically asking about the algorithm used in finding relevant content without specific selectors. – RobertPitt Oct 19 '10 at 13:50 @Robert if you find the algorithm Google or other search engines hold in secret, please start you own search engine company as they do not share this information as it would be considered a "Trade Secret". – Phill Pafford Oct 19 '10 at 14:12 Im not looking for Googles cluster bots source code here, im looking for community based methods on having programmatic way of finding relevant data by following a trend of content layout, read my examples please, and this is not for a search engine, this is for a content sharing network – RobertPitt Oct 19 '10 at 14:15 1 I think you should re-title you question as it reads "How does Search Engines find relevant content?
" to something like "Algorithm/Logic used for Content Sharing Network" just my 2 cents – Phill Pafford Oct 19 '10 at 14:19 Well I specifically wish to know about the search engine side of it because thats exactly what is needed, Content Sharing Network is basically a search engine as its a network that shares content. – RobertPitt Oct 19 '10 at 14:29.
Google also uses a system called Page Rank, where it examines how many links to a site there are. Let's say that you're looking for a C++ tutorial, and you search Google for one. You find one as the top result, an it's a great tutorial.
Google knows this because it searched through its cache of the web and saw that everyone was linking to this tutorial, while ranting how good it was. Google deceides that it's a good tutorial, and puts it as the top result. It actually does that as it caches everything, giving each page a Page Rank, as said before, based on links to it.
Hope this helps!
The question wasn't about ranking pages but rather about finding relevance to a search term. – musicfreak Oct 25 '10 at 4:40 Incorrect, Not finding relevance to a search term but finding relevant content within any site that our system finds, basically create and engine that finds content as the visual eye would – RobertPitt Oct 27 '10 at 22:12.
To answer one of your questions, I am reading the following book right now, and I recommend it: Google's PageRank and Beyond, by Amy Langville and Carl Meyer. Mildly mathematical. Uses some linear algebra in a graph theoretic context, eigenanalysis, Markov models, etc.I enjoyed the parts that talk about iterative methods for solving linear equations.
I had no idea Google employed these iterative methods. Short book, just 200 pages. Contains "asides" that diverge from the main flow of the text, plus historical perspective.
Also points to other recent ranking systems.
This is not what my question is asking! – RobertPitt Oct 25 '10 at 7:59 @RobertPitt: It sure looks like it is, to me. "How does Google find relevant content when its parsing the web?"
This is how Google does it. "Are there any books..." This is a book. – Merlyn Morgan-Graham Oct 28 '10 at 6:07 Thank you, Merlyn Morgan-Graham.Re-reading the question, I admit my answer may have missed the point.
However, RobertPitt, (1) is it necessary to chide people volunteering their knowledge to help you solve your problem, and (2) if five answerers misunderstood your question in the same way, could it be possible that, perhaps, just perhaps, the question itself could be improved? – Steve Tjoa Oct 28 '10 at 13:20.
There are some good answers on here, but it sounds like they don't answer your question. Perhaps this one will. What your looking for is called Information Retrieval It usually uses the Bag Of Words model Say you have two documents: DOCUMENT A Seize the time, Meribor.
Live now; make now always the most precious time. Now will never come again and this one DOCUMENT B Worf, it was what it was glorious and wonderful and all that, but it doesn't mean anything and you have a query, or something you want to find other relevant documents for QUERY aka DOCUMENT C precious wonderful life Anyways, how do you calculate the most "relevant" of the two documents? Here's how: tokenize each document (break into words, removing all non letters) lowercase everything remove stopwords (and, the etc) consider stemming (removing the suffix, see Porter or Snowball stemming algorithms) consider using n-grams You can count the word frequency, to get the "keywords".
Then, you make one column for each word, and calculate the word's importance to the document, with respect to its importance in all the documents. This is called the TF-IDF metric. Now you have this: Doc precious worf life... A 0.5 0.0 0.2 B 0.0 0.9 0.0 C 0.7 0.0 0.9 Then, you calculate the similarity between the documents, using the Cosine Similarity measure.
The document with the highest similarity to DOCUMENT C is the most relevant. Now, you seem to want to want to find the most similar paragraphs, so just call each paragraph a document, or consider using Sliding Windows over the document instead. You can see my video here.It uses a graphical Java tool, but explains the concepts: http://vancouverdata.blogspot.com/2010/11/text-analytics-with-rapidminer-part-4.html here is a decent IR book: http://nlp.stanford.edu/IR-book/pdf/irbookonlinereading.pdf.
I cant really gove you an answer,but what I can give you is a way to a solution, that is you have to find the anglde that you relate to or peaks your interest. A good paper is one that people get drawn into because it reaches them ln some way.As for me WW11 to me, I think of the holocaust and the effect it had on the survivors, their families and those who stood by and did nothing until it was too late.