Evolution of Google's algorithm since the origin
Developments noted or announced on the Google algorithm and its means of ranking pages.
We follow the evolution of the algorithm of Google and consequences for webmasters. The evolution of SERPs and interface are detailed in another article.
September 24, 2018 - Ranking images
Google announces a change in the image ranking algorithm, which is now closer to that of textual content. The authority of the site becomes a more important factor and is taken into account the "freshness" of the page (the fact that it has been updated or created recently) and the placement of the image in it.
Authority is related to relevance: if the subject of the image is also that of the site, this image is better positioned. The placement is considered more favorable if the image is in the center and at the beginning of the page.
March 8, 2017 - "Fred" Update
Many sites are penalized for not genuine content, especially if they have pages intended to promote other commercial sites. The filtering criteria are rather fuzzy even if the effect has been observed widely, and mark an attempt to downgrade not sites by content, like did Panda.
February 2017 - Phantom 5
Phantom updates are not officially recognized by the search engine. Their dates are estimated to be May 2013, May 2015, November 2015 and June 2016.
Still fuzzy criteria in trying to promote "quality", ie popularity with a minimum of content.
January 10, 2017 - For mobiles
This had been announced well in advance, sites that make popups and interstitial advertisements appear penalized.
September 23, 2016 - Penguin is integrated in the main algorithm
This was announced by Google's blog for webmasters. Penguin, is an algorithm launched in 2012 to penalize sites that want to increase the PageRank artificially by creating links on satellite sites to the site to promote. This also applies to links in footer or signature.
The algorithm was launched periodically, but now it operates in real time with each recrawl of the robot on the site. And it is also perfected, it affects the pages linked rather that a site in a whole.
July 1, 2016 - RankBrain, an increasing share of the queries
RankBrain is an element of the general algorithm and provides an answer when a search has no obvious answer among the pages of the stock index.
It is a machine learning algorithm on large amounts of data (Deep learning), so an algorithm that learns from user behavior to improve its performance.
February 5, 2016 - A whole deep learning search engine
Replacing Amit Singhal by John Giannandrea at head of Google is a symbol of the ongoing evolution of the search engine. While the second is a specialist in artificial intelligence and deep learning, learning by computer based on an amount of data and repeated experiments, the first was more traditionalist and relied more on programmers to make links ranking algorithms. We are used to see the search engine to follow a set of rules to rank pages and it is a delight for SEO specialists. It is disappearing and now we have no idea of what is happening when deep learning makes the ranking of result pages.
Artificial intelligence with RankBrain is already used to understand certain queries, it now represents 15% of the work. Under the influence of the new CEO, it will gradually do the whole task. We can predict here the end of SEO because Google engineers themselves do not understand how the machine selects the pages.
October 28, 2015 - Brain Rank
This update is several months old and has just been revealed: it would be the third most important ranking criterion! (After keywords and links). It was given the nickname RankBrain, this artificial intelligence algorithm based on deep learning tries to interpret an unprecedented request to give it sense and select the pages containing the answer.
April 2015 - Mobile friendness becomes a ranking factor
"Starting April 21, we will be expanding our use of mobile-friendliness as a ranking signal."
This is announced by Webmaster Central, the blog of Google about SEO. And also that it will have a big impact on search result on mobiles.
Google invites you to go on your Webmaster Tools account to see problems with your site and also provides a test site.
March 2015 - Wikipedia less favored
Since early 2014, the Wikipedia hearing stopped to grow and even is in slow decline. The algorithm promotes the site less than it did before.
This may be related to the Knowledge Graph, which since early 2014 provides more and more encyclopedic information and therefore makes Wikipedia less relevant.
View Wikimedia Report Card.
February 27, 2015 - App indexing
To let index an application becomes possible, see the Google guide.
Octobre 2014 - Penalizing copier sites
Sites for which Google receives numerous requests for removal of content (DMCA) and whose validity is checked, are penalized in the results. Only 1% of applications were deemed unjustified. Otherwise the link on the content is removed and now the entire site is penalized.
August 28, 2014 - End of the author tag
Webmasters have not really followed and have not widely adopted the rel=author tag to better position their pages in SERPs. Google has decided to no longer take it into account. Also displaying author information that had already reduced over time, will disappear completely. In June 2014, photos had already disappeared. It would seem that links with photos do not get more clicks than others.
But Google says that other types of structured data will still be used.
July 24, 2014 - Pigeon update
Modification to improve local search by applying a preference to positive signals (such as backlinks) when they have a local source.
May 20, 2014 - New Panda
A new version of the Panda mathematical formula that modify the score of ranking of websites, is applied on May 21. It seems that this patch is less harsh to the medium-sized or small sites. It would have affected 7.5% of English sites (according to SearchEngineLand).
May 18, 2014 - Payday Load Algorithm
This change in the algorithm is aimed at groups of keywords most affected by spam, ie pages made only to display ads. It has nothing to do with Panda or Penguin. This is an update of an algorithm implemented last summer (thus in 2013).
February 28, 2014 - Penalty to scraper sites
Google launches hunt to scraper website that are better ranked in its results that the sites they copy. Apparently, the algorithm is not able to identify them, so a call to witnesses was launched (now closed).
The site which is most denounced for the moment is Google itself, which incorporates paragraphs from Wikipedia for its knowledge graph!
In fact we see in the archives on August 29, 2011 that it is not the first time that Google launched this initiative.
September 27, 2013, Hummingbird takes flight
It's the end of the black and white serie, after Panda and Penguin, here is the hummingbird. This new algorithm applies to both the frontend and the backend: it can handle question as a whole rather than as a sequence of keywords and makes the link with the page contents, ie, to find those in the database that meet the issue better. This algorithm, has just been revealed by Google, but it is in place for several weeks.
July 18, 2013, Panda softened
An update of Panda which adds signals to the authority of a site in a niche to prevent useful sites are penalized.
The rollout takes about 10 days to be finished worldwide.
May 22, 2013, Penguin 2
While the previous version of Penguin affected only the home page of a site, the new iteration directly affects all pages (and not indirectly because they depend on the home page). This penalizes sites that have artificial backlinks, usually with optimized anchors.
May 15, 2013. The program for the year to come.
In a video, the head of webspam at Google says that we should expect from changes in the algorithm in the coming months:
- Better detection of the authority in various niches with better signals, so to moderate the negative effect of Panda.
It is claimed that the sites that are on the borderline will benefit of the doubt and will no longer be penalized. Here, as always, it's all about signals as these are signals that demonstrate authority in an area that will be improved. Hopefully this does not mean enrollment in social sites!
- More sophisticated methods to analyze links and to remove any value to the activity of spammers.
- Reduction of too many links to sites that have too much in the SERPs.
In this regard, there is a back and forth with Google in reduction and increase depending on the times, which leaves little credibility to this intention.
- Improving categories of results that have too much spam. Once again reduce spam.
- Do not pass PageRank by advertisements.
- Better information to the webmaster when a site is hacked.
This announcement confirms the weight given to "signals" to assess the pages. Besides the extraction of keywords, the algorithm ignores their content, which is the source of all spam. In essence, the list shows the intention to reduce spam, which is nothing new and it is recognized that Panda is approximate.
November 17, 2012. Mysterious update
There is no communication on the change that has affected many sites, but that if it is not an iteration of Panda. My theory is that it relates to the consideration of likes in social networking sites and comments, but it is purely personal. Note that the count of likes on Google Plus disappeared from GWT over the same time.
Panda updates occured on November 5 and 21.
October 10, 2012. The footer links are a penalty factor
The Webmaster Guidelines have recently changed. As always they state that exchanges and sales of links are prohibited, but a new line appears:
Links widely distributed in the footers of various sites.
These links are not considered natural and therefore violate the webmaster guidelines.
September 28, 2012. Domain names with exact-matching keywords
Sites that have a domain name chosen for its keywords, without content is considered of quality (see Panda), are now penalized. This affects less than 1% of sites.
September 14, 2012. Return to diversity
After the change of 17 November 2010 which allowed a single domain spamming search results pages, Google make a back and limit again the number of links to the same domain.
August 10, 2012. Complaints of copyright infringement penalize a site
Even if Google does not remove content from its index that is the subject of a complaint, it penalizes the page now, when a site is too often the subject of copyright infringement complaints. We wonder if Scribd will be penalized!
Probably not because Google says that many popular sites will escape. Youtube uses a different DMCA form that is not addressed by this criterion of the algorithm: it can not be penalized.
April 24, 2012. Change in algorithm against webspam affecting 3% of websites : Penguin Update!
And obviously related to suroptimisation as announced, even if the term is refined. A site is targeted when it accumulates negative signals, such as link exchanges, backlinks created by oneself. Even internal links if they have the same keywords in the anchor. Irrelevant external links, ie unrelated to the surrounding text, it is also a negative signal. Filling a page with the same keyword too, as we already know.
Studies have shown that a site is penalized when the majority of the backlinks are not from sites in the same topic and when anchors contain keywords that are not in the topic of the page that makes them.
It's the sum of negative signals causing the penalty. These are criteria that has always been taken into account, as we know, but the new algorithm performs a deeper analysis to better penalyze black hat techniques.
In fact Google has communicated about the changes and is talking specifically about "black hat webspam". This change also has a name, it's the penguin update, referring to the black-hat or white-hat sides (it's not a joke).
In April also the algorithm was modified to stop granting freshness boost to new articles when the site is considered of poor quality. This among other changes affecting mainly the presentation of results.
March 2012. Over-optimization to be penalized - Anchor of link revisited - Graphical interface, a ranking factor
After having penalized sites that the algorithm finds the content not original and substancial enough (resulting mostly by promoting verbose sites and doing in the digression), Google is preparing to take the next step and attack sites that drive the optimization too far, that's what Matt Cutts just announced at SXSW.
What optimizations will be penalized?
- The presence of many artificial keywords, not useful to the reader. One should also avoid repeated phrases in a page and all pages.
- Link exchanges that provide useless backlinks. Directories requesting a link back, if there are still one around, are targetted, as paid links and link exchanges in footer.
- Too much manipulation in general. For example too many nofollow links on external sites, for no reason.
All this was already countered by the algorithm but it seems that Google wants to improve recognition of these actions and penalize more the sites. The effect will appear in a few weeks.
Some already believe that de-optimizing a site (ignoring all the optimization rules) to make it look more natural and avoid penalty can have a positive effect, it may be confirmed with the changes of algo expected. But Google says that a good optimization, made only to help engiens to find the content, is recommended.
Google announces that how to interpret anchor links has been modified. Unspecified, a classifier has been removed. The interpretation of the anchor according to the query has been refined. Other changes relate synonyms, dates of threads, freshness, quality sites.
We also learn that the algorithm takes into account the interface and rendering on mobile for the ranking. We understand that the presence of icons, stars, etc ... is an indication of quality.
February 2012. After several years, how to analyze links changed
"We often use the characteristics of links to help us understand the topic of a linked page. We have changed the way we analyze the links. In particular we cancel a method of link analysis used for years."
Even if Google does not say precisely what has changed in its approach to evaluating the links, this sentence implies strongly that there are signals related to the relevance of the link that are concerned and that one is no longer taken into account. Here is the list of these signals:
- The link anchor.
- The text surrounding the link.
- The position in the page. Depending on whether the link is in the body of the text or at bottom, in footer, it has a different role and a different relationship with the linked page.
- The title, rel and nofollow attributes. The latter being unaware of the link (while consuming its share of PageRank), the only possible change would be that it is itself ignored.
- The PageRank of the page containing the link.
- Social link or link inside an article.
One could refer to various Google' patents on link analysis to find out what was changed. But without significant change in the SERPs, we can not really judge. Each of these factors may be written down except one: a relevant link in the body of the text.
The announcement on the changes in March contradicts the fact that this is the link anchor which is now devalued. Maybe it will have to be?
- This criterion can be replaced by other signals, in particular the content of the page containing the link.
- It is mostly SEOs who work on the choice of words in the anchor.
- It is often found that webmasters choose the best keywords for anchors when it concerns their own sites and put anything for others.
- The more correct is actually using the title of the linked article in the anchor. But this brings nothing to search engines, thus the anchor may be ignored.
January 19, 2012. Content visible above the fold is now a ranking factor
As announced in November last year, the pages that show at first advertising content and then actual content visible only when you scroll the text, will be penalized.
It affects 1% of searches.
Users complained that to find content that meets their query, they must scroll the page and skip ads.
But how to determine what is "above the fold", as it depends on the resolution of the screen. On a mobile, depending on whether one holds it in portrait or landscape mode, it is not the same. This is intended to affect pages that have two 280 pixels high ads side by side. In fact Google provides a statistical measure of what is the page height without scrolling with the tool Browser Size. 550 pixels is an acceptable value.
The size of the header is it a part of the equation? If it is not considered part of the content.
The site Browsersize from Google (now closed) provided a measure of what is "above the fold " for a web page.
Announcement in Inside Search.
December 2, 2011. Detection of parked domains.
And whose home page is filled with advertising. A new algorithm is added to detect and exclude them from results. This is part of a dozen measures announced for the month of November, includind also ones about the freshness of content, to promote the most recent pages.
November 14, 2011. Bonus for official sites.
Sites related to a product, a person, when identified as official sites (made by the brand owner itself), will now receive a preferential treatment in ranking, according to a modification of the algorithm announced November 14, 2011.
November 10, 2011. Too many ads in a page: a direct ranking factor in the algo.
At the 2011 PubCon, Matt Cutts said that having too many ads on a page was going to be a (negative) direct ranking factor.
This has always been an indirect factor in so far as this can encourage visitors to leave the site and increase the bounce rate and reduce the time of visit. But it will now be taken directly into account.
This also confirms that this was not a criterion in Panda.
Note that "too much ads" depends upon the size of the page and he said also its placement above the fold is taken into account.
November 3, 2011. New ranking about freshness of pages.
A change in the algorithm affects 35% of queries on the search engine. This concerns the novelty of pages that can be promoted according to the search context.
It is about recent events and hot topics, or topics that come up regularly in the news (Ex: F1 Grand Prix), or which is continually updated without current (Ex: A software).
Other topics such as cooking recipes should not be affected by this change.
August 29, 2011. Better recognition of scrapers.
Sites that duplicate verbatim the pages of other sites to display advertisements should be better identified. They are sometimes better positioned in search results pages as the originals!
Google is testing a new algorithm and asks users to report such sites to help develop.
Report a scraper.
This is not for copyright infringement but for sites that use some tool to extract a content and put in their pages.
August 12, 2011. Panda in all languages.
Apart Chinese, Korean, Japanese Panda now applies to the whole World. The impact is between 6 and 9% in each language.
Panda. In the time, Google changed the way Analytics calculates the bound factor.
June 20, 2011. Post-Panda recovery.
Since June 15, some sites recovered from Panda penalty, when they were modified to remove duplicate content.
This seems to address mainly site hits because duplicate content. The sites which were victim of scrapers recovered and the latter now often removed from SERPs.
June 8, 2011. The author attribute.
Several tags to place award in the body of the page are now recognized by Google:
<a rel="author" href="profile.html">Myself </a>
<a rel="me" href="profile.html">Myself </a>
This will help to classify the pages per author.
The profile page so designated must be on the site that contains this attribute.
April 11, 2011. Panda action extended to the World.
The Panda action against poor quality is now rolled out in the whole World.
But this target only English language queries (on local versions of the search engine). Google is also starting to take into account the fact that some sites are blocked by users. This is one more criterion but minor.
New big sites like eHow were affected by the update, but a lot of much smaller sites with indirect results as links from these sites are written down and this affects also other site, not directly affected.
Understanding Panda Update
Update March 3, 2011. Important change against content farms: Panda Update.
Called internally "Panda" (from the name of an engineer), this action impacted 11.8% of queries by reducing the presence in results pages of poor content, not original or not very useful. On the contrary those which provide detailed articles resulting from a original research will be favored.
"We want to encourage a healthy ecosystem..." Google said.
Google says that these changes does not come from the new extension for Chrome that allows you to block sites. But a comparison with data collected shows that 84% of the sites concerned are included in the list of blocked sites.
The effects appear today only in the U.S.A. Subsequently, this will involve the rest of the world. One main result will be an increase in Adsense revenue for other sites because these content farms are mostly intended to display advertisements.
Remain to see how content farms will be affected, on Alexa or Google trends and if it is a Farmer Day.
Finding more quality sites.
February 24, 2011.
Interview of Google's staff.
January 28, 2011. Change against copied content.
To fight against sites which take content from other sites or of wich the content has no originality, a change was made in the algorithm earlier this week, from 24 January.
This only affects 2% of queries but according to Matt Cutts, that's enough for you to experience a change in the positioning (in the case of Scriptol, the audience grew by 10%).
It is a further improvement affecting the long tail. This can affect content farms that produce line articles, necessarily not original.
Announcement by Matt Cutts.
How organized spam works. By SEOMoz.
January 21, 2011. New ranking formula.
The new algorithm is more efficient to detect spam in the page content, represented by a repetition of words, with the obvious intention of being ranked on those words.
They can be found in an article or blog comments.
See link below.
January 21, 2011. Algorithm better than ever against spam.
So says Google in a letter responding to criticisms about the quality of its search engine particularly in the fight against spam.
Google says ad posters Adsense does not a site prevent without any meaningful content to be downgraded nor participation in the Adwords program.
In 2010, the algorithm has undergone two major changes to cons spam. We remin the change that has affected the long tail at the expense of sites with no content.
Google want to go further in 2011 and invites webmasters to give their opinion. The target is mainly the "content farm" that provide interest-free pages filled with keywords to position themselves in the results.
Some sites as Demand Media (eHow, Answerbag), Associated Content, Suite101, could match the definition of content farm: lot of pages each day with few or no interest targetting the main asked keywords.
The algorithm will be enhanced to recognize the content copied or with no original content.
Google search and search engine spam.
December 2, 2010. Sentiment analysis added to the algorithm.
Following an article on the New York Times, denouncing the fact that a merchant who causes the dissatisfaction of its customers and generates many complaints in blogs and forum gains an advantage with search engines, Google reacted.
Indeed, when we denounce the practices or content of a site, we put links on it to provide examples, and these backlinks are treated as a popularity indice by search engines, which translates into better ranking in the results!
Google therefore developed an algorithm for sentiment analysis, which aims to recognize if the text surrounding a link is positive or negative towards it, depending keywords it contains to penalize sites that we complained.
Google also advises the nofollow attribute to put a link on a site without wishing to contribute to its positioning.
Being bad to your customers is bad for business.
Large-Scale Sentiment Analysis for News and Blogs. Analysis in English of the algorithm.
November 5, 2010. Black Friday.
Since 21 and 22 October depending on the region, a modification of the algorithm on the ranking in results affected a lot of sites, some losing up to 80% of their traffic. The Alexa search engine, has published graphs showing huge losses or gains equivalent to some sites.
These changes seem permanent. The purpose of these changes appears to be intended to improve the relevance of results.
August 31 2010. SVG indexed.
SVG content now indexed either it is in a file to include or embedded into HTML code.
List of files formats supported by Google.
August 20, 2010. Harmful internationalization?
Some webmasters have seen their traffic increased from search engines other than Google Google.com or that of their country.
So Americans can see the arrival of visitors who visit other ccTLD engines, such as google.co.uk or google.fr for example, implying that the engine of other ccTLD includes U.S. sites in the results.
This could reduce the audience for the sites in these countries.
June 8, 2010. Caffeine builds a fresher index.Google announced June 8 that the new indexing engine, Caffeine is finalized. It offers a new index with 50% fresher results.
Its operation differs from the previous system which was updating as a whole, by waves. Caffeine updates the index incrementally. New pages can be added and made available for search as soon as they are discovered.
The new architecture allows also to associate a page to several countries.
Caffeine vs. previous system.
May 27, 2010. MayDay: The long tail evolves.
This was confirmed by Matt Cutts at the Google I/O in May, the radical evolution in the month of April comes from the change in the algorithm, to promote quality content on the long trail.
This is an algorithmic change in Google, looking for higher quality sites to surface for long tail queries. It went through vigorous testing and isn’t going to be rolled back.
Remind that the long tail is the set of queries with multiple keywords, each being rare, but which together form the bulk of traffic to a site.
Webmasters gave the evolution the name of MayDay. I have previously called Black Tuesday. This has been disastrous for some sites well established but having not enough content in deep pages. This happened in late April and early May depending on the sites, even though other sites have experienced a loss of traffic for other reasons.
Matt Cutts explains Mayday in a video. It is not related to Caffeine and is definitive. Webmasters must add content to their page to retrieve the traffic lost.
April 27, 2010. Black Tuesday: Ranking changes on the long tail.
The long tail is the set of pages on a site that make few visit each but all together have a large traffic.
Queries on multiple keywords, make the long tail.
Many sites have seen a change in traffic of these pages since April 27. Some have lost up to 90% of their traffic. They attributed this change to Caffeine, the new infrastructure of Google indexes more pages and creates more competition, but it has been confirmed later by Google it is a change in the algorithm (see May 27).
April 9, 2010. Site speed.
It is officially a ranking factor. This was announced a few months ago, it became reality: a site that is too slow is now downgraded in SERPs or at least has a chance to be in conjunction with other factors.
Today we're Including a new signal in our search ranking algorithms: site speed.
It is possible to know if your site is too slow from Google Webmaster Tools (Labs -> Site performance).
Using site speed in web search ranking.
November 19, 2009. The speed of a site will be a ranking factor in 2010.
This is what Matt Cutts has just said in an interview.
"Historically, we haven't had to use it in our search rankings, but a lot of people within Google think that the web should be fast.
It should be a good experience, and so it's sort of fair to say that if you're a fast site, maybe you should get a little bit of a bonus. If you really have an awfully slow site, then maybe users don't want that as much.
December 15, 2009. Canonical Cross-Domain.
Taking into account the attribute rel="canonical" which was implemented some months ago to avoid duplicate content between pages within a site, has been extended to similar pages on different domain names.
It is still preferable to use 301 redirects when you migrate a site on another domain.
To protect your site against other sites that might copy your content without permission, see how to build a generic canonical tag in PHP.
August 11, 2009.New search engine, Caffeine.
Google is trying a new search engine that is intended to be faster and to provide more relevant results.
July 2, 2009. Less weight for irrelevant backlinks.
This is not confirmed officially by Google (who spoke little of its algorithm in any case), but webmasters believe that the results have changed and that positions in the SERPS are lost because they came from quantities of lower quality backlinks.
What are irrelevant links? There are:
- Backinks from social sites.
- From directories.
- Backlinks in footers in partner sites.
- Links included in CMS templates.
In fact Google recently announced that it would take no more account of blogrolls. It is without doubt the result. And it is not just a loss of importance to these links: they are no longer taken into account!
With regard to social sites (such as Delicious, Stumbleupon), in contrast, Google said in a roundtable with webmasters: "They are adressed as other sites".
June 19, 2009. Flash resources indexed.
The crawler is able to index Flash application, but now, it can index images and texts uses by these applications too. Source Webmaster Central Blog.
June 2, 2009. New effect of the nofollow attribute - Onclick links.
The nofollow attribute let crawlers to ignore a link in a page. So the PR is distributed among the remaining links.
It now appears that the PR is first distributed among all the links (with or without nofollow) and then not distributed to the nofollowed links.
Example: You have 10 points and 5 PR links, 2 points are awarded for each. If two links are nofollow, no PR is passed through them, but others will not receive more points, they will receive only 6 points shared in 3.
The consequences are dramatic, links in comments on a blog would result in lost of PR for other links.
Quoting Matt Cutts:
"Suppose you have 10 links and 5 of them are nofollowed. There’s this assumption that that the other 5 links get ALL that PageRank and that may not be as true anymore."
More on PageRank and nofollow.
Also, Google takes into account links assigned in the onclick event.
April 4, 2009. Local Search.
Google improves local search, based on the IP address, which allows it to find the country and the city of a visitor. From it, Google tries to show in results sites that are located as closely as possible.
To take advantage of this option requires that the research includes a place name, in which case a map is displayed.
February 26, 2009. Brand names.
The algorithm gives now more weight to brand name and therefore promotes related sites. This is confirmed by Matt Cutts (head of staff and spokesman of Google) in a video.
February 25, 2009. The canonical tag.
A new tag tells to crawlers which URL it should index when a page is accessible with multiple addresses.
The duplicate content problem solved.
July 16, 2008.
Google introduced on an experimental basis in some of Wikia's search engine. Users can mark results as good or spam.
The engine takes into account, but for the user who has scored only. For now ...
Google announces that it has one trillion URLs of Web pages in its database.
The pages are all indexed.
June 2008. Nofollow and PageRank.
Nofollow links do not count for the transmission of PageRank, but their PR is not spread over the normal links.
So the PR sent to related pages is divided by the number of links first, then when it evaporates links to nofollow.
Source: PageRank Sculpting.
May 17, 2007. Universal search
New algorithm and architecture to populate the results pages of diverse content such as images, videos, maps, news...
October 19, 2005. Jagger Update
This update adds more weight to relevance in the links. Important sites appear as fortunate.
Spam is fought, especially techniques using CSS for hidden content for visitors.
May 20, 2005. Bourbon Update
An update to penalize sites with duplicate content, links to irrelevant pages (unrelated to the linked page), reciprocal links in quantity, quantity links to a nearby site. This has affected many sites with collateral damage.
2003. Florida Update.
It upset the SERPs. One of the key changes and that the algorithm works differently for different types of queries, and the SERPs are populated with the results of different and complementary types.
The Google search engine appears on the Web.