WebProNews

Tag: SEO

  • Google Adds Copyright Removal Notices To Its Search Algorithm

    Google announced a new change to its search algorithm today. Starting next week, the search engine will begin taking into account the number of valid copyright removal notices in rankings.

    That is the number of valid notices Google receives itself. This should get interesting.

    “Sites with high numbers of removal notices may appear lower in our results,” says Google SVP, Engineering, Amit Singhal. “This ranking change should help users find legitimate, quality sources of content more easily—whether it’s a song previewed on NPR’s music website, a TV show on Hulu or new music streamed from Spotify.”

    “Since we re-booted our copyright removals over two years ago, we’ve been given much more data by copyright owners about infringing content online,” Singhal adds. “In fact, we’re now receiving and processing more copyright removal notices every day than we did in all of 2009—more than 4.3 million URLs in the last 30 days alone. We will now be using this data as a signal in our search rankings.”

    “Only copyright holders know if something is authorized, and only courts can decide if a copyright has been infringed; Google cannot determine whether a particular webpage does or does not violate copyright law,” Singhal notes. “So while this new signal will influence the ranking of some search results, we won’t be removing any pages from search results unless we receive a valid copyright removal notice from the rights owner.”

    Google says it will continue to provide “counter notice” tools that people can use, when they believe their content has been wrongfully removed, so they can get it reinstated.

    More reconsideration request-like things to file. Why do I get the feeling that fair use is going to be challenged more than ever?

    Google says it will continue to be transparent about copyright removals. Speaking of transparency, Google was putting out monthly lists of algorithm changes in an effort to be more transparent, but seems to have fallen behind on that, despite the occasional one-off announcement such as this one.

  • Google Algorithm Changes: Seriously, Is Google Not Sharing Them Anymore? [Updated]

    Google Algorithm Changes: Seriously, Is Google Not Sharing Them Anymore? [Updated]

    Update: Google has finally released the lists. I guess the “transparency” continues.

    In “Google Algorithm Changes For June: What’s On The List?” I pointed out that Google, by the end of July, had still not provided its monthly list of algorithm changes for June. It’s now August 9, and they still have not provided the list for June or July.

    Are you interested in seeing these lists or do you find them irrelevant? Share your thoughts here.

    Since Google started sharing these lists, the company has been posting them to its Inside Search blog. The last post to that blog, as of the time of this writing, is still “Make your mark on Google with Handwrite for Mobile and Tablet Search” from July 26. I was starting to wonder if they were simply no longer using this blog.

    Google announced at the end of last week that it would be shutting down a number of its blogs. Google has over 150 of them, and has decided that it would be better to consolidate the information into fewer blogs, while shutting down the ones that are either updated infrequently or are redundant with other blogs. Google does do a lot of cross-posting.

    That does not, however, appear to be the case with the Inside Search blog. Today, Google announced on its Google Finance blog that it is shutting down that blog, and that updates that would have otherwise appeared there will start appearing on the Inside Search blog. The Inside Search blog lives on.

    But where are those lists?

    Last November, Google started the tradition with this post, looking at some changes it had recently implemented. In December, Google announced that it would just make this a monthly series.

    “We know people care about how search works, so we always want to push the envelope when it comes to transparency,” Google engineering director Scott Huffman wrote at the time.

    “We’ve been wracking our brains trying to think about how to make search even more transparent,” he wrote. “The good news is that we make roughly 500 improvements in a given year, so there’s always more to share. With this blog series, we’ll be highlighting many of the subtler algorithmic and visible feature changes we make.”

    That’s a lot of changes Google makes each year, and I doubt there haven’t been any made over the last two months that Google hasn’t publicized. So what happened to the transparency?

    Huffman noted that the changes that appear in these lists aren’t necessarily big enough to warrant blog posts of their own. In other words, the changes are typically not as Earth shattering as the Panda update or the Penguin update. The lists have, however, given us insight into the kinds of improvements Google is looking to make on an ongoing basis. For example, month to month, we can usually see Google tweaking how it deals with freshness of content.

    Here’s a list of all of the things that have appeared on these lists since Google been releasing them (in order, since November). It’s a huge list, so feel free to skip past it, but that’s kind of the point. If Google stop releasing these lists, that seems like a whole lot less transparency.

    • Cross-language information retrieval updates: For queries in languages where limited web content is available (Afrikaans, Malay, Slovak, Swahili, Hindi, Norwegian, Serbian, Catalan, Maltese, Macedonian, Albanian, Slovenian, Welsh, Icelandic), we will now translate relevant English web pages and display the translated titles directly below the English titles in the search results. This feature was available previously in Korean, but only at the bottom of the page. Clicking on the translated titles will take you to pages translated from English into the query language.
    • Snippets with more page content and less header/menu content: This change helps us choose more relevant text to use in snippets. As we improve our understanding of web page structure, we are now more likely to pick text from the actual page content, and less likely to use text that is part of a header or menu.
    • Better page titles in search results by de-duplicating boilerplate anchors: We look at a number of signals when generating a page’s title. One signal is the anchor text in links pointing to the page. We found that boilerplate links with duplicated anchor text are not as relevant, so we are putting less emphasis on these. The result is more relevant titles that are specific to the page’s content.
    • Length-based autocomplete predictions in Russian: This improvement reduces the number of long, sometimes arbitrary query predictions in Russian. We will not make predictions that are very long in comparison either to the partial query or to the other predictions for that partial query. This is already our practice in English.
    • Extending application rich snippets: We recently announced rich snippets for applications. This enables people who are searching for software applications to see details, like cost and user reviews, within their search results. This change extends the coverage of application rich snippets, so they will be available more often.
    • Retiring a signal in Image search: As the web evolves, we often revisit signals that we launched in the past that no longer appear to have a significant impact. In this case, we decided to retire a signal in Image Search related to images that had references from multiple documents on the web.
    • Fresher, more recent resultsAs we announced just over a week ago, we’ve made a significant improvement to how we rank fresh content. This change impacts roughly 35 percent of total searches (around 6-10% of search results to a noticeable degree) and better determines the appropriate level of freshness for a given query.
    • Refining official page detection: We try hard to give our users the most relevant and authoritative results. With this change, we adjusted how we attempt to determine which pages are official. This will tend to rank official websites even higher in our ranking.
    • Improvements to date-restricted queries: We changed how we handle result freshness for queries where a user has chosen a specific date range. This helps ensure that users get the results that are most relevant for the date range that they specify.
    • Prediction fix for IME queries: This change improves how Autocomplete handles IME queries (queries which contain non-Latin characters). Autocomplete was previously storing the intermediate keystrokes needed to type each character, which would sometimes result in gibberish predictions for Hebrew, Russian and Arabic.
    • Related query results refinements: Sometimes we fetch results for queries that are similar to the actual search you type. This change makes it less likely that these results will rank highly if the original query had a rare word that was dropped in the alternate query. For example, if you are searching for [rare red widgets], you might not be as interested in a page that only mentions “red widgets.”
    • More comprehensive indexing: This change makes more long-tail documents available in our index, so they are more likely to rank for relevant queries.
    • New “parked domain” classifier: This is a new algorithm for automatically detecting parked domains. Parked domains are placeholder sites with little unique content for our users and are often filled only with ads. In most cases, we prefer not to show them.
    • More autocomplete predictions: With autocomplete, we try to strike a balance between coming up with flexible predictions and remaining true to your intentions. This change makes our prediction algorithm a little more flexible for certain queries, without losing your original intention.
    • Fresher and more complete blog search results: We made a change to our blog search index to get coverage that is both fresher and more comprehensive.
    • Original content: We added new signals to help us make better predictions about which of two similar web pages is the original one.
    • Live results for Major League Soccer and the Canadian Football League: This change displays the latest scores & schedules from these leagues along with quick access to game recaps and box scores.
    • Image result freshness: We made a change to how we determine image freshness for news queries. This will help us find the freshest images more often.
    • Layout on tablets: We made some minor color and layout changes to improve usability on tablet devices.
    • Top result selection code rewrite: This code handles extra processing on the top set of results. For example, it ensures that we don’t show too many results from one site (“host crowding”). We rewrote the code to make it easier to understand, simpler to maintain and more flexible for future extensions.
    • Image Search landing page quality signals. [launch codename “simple”] This is an improvement that analyzes various landing page signals for Image Search. We want to make sure that not only are we showing you the most relevant images, but we are also linking to the highest quality source pages.
    • More relevant sitelinks. [launch codename “concepts”, project codename “Megasitelinks”] We improved our algorithm for picking sitelinks. The result is more relevant sitelinks; for example, we may show sitelinks specific to your metropolitan region, which you can control with your location setting.
    • Soft 404 Detection. Web servers generally return the 404 status code when someone requests a page that doesn’t exist. However, some sites are configured to return other status codes, even though the page content might explain that the page was not found. We call these soft 404s (or “crypto” 404s) and they can be problematic for search engines because we aren’t sure if we should ignore the pages. This change is an improvement to how we detect soft 404s, especially in Russian, German and Spanish. For all you webmasters out there, the best practice is still to always use the correct response code.
    • More accurate country-restricted searches. [launch codename “greencr”] On domains other than .com, users have the option to see only results from their particular country. This is a new algorithm that uses several signals to better determine where web documents are from, improving the accuracy of this feature.
    • More rich snippets. We improved our process for detecting sites that qualify for shopping, recipe and review rich snippets. As a result, you should start seeing more sites with rich snippets in search results.
    • Better infrastructure for autocomplete. This is an infrastructure change to improve how our autocomplete algorithm handles spelling corrections for query prefixes (the beginning part of a search).
    • Better spam detection in Image Search. [launch codename “leaf”] This change improves our spam detection in Image Search by extending algorithms we already use for our main search results.
    • Google Instant enhancements for Japanese. For languages that use non-Latin characters, many users use a special IME (Input Method Editor) to enter queries. This change works with browsers that are IME-aware to better handle Japanese queries in Google Instant.
    • More accurate byline dates. [launch codename “foby”] We made a few improvements to how we determine what date to associate with a document. As a result, you’ll see more accurate dates annotating search results.
    • Live results for NFL and college football. [project codename “Live Results”] We’ve added new live results for NFL.com and ESPN’s NCAA Football results. These results now provide the latest scores, schedules and standings for your favorite football teams.
    • Improved dataset for related queries. We are now using an improved dataset on term relationships to find related queries. We sometimes include results for queries that are related to your original search, and this improvement leads to results from more relevant related queries.
    • Related query improvements. [launch codename “lyndsy”] Sometimes we fetch results for queries that are related to the original query but have fewer words. We made several changes to our algorithms to make them more conservative and less likely to introduce results without query words.
    • Better lyrics results. [launch codename “baschi”, project codename “Contra”] This change improves our result quality for lyrics searches.
    • Tweak to +1 button on results page. As part of our continued effort to deliver a beautifully simple user experience across Google products, we’ve made a subtle tweak to how the +1 button appears on the results page. Now the +1 button will only appear when you hover over a result or when the result has already been +1’d.
    • Better spell correction in Vietnamese. [project codename “Pho Viet”] We launched a new Vietnamese spelling model. This will help give more accurate spelling predictions for Vietnamese queries.
    • Upcoming events at venues. We’ve improved the recently released places panel for event venues. For major venues, we now show up to three upcoming events on the right of the page. Try it for [staples center los angeles] or [paradise rock club boston].
    • Improvements to image size signal. [launch codename “matter”] This is an improvement to how we use the size of images as a ranking signal in Image Search. With this change, you’ll tend to see images with larger full-size versions.
    • Improved Hebrew synonyms. [launch codename “SweatNovember”, project codename “Synonyms”] This update refines how we handle Hebrew synonyms across multiple languages. Context matters a lot for translation, so this change prevents us from using translated synonyms that are not actually relevant to the query context.
    • Safer searching. [launch codename “Hoengg”, project codename “SafeSearch”] We updated our SafeSearch tool to provide better filtering for certain queries when strict SafeSearch is enabled.
    • Encrypted search available on new regional domains. Google now offers encrypted search by default on google.com for signed-in users, but it’s not the default on our other regional domains (eg: google.fr for France). Now users in the UK, Germany and France can opt in to encrypted search by navigating directly to an SSL version of Google Search on their respective regional domains: https://www.google.co.ukhttps://www.google.de andhttps://www.google.fr.
    • Faster mobile browsing. [launch codename “old possum”, project codename “Skip Redirect”] Many websites redirect smartphone users to another page that is optimized for smartphone browsers. This change uses the final smartphone destination url in our mobile search results, so you can bypass all the redirects and load the target page faster.
    • Fresher results. [launch codename “nftc”] We made several adjustments to the freshness algorithm that we released in November. These are minor updates to make sure we continue to give you the freshest, most relevant results.
    • Faster autocomplete. [launch codename “Snappy Suggest”, project codename “Suggest”] We made improvements to our autocomplete system to deliver your predicted queries much faster.
    • Autocomplete spelling corrections. [launch codename “Trivial”, project codename “Suggest”] This is an improvement to the spelling corrections used in autocomplete, making those corrections more consistent with the spelling corrections used in search. This launch targets corrections where the spelling change is very small.
    • Better spelling full-page replacement. [launch codenames “Oooni”, “sgap”, project codename “Full-Page Replacement”] When we’re confident in a spelling correction we automatically show results for the corrected query and let you know we’re “Showing results for [cheetah]” (rather than, say, “cheettah”). We made a couple of changes to improve the accuracy of this feature.
    • Better spelling corrections for rare queries. This change improves one of the models that we use to make spelling corrections. The result is more accurate spell corrections for a number of rare queries.
    • Improve detection of recurrent event pages. [launch codename “neseda”] We made several improvements to how we determine the date of a document. As a result, you’ll see fresher, more timely results, particularly for pages discussing recurring events.
    • High-quality sites algorithm improvements. [launch codenames “PPtl” and “Stitch”, project codename “Panda”] In 2011, we launched the Panda algorithm change, targeted at finding more high-quality sites. We improved how Panda interacts with our indexing and ranking systems, making it more integrated into our pipelines. We also released a minor update to refresh the data for Panda.
    • Cross-language refinements. [launch codename Xiangfan] Previously, we only generated related searches based on the display language. With this change, we also attempt to auto-detect the language of the original query to generate related search queries. Now, a user typing a query in French might see French query refinements, even if her language is set to English.
    • English on Google Saudi Arabia. Users in Saudi Arabia can now more easily choose an English interface to search on google.com.sa.
    • Improved scrolling for Image Search. Previously when you scrolled in Image Search, only the image results would move while the top and side menus were pinned in place. We changed the scrolling behavior to make it consistent with our main search results and the other search modes, where scrolling moves the entire page.
    • Improved image search quality. [launch codename “endearo”, project codename “Image Search”] This is a small improvement to our image search ranking algorithm. In particular, this change helps images with high-quality landing pages rank higher in our image search results.
    • More relevant related searches. Sometimes at the bottom of the screen you’ll see a section called “Searches related to” with other queries you may want to try. With this change, we’ve updated the model for generating related searches, resulting in more useful query refinements.
    • Blending of news results. [launch codename “final-destination”, project codename “Universal Search”] We improved our algorithm that decides which queries should show news results, making it more responsive to realtime trends. We also made an adjustment to how we blend news results in Universal Search. Both of these changes help news articles appear in your search results when they are relevant.
    • Automatically disable Google Instant based on computer speed. [project codename “Psychic Search”] Google Instant has long had the ability to automatically turn itself off if you’re on a slow internet connection. Now Instant can also turn itself off if your computer is slow. If Instant gets automatically disabled, we continue to check your computer speed and will re-enable Instant if your performance improves. We’ve also tweaked search preferencesso you can always have Instant on or off, or have it change automatically.
    • More coverage for related searches. [launch codename “Fuzhou”] This launch brings in a new data source to help generate the “Searches related to” section, increasing coverage significantly so the feature will appear for more queries. This section contains search queries that can help you refine what you’re searching for.
    • Tweak to categorizer for expanded sitelinks. [launch codename “Snippy”, project codename “Megasitelinks”] This improvement adjusts a signal we use to try and identify duplicate snippets. We were applying a categorizer that wasn’t performing well for our expanded sitelinks, so we’ve stopped applying the categorizer in those cases. The result is more relevant sitelinks.
    • Less duplication in expanded sitelinks. [launch codename “thanksgiving”, project codename “Megasitelinks”] We’ve adjusted signals to reduce duplication in the snippets forexpanded sitelinks. Now we generate relevant snippets based more on the page content and less on the query.
    • More consistent thumbnail sizes on results page. We’ve adjusted the thumbnail size for most image content appearing on the results page, providing a more consistent experience across result types, and also across mobile and tablet. The new sizes apply to rich snippet results for recipes and applications, movie posters, shopping results, book results, news results and more.
    • More locally relevant predictions in YouTube. [project codename “Suggest”] We’ve improved the ranking for predictions in YouTube to provide more locally relevant queries. For example, for the query [lady gaga in ] performed on the US version of YouTube, we might predict [lady gaga in times square], but for the same search performed on the Indian version of YouTube, we might predict [lady gaga in India].
    • More accurate detection of official pages. [launch codename “WRE”] We’ve made an adjustment to how we detect official pages to make more accurate identifications. The result is that many pages that were previously misidentified as official will no longer be.
    • Refreshed per-URL country information. [Launch codename “longdew”, project codename “country-id data refresh”] We updated the country associations for URLs to use more recent data.
    • Expand the size of our images index in Universal Search. [launch codename “terra”, project codename “Images Universal”] We launched a change to expand the corpus of results for which we show images in Universal Search. This is especially helpful to give more relevant images on a larger set of searches.
    • Minor tuning of autocomplete policy algorithms. [project codename “Suggest”] We have a narrow set of policies for autocomplete for offensive and inappropriate terms. This improvement continues to refine the algorithms we use to implement these policies.
    • “Site:” query update [launch codename “Semicolon”, project codename “Dice”] This change improves the ranking for queries using the “site:” operator by increasing the diversity of results.
    • Improved detection for SafeSearch in Image Search. [launch codename “Michandro”, project codename “SafeSearch”] This change improves our signals for detecting adult content in Image Search, aligning the signals more closely with the signals we use for our other search results.
    • Interval based history tracking for indexing. [project codename “Intervals”] This improvement changes the signals we use in document tracking algorithms.
    • Improvements to foreign language synonyms. [launch codename “floating context synonyms”, project codename “Synonyms”] This change applies an improvement we previously launched for English to all other languages. The net impact is that you’ll more often find relevant pages that include synonyms for your query terms.
    • Disabling two old fresh query classifiers. [launch codename “Mango”, project codename “Freshness”] As search evolves and new signals and classifiers are applied to rank search results, sometimes old algorithms get outdated. This improvement disables two old classifiers related to query freshness.
    • More organized search results for Google Korea. [launch codename “smoothieking”, project codename “Sokoban4”] This significant improvement to search in Korea better organizes the search results into sections for news, blogs and homepages.
    • Fresher images. [launch codename “tumeric”] We’ve adjusted our signals for surfacing fresh images. Now we can more often surface fresh images when they appear on the web.
    • Update to the Google bar. [project codename “Kennedy”] We continue to iterate in our efforts to deliver a beautifully simple experience across Google products, and as part of that this month we made further adjustments to the Google bar. The biggest change is that we’ve replaced the drop-down Google menu in the November redesign with a consistent and expanded set of links running across the top of the page.
    • Adding three new languages to classifier related to error pages. [launch codename “PNI”, project codename “Soft404”] We have signals designed to detect crypto 404 pages (also known as “soft 404s”), pages that return valid text to a browser but the text only contain error messages, such as “Page not found.” It’s rare that a user will be looking for such a page, so it’s important we be able to detect them. This change extends a particular classifier to Portuguese, Dutch and Italian.
    • Improvements to travel-related searches. [launch codename “nesehorn”] We’ve made improvements to triggering for a variety of flight-related search queries. These changes improve the user experience for our Flight Search feature with users getting more accurate flight results.
    • Data refresh for related searches signal. [launch codename “Chicago”, project codename “Related Search”] One of the many signals we look at to generate the “Searches related to” section is the queries users type in succession. If users very often search for [apple] right after [banana], that’s a sign the two might be related. This update refreshes the model we use to generate these refinements, leading to more relevant queries to try.
    • International launch of shopping rich snippets. [project codename “rich snippets”]Shopping rich snippets help you more quickly identify which sites are likely to have the most relevant product for your needs, highlighting product prices, availability, ratings and review counts. This month we expanded shopping rich snippets globally (they were previously only available in the US, Japan and Germany).
    • Improvements to Korean spelling. This launch improves spelling corrections when the user performs a Korean query in the wrong keyboard mode (also known as an “IME”, or input method editor). Specifically, this change helps users who mistakenly enter Hangul queries in Latin mode or vice-versa.
    • Improvements to freshness. [launch codename “iotfreshweb”, project codename “Freshness”] We’ve applied new signals which help us surface fresh content in our results even more quickly than before.
    • Web History in 20 new countries. With Web History, you can browse and search over your search history and webpages you’ve visited. You will also get personalized search results that are more relevant to you, based on what you’ve searched for and which sites you’ve visited in the past. In order to deliver more relevant and personalized search results, we’ve launched Web History in Malaysia, Pakistan, Philippines, Morocco, Belarus, Kazakhstan, Estonia, Kuwait, Iraq, Sri Lanka, Tunisia, Nigeria, Lebanon, Luxembourg, Bosnia and Herzegowina, Azerbaijan, Jamaica, Trinidad and Tobago, Republic of Moldova, and Ghana. Web History is turned on only for people who have a Google Account and previously enabled Web History.
    • Improved snippets for video channels. Some search results are links to channels with many different videos, whether on mtv.com, Hulu or YouTube. We’ve had a feature for a while now that displays snippets for these results including direct links to the videos in the channel, and this improvement increases quality and expands coverage of these rich “decorated” snippets. We’ve also made some improvements to our backends used to generate the snippets.
    • Improvements to ranking for local search results. [launch codename “Venice”] This improvement improves the triggering of Local Universal results by relying more on the ranking of our main search results as a signal.
    • Improvements to English spell correction. [launch codename “Kamehameha”] This change improves spelling correction quality in English, especially for rare queries, by making one of our scoring functions more accurate.
    • Improvements to coverage of News Universal. [launch codename “final destination”] We’ve fixed a bug that caused News Universal results not to appear in cases when our testing indicates they’d be very useful.
    • Consolidation of signals for spiking topics. [launch codename “news deserving score”, project codename “Freshness”] We use a number of signals to detect when a new topic is spiking in popularity. This change consolidates some of the signals so we can rely on signals we can compute in realtime, rather than signals that need to be processed offline. This eliminates redundancy in our systems and helps to ensure we can continue to detect spiking topics as quickly as possible.
    • Better triggering for Turkish weather search feature. [launch codename “hava”] We’ve tuned the signals we use to decide when to present Turkish users with the weather search feature. The result is that we’re able to provide our users with the weather forecast right on the results page with more frequency and accuracy.
    • Visual refresh to account settings page. We completed a visual refresh of the account settings page, making the page more consistent with the rest of our constantly evolving design.
    • Panda update. This launch refreshes data in the Panda system, making it more accurate and more sensitive to recent changes on the web.
    • Link evaluation. We often use characteristics of links to help us figure out the topic of a linked page. We have changed the way in which we evaluate links; in particular, we are turning off a method of link analysis that we used for several years. We often rearchitect or turn off parts of our scoring in order to keep our system maintainable, clean and understandable.
    • SafeSearch update. We have updated how we deal with adult content, making it more accurate and robust. Now, irrelevant adult content is less likely to show up for many queries.
    • Spam update. In the process of investigating some potential spam, we found and fixed some weaknesses in our spam protections.
    • Improved local results. We launched a new system to find results from a user’s city more reliably. Now we’re better able to detect when both queries and documents are local to the user.
    • Autocomplete with math symbols. [launch codename “Blackboard”, project codename “Suggest”] When we process queries to return predictions in autocomplete, we generally normalize them to match more relevant predictions in our database. This change incorporates several characters that were previously normalized: “+”, “-”, “*”, “/”, “^”, “(“, “)”, and “=”. This should make it easier to search for popular equations, for example [e = mc2] or [y = mx+b].
    • Improvements to handling of symbols for indexing. [launch codename “Deep Maroon”] We generally ignore punctuation symbols in queries. Based on analysis of our query stream, we’ve now started to index the following heavily used symbols: “%”, “$”, “\”, “.”, “@”, “#”, and “+”. We’ll continue to index more symbols as usage warrants.
    • Better scoring of news groupings. [launch codename “avenger_2”] News results on Google are organized into groups that are about the same story. We have scoring systems to determine the ordering of these groups for a given query. This subtle change slightly improves our scoring system, leading to better ranking of news clusters.
    • Sitelinks data refresh. [launch codename “Saralee-76”] Sitelinks (the links that appear beneath some search results and link deeper into the respective site) are generated in part by an offline process that analyzes site structure and other data to determine the most relevant links to show users. We’ve recently updated the data through our offline process. These updates happen frequently (on the order of weeks).
    • Improvements to autocomplete backends, coverage. [launch codename “sovereign”, project codename “Suggest”] We’ve consolidated systems and reduced the number of backend calls required to prepare autocomplete predictions for your query. The result is more efficient CPU usage and more comprehensive predictions.
    • Better handling of password changes. Our general approach is that when you change passwords, you’ll be signed out from your account on all machines. This change ensures that changing your password more consistently signs your account out of Search, everywhere.
    • Better indexing of profile pages. [launch codename “Prof-2”] This change improves the comprehensiveness of public profile pages in our index from more than two-hundred social sites.
    • UI refresh for News Universal. [launch codename “Cosmos Newsy”, project codename “Cosmos”] We’ve refreshed the design of News Universal results by providing more results from the top cluster, unifying the UI treatment of clusters of different sizes, adding a larger font for the top article, adding larger images (from licensed sources), and adding author information.
    • Improvements to results for navigational queries. [launch codename “IceMan5”] A “navigational query” is a search where it looks like the user is looking to navigate to a particular website, such as [New York Times] or [wikipedia.org]. While these searches may seem straightforward, there are still challenges to serving the best results. For example, what if the user doesn’t actually know the right URL? What if the URL they’re searching for seems to be a parked domain (with no content)? This change improves results for this kind of search.
    • High-quality sites algorithm data update and freshness improvements. [launch codename “mm”, project codename “Panda”] Like many of the changes we make, aspects of our high-quality sites algorithm depend on processing that’s done offline and pushed on a periodic cycle. In the past month, we’ve pushed updated data for “Panda,” as we mentioned in a recent tweet. We’ve also made improvements to keep our database fresher overall.
    • Live results for UEFA Champions League and KHL. We’ve added live-updating snippets in our search results for the KHL (Russian Hockey League) and UEFA Champions League, including scores and schedules. Now you can find live results from a variety of sports leagues, including the NFLNBANHL and others.
    • Tennis search feature. [launch codename “DoubleFault”] We’ve introduced a new search feature to provide realtime tennis scores at the top of the search results page. Try [maria sharapova] or [sony ericsson open].
    • More relevant image search results. [launch codename “Lice”] This change tunes signals we use related to landing page quality for images. This makes it more likely that you’ll find highly relevant images, even if those images are on pages that are lower quality.
    • Fresher image predictions in all languages. [launch codename “imagine2”, project codename “Suggest”] We recently rolled out a change to surface more relevant image search predictions in autocomplete in English. This improvement extends the update to all languages.
    • SafeSearch algorithm tuning. [launch codenames “Fiorentini”, “SuperDyn”; project codename “SafeSearch”] This month we rolled out a couple of changes to our SafeSearch algorithm. We’ve updated our classifier to make it smarter and more precise, and we’ve found new ways to make adult content less likely to appear when a user isn’t looking for it
    • Tweaks to handling of anchor text. [launch codename “PC”] This month we turned off a classifier related to anchor text (the visible text appearing in links). Our experimental data suggested that other methods of anchor processing had greater success, so turning off this component made our scoring cleaner and more robust.
    • Simplification to Images Universal codebase. [launch codename “Galactic Center”] We’ve made some improvements to simplify our codebase for Images Universal and to better utilize improvements in our general web ranking to also provide better image results.
    • Better application ranking and UI on mobile. When you search for apps on your phone, you’ll now see richer results with app icons, star ratings, prices, and download buttons arranged to fit well on smaller screens. You’ll also see more relevant ranking of mobile applications based on your device platform, for example Android or iOS.
    • Improvements to freshness in Video Universal. [launch codename “graphite”, project codename “Freshness”] We’ve improved the freshness of video results to better detect stale videos and return fresh content.
    • Fewer undesired synonyms. [project codename “Synonyms”] When you search on Google, we often identify other search terms that might have the same meaning as what you entered in the box (synonyms) and surface results for those terms as well when it might be helpful. This month we tweaked a classifier to prevent unhelpful synonyms from being introduced as content in the results set.
    • Better handling of queries with both navigational and local intent. [launch codename “ShieldsUp”] Some queries have both local intent and are very navigational (directed towards a particular website). This change improves the balance of results we show, and helps ensure you’ll find highly relevant navigational results or local results towards the top of the page as appropriate for your query.
    • Improvements to freshness. [launch codename “Abacus”, project codename “Freshness”] We launched an improvement to freshness late last year that was very helpful, but it cost significant machine resources. At the time we decided to roll out the change only for news-related traffic. This month we rolled it out for all queries.
    • Improvements to processing for detection of site quality. [launch codename “Curlup”] We’ve made some improvements to a longstanding system we have to detect site quality. This improvement allows us to get greater confidence in our classifications.
    • Better interpretation and use of anchor text. We’ve improved systems we use to interpret and use anchor text, and determine how relevant a given anchor might be for a given query and website.
    • Better local results and sources in Google News. [launch codename “barefoot”, project codename “news search”] We’re deprecating a signal we had to help people find content from their local country, and we’re building similar logic into other signals we use. The result is more locally relevant Google News results and higher quality sources.
    • Deprecating signal related to ranking in a news cluster. [launch codename “decaffeination”, project codename “news search”] We’re deprecating a signal that’s no longer improving relevance in Google News. The signal was originally developed to help people find higher quality articles on Google News. (Note: Despite the launch codename, this project has nothing to do with Caffeine, our update to indexing in 2010).
    • Fewer “sibling” synonyms. [launch codename “Gemini”, project codename “Synonyms”] One of the main signals we look at to identify synonyms is context. For example, if the word “cat” often appears next to the term “pet” and “furry,” and so does the word “kitten”, our algorithms may guess that “cat” and “kitten” have similar meanings. The problem is that sometimes this method will introduce “synonyms” that actually are different entities in the same category. To continue the example, dogs are also “furry pets” — so sometimes “dog” may be incorrectly introduced as a synonym for “cat”. We’ve been working for some time to appropriately ferret out these “sibling” synonyms, and our latest system is more maintainable, updatable, debuggable, and extensible to other systems.
    • Better synonym accuracy and performance. [project codename “Synonyms”] We’ve made further improvements to our synonyms system by eliminating duplicate logic. We’ve also found ways to more accurately identify appropriate synonyms in cases where there are multiple synonym candidates with different contexts.
    • Retrieval system tuning. [launch codename “emonga”, project codename “Optionalization”] We’ve improved systems that identify terms in a query which are not necessarily required to retrieve relevant documents. This will make results more faithful to the original query.
    • Less aggressive synonyms. [launch codename “zilong”, project codename “Synonyms”] We’ve heard feedback from users that sometimes our algorithms are too aggressive at incorporating search results for other terms. The underlying cause is often our synonym system, which will include results for other terms in many cases. This change makes our synonym system less aggressive in the way it incorporates results for other query terms, putting greater weight on the original user query.
    • Update to systems relying on geographic data. [launch codename “Maestro, Maitre”] We have a number of signals that rely on geographic data (similar to the data we surface in Google Earth and Maps). This change updates some of the geographic data we’re using.
    • Improvements to name detection. [launch codename “edge”, project codename “NameDetector”] We’ve improved a system for detecting names, particularly for celebrity names.
    • Updates to personalization signals. [project codename “PSearch”] This change updates signals used to personalize search results.
    • Improvements to Image Search relevance. [launch codename “sib”] We’ve updated signals to better promote reasonably sized images on high-quality landing pages.
    • Remove deprecated signal from site relevance signals. [launch codename “Freedom”] We’ve removed a deprecated product-focused signal from a site-understanding algorithm.
    • More precise detection of old pages. [launch codename “oldn23″, project codename “Freshness”] This change improves detection of stale pages in our index by relying on more relevant signals. As a result, fewer stale pages are shown to users.
    • Tweaks to language detection in autocomplete. [launch codename “Dejavu”, project codename “Suggest”] In general, autocomplete relies on the display language to determine what language predictions to show. For most languages, we also try to detect the user query language by analyzing the script, and this change extends that behavior to Chinese (Simplified and Traditional), Japanese and Korean. The net effect is that when users forget to turn off their IMEs, they’ll still get English predictions if they start typing English terms.
    • Improvements in date detection for blog/forum pages. [launch codename “fibyen”, project codename “Dates”] This change improves the algorithm that determines dates for blog and forum pages.
    • More predictions in autocomplete by live rewriting of query prefixes. [launch codename “Lombart”, project codename “Suggest”] In this change we’re rewriting partial queries on the fly to retrieve more potential matching predictions for the user query. We use synonyms and other features to get the best overall match. Rewritten prefixes can include term re-orderings, term additions, term removals and more.
    • Expanded sitelinks on mobile. We’ve launched our expanded sitelinks feature for mobile browsers, providing better organization and presentation of sitelinks in search results.
    • More accurate short answers. [project codename “Porky Pig”] We’ve updated the sources behind our short answers feature to rely on data from Freebase. This improves accuracy and makes it easier to fix bugs.
    • Migration of video advanced search backends. We’ve migrated some backends used in video advanced search to our main search infrastructure.
    • +1 button in search for more countries and domains. This month we’ve internationalized the +1 button on the search results page to additional languages and domains. The +1 button in search makes it easy to share recommendations with the world right from your search results. As we said in our initial blog post, the beauty of +1’s is their relevance—you get the right recommendations (because they come from people who matter to you), at the right time (when you are actually looking for information about that topic) and in the right format (your search results).
    • Local result UI refresh on tablet. We’ve updated the user interface of local results on tablets to make them more compact and easier to scan.
    • Categorize paginated documents. [launch codename “Xirtam3”, project codename “CategorizePaginatedDocuments”] Sometimes, search results can be dominated bydocuments from a paginated series. This change helps surface more diverse results in such cases.
    • More language-relevant navigational results. [launch codename “Raquel”] For navigational searches when the user types in a web address, such as [bol.com], we generally try to rank that web address at the top. However, this isn’t always the best answer. For example, bol.com is a Dutch page, but many users are actually searching in Portuguese and are looking for the Brazilian email service, http://www.bol.uol.com.br/. This change takes into account language to help return the most relevant navigational results.
    • Country identification for webpages. [launch codename “sudoku”] Location is an important signal we use to surface content more relevant to a particular country. For a while we’ve had systems designed to detect when a website, subdomain, or directory is relevant to a set of countries. This change extends the granularity of those systems to the page level for sites that host user generated content, meaning that some pages on a particular site can be considered relevant to France, while others might be considered relevant to Spain.
    • Anchors bug fix. [launch codename “Organochloride”, project codename “Anchors”] This change fixed a bug related to our handling of anchors.
    • More domain diversity. [launch codename “Horde”, project codename “Domain Crowding”] Sometimes search returns too many results from the same domain. This change helps surface content from a more diverse set of domains.
    • More local sites from organizations. [project codename “ImpOrgMap2”] This change makes it more likely you’ll find an organization website from your country (e.g. mexico.cnn.com for Mexico rather than cnn.com).
    • Improvements to local navigational searches. [launch codename “onebar-l”] For searches that include location terms, e.g. [dunston mint seattle] or [Vaso Azzurro Restaurant 94043], we are more likely to rank the local navigational homepages in the top position, even in cases where the navigational page does not mention the location.
    • Improvements to how search terms are scored in ranking. [launch codename “Bi02sw41”] One of the most fundamental signals used in search is whether and how your search terms appear on the pages you’re searching. This change improves the way those terms are scored.
    • Disable salience in snippets. [launch codename “DSS”, project codename “Snippets”] This change updates our system for generating snippets to keep it consistent with other infrastructure improvements. It also simplifies and increases consistency in the snippet generation process.
    • More text from the beginning of the page in snippets. [launch codename “solar”, project codename “Snippets”] This change makes it more likely we’ll show text from the beginning of a page in snippets when that text is particularly relevant.
    • Smoother ranking changes for fresh results. [launch codename “sep”, project codename “Freshness”] We want to help you find the freshest results, particularly for searches with important new web content, such as breaking news topics. We try to promote content that appears to be fresh. This change applies a more granular classifier, leading to more nuanced changes in ranking based on freshness.
    • Improvement in a freshness signal. [launch codename “citron”, project codename “Freshness”] This change is a minor improvement to one of the freshness signals which helps to better identify fresh documents.
    • No freshness boost for low-quality content. [launch codename “NoRot”, project codename “Freshness”] We have modified a classifier we use to promote fresh content to exclude fresh content identified as particularly low-quality.
    • Tweak to trigger behavior for Instant Previews. This change narrows the trigger area forInstant Previews so that you won’t see a preview until you hover and pause over the icon to the right of each search result. In the past the feature would trigger if you moused into a larger button area.
    • Sunrise and sunset search feature internationalization. [project codename “sunrise-i18n”] We’ve internationalized the sunrise and sunset search feature to 33 new languages, so now you can more easily plan an evening jog before dusk or set your alarm clock to watch the sunrise with a friend.
    • Improvements to currency conversion search feature in Turkish. [launch codename “kur”, project codename “kur”] We launched improvements to the currency conversion search feature in Turkish. Try searching for [dolar kuru], [euro ne kadar], or [avro kaç para].
    • Improvements to news clustering for Serbian. [launch codename “serbian-5”] For news results, we generally try to cluster articles about the same story into groups. This change improves clustering in Serbian by better grouping articles written in Cyrillic and Latin. We also improved our use of “stemming” — a technique that relies on the “stem” or root of a word.
    • Better query interpretation. This launch helps us better interpret the likely intention of your search query as suggested by your last few searches.
    • News universal results serving improvements. [launch codename “inhale”] This change streamlines the serving of news results on Google by shifting to a more unified system architecture.
    • UI improvements for breaking news topics. [launch codename “Smoothie”, project codename “Smoothie”] We’ve improved the user interface for news results when you’re searching for a breaking news topic. You’ll often see a large image thumbnail alongside two fresh news results.
    • More comprehensive predictions for local queries. [project codename “Autocomplete”] This change improves the comprehensiveness of autocomplete predictions by expanding coverage for long-tail U.S. local search queries such as addresses or small businesses.
    • Improvements to triggering of public data search feature. [launch codename “Plunge_Local”, project codename “DIVE”] This launch improves triggering for the public data search feature, broadening the range of queries that will return helpful population and unemployment data.
    • Adding Japanese and Korean to error page classifier. [launch codename “maniac4jars”, project codename “Soft404”] We have signals designed to detect crypto 404 pages (also known as “soft 404s”), pages that return valid text to a browser, but the text only contains error messages, such as “Page not found.” It’s rare that a user will be looking for such a page, so it’s important we be able to detect them. This change extends a particular classifier to Japanese and Korean.
    • More efficient generation of alternative titles. [launch codename “HalfMarathon”] We use a variety of signals to generate titles in search results. This change makes the process more efficient, saving tremendous CPU resources without degrading quality.
    • More concise and/or informative titles. [launch codename “kebmo”] We look at a number of factors when deciding what to show for the title of a search result. This change means you’ll find more informative titles and/or more concise titles with the same information.
    • Fewer bad spell corrections internationally. [launch codename “Potage”, project codename “Spelling”] When you search for [mango tea], we don’t want to show spelling predictions like “Did you mean ‘mint tea’?” We have algorithms designed to prevent these “bad spell corrections” and this change internationalizes one of those algorithms.
    • More spelling corrections globally and in more languages. [launch codename “pita”, project codename “Autocomplete”] Sometimes autocomplete will correct your spelling before you’ve finished typing. We’ve been offering advanced spelling corrections in English, and recently we extended the comprehensiveness of this feature to cover more than 60 languages.
    • More spell corrections for long queries. [launch codename “caterpillar_new”, project codename “Spelling”] We rolled out a change making it more likely that your query will get a spell correction even if it’s longer than ten terms. You can watch uncut footage of when we decided to launch this from our past blog post.
    • More comprehensive triggering of “showing results for” goes international. [launch codename “ifprdym”, project codename “Spelling”] In some cases when you’ve misspelled a search, say [pnumatic], the results you find will actually be results for the corrected query, “pneumatic.” In the past, we haven’t always provided the explicit user interface to say, “Showing results for pneumatic” and the option to “Search instead for pnumatic.” We recently started showing the explicit “Showing results for” interface more often in these cases in English, and now we’re expanding that to new languages.
    • “Did you mean” suppression goes international. [launch codename “idymsup”, project codename “Spelling”] Sometimes the “Did you mean?” spelling feature predicts spelling corrections that are accurate, but wouldn’t actually be helpful if clicked. For example, the results for the predicted correction of your search may be nearly identical to the results for your original search. In these cases, inviting you to refine your search isn’t helpful. This change first checks a spell prediction to see if it’s useful before presenting it to the user. This algorithm was already rolled out in English, but now we’ve expanded to new languages.
    • Spelling model refresh and quality improvements. We’ve refreshed spelling models and launched quality improvements in 27 languages.
    • Fewer autocomplete predictions leading to low-quality results. [launch codename “Queens5”, project codename “Autocomplete”] We’ve rolled out a change designed to show fewer autocomplete predictions leading to low-quality results.
    • Improvements to SafeSearch for videos and images. [project codename “SafeSearch”] We’ve made improvements to our SafeSearch signals in videos and images mode, making it less likely you’ll see adult content when you aren’t looking for it.
    • Improved SafeSearch models. [launch codename “Squeezie”, project codename “SafeSearch”] This change improves our classifier used to categorize pages for SafeSearch in 40+ languages.
    • Improvements to SafeSearch signals in Russian. [project codename “SafeSearch”] This change makes it less likely that you’ll see adult content in Russian when you aren’t looking for it.
    • Increase base index size by 15%. [project codename “Indexing”] The base search index is our main index for serving search results and every query that comes into Google is matched against this index. This change increases the number of documents served by that index by 15%. *Note: We’re constantly tuning the size of our different indexes and changes may not always appear in these blog posts.
    • New index tier. [launch codename “cantina”, project codename “Indexing”] We keep our index in “tiers” where different documents are indexed at different rates depending on how relevant they are likely to be to users. This month we introduced an additional indexing tier to support continued comprehensiveness in search results.
    • Backend improvements in serving. [launch codename “Hedges”, project codename “Benson”] We’ve rolled out some improvements to our serving systems making them less computationally expensive and massively simplifying code.
    • “Sub-sitelinks” in expanded sitelinks. [launch codename “thanksgiving”] This improvementdigs deeper into megasitelinks by showing sub-sitelinks instead of the normal snippet.
    • Better ranking of expanded sitelinks. [project codename “Megasitelinks”] This change improves the ranking of megasitelinks by providing a minimum score for the sitelink based on a score for the same URL used in general ranking.
    • Sitelinks data refresh. [launch codename “Saralee-76”] Sitelinks (the links that appear beneath some search results and link deeper into the site) are generated in part by an offline process that analyzes site structure and other data to determine the most relevant links to show users. We’ve recently updated the data through our offline process. These updates happen frequently (on the order of weeks).
    • Less snippet duplication in expanded sitelinks. [project codename “Megasitelinks”] We’ve adopted a new technique to reduce duplication in the snippets of expanded sitelinks.
    • Movie showtimes search feature for mobile in China, Korea and Japan. We’ve expanded our movie showtimes feature for mobile to China, Korea and Japan.
    • MLB search feature. [launch codename “BallFour”, project codename “Live Results”] As the MLB season began, we rolled out a new MLB search feature. Try searching for [sf giants score] or [mlb scores].
    • Spanish football (La Liga) search feature. This feature provides scores and information about teams playing in La Liga. Try searching for [barcelona fc] or [la liga].
    • Formula 1 racing search feature. [launch codename “CheckeredFlag”] This month we introduced a new search feature to help you find Formula 1 leaderboards and results. Try searching [formula 1] or [mark webber].
    • Tweaks to NHL search feature. We’ve improved the NHL search feature so it’s more likely to appear when relevant. Try searching for [nhl scores] or [capitals score].
    • Keyword stuffing classifier improvement. [project codename “Spam”] We have classifiers designed to detect when a website is keyword stuffing. This change made the keyword stuffing classifier better.
    • More authoritative results. We’ve tweaked a signal we use to surface more authoritative content.
    • Better HTML5 resource caching for mobile. We’ve improved caching of different components of the search results page, dramatically reducing latency in a number of cases.
    • Deeper detection of hacked pages. [launch codename “GPGB”, project codename “Page Quality”] For some time now Google has been detecting defaced content on hacked pages and presenting a notice on search results reading, “This site may be compromised.” In the past, this algorithm has focused exclusively on homepages, but now we’ve noticed hacking incidents are growing more common on deeper pages on particular sites, so we’re expanding to these deeper pages.
    • Autocomplete predictions used as refinements. [launch codename “Alaska”, project codename “Refinements”] When a user types a search she’ll see a number of predictions beneath the search box. After she hits “Enter”, the results page may also include related searches or “refinements”. With this change, we’re beginning to include some especially useful predictions as “Related searches” on the results page.
    • More predictions for Japanese users. [project codename “Autocomplete”] Our usability testing suggests that Japanese users prefer more autocomplete predictions than users in other locales. Because of this, we’ve expanded the number or predictions shown in Japan to as many as eight (when Instant is on).
    • Improvements to autocomplete on Mobile. [launch codename “Lookahead”, project codename “Mobile”] We made an improvement to make predictions work faster on mobile networks through more aggressive caching.
    • Fewer arbitrary predictions. [launch codename “Axis5”, project codename “Autocomplete”] This launch makes it less likely you’ll see low-quality predictions in autocomplete.
    • Improved IME in autocomplete. [launch codename “ime9”, project codename “Translation and Internationalization”] This change improves handling of input method editors (IMEs) in autocomplete, including support for caps lock and better handling of inputs based on user language.
    • New segmenters for Asian languages. [launch codename “BeautifulMind”] Speech segmentation is about finding the boundaries between words or parts of words. We updated the segmenters for three asian languages: Chinese, Japanese, and Korean, to better understand the meaning of text in these languages. We’ll continue to update and improve our algorithm for segmentation.
    • Scoring and infrastructure improvements for Google Books pages in Universal Search.[launch codename “Utgo”, project codename “Indexing”] This launch transitions the billions of pages of scanned books to a unified serving and scoring infrastructure with web search. This is an efficiency, comprehensiveness and quality change that provides significant savings in CPU usage while improving the quality of search results.
    • Unified Soccer feature. [project codename “Answers”] This change unifies the soccer search feature experience across leagues in Spain, England, Germany and Italy, providing scores and scheduling information right on the search result page.
    • Improvements to NBA search feature. [project codename “Answers”] This launch makes it so we’ll more often return relevant NBA scores and information right at the top of your search results. Try searching for [nba playoffs] or [heat games].
    • New Golf search feature. [project codename “Answers”] This change introduces a new search feature for the Professional Golf Association (PGA) and PGA Tour, including information about tour matches and golfers. Try searching for [tiger woods] or [2012 pga schedule].
    • Improvements to ranking for news results. [project codename “News”] This change improves signals we use to rank news content in our main search results. In particular, this change helps you discover news content more quickly than before.
    • Better application of inorganic backlinks signals. [launch codename “improv-fix”, project codename “Page Quality”] We have algorithms in place designed to detect a variety of link schemes, a common spam technique. This change ensures we’re using those signals appropriately in the rest of our ranking.
    • Improvements to Penguin. [launch codename “twref2”, project codename “Page Quality”] This month we rolled out a couple minor tweaks to improve signals and refresh the data used by the penguin algorithm.
    • Trigger alt title when HTML title is truncated. [launch codename “tomwaits”, project codename “Snippets”] We have algorithms designed to present the best possible result titles. This change will show a more succinct title for results where the current title is so long that it gets truncated. We’ll only do this when the new, shorter title is just as accurate as the old one.
    • Efficiency improvements in alternative title generation. [launch codename “TopOfTheRock”, project codename “Snippets”] With this change we’ve improved the efficiency of title generation systems, leading to significant savings in cpu usage and a more focused set of titles actually shown in search results.
    • Better demotion of boilerplate anchors in alternate title generation. [launch codename “otisredding”, project codename “Snippets”] When presenting titles in search results, we want to avoid boilerplate copy that doesn’t describe the page accurately, such as “Go Back.” This change helps improve titles by avoiding these less useful bits of text.
    • Internationalizing music rich snippets. [launch codename “the kids are disco dancing”, project codename “Snippets”] Music rich snippets enable webmasters to mark up their pages so users can more easily discover pages in the search results where you can listen to or preview songs. The feature launched originally on google.com, but this month we enabled music rich snippets for the rest of the world.
    • Music rich snippets on mobile. [project codename “Snippets”] With this change we’ve turned on music rich snippets for mobile devices, making it easier for users to find songs and albums when they’re on the go.
    • Improvement to SafeSearch goes international. [launch codename “GentleWorld”, project codename “SafeSearch”] This change internationalizes an algorithm designed to handle results on the borderline between adult and general content.
    • Simplification of term-scoring algorithms. [launch codename “ROLL”, project codename “Query Understanding”] This change simplifies some of our code at a minimal cost in quality. This is part of a larger effort to improve code readability.
    • Fading results to white for Google Instant. [project codename “Google Instant”] We made a minor user experience improvement to Google Instant. With this change, we introduced a subtle fade animation when going from a page with results to a page without.
    • Better detection of major new events. [project codename “Freshness”] This change helps ensure that Google can return fresh web results in realtime seconds after a major event occurs.
    • Smoother ranking functions for freshness. [launch codename “flsp”, project codename “Freshness”] This change replaces a number of thresholds used for identifying fresh documents with more continuous functions.
    • Better detection of searches looking for fresh content. [launch codename “Pineapples”, project codename “Freshness”] This change introduces a brand new classifier to help detect searches that are likely looking for fresh content.
    • Freshness algorithm simplifications. [launch codename “febofu”, project codename “Freshness”] This month we rolled out a simplification to our freshness algorithms, which will make it easier to understand bugs and tune signals.
    • Updates to +Pages in right-hand panel. [project codename “Social Search”] We improved our signals for identifying relevant +Pages to show in the right-hand panel.
    • Performance optimizations in our ranking algorithm. [launch codename “DropSmallCFeature”] This launch significantly improves the efficiency of our scoring infrastructure with minimal impact on the quality of our results.
    • Simpler logic for serving results from diverse domains. [launch codename “hc1”, project codename “Other Ranking Components”] We have algorithms to help return a diverse set of domains when relevant to the user query. This change simplifies the logic behind those algorithms.
    • Precise location option on tablet. [project codename “Mobile”] For a while you’ve had the option to choose to get personalized search results relevant to your more precise location on mobile. This month we expanded that choice to tablet. You’ll see the link at the bottom of the homepage and a button above local search results.
    • Improvements to local search on tablet. [project codename “Mobile”] Similar to thechanges we released on mobile this month, we also improved local search on tablet as well. Now you can more easily expand a local result to see more details about the place. After tapping the reviews link in local results, you’ll find details such as a map, reviews, menu links, reservation links, open hours and more.
    • Internationalization of “recent” search feature on mobile. [project codename “Mobile”] This month we expanded the “recent” search feature on mobile to new languages and regions.

    This is the kind of stuff we’re missing out on.

    Publishers, webmasters and search industry enthusiasts would no doubt like to see the lists continue. Will they? That remains to be seen.

    We’ve reached out to Google more than one about the lists, trying to find out if we should still expect them or not, but the company has not responded.

    Update: Since this was originally written, Google did make another post to the Inside Search blog about updates to quick answer results on mobile. More on that here. Still no update lists.

    Update 2: Google has posted to the Inside Search blog a second time since this piece was originally written, and this time it is actually about an algorithmic update, though it’s just one new update – they’ll be taking into account copyright removal notices – and no sign of the lists.

    Should Google continue to put out these lists? Do you care? Do you feel like Google is being less transparent? Tell us what you think in the comments.

  • Google Shows That It Cares If Your Site Is Optimized For Smartphones

    Google is testing a new mobile search feature, which displays a small smartphone icon next to results for pages that have been optimized for smartphones.

    The feature was first spotted by Bryson Meunier, the Director of Content Solutions at Resolution Media, who blogged about it, showing the following screen cap:

    Google SERP with Mobile-optimized icon

    He says he was seeing the icon next to Wikipedia and IMDB results, as well as for other sites with mobile content.

    Meunier also updated his post with an official confirmation from Google about the test, saying, “We’re experimenting with ways to optimize the mobile search experience, including helping users identify smartphone-optimized sites.”

    Perhaps there are more approaches out there in the wild.

    Google has been encouraging webmasters to optimize their sites for mobile for quite some time. In recent months, they’ve really ramped up this encouragement with their GoMo campaign.

    It’s in Google’s best interest to see that content is optimized for mobile phones as mobile search queries begin to grow. It doesn’t reflect very well on Google or the Google user experience when the search engine points to pages that are awkward for users.

    It’s obviously in webmasters’ best interest to provide users with optimized experiences to keep them from leaving their sites before converting.

    Hat tip to Barry Schwartz for pointing to Meunier’s post.

  • Google Algorithm Changes For June: What’s On The List?

    In what Google called an effort to be more transparent, the company began posting monthly lists of algorithm changes, highlighting some of the things it has done to improve the quality of search results and/or the user experience. There are typically a slew of changes that users can take a look at and use to at least get a feel for some of the kinds of tweaks Google is making on an ongoing basis.

    For example, we usually see multiple changes to how Google deals with freshness of search results.

    Google usually releases the list fairly early in the month, for the previous month’s changes. Once, they even released the list before the month was actually over. But here we are with one day left in July, and we have not yet seen a list for June’s changes, let alone July’s.

    Google always posts the list to its Inside Search blog, but the last list on the blog looks at May’s changes. There were 39 of them.

    Posts from Google’s Inside Search Blog since then:

    Make your mark on Google with Handwrite for Mobile and Tablet Search

    Find smarter, more comprehensive Search by Image results

    Become a Google power searcher

    Find out what people are searching for with the updated Hot Searches list

    What happened to June. I tweeted at Google’s Matt Cutts about it last week, but received no reply. So here we are just waiting to see if Google releases a list. Perhaps at this point, Google will put out one giant list for June and July. Or maybe we’re not going to see these lists anymore.

    That wouldn’t be very transparent though.

    I’m going to go out on a limb and speculate that there were more adjustments made to how Google deals with freshness.

    We do know that there were two Panda data refreshes during June.

  • What To Do About Google’s Confusing Link Warnings

    Editor’s Note: This is a post from a guest author in the search engine marketing field. Any advice within is that of the author’s, and does not necessarily reflect the view of our organization.

    If you read Chris Crum’s post last Monday, you may be one of the thousands of webmasters who received Google’s indistinguishable link notifications in Google Webmaster Tools (GWT) last week. As Chris’ article mentioned, Google’s “head of the webspam team”, Matt Cutts promised on Friday that a revised notification that more clearly stated the purpose of the original message would be distributed soon.

    Update: Since this article was originally written, Cutts has talked about the messages in a blog post.

    That notification arrived in the GWT message inbox Monday afternoon and it reads:

    “We’ve detected that some of the links pointing to your site are using techniques outside Google’s Webmaster Guidelines.

    We don’t want to put any trust in links that are artificial or unnatural. We recommend removing any unnatural links to your site. However, we do realize that some links are outside of your control. As a result, for this specific incident we are taking very targeted action on the unnatural links instead of your site as a whole. If you are able to remove any of the links, please submit a reconsideration request, including the actions that you took.

    If you have any questions, please visit our Webmaster Help Forum.”

    The second message helps explain that the website receiving the message is not distrusted by Google as a whole, but instead that only “some” of the links pointing to the website did not meet Google’s Webmaster Guidelines.

    It recommends that webmasters “remove” any “unnatural links” and submit a reconsideration request. Reconsideration requests, historically, have been reserved for websites that have been penalized by Google in any number of ways, including deindexing and/or ranking drops. Remember though, the updated notification said that Google did not distrust the website as a whole.

    It begs the question that if Google is not distrusting your site as a whole and Matt Cutts said that the notifications are“… not necessarily something that you automatically need to worry about,” why would Google suggest that you submit a reconsideration request at all?

    Therein lays the confusion.

    What does one do now?

    There is a lot of speculation circulating online about how to treat these warnings. Some are suggesting you simply ignore them and carry on with business as usual. Some believe that submitting a reconsideration request is an admission of guilt.

    It is always my suggestion to follow Google’s instructions when it comes to these notifications. You should try to remove unnatural backlinks and you should submit a reconsideration request. After all, Google did say “please.”

    Where do I start? Start by pulling an inventory of backlinks. You can pull linking domains from the GWT console under the traffic section. However, this doesn’t give you the specific URLs that are linking to you, only the domains that have links pointing to your site. I recommend using a third party tool like SEOMoz’s Open Site Explorer or the Majestic SEO Site Explorer to inventory your links. These are paid tools but will provide you with all inbound link URLs, which will make your analysis considerably easier.

    What am I looking for? Google does not want to trust “unnatural” or “artificial” links, as they said in their (equally confusing) second message. Yet, what does “unnatural” and “artificial” really mean? According to their Webmaster Quality Guidelines, they suggest following these quality tests:

    “A good rule of thumb is whether you’d feel comfortable explaining what you’ve done to a website that competes with you. Another useful test is to ask, ‘Does this help my users? Would I do this if search engines didn’t exist?’”

    I’ve always liked the latter approach – focus on your users. You should promote your businesses on sites and in content that is relevant to your audience. You should ask yourself, “does this link provide value to a potential customer?” If not, it should be removed.

    Create a spreadsheet of questionable sites and specific URLs with links to you that you would like to be removed. This spreadsheet will function as your tracking document when you file your reconsideration request.

    How do I get these links removed? There is no easy, one click solution here. It takes time to gather contact information and submit a request to webmasters to remove the link. When possible, include the specific URL that the link resides on to simplify the process on the webmaster’s side.

    There is no guarantee that the links will be removed. In fact, the success rate of removing links is fairly low. There is no incentive for the webmaster to take the time to find the page in their CMS and remove the link. However, this is what Google wants us to do, so we do it.

    It is extremely important that you track your analysis and outreach in detail because you will need this information when submitting your reconsideration request.

    On your spread sheet, add a date column for “1st removal request” and “2nd removal request” so you can track all correspondence that you send. This can help support your case when you submit a reconsideration request.

    How many rounds of outreach you perform is really up to you. Some webmasters will go as far as having a Cease and Desist letter drafted by their legal department and sent to webmasters, which has proven successful in motivating publishers to remove the links.

    It is my recommendation to be thorough as possible with your backlink analysis in order to avoid any further notifications from Google regarding links that do not match their Webmaster Guidelines.

    What should a reconsideration request look like? First off, don’t stress about this. You are not drafting legal document, it’s an email to Google letting them know that you’re following their rules, they should check your work and award you a proverbial gold star. Use the facts you’ve gathered on your tracking sheet to help you craft your message. An example request may look something like this:

    Hello,

    After receiving the Unnatural Link notification in Webmaster Tools on July 19th, we have proactively reviewed [total number of linking URLs] backlinks and identified [total number of links you wanted removed] unnatural links, then contacted each webmaster and requested removal. We conducted [1, 2, 3, etc.] rounds of outreach to these webmasters to request removal, concluding with a Cease and Desist letter in our final round of outreach [only if you sent a C&D]. We believe we have taken the necessary steps to remove these unnatural links and would like to have our Unnatural Link notification reconsidered.

    Thank you,

    [website] webmaster

    You can obviously expand upon this template as you see fit, but beware that there is a 500 word and >2850 character (including space) limit for the reconsideration request form in GWT.

    Conducting this entire process should guarantee that your Unnatural Link notification is removed. Furthermore, taking the time to analyze your backlink profile at this time should ensure that your inbound links meet Google’s Webmaster Guidelines and will prevent you from receiving future notifications and negative impacts from algorithm updates.

  • Google Gives You A New Way To See How Many Pages You Have In Its Index

    Google has released a new feature in Webmaster Tools called Index Status. The feature shows you how many pages you have included in Google’s index at any given time.

    The feature appears in the Health menu in Webmaster Tools. It shows how many pages are currently indexed, but also shows you a graph dating back a year:

    Index Status

    “If you see a steadily increasing number of indexed pages, congratulations! This should be enough to confirm that new content on your site is being discovered, crawled and indexed by Google,” Google says in a blog post. “However, some of you may find issues that require looking a little bit deeper. That’s why we added an Advanced tab to the feature.”

    “The advanced section will show not only totals of indexed pages, but also the cumulative number of pages crawled, the number of pages that we know about which are not crawled because they are blocked by robots.txt, and also the number of pages that were not selected for inclusion in our results,” Google adds.

    When you click on the Advanced tab, you will see something like this:

    Index Status Advanced

    Google says the data you get from the tool can be used to identify and debug numerous indexing-related issues your site may have.

    Of course, the company also touts the feature as a way to bring “more transparency” to the table, much like its recent, confusing messages about links.

  • Google Panda Update: New Data Refresh Rolls Out

    Google tweeted late last night that it has begun rolling out a new data refresh of the Panda update. The refresh affects about 1% of search results in a noticeable way, according to the company.

    The last Panda update data refresh we saw was in late June, and that was the second one during that month. It was a rarity for Google to roll out two refreshes to this update in a single month.

    Google did, however, announce last week, that it was launching Panda for Japanese and Korean, affecting about 5% of queries (with no other languages affected at all).

    Remember, the latest announcement was just a data refresh. For more on what that means (vs. an update), see what Google’s Matt Cutts had to say about it here.

    For more information about the Panda update, including Google’s advice, recovery stories and other commentary, please peruse our Panda coverage section.

  • Google Gives Webmasters Just What They Need: More Confusion

    Last week, Google began sending out messages to webmasters, warning them of bad links, much like the ones that many webmasters got prior to the infamous Penguin update. Google said, however, that these messages were different. Whereas the company’s advice in the past was to pay attention to these warnings, Google’s said this time, that they’re not necessarily something you need to worry about it.

    Google’s head of webspam, Matt Cutts, wrote on Google+,”If you received a message yesterday about unnatural links to your site, don’t panic. In the past, these messages were sent when we took action on a site as a whole. Yesterday, we took another step towards more transparency and began sending messages when we distrust some individual links to a site. While it’s possible for this to indicate potential spammy activity by the site, it can also have innocent reasons. For example, we may take this kind of targeted action to distrust hacked links pointing to an innocent site. The innocent site will get the message as we move towards more transparency, but it’s not necessarily something that you automatically need to worry about.”

    “If we’ve taken more severe action on your site, you’ll likely notice a drop in search traffic, which you can see in the ‘Search queries’ feature Webmaster Tools for example,” Cutts added. “As always, if you believe you have been affected by a manual spam action and your site no longer violates the Webmaster Guidelines, go ahead and file a reconsideration request. It’ll take some time for us to process the request, but you will receive a followup message confirming when we’ve processed it.”

    Obviously, this all caused a great deal of confusion, and panic among webmasters and the SEO community. Barry Schwartz, who spends a lot of time monitoring forum discussions, wrote, “It caused a major scare amongst SEOs, webmasters and those who owned web sites, never bought a link in their life, didn’t even know what link buying was and got this severe notification that read, ‘our opinion of your entire site is affected.’

    Even SEOmoz was getting these warnings. The company’s lead SEO, Ruth Burr, wrote,”We’ve got the best kind of links: the kind that build themselves. Imagine the sinking feeling I got in the pit of my stomach, then, when a Google Webmaster Tools check on Thursday revealed that we’d incurred an unnatural link warning.”

    Cutts eventually updated his post to indicate that Google has changed the wording of the messages it is sending, in direct response to webmaster feedback.

     
     

    Google has also removed the yellow caution sign that accompany the messages from the webmaster console. According to Cutts, this illustrates that action by the site owner isn’t necessarily required.

  • Google Is Considering Discounting Infographic Links

    Matt Cutts spoke with Eric Enge at SMX Advanced, and Enge has now published the entire interview. In that interview, Cutts reveals that Google may start looking at discounting infographic links.

    That doesn’t mean Google is doing this right now, or that they definitely will, but…come on.

    “In principle, there’s nothing wrong with the concept of an infographic,” Cutts says in the interview. “What concerns me is the types of things that people are doing with them. They get far off topic, or the fact checking is really poor. The infographic may be neat, but if the information it’s based on is simply wrong, then it’s misleading people.”

    “The other thing that happens is that people don’t always realize what they are linking to when they reprint these infographics,” he adds. “Often the link goes to a completely unrelated site, and one that they don’t mean to endorse. Conceptually, what happens is they really buy into publishing the infographic, and agree to include the link, but they don’t actually care about what it links to. From our perspective this is not what a link is meant to be.”

    I don’t think it’s much of a surprise to a lot of people that Google would consider not counting these kinds of links. In fact, last month, we ran an article from David Leonhardt, who talked about this very thing.

    There are certainly legitimate infographics, just as there are legitimate directories, but there is always that room for abuse, and it could represent something like what Google considers to be a linking scheme (which is against its quality guidelines).

    “I would not be surprised if at some point in the future we did not start to discount these infographic-type links to a degree,” Cutts told Enge. “The link is often embedded in the infographic in a way that people don’t realize, vs. a true endorsement of your site.”

    I think that says it all. If you have a major infograhpic strategy that’s built for SEO purposes, I wouldn’t put too much stock into it moving forward. That doesn’t mean, however, that infographics can’t still provide value, and certainly spark some quality social traffic.

    That’s only a small part of Enge’s interview with Cutts. Read the whole thing here.

    Hat tip: Barry Schwartz

  • Google Panda Update: Matt Cutts Talks About Recovery (And A Bunch Of Other Stuff)

    It’s pretty common for Google’s Matt Cutts to appear in Webmaster Help videos, but they’re usually only a few minutes long. This time, he’s treated webmasters to an hour-long Google+ Hangout (from India), with some other members of Google’s search quality team.

    In the video, Cutts responds to a user question, asking if it’s possible to make a one hundred percent recovery from the Panda update.

    “And the answer is yes,” says Cutts. “It is possible to recover a hundred percent from Panda….So, it is possible to recover from Panda in the following ways. Remember, Panda is a hundred percent algorithmic. There’s nothing manual involved in that. And we haven’t made any manual exceptions. And the Panda algorithm, we tend to run it every so often. It’s not like it runs every day. It’s one of these things where you might run it once a month or something like that. Typically, you’re gonna refresh the data for it. And at the same time, you might pull in new signals. And those signals might be able to say, ‘Ah, this is a higher-quality site.’”

    So, there’s a solid group of engineers that I had the chance to work with who have been looking for signals that differentiate the higher quality sites from the sites that might be slightly lower quality,” he continues. “And if look at–even in the last, say, two or three months–we’ve done a better job about differentiating all those. And so, when we rerun the pipeline to recompute the data and then we push that back out–I think the most recent one was earlier this month–there was one that was probably about two weeks ago.”

    Google pushed two Panda refreshes in June. More on that here.

    “And so, when that happens, if, according to our signals, it looks like the site is high-quality or there’s new data or there’s new signals, then you just wouldn’t, you would pop out of being affected and you wouldn’t have to worry about it at all,” Cutts says. “So, in order to recover from Panda, take a fresh look and basically ask yourself, ‘How compelling is my site?’ We’re looking for high quality. We’re looking for something where you land on it, you’re really happy, the sort of thing where you wanna tell your friends about it and come back to it, bookmark it. It’s just incredibly useful. That’s the sort of thing that we don’t want to get affected. So, yes. It is possible to recover.”

    More on Google’s Panda update here.

  • Who Will Win the Social Search Engine War? Facebook, Google, Or Someone Else?

    The search industry has witnessed multiple changes over the past couple of years. Among the most notable changes are social, mobile, and local being integrated into search results. While each of these play a big role in the search evolution that has occurred, I think it’s fair to say that social tops the list.

    How has social media influenced your search experience? Do you prefer its integration, or would you rather have the “old search” model back? Let us know.

    The influx of Facebook, Twitter, Google+, and now players such as Pinterest has all dramatically impacted how users search. With Google as the dominant search engine, it has obviously pushed its social endeavors aggressively to become the leader in both sectors. After Larry Page became Google’s CEO, reports circulated that he was tying all employees’ bonuses to the success of Google’s social strategy.

    What’s more, in a recent interview with Charlie Rose, when asked if he was concerned about Facebook’s competition in search, Page answered with: “It’s something that we take seriously just like we do social media.”

    According to Grant Simmons, the Director of SEO and Social Product at The Search Agency, Google has been worried for some time about competition, especially in social. Although the latest statistics show that Google+ has over 100 million users, it’s clear that it is not gaining the same traction that Facebook did.

    In addition, Google has made some drastic moves including Search Plus Your World and the recent launch of its Knowledge Graph, both of which are designed to improve users’ search experience by making it simpler, personal, and more social. Interestingly though, as Simmons pointed out to us, these recent announcements from Google are similar to products from other companies, AKA competitors.

    Grant Simmons, Director of SEO and Social Product at The Search Agency He told us that Google has rolled out lots of “me too” announcements and that it is “trying to differentiate itself by copying.”

    “The Knowledge Graph is certainly to address some of that,” he went on to say, “but I think also, it’s more like to them looking forward to how they can get Google to be more sticky and also more competitive.”

    The push, however, of Google’s social products into search has not exactly set well with users or the search industry as a whole. Although the tone could change and users might begin to find Google’s products useful, it could also turn users away.

    “With the alternative search engines whether it’s social search or whether it’s media search or whether it’s just mobile search, I think that Google is a little bit concerned,” explained Simmons,” and tying in their 400-500 million users into search results makes sense, it’s just whether users will get a little bit overcome by Google products always being pushed; or, whether it’s gonna be something they adopt because it’s useful to them.”

    When talking about social and the possibilities of a search engine though, one can’t rule out Facebook. At this point, Bing powers the search function on Facebook, but given the social giant’s wide range of data, it’s likely got something up its own sleeve in the search department.

    “I think Facebook has to get into some type of search leveraging the data they have,” said Simmons. “They’re already working on it, I’m sure.”

    But, as Jessica Lee of The Search Agency pointed out on GigaOM, Facebook can’t exactly compete with Google’s more than 8.6 billion indexed pages. Simmons believes Facebook will use its “Like” system, Open Graph, user profiles, and other data it has acquired toward its search curation product.

    “What they do have is they have a core or a connection between lots of pages that have likes on them,” he said. “They have, within Facebook, an understanding of what you do like from a page standpoint, what your connections are, what you talk about within Faceboook, and I think by leveraging that data, they can get a pretty good understanding of what you’re looking for, what you like, [and] what you might connect with on a more relevant basis.”

    He sees Facebook’s partnership with Bing continuing at least in the short-term, since both parties benefit. In addition, it addresses Google’s integration with Google+ into its search results.

    Incidentally, Danny Sullivan in a piece on Search Engine Land in April expressed his doubt for Facebook starting a search engine:

    I’ve been asked a lot recently about whether I believe Facebook will create its own search engine, because the rumors that never die have started again. I don’t, not a web-wide one.

    Among other reasons, I think Facebook won’t because it understands that doing social right, when you’re at the scale Facebook handles, is ensuring that people discover what’s interesting to them rather than having to search for it.

    Social = discovery, a kissing cousin of search, but not the same. And social is what Facebook is focused on. That’s plenty to chew on. Facebook CEO Mark Zuckerberg has said as much repeatedly, such as to Charlie Rose last year, about wanting to do “one thing incredibly well.” That means continuing to perfect social.

    Twitter has also been mentioned in the talks of social search engines, but according to Simmons, it already has “robust search,” but it is challenged by a struggling business model.

    Going forward, it’s essentially inevitable that social search will play an even greater role in the game of search. The winner, however, is yet to be determined.

    Simmons did tell us that SEOs needed to begin thinking about social search and where their audiences are. Most importantly, he believes that brands need to understand that there are different venues and that their audiences will likely need to be approached in different ways.

    “We talk about The Search Agency Search Anywhere Optimization or Search Everywhere Optimization as the new SEO, and it really is about where people are searching and where they’re searching and niche-type stuff,” he said.

    “The single search venue as the solution to social search probably won’t happen,” he continued. “It’s gonna be more about these connections, affiliations, [and] associations that make sense to me as an individual as opposed to just a personalized search engine like Google.”

    How do envision the future of social search? Do you see Google or Facebook playing a role, or could it be a completely different player? We’d love to hear your thoughts.

  • Matt Cutts: Nofollow Links Are Small, Single Digit Percentage Of Links On The Web

    Google’s Matt Cutts recently downplayed the significance of social signals in search, compared to links. Search Marketing Expo uploaded a new video to YouTube, featuring a discussion between Cutts and moderator Danny Sullivan, in which he talks about the notion that social signals have replaced links. In short, while social signals may gain power in time, links are still pretty important.

    “If you look at the library of congress, they say they have 151.4 million items,” says Cutts. “That’s roughly 34 million books, and if you convert that to just pure text like OCR, that’s something like ten terabytes. The web capture team at the library of congress says they have 235 terabytes. Now everybody in this room probably ought to be saying to themselves: 235 terabytes for the largest library in the world is not that much data. YouTube gets 72 hours of video uploaded every minute. So the web is the largest source of data we’ve ever seen before.”

    “There’s more data being generated on the web, compared to any other source of data around the web, and I think, the fact is, a lot of people think, ‘Links are dying,’ or ‘Links are not a democracy,’ or ‘It’s impossible to get links that aren’t nofollow,’ or whatever,” says Cutts. “And the fact is, that’s a little bit of a bubble in my opinion, in the SEO industry, because if you look at the actual percentage of Nofollow links on the web, it’s a single digit percentage. In fact, it’s a pretty small single digit percentage. So there’s this perception that, ‘Yes, everything will go social,’ or ‘Links are completely obsolete,’ and I think it’s premature to reach that conclusion.”

    “I don’t doubt that in ten years, things will be more social, and those will be more powerful signals, but I wouldn’t write the epitaph for links quite yet,” he adds.

    You would think that social signals are pretty damn important, looking at Google’s results on any given day, if you’re using Search Plus Your World (and there’s a good chance you are, as it’s the default experience for signed in users). How often have you seen results appear simply because someone you’re connected to through Google+ has +1’d something?

    I don’t necessarily think social is the best indicator of relevance, as I’ve discussed in the past, but I do believe they can carry a lot of weight, and perhaps more importantly, will help you diversify your traffic sources, and not have to depend on the search giant for so much of your traffic.

  • Blekko Gets New Crawl Technology, Adds Real Time SEO Updates

    Alternative search engine blekko announced today that it has integrated new crawl technology to power its search index, and is now claiming to offer more comprehensive, real time updates to its SEO pages. This, a representative for the company tells WebProNews, is “transparent data that no other search engine offers for free.”

    “Blekko is now one of only a handful of search companies that has more than 4 billion pages indexed and crawls more than 100 million pages daily, with top sites being indexed hourly,” the representative says.

    Blekko shared the following graphic with us:

    Blekko fast SEO

    “When it comes to pages crawled, the sweet spot for blekko is a little more than 4 billion pages and we update at least 100 million pages each day,” says blekko’s director of engineering, Robert Saliba. “As soon as our crawler, Scoutjet, crawls a webpage, users have access to information about it through blekko’s SEO product. We want to allow people to see the Internet the way a search engine sees it, especially what the rest of the internet is saying about an url.”

    “Scoutjet updates the top ranked starting pages on the Internet around every hour, while other high quality pages are checked at least every week,” he explains. “The continuous updates to blekko’s SEO data include page content, meta data, duplicate text, and inbound link counts. Since staying up-to-date is as much about forgetting the old as finding the new, inbound links that are no longer live and duplicate content that is no longer available disappear.”

    Saliba says blekko is bringing more machines into the service, since traffic continues to “grow rapidly”.

    Blekko says it is is getting over 5.5 million unique monthly visitors (a 500% increase over the beginning of the year) and over 100 million queries per month.

  • Google Doesn’t Care How Many Nofollow Links You Have

    Adding the nofollow attribute to links prevents PageRank from being passed. This is something that Google wants webmasters to do for any links that have been purchased. To do otherwise is strictly against Google’s quality guidelines. Violating these guidelines can either get you hit with a manual penalty in Google, or get you snagged by Google’s Penguin updated, which will continue to see regular data refreshes.

    Some webmasters have wondered if having a large amount of links with the nofollow attribute pointing to a page could hurt that page in search. Barry Schwartz at Search Engine Roundtable points to an interesting Google webmaster forum discussion, in which Google Webmaster Trends analyst John Mueller sets the record straight.

    In that thread, user rickweiss writes, “Bloggers have apparently taken the issue of never having a dofollow on any link that is tied to something you are compensated for so seriously that they are putting nofollow on all links in their posts. In other words, the legitimate link to the products page is getting a nofollow.”

    Later, a user going by the name Bens Mom, asks, “I am incorrect in the belief that having too many rel=nofollow links can actually hurt a site? Because that is the impression I’m under.”

    Mueller responded:

    I’d like to back up what others said — having links (even a large number of them) with rel=nofollow pointing to your site does not negatively affect your site. We take these links out of our PageRank calculations, and out of our algorithms when they use links.

    If you’ve been doing this for a longer time, then it might even make sense to work to clean up those older links, and to have them either removed or the rel=nofollow attached (given that those kinds of paid posts would be against our Webmaster Guidelines).

    This isn’t much of a surprise, considering that nofollow is designed to do what its name would imply: keep the search engine algorithms from following these links. That would indicate that these links carry absolutely no weight one way or another.

    Google’s constantly changing algorithm has a lot of people paranoid about their linking strategies, and it seems that some are so worried about Google’s actions that they’re taking unneeded actions of their own, and ironically, possibly hurting SEO in the process.

    Image: John Mueller, from Google+

  • The Everywhereist Has A Brain Tumor, Names It Steve, And Blogs About It

    Geraldine, the travel blogger behind The Everywhereist, has a brain tumor.

    As we’ve been covering the SEO industry for many years, many of our readers may know her best as Rand Fishkin’s wife. Rand, as you probably know, is the CEO of SEOmoz, and as Geraldine explains, The Everywhereist is as much a love letter to her husband as it is a travel blog.

    That love is on display, probably as much as it has ever been, as Geraldine has blogged about her tumor, which she has named Steve.

    In her post, Geraldine writes:

    As for why I named it Steve, … well, duh. What else was I going to name it? There is no one to whom I am particularly close who is named Steve. I’ve never kissed a boy named Steve. I’ve never uttered the phrase, “Steve, I love you.” And Steve is nice and short and easy to add to a long list of unrepeatable words. Behold:

    “Fucking goddamn miserable piece-of-shit Steve.”

    See how well that works? It kind of rolls off the tongue, really. And considering how many big words we’ve had to deal with over the last couple of weeks, I’m inclined to stick to something short and sweet and monosyllabic (this must be how the Kardashians feel).

    Rand comments on the post:

    Fucking goddamn miserable piece-of-shit Steve.

    Hey, look at that. It DOES roll off the… er… keyboard.

    I’m really proud of you KTL. You’ve been a trooper, and you’ve been so awesome to me these last few weeks. I love that you wrote this, too. You know it’s been tough for me to keep it secret, and I almost feel like part of the reason you’re publishing is to make me feel better. Thank you. I love you. I promise to be (mostly) nice and patient with your Mom while we wait at the hospital.

    p.s. We’re not religious, so mentions of various deities may confuse us.

    Rand has also been tweeting about the situation:

    Geraldine says the doctors are confident the tumor is not a not a glioma, but rather pilocytic astrocytoma. Characteristics of this, according to the National Brain Tumor Society, which Geraldine linked to, include:

    • Slow growing, with relatively well-defined borders
    • Grows in the cerebrum, optic nerve pathways, brain stem and cerebellum
    • Occurs most often in children and teens
    • Accounts for two percent of all brain tumors

    The good news, Geraldine says, is that that the doctors say there’s an 80% chance that Steve is benign, and that even if the tumor is not benign, “odds are he’s still very easily treatable.”

    Finding out a loved one has a brain tumor is tough news to hear. I know. I’ve lived it. I can’t imagine what it must be like to find out you have one yourself. These two have clearly kept in good spirits about the whole thing, or at least as good as anyone could keep. Considering the circumstances, the outlook seems pretty positive.

    Best of luck to Geraldine and Rand.

  • Everything Google Thinks Small Businesses Should Know About SEO In Under 10 Minutes

    Google has put out a new video called “SEO for startups in under 10 minutes”.

    Google says the video would be good to watch if you work on a company site with under 50(ish) pages, hope to rank well for your company name (and a “handful” of related terms – “not lots of terms”), and want to be smart about search engines and attracting searchers, but haven’t kept up with the latest search news.

    If you are interested in the latest search news, however, you might want to know that Google just launched a data refresh for the Panda update.

    Google also updated its Webmaster Academy this week to include more videos. These are the ten videos webmasters should watch, according to Google.

  • Google Just Rocked Nearly 1% Of Search Queries With A Panda Refresh

    Google announced that it has pushed a new data refresh for the Panda update, which originally launched in February of last year. This is the second such refresh this month.

    On June 12, Google tweeted that a refresh was rolling out, which affected less than 1% of queries in the U.S. and 1% worldwide. Now, they’ve tweeted again, saying:

    On a somewhat related note, one webmaster claimed to have recovered from Google’s Penguin update over the weekend. A Google spokesperson tells us that there was no Penguin refresh that they’re aware of. When we saw WPMU recover from Penguin, it had happened as Google rolled out a refresh.

    To learn more about the difference between an algorithm update and a data refresh, see what Google’s Matt Cutts has to say about it here.

    Image: Pandara – The Heavy Metal Panda! (YouTube)

  • These Are The 10 Videos Webmasters Need To Watch, According To Google

    Google has updated its Webmaster Academy site to feature more videos. Some of the videos on the site are old, but some are brand new. None fo them are incredibly long, so if you have a few minutes to spare, I recommend watching all of them.

    Some webmasters are pretty used to videos from Google’s Matt Cutts, and he does appear in some of these, but there are also some other faces in the mix.

    Of course the site itself has complementary information to go along with the videos, but watching the videos themselves is a good start. Here they are:

    Google’s Matt Cutts explains how search works:

    Jen Lee of Google’s search quality team explains how to find your site on Google:

    Cutts talks about snippets:

    Alexi Douvas from Google’s search quality team talks about creating content that performs well in Google search results. It’s worth noting that this one was uploaded just today (post Panda and Penguin):

    Michael Wyszomierski from Google’s search quality team talks about webspam content violations:

    Betty Huang from Google’s search quality team talks about how malicious parties can spam your site:

    A hairless Cutts (for more on that story, see here) discusses how a site that focuses on video or images can improve its rankings:

    Lee talks about using sitemaps to help Google find content hosted on your site:

    An introduction to Google+:

    Cutts and colleague Othar Hansson discuss authorship markup:

  • Here’s What Google’s Matt Cutts Says About Affiliate Links And Nofollow

    Google’s Matt Cutts participated in a keynote discussion at SMX Advanced earlier this month. Among various other topics, Cutts talked briefly about affiliate links with moderator Danny Sullivan.

    SMX just uploaded the relevant clip of the discussion to its YouTube channel today, and to reiterate the point Cutts made, fellow Googler John Mueller posted the video to Google+:

    John Mueller

    Regarding affiliate links and "nofollow" – here's what Matt had to say:

    "We handle the vast majority of affiliate stuff correctly because if it is a large enough affiliate network we know about it and we handle it on our side. Even though we handle I believe the vast majority of affiliate links appropriately if you are at all worried about it, I would go ahead and just add the nofollow because you might be earning money from that."

    In Google’s quality guidelines (the basis for the Penguin update), affiliate programs come up more than once.

    “Avoid ‘doorway’ pages created just for search engines, or other ‘cookie cutter’ approaches such as affiliate programs with little or no original content,” Google says. “If your site participates in an affiliate program, make sure that your site adds value. Provide unique and relevant content that gives users a reason to visit your site first.”

    Google has a page about what it means by “little or no original content,” which talks about “thin affiliate sites”. There, Google says, “These sites collect pay-per-click (PPC) revenue by sending visitors to the sites of affiliate programs, while providing little or no value-added content or service to the user. These sites usually have no original content and may be cookie-cutter sites or templates with no unique content.”

  • This Webmaster Changed A Page’s Internal Linking Structure, Recovered From Google’s Penguin Update

    Another webmaster has claimed to have recovered from Google’s Penguin update.

    In a WebmasterWorld forum post (via Barry Scwhartz), member neildt said that they noticed a page they thought had been hit by Penguin, had returned to ranking for the keyword phrase they were targeting. “Now it’s either just a coincidence that it has returned, or it is due to the changes we made throughout our site for this page based on why we were hit by Penguin,” neildt wrote.

    “Basically from the 24th April when we appeared to be affected by Google’s Penguin update, I took a totally random page that was affected and changed the internal linking structure that pointed to the page,” neildt wrote. “Before the 24th April this page would rank on page 1 of Google for ‘hotel name’ and ‘hotel name city’ as example phrases we were targeting. After this time, the ranking for those phrases was beyond page 30 of Google’s SERPS.”

    “Until yesterday (Sunday 24 June) when I checked if these phrases had made any progress and there were some changes in Google’s SERPs,” neildt continued. “To my surprise we are ranking on page 1 for ‘hotel name city’ and page 3 for ‘hotel name’.

    At this point, it’s unclear whether or not Google launched a data refresh for the Penguin update over the weekend. We’ve reached out to the company, and will update when we learn more.

    Schwartz, who linked to the forum thread where this Penguin recovery is being discussed, says he has heard other rumors of a Google update, that may have occurred on Thursday night.

    As far as Penguin goes, clearly it is possible to recover. You may or may not need to start with a freshly designed site, but make sure you’re in total compliance with Google’s quality guidelines. For more details on another recent recovery, read here.

    When the webmaster from that story recovered, there had been a Penguin data refresh.

    Image: The Batman Season 4 Episode 2 (Warner Bros.)

  • Are Short Title Tags More Important In Google Now?

    Ruth Burr, the lead SEO at SEOmoz put out an interesting article, suggesting that shorter title tags may be more important in Google these days. SEOmoz was finding that some of its posts with longer titles were not displaying the right titles in search results. Rather, Google was returning the keyword-heavy part of the content’s URLs as the titles.

    Pretty weird.

    “It looks like having a short, search-friendly title tag has increased in importance – without it, Google could replace your title with just about anything, including part of your URL,” she concludes. “This doesn’t exactly create the user experience we want, and a replaced title tag is a lost opportunity to encourage searchers to click.”

    We’ve been unable to reproduce these kinds of results. We’ve had some pretty long title tags, but they don’t seem to be affecting the results in this way.

    Here’s an example of a title tag, and the way it shows up in Google:

    Title tag

    search result

    Google talked about page titles in search results a bit on its Webmaster Central blog earlier this year. I’m not sure it helps to explain much about this exact situation, but it does provide some insight into how Google handles titles:

    We use many signals to decide which title to show to users, primarily the <title> tag if the webmaster specified one. But for some pages, a single title might not be the best one to show for all queries, and so we have algorithms that generate alternative titles to make it easier for our users to recognize relevant pages. Our testing has shown that these alternative titles are generally more relevant to the query and can substantially improve the clickthrough rate to the result, helping both our searchers and webmasters. About half of the time, this is the reason we show an alternative title.

    Other times, alternative titles are displayed for pages that have no title or a non-descriptive title specified by the webmaster in the HTML. For example, a title using simply the word “Home” is not really indicative of what the page is about. Another common issue we see is when a webmaster uses the same title on almost all of a website’s pages, sometimes exactly duplicating it and sometimes using only minor variations. Lastly, we also try to replace unnecessarily long or hard-to-read titles with more concise and descriptive alternatives.

    Matt Cutts talked about snippets in titles in a video a few years ago:

    As Barr notes, some of Google’s algorithm changes for the month of May had to do with how Google handles titles as well. These included:

  • Trigger alt title when HTML title is truncated. [launch codename “tomwaits”, project codename “Snippets”] We have algorithms designed to present the best possible result titles. This change will show a more succinct title for results where the current title is so long that it gets truncated. We’ll only do this when the new, shorter title is just as accurate as the old one.
  • Efficiency improvements in alternative title generation. [launch codename “TopOfTheRock”, project codename “Snippets”] With this change we’ve improved the efficiency of title generation systems, leading to significant savings in cpu usage and a more focused set of titles actually shown in search results.
  • Better demotion of boilerplate anchors in alternate title generation. [launch codename “otisredding”, project codename “Snippets”] When presenting titles in search results, we want to avoid boilerplate copy that doesn’t describe the page accurately, such as “Go Back.” This change helps improve titles by avoiding these less useful bits of text.
  • We talked about these a bit in the article: Google Calls Upon Tom Waits and Otis Redding To Help With Your Site’s Titles.

    Barr’s advice is to use short, but keyword-rich titles in title tags, even when the titles themselves are longer. She’s even going back on “years of posts” and adding shorter titles.