Friday, June 10, 2016

Columnist Patrick Stox takes a comprehensive look at what Google might consider to be "quality content" and adds his own thoughts and tips based on his experience in the SEO industry.

We’ve all heard that content is king and that you need to write high-quality content, or now “10x content,” as coined by Rand Fishkin. Ask SEOs what “quality content” is and you’ll receive a lot of varied and opinionated answers. Quality is subjective, and each person views it differently.
Ask SEOs what Google considers to be quality content, and you will get a lot of blank stares. I know because I like to ask this a lot.
The number one answer I get, sadly, is that content should be x number of words, where x is usually 200, 300, 500, 700, 1,000, 1,500, or 2,000. More content does not mean better content. A simple query about the age of an actor can be fully answered in a sentence and doesn’t require their life story and filmography.
Another answer I receive is that the content should be “relevant.” The problem with this is that low-quality pages can be relevant as well.
Other SEOs I’ve asked have given amazingly detailed answers from patents or ideas from machine learning about word2vec, RankBrain, deep learning, count-based methods and predictive methods.
Is there a right answer?

Google Webmaster Quality Guidelines

Google has quality guidelines here. However, you may notice that there are many guidelines around negative signals but few around positive signals. When reading these, think for a minute what happens when two, ten or a hundred websites aren’t doing anything bad. How do you determine the quality difference if no one does anything wrong?

Basic principles

  • Make pages primarily for users, not for search engines.
  • Don’t deceive your users.
  • Avoid tricks intended to improve search engine rankings. A good rule of thumb is whether you’d feel comfortable explaining what you’ve done to a website that competes with you, or to a Google employee. Another useful test is to ask, “Does this help my users? Would I do this if search engines didn’t exist?”
  • Think about what makes your website unique, valuable or engaging. Make your website stand out from others in your field.

Specific guidelines

Avoid the following techniques:
  • Automatically generated content
  • Participating in link schemes
  • Creating pages with little or no original content
  • Cloaking
  • Sneaky redirects
  • Hidden text or links
  • Doorway pages
  • Scraped content
  • Participating in affiliate programs without adding sufficient value
  • Loading pages with irrelevant keywords
  • Creating pages with malicious behavior, such as phishing or installing viruses, trojans or other badware
  • Abusing rich snippets markup
  • Sending automated queries to Google
Follow good practices like these:
  • Monitoring your site for hacking and removing hacked content as soon as it appears
  • Preventing and removing user-generated spam on your site

Google on how to create valuable content

Then there’s this section from Google’s Webmaster Academy course, which tells you how to “create valuable content.” There are a few good tips here on what to avoid: broken links, wrong information, grammar or spelling mistakes, excessive ads and so on. These are useful tips, but again, they focus on what not to do.
There are some tips on how to make your site useful, credible and engaging; however, when it comes to being more valuable or high-quality, Google basically says, “be more valuable or high-quality.”
As you begin creating content, make sure your website is:
Useful and informative: If you’re launching a site for a restaurant, you can include the location, hours of operation, contact information, menu and a blog to share upcoming events.
More valuable and useful than other sites: If you write about how to train a dog, make sure your article provides more value or a different perspective than the numerous articles on the web on dog training.
Credible: Show your site’s credibility by using original research, citations, links, reviews and testimonials. An author biography or testimonials from real customers can help boost your site’s trustworthiness and reputation.
High-quality: Your site’s content should be unique, specific and high-quality. It should not be mass-produced or outsourced on a large number of other sites. Keep in mind that your content should be created primarily to give visitors a good user experience, not to rank well in search engines.
Engaging: Bring color and life to your site by adding images of your products, your team or yourself. Make sure visitors are not distracted by spelling, stylistic and factual errors. An excessive number of ads can also be distracting for visitors. Engage visitors by interacting with them through regular updates, comment boxes or social media widgets.

Google’s Panda algorithm



Panda algorithmically assessed website quality. The algorithm targeted many signals of low-quality sites but again didn’t provide much in the way of useful information for positive signals.
Angry Google Panda

Google’s Search Quality Rating Guidelines

There were a lot of signals for both high- and low-quality content and websites in the Google Search Quality Ratings Guidelines. It is worth reading in its entirety multiple times, but I pulled out some of the important parts here:
What makes a High-quality page? A High-quality page may have the following characteristics:
  • High level of Expertise, Authoritativeness and Trustworthiness (E-A-T)
  • A satisfying amount of high quality MC (Main Content)
  • Satisfying website information and/or information about who is responsible for the website, or satisfying customer service information if the page is primarily for shopping or includes financial transactions
  • Positive website reputation for a website that is responsible for the MC on the page
They expand further on the concept of E-A-T. This was the part of the guidelines I found the most interesting and relevant in determining quality of content (or a website in general).

6.1 Low Quality Main Content

One of the most important criteria in PQ (Page Quality) rating is the quality of the MC, which is determined by how much time, effort, expertise and talent/skill have gone into the creation of the page and also informs the E-A-T of the page.
Consider this example: Most students have to write papers for high school or college. Many students take shortcuts to save time and effort by doing one or more of the following:
  • Buying papers online or getting someone else to write for them
  • Making things up
  • Writing quickly, with no drafts or editing
  • Filling the report with large pictures or other distracting content
  • Copying the entire report from an encyclopedia or paraphrasing content by changing words or sentence structure here and there
  • Using commonly known facts, for example, “Argentina is a country. People live in Argentina. Argentina has borders.”
  • Using a lot of words to communicate only basic ideas or facts, for example, “Pandas eat bamboo. Pandas eat a lot of bamboo. Bamboo is the best food for a Panda bear.”
I found the part of about large images amusing. I’m not a fan of hero images unless they are exceptional. Unfortunately, most end up being generic. Some publications make it worse and use generic hero sliders. Remember, there is an algorithm for “above-the-fold,” and I feel like hero images are completely against this. Most hero images provide little to no useful content without having to scroll.
In section 7.0, “Lowest Quality Pages,” Google notes that the following types of pages/websites should receive the Lowest quality rating:
  • Harmful or malicious pages or websites
  • True lack of purpose pages or websites
  • Deceptive pages or websites
  • Pages or websites which are created to make money with little to no attempt to help users
  • Pages with extremely low or lowest-quality MC
  • Pages on YMYL websites that are so lacking in website information that it feels untrustworthy
  • Hacked, defaced or spammed pages
  • Pages or websites created with no expertise or pages which are highly untrustworthy, unreliable, unauthoritative, inaccurate or misleading
  • Websites which have extremely negative or malicious reputations
  • Violations of the Google Webmaster Quality Guidelines
Speaking more specifically about page content in section 7.4, “Lowest Quality Main Content,” the guidelines note that the following types of Main Content (MC) should be judged as Lowest quality:
  • No helpful MC at all or so little MC that the page effectively has no MC
  • MC which consists almost entirely of “keyword stuffing”
  • Gibberish or meaningless MC
  • “Auto-generated” MC, created with little to no time, effort, expertise, manual curation or added value for users
  • MC which consists almost entirely of content copied from another source with little time, effort, expertise, manual curation or added value for users.
Finally, in section 7.2, “Lack of Purpose Pages,” Google notes that:
Sometimes it is impossible to figure out the purpose of the page. Such pages serve no real purpose for users. For example, some pages are deliberately created with gibberish or meaningless (nonsense) text. No matter how they are created, true lack of purpose pages should be rated Lowest quality.
I love how these sections are all basically saying that your page needs to have a purpose and be understood. I’ve seen many marketing pages that use so much lingo, jargon or marketing-speak that even people at the company can’t tell you what the page is about. What’s worse is when good content is stripped away to make more of these kinds of pages.
There are also some interesting snippets regarding the different elements and signals of trust that might need to be included based on the type of website. This information is extremely important, and it’s easy to brainstorm the different website elements that a local business would need (such as “about us” or “contact”), compared to an e-commerce store that might need reviews, pricing and so forth.
The point is that you need to understand the questions your customers are asking and provide that information to them.

12.7 Understanding User Intent

It can be helpful to think of queries as having one or more of the following intents.
  • Know query, some of which are Know Simple queries
  • Do query, some of which are Device Action queries
  • Website query, when the user is looking for a specific website or webpage
  • Visit-in-person query, some of which are looking for a specific business or organization, some of which are looking for a category of businesses
The above is very similar to the standard “informational, navigational and transactional” system, but I like this better.
Google elaborates on the idea of matching user intent with the purpose of the page elsewhere in the document — section 2.2, “What is the Purpose of a Webpage?” lists the following common page purposes:
  • To share information about a topic
  • To share personal or social information
  • To share pictures, videos or other forms of media
  • To express an opinion or point of view
  • To entertain
  • To sell products or services
  • To allow users to post questions for other users to answer
  • To allow users to share files or to download software


explosion boom
Boom! Jackpot. Matching the user intent with the purpose of a page and type of content expected is exactly what I’m looking for in trying to determine quality.
This makes sense if you think about it from the standpoint of semantic search. If I’ve got a product page, and the top results for the keyword I’m targeting are all informational in nature, then I obviously need to either create an informational page or add more information to my product page if I even want to compete.
I see this mismatch often when people ask why they’re not ranking for a specific term.

Google’s guidance on building high-quality websites

Even before the Quality Raters Guidelines, way back in 2011, there was this gem on the Google Webmaster Central Blog that told us the questions Google engineers asked themselves when building the algorithm.
  • Would you trust the information presented in this article?
  • Is this article written by an expert or enthusiast who knows the topic well, or is it more shallow in nature?
  • Does the site have duplicate, overlapping or redundant articles on the same or similar topics with slightly different keyword variations?
  • Would you be comfortable giving your credit card information to this site?
  • Does this article have spelling, stylistic or factual errors?
  • Are the topics driven by genuine interests of readers of the site, or does the site generate content by attempting to guess what might rank well in search engines?
  • Does the article provide original content or information, original reporting, original research or original analysis?
  • Does the page provide substantial value when compared to other pages in search results?
  • How much quality control is done on content?
  • Does the article describe both sides of a story?
  • Is the site a recognized authority on its topic?
  • Is the content mass-produced by or outsourced to a large number of creators or spread across a large network of sites, so that individual pages or sites don’t get as much attention or care?
  • Was the article edited well, or does it appear sloppy or hastily produced?
  • For a health-related query, would you trust information from this site?
  • Would you recognize this site as an authoritative source when mentioned by name?
  • Does this article provide a complete or comprehensive description of the topic?
  • Does this article contain insightful analysis or interesting information that is beyond the obvious?
  • Is this the sort of page you’d want to bookmark, share with a friend or recommend?
  • Does this article have an excessive amount of ads that distract from or interfere with the main content?
  • Would you expect to see this article in a printed magazine, encyclopedia or book?
  • Are the articles short, unsubstantial or otherwise lacking in helpful specifics?
  • Are the pages produced with great care and attention to detail vs. less attention to detail?
  • Would users complain when they see pages from this site?
Once again, spelling, factual errors and content quality control are mentioned, just like in the Google Search Quality Rating Guidelines. There are also a couple of questions about a site being recognized as an authority on the topic or an authority in general.
Additionally, there are questions that seek to answer if the person knows the topic well, if the content is unique and how comprehensively the topic is covered. This matches up perfectly with the E-A-T concept from the Search Quality Rating Guidelines.

Some content quality signals you can control

  • Broken links. Crawl your site with a program like Screaming Frog and fix them.
  • Wrong information. Do research and find the right sources.
  • Grammatical mistakes. You can use a tool like Grammarly or have someone proofread your writing.
  • Spelling mistakes. Use spell-check or an editor.
  • Reading level. The Hemingway App is good for this. You should be adjusting your reading level based on your target audience and the intent of the query.
  • Excessive ads. Just don’t.
  • Page load speed. Go read this.
  • Website features. The features you should have will change depending on the type of website and the intent of the query.
  • Matching the user intent with the purpose of a page and type of content expected. Take a look at the search results to see what is already ranking.
  • Authority and comprehensiveness. Keep reading.
There are things outside of your control in the short term, but you can play the long game and continue to build your authority over time by consistently creating comprehensive content.
At SMX West, I briefly showed a way of identifying all topics/subtopics in an industry and how to completely cover these based on keyword groupings. I believe that if you’re covering everything that’s being searched for and answering every question that people are asking about a topic, then you have a complete answer, and it will be the best answer for a search engine to return in the results.

How do I determine quality content?

I want to share a little more about my actual process and what I look for on a page, or a section of the site as it relates to the content of the page. Besides technical on-page elements, in the content itself what I’m usually looking for are:
  • Concepts and entities
  • Co-occurrence of keywords/phrases
  • Topical completeness

Concepts and entities

We know that Google looks for concepts and entities in the content, so I usually start here. I useAlchemy API for this.
If I enter the page from Google about creating valuable content —https://support.google.com/webmasters/answer/6001093?hl=en — I get back some information on entities such as Search Console, search engines, Google and social media. Concepts returned are for website, Google search, PageRank, web search engine, Bing and Google. Keyword relevance is also returned through Alchemy:
Alchemy API results for keyword relevance for a page about quality content
If you run many of the top ranking websites for a search query through Alchemy API, you will find a lot of overlap that indicates useful data. There are likely consistent concepts and entities that you would want to include in the body of your text. Alchemy has a JSON output, and I know a lot of people use Blockspring to pull into Google Sheets.

Co-occurrence of keywords and phrases

Ultimate Keyword Hunter provides words or phrases that are used on the pages the most. I normally sort by co-occurrence across websites and find that usually two-, three- and four-keyword phrases are the most useful. I set this to pull data from the top 50 search results.
Moz’s new Keyword Explorer has an interesting filter, “related to keywords with similar results pages,” that looks at pages that rank highly for the query entered and looks for other searches that contain the same pages. For example, a quick glance shows me that the pages ranking for “quality content” also rank for different terms around blogs, websites, content marketing and content strategy — all of which I may want to include on my page.
Moz Keyword Explorer result for quality content

Topical completeness

I like to pull all auto-suggested keywords around a topic with Keyword Sh**ter (terrible name, but it’s very useful) and then put the resulting terms back into the AdWords Keyword Planner, which groups them. These groups are the main ideas I want to cover around a topic, whether all on the same page in subsections or on their own pages.
You can see the pivot table I created for auto-suggested terms based on “content quality” here. On a side note, I almost always put the original topic into the Keyword Planner as well, and I will often stem off the original topic into other topics based on the results.
Another tool I like is Answer The Public, which I first heard about from Wil Reynolds. Remember to change the country if you’re not from the UK. The tool is scraping auto-suggested terms and displaying them nicely in a grouped fashion by questions people are asking.
These create the silo of pages I need around each topic to really make sure I’m covering it in-depth, providing answers to all questions being asked and catching people in every part of their journey. I like to think it makes a website the best answer. The more of these are covered, the more expertise and authority you and your website are building around a topic.

Final thoughts

It all really starts with the query intent. Then it’s matching your information and your website to the kinds of information that someone would need to be a good result for them.
This is the data I use to determine what I need to include in my content for completeness and relevancy. I like to inject my own expertise and opinions into the content as well — after all, it’s important to know what has been said, but it’s more important to add insights into things that might not have been said.
I know everyone has their own processes and ways of doing things, and I would love to hear from some of you about how you approach quality content. Let me know what you look for, what tools you use or what your process is for determining quality of content.



 
Ruben Gomez spotted another test, this one is from Google, he posted about it on Twitterand called it material design for desktop. It looks like the card style design we reported about a few weeks ago.
But Ruben explained it is also visible on the home page and shared screen shots showing how the Google home page is gray in the new test design. It also has a bigger search box in the center.
Here is the new home page test in gray (click to enlarge):
click for full size
Here is the current white home page (click to enlarge):
click for full size
Ruben shared more examples as well as a video.
Forum discussion at Twitter.

 
google us map
GoogleBot typically crawls from the United States, not 100% of the time. In fact, Google recently began crawling on a limited basis from other countries but only to check on local-aware features.
In any event, what if you have a web site that is not accessible for US users for legal or other reasons? Google says that GoogleBot from the US wouldn't be able to access it and it probably will cause major indexing issues.
Google's John Mueller said this in a Google Webmaster Help thread yesterday. He wrote, "In general, our cloaking guidelines say that you must show Googlebot the same content as you would show other users from the region that it's crawling from. So if you're blocking users in the US, then you'd need to block Googlebot when it's crawling from the US (as is generally the case)."
He did offer advice that you can allow some legal content to be shown to US users and thus GoogleBot can index the legal content. But without allowing US users to your web site, you have to imagine GoogleBot won't access it - unless you do things that are against Google's Webmaster Guidelines. John said, " one suggestion would be to have content that's globally accessible, for both users & Googlebot from the US, which can then be indexed in search."
This isn't a new topic, we actually wrote about it a few times including in 2008 and 2011 - the interesting part is the advice didn't change even since the January 2015 news that GoogleBot has locale aware GoogleBot smarts.
Forum discussion at Google Webmaster Help.

 
google gold stars
Joy Hawkins asked in a Local Search Forum thread if Google will filter the gold stars shown in the organic search results for local results for specific industries. She said she can't get the stars to come up for specific industries and she is wondering if Google filters them out for specific types of industry queries.
Tim Capper responded that he asked Google's John Mueller (he didn't share the source of this conversation) and John said they do not filter by industry. Here is what Tim wrote:
Just asked John Mu and he is not aware of any filter being applied to industry. He did say that they don't like if the review markup is sitewide or on irrelevant pages. Also they don't like if testimonials are used within the markup.
There was a time where Google would not show local results for SEO or web design companies back in 2009, but it came back a few years later. So it isn't too far fetched to think there might be a gold star review filter.
I do see gold review stars for my company:
click for full size
Forum discussion at Local Search Forum.

 
Bing announced you can now submit your news site, like this one, to Bing News at the new Bing News PubHub.
When you go to Bing News PubHub, you can fill out the form as I screen shot below, to submit your news site to Bing. Here is a screen shot of the form:
click for full size
To be accepted, first you need to make sure your site is verified in Bing Webmaster Toolsand that the site is in accordance with Bing Webmaster guidelines.
Then Bing will review your site and see if they want to include it in Bing News.
Here is their internal criteria:
  • Newsworthiness – Report on timely events and topics that are interesting to users. Content that doesn’t focus on reporting, such as how-to articles, job postings, advice columns, product promotions, is not considered newsworthy. Similarly, content that consists strictly of information without including original reporting or analysis, such as stock data and weather forecasts, is not considered newsworthy.
  • Originality - Provide unique facts or points of view. Faced with numerous sources frequently reporting similar or identical content, originality or uniqueness becomes a critical way to determine the value to a user of an individual story.
  • Authority – Identify sources, authors, and attribution of all content. News sites with authority maintain the highest level of trust and respect from our users.
  • Readability – Create content with correct grammar and spelling, and keep site design easy for users to navigate. Advertising should never interfere with the user experience.
Forum discussion at Twitter.

 
apple
Apple is bringing search ads to their App Store. Apple announced it last night saying, "starting this summer, you'll be able to participate in the Search Ads beta and see the ads in action."
In short, people search in the App Store and if you want to advertise your apps there by keyword, you can using Apple's Search Ads platform. "Search Ads is an efficient and easy way for you to promote your app directly within the U.S. App Store search results, helping customers discover or reengage with your app, while respecting their privacy," Apple added.
Here are screen shots of what it looks like:
Apple App Store Search Ads
See the ad highlighted at the top?
Here is the campaign builder:
click for full size
Here is the reporting engine:
click for full size
If you want to try the beta apply here.
There are more details on this product over here.
Forum discussion at Twitter.

 
Like it or not, many SEOs use the disavow backlinks feature within Google which launched in 2012 mostly based on webmaster feedback.
But sometimes the disavow file can cause a headache for webmasters. When you have weird characters in your URLs, sometimes Google can confuse them and not disavow them or show errors.
John Mueller of Google responded to one such complaint with advice on how to test it. He said on Twitter to "try to submit a file with just that line, then rewrite the line." Use this method to check which URLs are causing you problems and then fix them one by one. Sounds like this method can take some time with debugging but I don't know if there are any better ways?
He added that "sometimes special characters are tricky" for Google's disavow system to handle.
Forum discussion at Twitter.

 
I always find it interesting when a Googler responds to a specific SEO question with an audit like response. By that I mean, if someone complains about their SEO efforts and rankings in Google and Google responds with additional qualification questions, it makes you wonder - what does that all mean.
Let me share an example. In a Google Webmaster Help thread a webmaster for slant.cois complaining his traffic is flat from Google. It has been flat for six months and he isn't sure why because the content improves and more content is added daily. His bullet points are:
  • This is a good example page: http://www.slant.co/topics/341/~2d-game-engines
  • Been around for about 3 years
  • Mostly works like a structured Q&A site with wikipedia elements to it as products are a rapidly changing area and our site keeps up with new releases etc.
  • Google traffic is about 350k a month and pretty flat.
  • We've done a lot of work on algorithmically noindexing content until it hits a quality threshold. A lot of currently noindexed pages are still in the index (such as our /comments links) unfortunately.
  • Our content/community is growing really quickly, around 30/40% each month. We also have 24/7 moderation.
Then he pinged John Mueller of Google about it on Twitter and John's response onTwitter was interesting. He asked "If it's flat for a while, look at the pages with traffic: do they change? do the new ones show up?"
The response is interesting.
Look at the pages that have traffic and look to see if they change. Look if your new pages are showing up in Google and have traffic.
Makes you wonder...
Forum discussion at Twitter & Google Webmaster Help.

Copyright © SEO Link Building List | OffPage SEO List
Social | Articles | Classified | Blog | Forum | Profile | Review | Business | Audio | Video | Image | Documents | Que-Ans | Ping | Proxy | Directory | RSS | Search | PR