Google has for a while has been holding a Google Webmaster Central Hangout on a weekly basis. The aim is to provide "you" an opportunity to ask questions and give feedback to Google about search related issues.
If you don't have time to watch the hour long video you can check out our transcript of it below. We have removed all the superfluous conversation, and where possible expanded on the various topics by adding commentary and other references for further reading. You will also see time references next to each Question title so that you can move to the relevant part of the video should you wish.
In order to assist in finding questions that of most interest to you can find a summary of all the questions below (with internal links to take you to the relevant part of the page):
- Question 1: 0:41 - Why do Indexed URLs in Sitemap Fluctuate in Webmaster tools, and are there any plans to indicate which URLs are not indexed?
- Questions 2: 2:47 - What new stuff is being released for the Search Queries Report?
- Questions 3: 4:30 - Could Gmail be used to store Webmaster Tools legacy data?
- Question 4: 5.34 - Any updates on Penguin? & Does Penguin Crawl Sites?
- Question 5: 7:04 - Does HTTPS boost SEO Performance or Traffic?
- Question 6: 8.37 - Seroundtable Panda Penalty Discussion - Bad Quality Comments to blame?
- Question 7: 27:21 - How much weight does Google give to Page Content Booking Engine in rankings?
- Question 8: 28:51 - How long after publishing more quality content will it take before rankings improve?
- Question 9: 30:28 - Question about merging 2 websites together.
- Question 10: 32:05 - Can
linksfrom bad websites have a negative effect on my website?
- Question 11: 33:52 - If a site has bad links would good links help without using the Disavow Tool?
- Question 12: 37:25 - Question about Negative SEO and Disavowing links daily
- Question 13: 40:08 - Where does the Knowledge Graph take its data?
- Question 14: 41:09 - How do you explain the extremely small changes in search results in the last three weeks?
- Question 15: 41:46 - Use Keyword Based Traffic to Assess User Satisfaction?
- Question 16: 46:35 - Discussion about the "Wrong" type of traffic.
- Question 17: 50:04 - Question about Hidden Tabbed Content. Is tabbed content bad for SEO?
- Question 18: 52:05 - A question about posts with large header images.
- Question 19: 53:20 - Can you tell us more about what's happening in Webmaster Tools next year
- Question 20: 54:57 - Does Panda or Penguin penalty trigger a mass crawl of your site?
- Question 21: 56:22 - Is Google Panda a penalty?
- Question 22: 58:10 - Discussion about an old penalty hit site's chance to recover
- Question 23: 60:06 - Site Discussion - Indicates sometimes Google needs to change their algorithm
Transcript Commentary of the Google Webmaster Central Hangout 30 Dec 2014
Introduction - 0:06
JOHN MUELLER - My name is John Mueller. I am a Webmaster Trends Analysts at Google in Switzerland and part of my role is to talk with webmasters like you guys and to make sure that we answer your questions and that we take your feedback on board when we work on our products when we work on our search algorithms.
Question 1: 0:41 - Why do Indexed URLs in Sitemap Fluctuate in Webmaster tools, and are there any plans to indicate which URLs are not indexed?
LYLE ROMER : Real quick on a Webmaster Tools question. I noticed something on the site maps where sometimes where it says how many are indexed out of how many are in the sitemap, that will seem to fluctuate even though nothing has changed as far as the URLs that are actually in the site map. So I guess a two-part question-- one - would you have any idea what causes that phenomenon, and two, might there be any plans in the future to actually indicate what items from the site map are and are not indexed?
JOHN MUELLER: It's hard to say what could be causing that. My guess is that we're just kind of dropping and re-picking up URLs. I could see that happening, for example, if you have a fairly large site. And sometimes we can pick up a little bit more of that. Sometimes we can't pick up all of that. So those kind of fluctuations
Best Host News Commentary - John Mueller is saying to split your Sitemaps up, but all the major sitemap plugins (for WordPress at least) will only let you divide your sitemap up by Post, Page, Category, Tag and in some cases Video. We use the Yoast SEO plugin that does exactly this. What might be more useful is to have a sitemap for each post category, so you can see whether any category is failing for the purposes of SEO but we could not find a WordPress plugin that did this automatically.
Having multiple sitemaps also has some benefits when it comes to crawling and indexing your site, and you can see an interesting case study on this here .
Question 2: 2:47 - What new stuff is being released for the Search Queries Report?
BARRY : John, since nobody's around,could you share with us what fun stuff is coming out in the Search Queries Report in the Webmaster Tools?
JOHN MUELLER: Uh, not really.
Best Host News Commentary: John Mueller went on a little here by simply stating that they are simply working to bring together all the information regarding search queries and view the parts relevant to your site but couldn't give any details as it was in its early stages. Plus they try not to release many changes over the Christmas period. You can read more about the Search Queries Report here . Or if you only have a couple of minutes to spare you can watch the video below:
Questions 3: 4:30 - Could Gmail be used to store Webmaster Tools legacy data?
JOSH BACHYNSKI : Did you happen to get my email? I did have one feature request that I thought would be interesting. I thought, why not use Gmail data for extra data space to save our legacy Webmaster Tools search query data? Or what if we could give you FTP information? And we'll gladly store it on our sites. If you need space, we have space, if that's the issue.
JOHN MUELLER : Yeah. [...] I don't think that would work from a technical point of
Best Host News Commentary: This can cause a real headache when trying to find a reliable source of data for your
Question 4: 5.34 - Any updates on Penguin? & Does Penguin Crawl Sites?
JOSH BACHYNSKI : So I guess we're obliged to ask about Penguin and if you have any new information.
JOHN MUELLER : I don't really have anything new to share there, no.[...] This is also, I guess, one of the areas where the engineers are trying to take a break and see how things go at the moment, so no big changes happening there.
JOSH BACHYNSKI : [...] Do you know-- does Penguin crawl sites? I know Panda will do a site crawl sometimes. But I was wondering if Penguin was doing a site
JOHN MUELLER : That would probably be completely separate from like an algorithm like Penguin or Panda. That's something where usually there are technical signals that we pick up when we kind of re-crawl a site in a big way. And those are the kind of things where we look at what we've seen from the recent crawls and kind of recognize that something significant has changed. And then we'll say, oh, well, maybe we should check the rest of the site as well. So it's not so much that a quality or a webspam algorithm would say we should re-crawl this website, but it's more like we see significant changes on this website, so we say, OK. Well, let's take a look at what else changed.
Best Host News Commentary: Actually there are a few reasons at the present time why you might not want to convert to
One such example is if you display advertising on your site. The way many advertisements work is that the advertisers will bid, and higher payouts (Cost per click, or Cost per impression) will be made the more demand there is for that advert space.
Another reason you might not want to transfer to HTTPs is due to the fact many shared hosting companies that offer sophisticated technologies such as Varnish based Caching (eg. SiteGround ) do not support HTTPs. Whilst your website will still work on those hosting providers, the technology that speeds up your site will not.
Question 5: 7:04 - Does HTTPS boost SEO Performance or Traffic?
Written Submitted Question: Here we have a question about HTTPS. I checked a few websites that made this change, and I didn't see a boost in their SEO performance or traffic. So why should we move our website to HTTPS if it's not going to help?
JOHN MUELLER: So I think first off it's important to keep in mind that this is a fairly small change at the moment. So it's not something where you would see a jump from number 10 to number one immediately. And on the other hand, this is also something where I think the long-term trend is just going to be towards HTTPS. So in the long run, it's not going to be something where users will go to a website and say, well, if only this website were available in an unencrypted, insecure way, then I'd love to visit it more often. Because there's really no downside to implementing HTTPS once you have that
8:04put online is the content that the users see. So it's something where I think that the general move is just going to happen more and more in that direction. I imagine over time, you'll see that this factor is something that we can improve on and where we can say, well, it really makes sense to kind of push this a little bit more so that we can perhaps count it a little bit higher in the rankings. So it's something where I imagine over the long run, you'll see a bigger change than you would see immediately at the moment.
Best Host News Commentary: When this was first announced there was a mad rush by many internet marketers to implement SSL on their sites. We too... took the hint from Google setup SSL. SSL companies went crazy and there was a slew of blog posts claiming that SSL will increase your rankings. Since August 2014 when the announcement was made people have begun to sober up
Question 6: 8.37 - Seroundtable Panda Penalty Discussion - Bad Quality Comments to blame?
Best Host News Commentary: We will just briefly mention a few things before going into the rather long, but very interesting conversation between Barry and John Mueller. We think it helps to have our commentary at the beginning so it can set the scene for the conversation that follows. The first relevant post was back on the 2nd October when Barry Schwartz of the Search Engine Round Table (Seroundtable.com) queried whether he had been hit by Panda 4.1. If you ever read Barry's site you will know that the posts tend to be short, but most posts attract quite a significant number of comments, often of low quality.
Barry has on several occasions discussed and invited comments at what might be causing his site to have been hit, and theories from Bad comments, indexed Categories Tags (with no value to the reader) and short content have been mentioned and discussed. We probably think that John Mueller has picked up the main issue, but we bet there
After Barry attended this Google Hangout, he then wrote this post , which he then took his interpretation on what John Mueller said to his readers. You can see the decline in his traffic coincided with the Panda Penalty below:
BARRY : John, can I ask about my own site?
JOHN MUELLER : Sure.
BARRY : So I assume you've seen some comments I've made or some people have made on the downward trend in traffic on the [INAUDIBLE]. And then somebody was like, what are you doing, what am I doing to improve things on the site to increase the traffic? And I kind of made like a joke that I'm trying to figure out how to become an authority in SEO. I'm trying to write content that people want to hear and people haven't heard other places, and I'm trying to make sure that people enjoy reading the site. So as you know-- I'm pretty sure you know. I mean, it looks like I was hit by Panda. I mean, I don't know if you can confirm that or not.
JOHN MUELLER : I haven't actually looked at your site recently,so I don't really know which part of the trend you're looking at.
BARRY : OK. [...] But I mean, I don't know what to do. Honestly, I don't care about the little technical issues.
JOHN MUELLER : Do you have old content?
BARRY : Of course I have old content that's dated. It's a news site that's 11, 12 years old. So I'm sure some of the stories that have from 11 years ago about page-writing updates or monthly Google crawls and stuff like that, that's outdated,but yet, it's relevant back then. I'm not going to go ahead and delete historical content because it's dated.
[Barry goes on a bit about this, about his wiki page and about how he won't remove old content because it is no longer relevant]
JOHN MUELLER : So I guess in general I wouldn't worry too much about the dated content, and say, this is-- I don't know. Google made this change five years ago, and it's made 17 changes in that part of the algorithm since then, so it's no longer relevant. I wouldn't worry about it from that point of view. In general, I also wouldn't look at it as too much of a technical issue if you're looking at Panda issues.
JOHN MUELLER : But instead, I'd look at, I guess, the website overall, the web pages overall. That includes things like the design,where I think your website is actually pretty good compared to a lot of other sites. So from that point of view, I see less of a problem. But it also includes things like the comments. It includes things like the unique and original content that you're putting out on your site that is being added through user-generated content-- all of that as well. So I don't really know exactly what our algorithms are looking at specifically with regards to your website.
JOHN MUELLER : It's something where sometimes you'll go through the articles. And you'll say, well, there's some useful information in this article that you're sharing here. But there's just a lot of other stuff happening on the bottom of these little posts, where when our algorithms look at these pages in an aggregated way across the whole page, then that's something where they might say, well, this is a lot of content that's unique to this page, but it's not really high-quality content that we'd want to kind of promote in a very visible way. So that's something where I could imagine maybe there's something you could do.
JOHN MUELLER : Otherwise, it's really tricky, I guess to look at specific changes that you could do when it comes to things like our quality algorithms.
BARRY : So you're telling me to look at the comments.
JOHN MUELLER : Well, I think you have to look at the pages in an overall way. You shouldn't look at the pages and say--I mean, we see this a lot in the forums. For example, people will say, well, my text is unique. 1You can copy and paste it, and it's unique to my website. But that doesn't make this website, this page, a high-quality page. So things like the overall design, how it comes across, how it looks as an authority-- kind of this information that's in general to a web page, to a website, that's something that all comes together, but also things like comments. Where a webmaster might say, well, this is user-generated content. I'm not responsible for what people are posting on my site.
[Continues to confirm it is the page overall, not just the main content]
JOSH BACHYNSKI : Hey, John. [...] What about the ads? Sometimes I notice there are ads on Barry's site that seem to me to be a little off-topic.
JOHN MUELLER : I hadn't noticed that. At least speaking for myself, personally, it's not something that I noticed. And if these ads are on these pages and they're essentially not in the way, they don't kind of break the user experience of these pages,then I see no real problem with that.
[Continues on to say it would be more tricky if the adverts were adult orientated]
MALE SPEAKER : Hey, John [...] Barry's site is one of those trusted sites among SEO communities. I mean, it's just right up there. And how could something that big be suppressed, and to me, I'm obviously a
JOHN MUELLER : I'd have to take a look at the details of what specifically is happening there. So it's not that I have anything magical where I can just like pop a URL in and say, oh. Well, this line of HTML needs to be changed, and then everything will be back to normal. To some extent, things always change on the web. So you'll always see some fluctuations. I imagine that kind of plays into this as well. But I guess looking at the site overall, I do see that it's a great website. It brings across a lot of really timely and important information. So it's not something where I would say, offhand, this is something that we would see as low-quality content.
JOHN MUELLER : But sometimes there are elements that are kind of a part of that overall image that also play a role, where we'd see
MALE SPEAKER : Well, I'm just curious for really any website, not just Barry's. I mean, all of us, I believe, except for you, John, are non-Google employees. What access do you and your team have back there to be able to really assess what's going on with the site, good and bad?
JOHN MUELLER : We have a lot of things.I mean, on the one hand, we have to kind of be able to kind of diagnose the search results and see what's happening there. That's something where we have to figure out which teams we have to contact internally to kind of find out more about this, to escalate problems that we see. So we do have quite a big insight into what's happening in search. And we try to kind of refine the information that we have into something that the webmaster can use, something that we can say publicly.
[Continues to explain that algorithms can
MALE SPEAKER : Hi, John. Previously you mentioned in regards to comments that if an article was a considerably long, original article, and then there
JOHN MUELLER : I think that's something you could look at if you wanted to look at it in a specific way like that. But in general, I just see comments as something that's a part of the pages that you're publishing. So if you look at the page overall and you see,well, this is the part that I wrote. This is the part that other people contributed. Then, of course, depending on the
JOHN MUELLER : So that's something we'll pick up, as well. So it's not something where I'd say you have to watch out more for your comments if you have short content. But just generally keep in mind that what you're publishing is a whole page. It's not just that specific article that you have on top, but
BARRY : You're not saying that Google doesn't understand the difference between the primary content versus the comments on the content. The algorithm does understand the difference, correct?
JOHN MUELLER : We do try to figure that out. And for some
BARRY : OK. And going back to the comments themselves-- so, at least in my case, I watch the comments. I have Disqus. It does a lot of the filtering of the pure spam stuff. It lets some things through. I have three different queues on howI actually monitor and block spam. I'm pretty good at it. So spam is clearly something that's not there. Maybe there's some that
JOHN MUELLER : It's hard to say. So I'd kind of have to look at your site in general to kind of figure out what I'd recommend there. I think definitely monitoring the comments for spam is something that should definitely be done. It sounds like you kind of have that covered. How you handle lower quality comments-- I guess it's tricky sometimes. And it's something where I don't have any default answer where I'd say you should do this or you should do that. I think rewriting other people's comments probably doesn't make that much sense. But maybe there
BARRY : OK. I'm just trying to go by the rule that I tell people-- forget about SEO, per se.
[Further discussion that just goes over what has been said before]
JOHN MUELLER : Yeah. [...] But another thing that I've seen some sites do is bring only a part of their comments
Question 7: 27:21 - How much weight does Google give to Page Content Booking Engine in rankings?
MALE SPEAKER : Hi, John.
JOHN MUELLER : Hi.
MALE SPEAKER : Talking about quality and high-quality content, in the e-commerce space, we are building
JOHN MUELLER : It kind of comes together. So if the system behind your website doesn't work that well, then essentially it kind of generates low-quality content.So that's something--
MALE SPEAKER : I was talking about the results for the people. If we bring different results of other booking engines, and they are more with more information or more quality information for the user-- not inside our machine. Only talking
JOHN MUELLER : Well, that's essentially content. [...] Whether it comes from your database and your algorithms that you use to compile that information or if that's a page that was manually written, that's essentially content that you're providing on those pages.
Best Host News Commentary: This comes down to the Google Webmaster Tools Quality Guidelines . Google will not say specifics, but when it comes to thinking about what Google wants just think "Quality Original Content". So if all you do is aggregate data from another source just think if Google would want to show your
Question 8: 28:51 - How long after publishing more quality content will it take before rankings improve?
MALE SPEAKER : And I'd like to know-- if we are doing a lot of more quality content in all of our websites, I'd like to know, if we started today publishing it, how much time I need to wait to have improved our rankings? We work in Spain, much or less.
JOHN MUELLER : Yeah. There is actually no fixed time for something like that. So we would probably re-crawl it fairly
Best Host News Commentary: A couple of points we feel needed to be added here. Firstly you can decrease the time it takes to get new posts index by using the Google Fetch feature in webmaster tools . As an example, when we use it on this site for new posts the new posts are indexed within a couple of minutes. Secondly, we don't believe it is a simple answer
Question 9: 30:28 - Question about merging 2 websites together.
MALE SPEAKER : And a little more question. I have a website. It's only a one-page website. And I'd like to merge with other websites in order to keep the backlinks and the content. How could I do this? If I want to merge this one site with one page to the other website, but the page already exists?
JOHN MUELLER : So you're combining two websites?
MALE SPEAKER : Yeah, but no creating a new web page, but merging it with an existing web page, if it's possible?
JOHN MUELLER : Sure. I mean, one thing you can do is redirect one of the URLs to the other one. And that helps us to combine the signals that we have for those pages. And then we can re-crawl and re-index that new page that you have based on whatever is there. So redirecting pages and combining them like that is something that I think is a normal part of the web. It's something that always happens, where you refine your content. And you
Best Host News Commentary: We would add a word of caution when redirecting one page
Question 10: 32:05 - Can
links from bad websites have a negative effect on my website?
WRITTEN SUBMITTED QUESTION : Does natural links from domains like directories with additional values like a blog forum on the same domain or other websites with many unnatural outbound links can be bad for my rankings? Google flagged bad link, bad pages, or bad domains. So I think this question is can links from bad websites have a negative effect on my website?
JOHN MUELLER : And in general, that's something where I'd say no. That doesn't have a bad effect on your website. If we know that a website, for example, is selling links, then that's something where we could say, well, we're just going to disregard all of the links from this website. So that's something where, in general, that's not going to cause any problems. On the other hand, if the majority of your links are from sites that we would classify as being problematic, then that can look a bit fishy to our algorithms, where it looks like, well, all of your links are just from directories, and none of your users are recommending your website. Then maybe there is something problematic. Maybe there is something that you need to look at there. So in general, I'd say no. Having individual links from sites like that is not a problem. If you see that all of your links are like that, and especially if you have kind of done this yourself, if you've hired
Best Host News Commentary: Basically John is saying you don't have to worry unless your backlink profile is unnatural. Everyone has a few bad links, but if your backlink profile wholly consists of bad backlinks then you have something to worry about. You can read more about the Google Penguin Penalty here . Essentially if you are worried, then you can use the Disavow tool which we mentioned earlier in this article. Another alternative, which is not really recommended by John Mueller, is to focus on building good quality backlinks. We think this will just get you in that
Question 11: 33:52 - If a site has bad links would good links help without using the Disavow Tool?
MALE SPEAKER : Let's take a hypothetical situation where a webmaster doesn't know about the Webmaster Tools disavow tool, and the majority of his links are directories or websites selling
JOHN MUELLER : That would definitely help. Yeah. So, I mean, we look at it on an aggregated level across everything that we have from your website. And if we see that things are picking up and things are going in the right direction, then that's something our algorithms will be able to take into account. So in the hypothetical situation of someone who doesn't know about any of this and they realized they did something wrong in the past and they're working to improve that in the future, then that's something that our algorithms will pick up on and will be able to use as well. Still, if you're in that situation, it wouldn't be that I'd say you should ignore the disavow tool and just focus on moving forward in a good way, but instead really trying to clean up those old issues as well. And it's not something where we'd say that using the disavow tool is a sign that you're a knowledgeable SEO and that you should know better about these links. It's essentially a technical thing on our side, where we don't take those links into account anymore. It doesn't count negatively for your website if you use a disavow tool. It's not something you should be ashamed of using. If you know about this tool, if you know about problematic links to your site, then I just recommend cleaning that up.
MALE SPEAKER : [...] I mean, would the penalty actually get removed if the majority of the percentage of low-quality links diminishes? The actual Penguin penalty-- would it be removed?
JOHN MUELLER : Yeah. That's something that our algorithms would take into account-- where if they look at the site overall and they see that this is essentially improving, if it looks like things are headed in the right way and the important links are really good links that are recommendations by other people, then they'll be able to take that into account and modify whatever adjustment there was made with that change there on that website. So they would take that into account. I wouldn't say that you have to have more than 50% and then the algorithm will disappear for your website. Let's say there are lots of shades of gray involved there, where the algorithm could say, well, this is looked really bad in the beginning. They worked a lot to kind of improve things overall. Things were improving significantly across the web with lots of good recommendations for this site. So it's kind of headed in the right direction. So it wouldn't be that it disappears completely, but maybe it'll kind of step-by-step improve.
MALE SPEAKER : OK, thank you.
Question 12: 37:25 - Question about Negative SEO and Disavowing links daily
MALE SPEAKER : John, if I can jump in just to follow-up on his last question.
JOHN MUELLER : Sure.
MALE SPEAKER : Regarding the disavow, I've posted a bunch of-- probably Barry's site, as well. People know me as the guy that disavows twice a day. I've been having to-- I do it in the morning and at night, because we're still undergoing a negative SEO attack-- I should say a negative link spam attack on the site. How will we ever know if the disavow tool is actually working? I only ask that because a month after the attack started,
38:00our trajectory was going-- let's see if I can do this perspectively right-- straight up, and we are literally just down one month after. And I was disavowing twice a day since day one. We have probably 1,700 domains disavowed. We only have like 800 and some actual domains to our site-- legitimate linking domains. Basically, what should we do, keep disavowing and hope it's going to work.
JOHN MUELLER : You sent me that out on Google+ yesterday, right?
MALE SPEAKER : Yes.
JOHN MUELLER : Yeah. So I looked at your site briefly, and there was really nothing from those problematic links causing any problem. So that's something where if you're seeing negative changes, then they wouldn't be resulting from those bad links that you're seeing there. So either you're handling the disavow properly, or we've been able to kind of ignore those links anyway. So from that point of view, I think you're probably doing the right thing by cleaning up
[Discussion continued by saying don't just concentrate on links, but on the site overall]
Best Host News Commentary: As someone who has in the past been hit by a negative SEO attack this is exactly what we did, and we were never hit by the recent Penguin Penalty updates. We used Cognitive SEO (it is quite pricey at around $99), but it does simplify the process in diagnosing bad backlinks, whilst identifying many of your backlinks using data from Moz,
Question 13: 40:08 - Where does the Knowledge Graph take its data?
WRITTEN SUBMITTED QUESTION : From where does the Knowledge Graph take its data? Google's not showing the Knowledge Graph of my website. What can I do in order to appear with a Knowledge Graph?
JOHN MUELLER : For the Knowledge Graph, I think we use a variety of public data that's available. And that's something that we show algorithmically. It's not something that we'd show for every website, for every query. So just because that information is available doesn't mean that we'd always show it for your website, as well. So with that in mind, what I'd recommend doing is just primarily focusing on your website and making sure that people are seeing it as something important, that they're writing about it, that they're talking about your website, that we do have this additional information to back up that it would make sense to actually show a Knowledge Graph entry for that website.
Best Host News Commentary: This is
The Knowledge Graph was built to help with that mission . It contains information about entities and their relationships to one another – meaning that Google is increasingly able to recognize a search query as a distinct entity rather than just a string of keywords.
Question 14: 41:09 - How do you explain the extremely small changes in search results in the last three weeks?
WRITTEN SUBMITTED QUESTION : How do you explain the extremely small changes in search results in the last three weeks?
JOHN MUELLER : I don't really know what specifically you're kind of pointing at there, but fluctuations are essentially normal. And these are things that always happen with our algorithms, even if there is nobody manually flipping switches at the moment. So these are the kind of fluctuations that always can happen-- my guess. I don't really know what specifically you're pointing at. Maybe there's also technical issues that are sometimes involved on these websites.
Question 15: 41:46 - Use Keyword Based Traffic to Assess User Satisfaction?
JOSH BACHYNSKI : So you said a couple times in the past that in terms of the quality algorithms, the traffic we should be looking at in terms of their satisfaction should be mostly keyword traffic. And you've also said to me before that organic traffic is most important to look at in terms of how satisfied they are. Is that still the case? Should we still be focusing mostly on the keyword-based traffic?
JOHN MUELLER : I'm not really sure what you're referring to there.
JOSH BACHYNSKI : Not in terms of I know you guys don't use the Analytics bounce rate or anything in the Analytics package. But in terms of our seeing it from our perspective, what traffic do you think we should focus on there to make sure that they're satisfied? Because you used to say keyword
JOHN MUELLER : I don't know what I could best recommend there. So my suggestion would be to really take into account anything that you have available to kind of understand how happy users are with your website. So that's something where you can take into account whatever metrics you have available on your side, which could be based on analytics. It could be based on social mentions. It could be things like +1's or thumbs up, thumbs down-- those kinds of things on a site. I wouldn't focus on any specific metric from that point of view. So anything that you can do to figure out which parts of your site are really well-received by your users, where they think this is great content, and which parts of your site are not so well-received by your users, that's the kind of thing where I would focus on. That's the kind of thing where our algorithms should be able to pick that up on, as well. So instead of focusing on specific metrics and saying only people coming through with keyword traffic from organic search results are the ones that we need to make happy-- I'd really think about what you can do overall on your website to find the parts of your site that work really well, improve the parts that don't work so well, and make sure that everyone who kind of comes to your site leaves as a happy user and is willing to kind of recommend your site to other people, as well.
JOSH BACHYNSKI : So it makes more
JOHN MUELLER : I wouldn't worry about that from an SEO point of view. I think it helps a lot to understand your users and kind of which parts of your site work well and which parts don't. And indirectly, that's something that we would try to pick up on, as well. But it's not that we primarily look at the users coming to your website and say, oh, these all typed the URL that was in a TV spot. And therefore, they're not as relevant or they're really relevant. It's something where I think internally you get a lot of information out of that, and indirectly, that can lead to us recognizing which parts are relevant on your website. But that's not something where we'd see a direct relationship between the traffic that you drive to your website through a TV spot and how we would be able to rank that, or that site in general, or maybe a related site to that. So a lot of times people have a different domain for a TV campaign because they want something short and memorable that has a specific landing page, and that's not something we'd be able to take into account from search anyway. So if these are things that are outside of search, then that's not something where I'd say there's a primary direct effect available for us to even look at there. But it does always give you more information about how people are looking at your site, and maybe how other people are looking at your site, and which parts you could improve on, which would, then, be something that we'd be able to kind of look into.
Question 16: 46:35 - Discussion about the "Wrong" type of traffic.
JOSH BACHYNSKI : That actually kind of goes back--dare I say-- to the comments question, because [LAUGH] there's a quality of traffic, right. There's very interested traffic that's going to love your site, or have a better percentage chance of loving your site, and then there's traffic which is at best uninterested and at worst cruel. And so who do we have to cater to? You know what I mean? Sometimes it doesn't make any sense to try to cater to that second
JOHN MUELLER : Good question. Yeah.I don't really have a direct answer for that one. Yeah. So I think this is something where it's always a balance between which people we send to your site through search and how you respond to that. And sometimes you might say, well, Google is sending me people who I don't really want on my site because they're all idiots, or because they actually don't like what I'm publishing, and I prefer to have people who kind of like what I'm publishing. And that's something where-- at least at the moment, there's no direct feedback loop where you could say, hey, Google, stop sending me these types of people, and send me more of these types of people. So it's kind of a tricky situation there. But I don't really have any, let's say,
[Discussion continued with one suggestion saying write more specific content, rather than generic content that might get more diverse traffic]
Question 17: 50:04 - Question about Hidden Tabbed Content. Is tabbed content bad for SEO?
MALE SPEAKER : Hi, John. I have a question about page layout.
JOHN MUELLER : OK.
MALE SPEAKER : I'd like to know-- I've seen some content organized with tabs with CSS. In order to hide some tabs, and showing only one of them, like the tabs that are seen. Is it a good solution or a bad solution in order to show a lot of content without making a truly big website, web page? What do you think about?
JOHN MUELLER : There
Best Host News Commentary: This is one that actually caught us out when we used a Review Plugin (they subsequently changed how they displayed the reviews after we brought it to their attention ) that stored the reviews in tabs. It was only after the loss of some rankings that we discovered by chance that the reviews left on our site were not indexed in Google. The wider public caught on to this around November 2014, and you can see a couple of posts on the issue here and here .
Question 18: 52:05 - A question about posts with large header images.
MALE SPEAKER : Now it's very typical having big headers with big images, like that you can see in media websites. And most of the content is below the fold, or mostly below the fold. What do you think about that? Because pretty nice design, but I don't know what Google thinks about that.
JOHN MUELLER : That's usually fine. That's not something where I'd say you have to worry about that. I'd only worry about something like that if the images are really unrelated to the primary content. So if the image is an ad, for example, if it's something from another website, if it's something pointing to other articles on this website, then the user will land on that, and they'll be kind of lost. But if the image is a part of the primary content where you say, going to this page and kind of having this nice header image on top encourages the user to kind of read the rest of the page, and it all matches the same general topic, the primary content, then that's something that is perfectly fine. That's not something I'd discourage.
Best Host News Commentary: This is quite interesting, as we removed our large
Question 19: 53:20 - Can you tell us more about what's happening in Webmaster Tools next year
WRITTEN SUBMITTED QUESTION : Can you tell us more about what's happening in Webmaster Tools next year?
JOHN MUELLER : We try not to preannounce too many things, so I don't really have that much more to let you know about that there. I think you'll see more and more happening with mobile. That's something that I believe is a really important topic. That's something that I think a lot of websites still aren't getting completely right. So that's obviously one area that you'll see more happening there. We also talked about the Search Query Report, I believe, in one of the last Hangouts, where we're working on some changes there. So I'd imagine in the course of the year you'd see some changes there. But lots of things can still happen, and depending on the feedback that we get from you guys, obviously we can kind of tailor the future as well and say, well, this is a more important topic than other things, so we should focus more on that. But otherwise, I don't really have anything interesting that I could preannounce for you guys. What I'll try to do where we can do that is to kind of provide beta versions for people if we have something that we'd like users to try out before we make it into a final version. I don't know how easily we'll be able to do that with anything that's immediately coming up, but I think that's something that's always really useful
Question 20: 54:57 - Does Panda or Penguin penalty trigger a mass crawl of your site?
BARRY : John, at the eight-minute mark, Josh asked you about Googlebot being triggered to crawl more of a site based off of an algorithm. If that's the case-- that Googlebot can be triggered, say, crawl more based off of the Panda
JOHN MUELLER : I don't think we do that. I don't think we do that based on those specific algorithms. Obviously, we have lots of algorithms, and some of these are algorithms specifically for crawling and indexing. So if we see, for instance, specific changes on a site, then we might start re-crawling that. And that's obviously something where if you make bigger quality changes on your website, then we'll see that and say, well, lots of things have changed on this website. Maybe we should take a look at it a little bit better. So maybe there is some kind of correlation there that, if you make specific changes on your website, then we'll start crawling more to kind of recognize those changes. But it wouldn't be specific to the quality algorithms where we'd say, well, your site has-- I don't know-- some Panda on it, therefore we'll crawl it more frequently or we'll crawl it less frequently. It's usually more of a technical issue where we say, well, things have changed on this website, therefore we should be crawling it to make sure that we pick up those changes.
Best Host News Commentary:
Question 21: 56:22 - Is Google Panda a penalty?
MALE SPEAKER : John, a quick follow-up on Panda in general. I was actually asking in the group chat, as well. Would Barry's issue be something like a penalty? So would you consider Panda actually a penalty? Or do you see a problem with the
JOHN MUELLER : Yeah. So we don't see the Panda or the Penguin algorithms as penalties, internally. We see these as essentially part of our search quality algorithms to kind of figure out which pages are relevant to specific queries. So that's not something where we'd say we're penalizing a site for being low quality. But rather, our algorithms are looking at these pages and saying, well, overall, maybe we were ranking this site or these pages in a way that we shouldn't have in the past, and we're kind of adjusting that ranking. Or we're adjusting the ranking so that we can make sure that other pages that we think might be higher quality content are more visible in the search results. So it's not that we would see this as a penalty, but rather as something just a part of our search algorithms that adjust the ranking based on the perceived relevance for these individual queries.
MALE SPEAKER : OK. So basically most are a relevance problem rather than the site is bad, there's something wrong with it, problem.
JOHN MUELLER : Yeah.
MALE SPEAKER : Interesting.
Best Host News Commentary: Well, they can call it whatever they like, but we and many others still think it is a penalty! But of course, it is only natural Google will take the less controversial term "re-adjustment".
Question 22: 58:10 - Discussion about an old penalty hit site's chance to recover
LYLE ROMER : Hey, John. For the last question of the year-- from me,at least-- if somebody basically is competitive, highly ranking on competitive keywords and getting good natural links along with the other people in the space, and then that site is suppressed by one of the quality algorithms for a year, a year and a half, two years, at which point during that period of time they're not gaining those organic links because they're not able to be found, once they're no longer suppressed and they're back kind of near the top, is it ever possible for that site to kind of catch up with the natural high-quality links of the other sites that were never suppressed?
JOHN MUELLER : Sure.Yeah. I don't see a problem with that. So I mean, things always change on the web. So if you're talking about a period of more than a year,
Best Host News Commentary: Well John Mueller in a previous Hangout has already said this:
If you've had a manual action on your website and that's been revoked, then essentially there's no bad history attached to your site . It's not harder to rank anymore….It's not the case that there is any kind of a grudge that our algorithms would hold against a site that has had a manual action.
If you have disavowed links, then it is always possible that you have removed some that Google was still counting as positive, although Google would like you to think that bad links had already been discounted anyway. For a detailed commentary about
Question 23: 60:06 - Site Discussion - Indicates sometimes Google needs to change their algorithm
MALE SPEAKER : John, since we've just finished on the sites that haven't ranked for 18 months to two years question, is there any chance you can give me an update on what's happening with ours?
JOHN MUELLER : I know some engineers are looking at it.
MALE SPEAKER : We saw a very slight increase in
JOHN MUELLER : I'd have to take a look at the details there. [...] I don't have anything specific. I know the engineers said they would look into your specific case to see what was happening there. And the last I heard, they were kind of looking at that as something that we need to improve on our side, as well. So from that point of view, I think it's on the right track, but I don't really have an update on how things will change there.
[Conversation continues for a minute or too, but no new information]
Best Host News Commentary: Well, it just goes to show it can be worth sending John Mueller a message on Google+ ( https://plus.google.com/+JohnMueller ) if you think something is not quite right with the Rankings.