On May 20th, 2014 Google’s Matt Cutts announced that Panda 4.0 was rolling out. Leading up to that tweet, there was a lot of chatter across the industry about an algorithm update rolling out (based on reports of rankings volatility and traffic gains/losses). I was also seeing lots of movement across clients that had been impacted by previous algorithm updates, while also having new companies contact me about massive changes in rankings and traffic. I knew something serious was happening, but didn’t know exactly what it was. I thought for a while that it could be the pre-rollout and testing of Penguin, but it ended up being a new Panda update instead.
When Panda 4.0 was officially announced, I had already been analyzing sites seeing an impact (starting on Saturday May 17th, 2014). I was noticing major swings in rankings and traffic with companies I’ve been helping with previous algo trouble. And like I said above, several companies started reaching out to me via email about new hits starting that weekend.
And I was glad to hear a confirmation from Matt Cutts about Panda 4.0 rolling out. That enabled me to hone my analysis. I’ve mentioned in the past how unconfirmed Panda updates can drive webmasters insane. When you have confirmation, it’s important to analyze the impact through the lens of a specific algorithm update (when possible). In other words, content quality for Panda, unnatural links for Penguin, ad ratio and placement for Top Heavy, etc.
And by the way, since Google named this update Panda 4.0, we must assume it’s a new algorithm. That means new factors could have been added or other factors refined. Needless to say, I was eager to dig into sites that had been impacted to see if I could glean any insights about our new bamboo-eating friend.
Digging into the Panda 4.0 Data (and the Power of Human Barometers)
I’ve written before about the power of having access to a lot of Panda data. For example, working with many sites that had been previously impacted by Panda. It’s often easier to see unconfirmed Panda updates when you can analyze many sites impacted previously by the algorithm update. I’ve helped a lot of companies with Panda hits since February of 2011 when Panda first rolled out. Therefore, I can often see Panda fluctuations, even when those updates aren’t confirmed. That’s because I can analyze the Panda data set I have access to in addition to new companies that reach out to me after getting hit by those Panda updates. The fresh hits enable me to line up dates with Panda recoveries to better understand when Google rolls out unconfirmed updates. I’ve documented several of the unconfirmed updates here on my blog (in case you wanted to go back and check the dates against your own data).
So, when Google announced Panda 4.0, I was able to quickly start checking all the clients I have helped with Panda recovery (in addition to the ones I was already seeing jump in the rankings). And it didn’t take long to see the impact. A number of sites were clearly being positively impacted by P4.0.
Then, I analyzed new sites that were negatively impacted, based on those companies reaching out to me after getting hit (starting on 5/17/14). Together, I have been able to analyze a boatload of Panda 4.0 data. And it’s been fascinating to analyze.
I have now analyzed 27 websites impacted by Panda 4.0. The sites I analyzed ranged from large sites receiving a lot of Google Organic traffic (1M+ visits per month) to medium-sized ecommerce retailers and publishers (receiving tens of thousands of visits per month) to niche blogs focused on very specific topics (seeing 5K to 10K visits per month). It was awesome to be able to see how Panda 4.0 affected sites across industries, categories, volume of traffic, etc. And as usual, I was able to travel from one Panda 4.0 rabbit hole to another as I uncovered more sites impacted per category.
What This Post Covers – Key Findings Based on Heavily Analyzing Websites That Were Impacted by Panda 4.0
I can write ten different posts about Panda 4.0 based on my analysis over the past few days, but that’s the not the point of this initial post. Instead, I want to provide some core findings based on helping companies with previous Panda or Phantom hits that recovered during Panda 4.0. Yes, I said Phantom recoveries. More on that soon.
In addition, I want to provide findings based on analyzing sites that were negatively impacted by Panda 4.0. The findings in this post strike a nice balance between recovery and negative impact. As many of you know, there’s a lot you can learn about the signature of an algorithm update from fresh hits.
Before I provide my findings, I wanted to emphasize that this is simply my first post about Panda 4.0. I plan to write several additional posts focused on specific findings and scenarios. There were several websites that were fascinating to analyze and deserve their own dedicated posts. If you are interested in learning about those cases, the definitely subscribe to my feed (and make sure you check my Search Engine Watch column). There’s a lot to cover for sure. But for now, let’s jump into some Panda 4.0 findings.
Panda 4.0 Key Findings
The Nuclear Option – The Power of Making Hard Decisions and Executing
When new companies contact me about Panda, they often want to know their chances of recovery. My answer sometimes shocks them. I explain that once the initial audit has been completed, there will be hard decisions to make. I’m talking about really hard decisions that can impact a business.
Beyond the hard decisions, they will need to thoroughly execute those changes at a rapid pace (which is critically important). I explain that if they listen to me, make those hard decisions, and execute fully, then there is an excellent chance of recovery. But not all companies make hard decisions and execute thoroughly. Unfortunately, those companies often sit in the grey area of Panda, never knowing how close they are to recovery.
Well, Panda 4.0 reinforced my philosophy (although there were some anomalies which I’ll cover later). During P4.0, I had several clients recover that implemented HUGE changes over a multi-month period. And when I say huge changes, I’m talking significant amounts of work. One of my Panda audits yielded close to 20 pages of recommendations in Word. When something like that is presented, I can tell how deflated some clients feel. I get it, but it’s at that critical juncture that you can tell which clients will win. They either take those recommendations and run, or they don’t.
To give you a feel for what I’m talking about, I’ve provided some of the challenges that those clients had to overcome below:
- Nuking low-quality content.
- Greatly improving technical SEO.
- Gutting over-optimization.
- Removing doorway pages.
- Addressing serious canonicalization problems.
- Writing great content. Read that again. :)
- Revamping internal linking structure and navigation.
- Hunting down duplicate content and properly handling it.
- Hunting down thin content and noindexing or nuking it.
- Removing manual actions (yep, I’ve included this here).
- Stop scraping content and remove the content that has been scraped.
- Creating mobile friendly pages or go responsive.
- Dealing with risky affiliate marketing setups.
- Greatly increasing page speed (and handling bloated pages, file size-wise).
- Hunting down rogue risky pages and subdomains and properly dealing with that content.
- And in extreme cases, completely redesigning the site. And several of my clients did just that. That’s the nuclear option by the way. More about that soon.
- And even more changes.
Now, when I recommend a boatload of changes, there are various levels of client execution. Some clients implement 75% of the changes, while some can only implement 25%. As you can guess, the ones that execute more have a greater chance at a quicker recovery.
But then there are those rare cases where clients implement 100% of the changes I recommend. And that’s freaking awesome from my standpoint. But with massive effort comes massive expectations. If you are going to make big changes, you want big results. And unfortunately, that can take time.
Important Note: This is an incredibly important point for anyone dealing with a massive Panda or Penguin problem. If you’ve been spamming Google for a long time (years), providing low-quality content, that’s over-optimized, using doorway pages to gain Google traffic, etc., then you might have to wait a while after changes have been implemented. John Mueller is on record saying you can expect to wait 6 months or longer to see recovery. I don’t think his recommendation is far off. Sure, I’ve seen some quicker recoveries, but in extreme spamming cases, it can take time to see recovery.
Fast forward to Panda 4.0. It was AWESOME to see clients that made massive changes see substantial recovery during P4.0. And several of those clients chose the nuclear option of completely redesigning their websites. One client is up 130% since 5/17, while another that chose the nuclear option is up 86%. Here’s a quick screenshot of the bump starting on 5/17:
Side Note: The Nuclear Option is a Smart One When Needed
For some of the companies I was helping, there were so many items to fix that a complete redesign was a smart option. And no, that doesn’t come cheap. There’s time, effort, resources, and budget involved versus just making changes to specific areas. It’s a big deal, but can pay huge dividends down the line.
One client made almost all of the changes I recommended, including going responsive. The site is so much better usability-wise, content-wise, and mobile-wise. And with Panda 4.0, they are up 110% since 5/18 (when they first started seeing improvement).
I’ve mentioned before that for Panda recovery, SEO band-aids won’t work. Well, the clients that fully redesigned their sites and are seeing big improvements underscore the point that the nuclear option may be your best solution (if you have massive changes to make). Keep that in mind if you are dealing with a massive Panda problem.
Phantom Victims Recover
On May 8th, 2013, I picked up a significant algorithm update. After analyzing a number of websites hit by the update, I decided to call it “Phantom”. It simply had a mysterious, yet powerful signature, so Phantom made sense to me. Hey, it stuck. :)
Phantom was a tough algorithm update. Some companies lost 60% of their traffic overnight. And after auditing a number of sites hit by Phantom, my recommendations were often tough to hear (for business owners). Phantom targeted low-quality content, similar to Panda. But I often found scraped content being an issue, over-optimized content, doorway pages, cross-linking of company-owned domains, etc. I’ve helped a number of Phantom victims recover, but there were still many out there that never saw a big recovery.
The interesting part about Panda 4.0 was that I saw six Phantom victims recover (out of the 27 sites I analyzed with previous content quality problems). It’s hard to say exactly what P4.0 took into account that led to those Phantom recoveries, but those victims clearly had a good day. It’s worth noting that 5 out of the 6 sites impacted by Phantom actively made changes to rectify their content problems.
One of the sites did nothing to fix the problems and ended up recovering anyway. This could be due to the softening of Panda, which is definitely possible. There were definitely some sites I analyzed that showed increases after Panda 4.0 that didn’t necessarily tackle many problems they were facing. But in this situation, the site was a forum, which I cover next. Note, you can read my post about the softening of Panda and what I saw during the March 24, 2014 Panda update to learn more about the situation.
Forums Rebound During Panda 4.0
My next finding was interesting, since I’ve helped a number of forums deal with previous Panda and/or Phantom hits. I came across four different forums that recovered during Panda 4.0. Three were relatively large forums, while one was a smaller niche forum run by an category expert.
One of the larger forums (1M+ visits per month) made a boatload of changes to address thin content, spammy user-generated content, etc. They were able to gut low-quality pages, noindex thinner ones, and hunt down user-generated spam. They greatly increased the quality of the forum overall (from an SEO perspective). And they are up 24% since Panda 4.0 rolled out.
A second forum (1.5M visits per month) tackled some of the problems I picked up during an audit, but wasn’t able to tackle a number of items (based on a lack of resources). And it’s important to know that they are a leader in their niche and have some outstanding content and advice. During my audit I found they had some serious technical issues causing duplicate and thin content, but I’m not sure they ever deserved to get hammered like they did. But after Panda 4.0, they are up 54%.
And the expert-run forum that experienced both Panda and Phantom hits rebounded nicely after Panda 4.0. The site has some outstanding content, advice, conversations, etc. Again, it’s run by an expert that knows her stuff. Sure, some of the content is shorter in nature, but it’s a forum that will naturally have some quick answers. It’s important to note that the website owner did nothing to address the previous Panda and Phantom problems. And that site experienced a huge uptick based on Panda 4.0. Again, that could be due to the softening of Panda or a fix to Panda that cut down on collateral damage. It’s hard to say for sure. Anyway, the site is up 119% since May 17th.
Industry Experts Rise
During my research, I saw several examples of individual bloggers that focus heavily on niche areas see nice bumps in Google Organic traffic after Panda 4.0 rolled out. Now, Matt Cutts explained Google was looking to boost the rankings of experts in their respective industries. I have no idea if what I was seeing during my research was that “expert lift”, but it sure looked like it.
Here’s an example of a marketing professional that saw a 38% lift after Panda 4.0:
And here’s a sports medicine expert that has shown a 46% lift:
It was great to see these bloggers rise in the rankings, since their content is outstanding, and they deserved to rank higher! They just didn’t have the power that some of the other blogs and sites in their industries had. But it seems Google surfaced them during Panda 4.0. I need to analyze more sites like this to better understand what’s going, but it’s worth noting.
Update: I reached out to Matt Cutts via Twitter to see if Panda 4.0 incorporated the “authority” algo update I mentioned earlier. Matt replied this afternoon and explained that they are working on that independently. So, it doesn’t seem like the bloggers I analyzed benefited from the “authority” algo, but instead, benefited from overall quality signals. It was great to get a response from Matt. See screenshot below.
An Indexation Reality Check – It’s Not The Quantity, But the Quality That Matters
After conducting a laser-focused Panda audit, it’s not uncommon for me to recommend nuking or noindexing a substantial amount of content. That is usually an uncomfortable decision for clients to make. It’s hard to nuke content that you created, that ranked well at one point, etc. But nuking low-quality content is a strong way to proceed when you have a Panda problem.
So, it was awesome to see clients that removed large amounts of content recover during Panda 4.0. As an extreme example, one client removed 83% of their content from Google’s index. Yes, you read that correctly. And guess what, they are getting more traffic from Google than when they had all of that low-quality and risky content indexed. It’s a great example about quality versus quantity when it comes to Panda.
On the other hand, I analyzed a fresh Panda 4.0 hit, where the site has 40M+ pages indexed. And you guessed it, it has serious content quality problems. They got hammered by Panda 4.0, losing about 40% of their Google organic traffic overnight.
If you have been impacted by Panda, and you have a lot of risky content indexed by Google, then have a content audit completed now. I’m not kidding. Hunt down thin pages, duplicate pages, low-quality pages, etc. and nuke them or noindex them. Make sure Google has the right content indexed.
Engagement and Usability Matter
While analyzing the fresh hits, it was hard to overlook the serious engagement issues I was coming across. For example, stimulus overload on the pages that were receiving a lot of Google organic traffic prior to the hit. There were ads that expanded into or over the content, double-serving of video ads, stacked “recommended articles” on the page, lack of white space, a long and confusing navigation, etc. All of this led to me wanting to bounce off the page faster than a superball on concrete. And again, high bounce rates and low dwell times can get you killed by Panda. Avoid that like the plague.
Check out the bounce rates and pages per session for a site crushed by Panda 4.0:
Side Note: To hunt down low-quality content, you can run this Panda report in Google Analytics. My post walks you through exporting data from GA and then using Excel to isolate problematic landing pages from Google Organic.
Downstream Matters
While analyzing fresh Panda 4.0 hits, it was also hard to overlook links and ads that drove me to strange and risky sites that were auto-downloading software, files, etc. You know, those sites where it feels like your browser is being taken over by hackers. This can lead to users clicking the back button twice and returning to Google’s search results. And if they do, that can send bad signals to Google about your site and content. In addition, risky downstream activity can lead to some people reporting your site to Google or to other organizations like Web of Trust (WOT).
And as I’ve said several times in this post, Panda is tied to engagement. Engagement is tied to users. Don’t anger users. It will come back to bite you (literally).
Summary – Panda 4.0 Brings Hope
As I said earlier, it was fascinating to analyze the impact of Panda 4.0. And again, this is just my first post on the subject. I plan to write several more about specific situations I’ve analyzed. Based on what I’ve seen so far, it seems Panda 4.0 definitely rewarded sites that took the time to make the necessary changes to improve content quality, engagement, usability, etc. And that’s awesome to see.
But on the flip side, there were sites that got hammered by P4.0. All I can say to them is pull yourself up by your bootstraps and get to work. It takes time, but Panda recovery is definitely possible. You just need to make hard decisions and then execute. :)
GG
I thought it was not a “bamboo-eating friend” but Mr. Navneet of Google that ‘Panda’ refers to …
Just saying :-)
Of course, but I’m just referring to the algo! :) And we don’t even know if Navneet is still working on the Panda algorithm. For all we know, he’s working on the next algo in the Google lab.
Quick question about forums: one of my sites has an Invision Board Script 3.x and it’s a pretty old board ~ 100 000 UV/day.
No spam, is well moderated and I’ve blocked the bot and guest access to profiles and other areas since about a month.
No signatures for bots/guests.
So, only the POSTED content is indexable, but I got hit by 20% after Panda 4.0.
Can you tell me what to do? I do not have “special” thin content to drop anymore.
I Know that IPB has schema.org markup problems and I’ve removed the markup.
Interesting, can you email me the domain? I can take a quick look. There are a lot of moving parts with Panda, and many times there are gremlins causing problems behind the scenes. You can use the email address on my contact page. Thanks.
Mail sent, thank you!
Great post Glenn. Have you had time to jump into eBay yet?
Like you, I’m anxiously digging through previous client work and analyzing top losers such as Ask.com, History.com, etc. Engagement and usability seems to have ratcheted up this round.
Thanks Art. I’ve focused on analyzing clients I’ve been helping with Panda (and Phantom), and then companies that have reached out to me with confirmed hits based on Panda 4.0. That said, I have started to look at RetailMeNot and ebay.
I definitely plan to publish additional posts in the near future based on what I’m finding (probably about specific scenarios and cases I’ve come across). I had to cut this post down because it was getting too long, and there are still some important findings to share! Stay tuned.
If you write a post about what you’ve found with any of the sites you listed above, definitely let me know. I’d be interested in checking out your analysis! Thanks again.
Thanks Glenn, definitely an informative piece. Could you elaborate further on bounce rate. I manage several blogs with high bounce rates e.g over 90% in some cases but time on site over 3 minutes. These are posts that address common questions so I expect those kind of results. Would the lack of interaction be hurting these sites? One was hit badly losing approx. 75%
Great question Marty and I’ve written several posts about bounce rate and SEO. Standard bounce rate is a flawed metric. It doesn’t take time on page into account. And time on site only takes non-bounce visits into account. So, webmasters are stuck looking at poor metrics.
That’s where adjusted bounce rate comes in handy. It takes time on site into account and can give you a stronger feel for actual bounce rate. You can read my post about ABR here -> http://www.hmtweb.com/marketing-blog/adjusted-bounce-rate-in-google-analytics/
And if you are using Google Tag Manager, then follow my tutorial on Search Engine Watch -> http://searchenginewatch.com/article/2322974/How-to-Implement-Adjusted-Bounce-Rate-ABR-via-Google-Tag-Manager-Tutorial
Blog posts will naturally have a higher bounce rate. E.g. Someone searches for a specific topic, finds a post, reads that post (hopefully for a decent amount of time), and then they leave. But if they engaged with the content, then that’s totally fine (even if it’s a standard bounce). Understanding adjusted bounce rate can tell you if people are hitting that post and quickly leaving. And that can be a very bad thing SEO-wise. Low dwell time in aggregate can be a killer Panda-wise. I hope that helps!
Thanks Glenn, I most definitely agree re: bounce rate being a flawed metric. Time on page should be an indication that the search provided a good result.
I’ll dig into those links, especially interested in adjusted bounce rate. You have only recently come onto my radar thanks to G+ so have some reading to catch up on!
One last point re: the site I mentioned previously. They client updated their theme a week prior to Panda, the theme update had an incorrect variable in the comments.php page which caused the replyto’s to become indexed. We worked to fix the issue as soon as we noticed it but it was unfortunate timing as Panda rolled out within that 3 day window. Terrible luck on the client’s end but I believe they can rest assured their content wasn’t the problem.
Thanks again.
Thanks Marty. I’m glad G+ led you to my blog! Let me know if you have any questions as you go through my posts about Panda, Penguin, etc.
Regarding the wordpress coding error, what was getting indexed that wasn’t supposed to (specifically)? I just want to make sure I understand what was happening, and what the result was content-wise. And if you want to email me the domain so I can take a quick look, you can use the address on my contact page. Thanks again Marty.
Solid advice Glenn. I especially appreciate the real examples of both positive and negative impacts. Your advice, though painful to digest, then implement for some, is paying off for the clients that invest in taking the necessary steps for Panda 4.0 recovery. Do you think Panda 4 was in any way responsible for or related to ebay penalty? I know that it is reported that they had manual penalty in addition to or perhaps instead of. Thanks for sharing.
Thanks Rick. I’m glad you found my post valuable. And yes, the nuclear option seemed to really help clients with severe Panda problems. They implemented a lot of changes over a relatively long period of time, but it paid off big-time.
Regarding ebay, I need to analyze the situation way more than I have already. I’ve been focused on clients that recovered during Panda 4.0 and also fresh hits. I have started looking into RetailMeNot and ebay, but not enough to come to any conclusions yet.
I will tell you, though, that the timing for ebay lines up perfectly with Panda 4.0. I see their first dip starting on Monday, May 19th (when many others first started seeing the impact of Panda 4.0). It could be a manual penalty, but the trending for ebay screams Panda. I’ll let you know if I dig in further.
Thanks Glenn. I appreciate your feedback.
Hi nice info on panda fix. Please check the mail i have recently sent to you.
Will do! Thanks Rijo.
An impressive analysis of the Panda 4.0 update! Many points were raised and some metrics were revealing.
If Panda 4.0 is going to bring hope then I guess webmasters really should not worry about many updates.
I am still of the opinion that despite careful link building strategies, Google has a way to make itself the final decision maker with link building.
I have left this comment in kingged.com – the content syndication and social marketing platform for Internet marketers, where this post was shared.
Sunday – kingged.com contributor
http://kingged.com/panda-40-analysis-nuclear-option-rewarded-phantom-victims-recover-industry-experts-rise/
Thanks Sunday. Glad you enjoyed my post. Stay tuned. I have more posts coming based on my latest analyses. You should definitely subscribe to my feed if you are interested in algorithm updates, Panda, Penguin, etc.
Glenn – I must say I am really impressed. Finally an in-detail and well explained case study of Panda 4.0. It was really a pleasure reading it!
Are you publishing somewhere your panda update dates? That would be great to confirm what I got from customers with you!
Thanks Bartosz. I’m glad you found my post valuable! It was fascinating to analyze sites impacted by Panda 4.0. I’ve now analyzed over 30 sites impacted and plan to write more posts in the coming weeks.
Regarding the unconfirmed Panda updates I’ve seen, I have written several posts documenting them, but I don’t have a central page detailing those updates. And I should. :)
In the meantime, you can check the following posts out:
http://www.hmtweb.com/marketing-blog/panda-update-january-2014/
http://www.hmtweb.com/marketing-blog/softer-panda-march-2014/
http://www.hmtweb.com/marketing-blog/google-algorithm-updates-february-2014/
Thank you Glenn – I will definitely read those. Glad I’ve found somebody to explain Panda problems similar to how I do that with customers. Also I share similar problems (1 out of 10 customers is implementing all the changes).
Hi Mr Glenn, My website is suffering drop in google search engine for all pages from google panda up date on October 2013, i did hard work (off page seo) but almost no recovery, i read your wonderful article and increased some thing content in few pages (Category pages), however i think i have more problems, can you please take a look and advise me!!
website: http://www.gosharmtours.com
This is a brilliant article. I just wanted to point out a very small error. You wrote,”It will come back to bit you (literally).” and I believe you meant “bite”. Thanks for publishing your panda research.
Thank you! I just fixed that. And glad you enjoyed my post. There’s more to come on the Panda front. I have several more posts coming. :)
Glenn, thanks for a great analysis of the Panda 4.0 update. Awesome stuff as usual.
When you mention nuking thin pages, duplicate pages and low-quality pages, what have you found is the most “Panda effective” way of getting rid of these pages: 301 redirecting them to some other page, 404ing them or 410ing them?
I have seen some success with returning 410 status codes, but wanted to hear your thoughts on what has worked best for you.
Thanks
If they are low-quality, thin, etc., then 404s or 410s work best. But it depends on the specific urls. For example, you might find a tech glitch that creates duplicate pages and maybe some of those pages are being linked to from outside the site. Then 301s could work. It’s hard to say without looking at the specific situation.
But overall, if you want the pages removed (to help with Panda), and you identify that they are low-quality, then 404 or 410 them. I hope that helps.
Add to the last pot: I would like to refer shortly that my website was hit exactly on 04 October 2013, so it must be by Penguin 2.1 !
Thank you and your support will be highly appreciated!
Good to know. Then you have a different black and white animal to deal with. I have several posts about Penguin. You should check them out. :)
Awesome insights Glenn. Definitely shed light to why I’m seeing increases in my (and my clients’) sites. Keep ’em coming.
Hey, thanks. I appreciate it. Glad you found my post valuable. And I hope you are seeing more increases than decreases! :) BTW, my Search Engine Watch column goes live today. I’m providing more insights based on analyzing another 14 websites impacted by Panda 4.0. You should check it out.
HI Glenn
Great Post, thanks for sharing your findings, I am currently dealing with a site that has been affected by the 4.0 update, I think a large part of the problem may be the forum pages.
And also the poor/thin quality content on site.
I would be most interested to know what your audit process steps are for evaluating whether a site has been hit by an algorithm update.
Could this be a future post, or indeed if you have already placed such a detailed audit post, where could I find it please.
Thanks Glenn.
Hi John. I’m sorry you were impacted by Panda 4.0. Regarding my audit process, I complete a heavy analysis of websites impacted by Panda through the lens of the algo update. That includes heavy manual analysis and analysis of multiple crawls. There’s not a list of steps… the data takes me down various paths (which often lead to Panda problems). I hope that helps.
Hey Glenn. My website (www.3rank.com) got hit by panda pretty hard. I lost 50% traffic.
I have removed all the tags, attachments and category pages from Google index. Can you tell me, how much time it usually takes to recover.
I’m sorry to hear you were hit by Panda 4.0. Regarding timeframe, Panda rolls out monthly, so you should have a shot of recovery once per month. But, you might not see full recovery all in one shot. It obviously depends on how bad of a situation you had in the first place. I haven’t analyzed your site, so it’s hard for me to say. I hope that helps.
Hi Glenn, could you please provide more details for the point “Dealing with risky affiliate marketing setups” ? From your point of view, which aff setups are especially risky when it comes to Panda ? Thanks in advance.
Depends on how content is pulled in and displayed. If you look at Glenn’s post above and the companion article he wrote below, there are hints at what needs to be done depending on where the content it is living/how it is displayed. Utilizing noindex, nofollow, canonical tags, avoiding thin pages, etc.
http://searchenginewatch.com/article/2348110/More-Panda-4.0-Findings-Syndication-User-Engagement-Indexation-Keyword-Hoarding
Thanks, I will try to implement suggested changes over the website..