Google Adsense introduced the Super Large leaderboards ad unit measuring 970×90 px and promises to overtake the whole width of your website. It is a welcome addition to the Adsense family as large ads are popular and the new ad formats are more compatible with modern site design formats and sizes.
I always found the leaderboard 728×90 a little short of the usual blog widths and designers had to find new ways to fit them in by center aligning or adding thicker borders. Adsense had earlier released the large skyscaper 300×600 ad units and you must have read our recent review of the same.
970×90 Super Large Leaderboard
Brands advertisers will particularly find it a great addition as the leaderboard is usually placed on top of websites and such a large visual ad promises a huge impact and high CTR. As a newly adopted IAB standard, I am sure this ad will get more popular with time. As with all new Adsense ad sizes, the image ads inventory is low and they mostly show text ads. This large leaderboard is customized to show upto 4 text ads in columns.
I tested it on our website and right now the 970px wide design cannot fit on our pages unless we widen the site design. A single text ad is showing on most pages

As there is a lack of image ads inventory, they are filling the image ad size with 728×90 leaderboards.

Will you be using this large super leaderboard on your site?
Original article: New 970×90 Super Large Leaderboard Adsense Units
©2013 QuickOnlineTips (QOT). All Rights Reserved.
Robots.txt is blocking a huge number of urls on our site! We delved further into causes of falling traffic since Panda 25 and especially after Reconsideration request revealed no manual penalty. As the traffic continuously falls for the past few weeks by over 50% since March 15, we looked closely at Webmaster tools setting and found that robots.txt was blocking excessive urls and this block was continuously increasing every week such that very few pages were still left unblocked.

Bad Robots
While we were aware robots.txt was set to block some urls, but how many of these indexed urls are being blocked by robots.txt? The answer is provided by this key report in Google Webmaster tools at Index Status > Advanced View. This gives a great view of all the indexed urls and robots.txt blocked urls. This is what our chart looks like … scary!

As you can see over the past few weeks the robots.txt has continued to block a huge number of urls, and more urls are getting blocked every week. In fact at this rate all urls will be blocked out in the next few weeks. I am not sure when we last tweaked this file, maybe a few months back. Maybe there was some noindex tag somewhere, though we do not noindex categories or tags for many months now. This explains the falling traffic over the past few weeks as the number of urls in the Google index are dropping.
Why did the Robots.txt suddenly start blocking so aggressively? Is it a Googlebot problem?
Old Robots.txt
Over the years we have been tweaking the robots.txt in an attempt to reduce indexing of duplicate content and large number of URL parameters – to fix possible Panda issues. This is what our robots.txt looked like for the past few months, as suggested by some popular websites.
User-agent: Mediapartners-Google Disallow: User-agent: * Disallow: /cgi-bin/ Disallow: /wp-admin/ Disallow: /wp-content/ Disallow: /wp-includes/ Disallow: /author/ Disallow: /search/ Disallow: /*/trackback Disallow: /?q* Disallow: /*? Disallow: /*.css$ Disallow: /*/?replytocom
This is partly modified to not show some private folders/sitemap. We allow Google Image search full access to the WP uploads. We blocked WordPress folders access; disallowed author, search, trackback pages; and pages with url parameters like q for search, replytocom for reply comments pages, css and other ? parameter pages. We also blocked replytocom URL parameters in GWT.
Can you identify what is so wrong in this robots.txt file? Your feedback will be valuable.
New Robots.txt
This is our robots now changed a few days back, as we really do not know what is blocking the urls. We also removed all replytocom URL parameter blocks, as Google warns this could also deindex urls. We still block access to the wp-admin, wp-includes and the wp-content plugins folder. Though probably I do not need to do block them as well because robots must be smart enough to not index them anyway across millions of WordPress blogs.
User-agent: * Disallow: /cgi-bin/ Disallow: /wp-admin/ Disallow: /wp-content/plugins/ Disallow: /wp-includes/
Now we are waiting and watching for the next weekly update to show how the Index status changes. The Blocked urls page has already started showing a decline in the number of blocked urls! Great.
Big Lessons learnt – I just wished I had reviewed the Advanced Index report before. Still confused about the relationship between Panda 25, March 15 and robots blocks. Its also reminds webmasters that robots.txt is a powerful tool and be sure what you add there. Also take care before you decide to add noindex tags indiscriminately. What are your thoughts on this?
Update: wondering if it is a Googlebot access / server connectivity issue which people are reporting.
Update: Our Knownhost VPS hosting support investigated the issue and suggested maybe our server firewall CSF was temporarily blocking googlebot. They have added rules in csf.rignore to ensure that the firewall does not block googlebot, bing or yahoo. They seemed pretty sure this might be the issue and think this should fix the robots blocking issue. Earlier we had reported CSF can block site traffic, and I think all really need to see what the server firewall is blocking or allowing.
Original article: Bad Robots.txt Causes Massive Blocked Urls?
©2013 QuickOnlineTips (QOT). All Rights Reserved.
No comments:
Post a Comment