Thanks for sharing these tips, Brian. Agree with all of these, except maybe #3 Delete zombie pages. A better strategy would be to update these pages with fresh content and convert them into a long form blog posts/guides. Deleting them entirely would mean either setting up a 404 or 301 redirect – both of which can hurt your organic traffic in the short run.
Kelly Main is a staff writer at Fit Small Business specializing in marketing. Before joining the team, she worked as an analyst at firms like Lincoln Financial Group. She has also founded a number of successful startups, including OpenOnion under the Google Tech Entrepreneurs Program, which was later acquired under the name Whisper. She holds an MS in International Marketing from Edinburgh Napier University.
But sometimes there are site-wide technical issues that get in your way of ranking on Google. Luckily, fixing technical issues is not a required step for every single piece of content you create. However, as you create more and more content you should be aware of duplicate content, broken links, or problems with crawling and indexing. These issues can set you back in search results.
Hi Brian, great list. I noticed you mentioned using Kraken to optimize your images. A couple other tools I’ve found to work really well are ImageOptim (an app that you can download for your computer) and Optimus (a WordPress plugin that will optimize them when uploaded to your site). I’m not sure what your policy is on including links in comments so I won’t link to them here (even though I have no affiliate with either one.) Hope those resources can help someone!
Robots.txt is not an appropriate or effective way of blocking sensitive or confidential material. It only instructs well-behaved crawlers that the pages are not for them, but it does not prevent your server from delivering those pages to a browser that requests them. One reason is that search engines could still reference the URLs you block (showing just the URL, no title or snippet) if there happen to be links to those URLs somewhere on the Internet (like referrer logs). Also, non-compliant or rogue search engines that don't acknowledge the Robots Exclusion Standard could disobey the instructions of your robots.txt. Finally, a curious user could examine the directories or subdirectories in your robots.txt file and guess the URL of the content that you don't want seen.
Creating high quality content takes a significant amount of at least one of the following: time, effort, expertise, and talent/skill. Content should be factually accurate, clearly written, and comprehensive. So, for example, if you describe your page as a recipe, provide a complete recipe that is easy to follow, rather than just a set of ingredients or a basic description of the dish.
As an alternative to tackling this on your own, try using a service like Mayple to match you with your industry’s best marketing experts. Mayple is a platform that connects business owners to vetted marketing experts, so you can focus on running your business and delegate the rest to experienced professionals — all you need to do is fill out a brief explaining your business’ goals. It even monitors your project’s progress and ensures your expert always delivers the best results. Get started today.