Scraper is a pretty good Chrome extension I use on a regular basis to quickly extract links from a page. Unfortunately, there can be rare instances where it actually takes more effort to use.
For example, if I wanted to retrieve all links from Hewlett-Packard’s HTML sitemap, I would need to create multiple Google spreadsheets to capture that data because of the way the page is structured. In this particular case, I’d have to scrape the page a total of 14 times to account for the different sections.
When I joined SEOgadget last year, my first blog post was about using webkit2png for site audits, stalking and more. What I didn’t mention was my 2013 new years resolution – to track the home page of three websites for the entire year with webkit2png.
The following videos come from the home pages of Macy’s, Yahoo and Amazon with a years worth of images compiled together. It’s nothing too crazy but feel free to turn on your favorite jam, sit back, relax and view them for your pleasure.
Enjoy and have an amazing 2014! =]
If you write a technical blog post about optimizing source code for SEO or programming scripts, I highly recommend the Crayon Syntax Highlighter for WordPress users – it gives your examples a nice, snazzy look to it. The plugin includes 25 default themes but I couldn’t find a good preview gallery for them anywhere so I decided to list them all out below. Enjoy!
This is the "Arduino Ide" theme.
This is the "Classic" theme.
This is the "Eclipse" theme.
I updated my aHrefs bulk link analysis script to improve its functionality by adding two features.
- The script now returns the results in a CSV file called ahrefs_results.csv
- Introduces the .map Ruby enumerable for a “cleaner” syntax
The source code for this Ruby script can be found at my Github repository.
My next task is defining individual functions to eliminate any code redundancy and ultimately speed up the API calls. Stay tuned!
This post is super late but if you didn’t know already I left my short tenure with Macy’s earlier this month and joined the SEOgadget family!
You can read my first SEOgadget blog post here.
I’m helping out Laura Lippay expand the US office so I’ll be getting a taste of both agency and startup life. Business is already booming and I’ve got so much work ahead we are looking to hire another Organic Search Strategist. Yes, that’s right – I need a partner in crime!
Just like the natural world SEO traffic adheres to a Power Law, more commonly known as the long-tail or the 80-20 rule for you MBAs. Applying this to search, it means approximately 80% of your organic traffic is attributed to the top 20% of your keywords.
When you visualize this type of data though an inherent problem occurs…
…the number of visits for the head terms far exceed that of the torso and tail keywords, thus rendering the graph useless. And if you wanted a YoY comparison of your SEO performance?
No actionable insight – a fallacy of linear graphs.
One of my top blog posts this year is the bulk URL checker and has become my staple tool for HTTP checks en masse when I don’t want to fire up third party software. This accomplishment got me hooked (on coding) and to keep my momentum going I decided to write a Ruby script which interfaced with the aHrefs API and emulate their batch analysis tool.
If you’re only interested in the bulk link analysis script then you can find it here.
It may look simple but it took me a couple hours to complete and needs some cleaning up on my part. This file is a good start for anyone who needs a quick analysis about a list (big or small) of URLs and their links. Regardless, I’ll continue to build upon it so it resembles that of the aHrefs tool.
What Does It Do?
The script is built with the Ruby programming language and analyzes a list of URLs in bulk by accessing the aHrefs API then responds with the appropriate data (exact match).
The current version returns the target URL and the following link metrics to the page: total backlinks, linking root domains, unique IP addresses, .COM links, .EDU links and .GOV links. It will also indicate the remaining API calls in your account upon completion.
When I first started doing keyword research I found myself easily downloading 10+ CSV files from the Google Adwords Keyword Tool. This became a (painful) problem because I had to manually copy and paste all the data into a single Excel file.
Fast forward a few months when I finally found the perfect solution: using the CAT function (short for CONCATENATE) within the command line. Mac users can get started immediately with their Terminal (located in the Utilities folder); Windows users will have to download Cygwin.
Are you ready for the super secret code to painlessly aggregate all your CSV files into a single document? Here you go…
cat *.csv >> OUTPUT_FILE.csv