Get page level Core Web Vital data from competitors in Search Console

Let’s say you have a lot of crawl budget assigned to your domain and you want to use it to benchmark your own Core Web Vital metrics against your direct competitors. Using Chrome UX report APIs is OKish for domain level metrics but you obviously want more data.  

Step 1: Crawl your competitors website to be able to map all URLs they have to get a full picture.

Step 2: Add a directory or subdomain on your own domain that contains all those URLs. So if the original URL is make something like

Step 3: Add 302 redirects of these URLs to the original competitors URL

Step 4: add a GSC profile for the specific competitor folder of URLs in, in this case I would add

Step 5: add a HTML sitemap containing all the URLs within the competitors folder so Google can actually find them organically by crawling oldskool. Buy some forum profile links, because yeah, we do oldskool SEO here.

Step 6: add a XML sitemap containing all those URLs to the newly created GSC profile so you have some visibility on how slow Google can be with crawling and indexing shit.

Step 7: once you had a burst of crawls from the ever happy Googlebot, head out to your CWS dashboards:

Why does this work?

Because with temporary 302 redirects, Google attributes everything from the final destination URL to original URL. That works with onpage elements like meta title and description, but also with reporting CWS data in Google Search Console. 302 redirect hijacking is not something new, so if you find it interesting, search some SEO blog archives for it.

What I learned about SEO from using the 10 most used JS frameworks #brightonseo

JavaScript will define and impact the future of most SEO consultants. A big chunk of websites has, is or will move over to a JS framework driven platform. Stack Overflow published an extensive study about the data gathered from an enquiry amongst more than 100.000 professional programmers’ most used Programming, Scripting and Markup languages: read more at Most Popular Technologies The outcome is quite clear, it’s all about JavaScript today:

But JavaScript and search engines are a tricky combination. It turns out there is a fine line between successful and disastrous implementations. Below I will share 10 tips to prevent SEO disasters to happen with your own or your clients sites.

Continue reading

Rendering on the Web – The SEO Version

Last week Google published a handy overview about Rendering on the Web in their developers section. It is aimed at developers needing to make a decision about the architecture they will use for their JavaScript driven applications. The main decision that affects logic within apps is the way rendering is set up. SSR, CSR, Rehydration and prerendering: they describe the different terminology in detail and align the advantages or disadvantages of using the different scenario’s. This overview is resulting in the following handy table:

Continue reading

How a medical site recovered from the Google Medic / EAT updates

  • Put all EAT related To-do’s on the backlog
  • Have a meeting with the developers
  • Developers are extremely busy with real improvements
  • Google will put the site back due to EAT-related To-do’s being present in the backlog
  • Happy holidays, Merry Christmas!

SEO Quick wins: finding meta data optimisation opportunities with Data Studio

Many of my (former) clients have great content teams working on optimizing websites. Sending them a spreadsheet of data with some fancy click through ratios for keywords or URLs that perform better or worse than other datapoints will not get them enthousiastic at all. Sending them a fancy interactive graph in Data Studio will get them to work (not a 100% success ratio!) on boring things like optimizing meta descriptions and titles. So how can you spend 5 minutes of your time on Data Studio and make your life as a marketer much easier? Follow the instructions below.

Continue reading