For a simple quickscan of a random website, you can’t use the standard Fetch as Googlebot functionality without verifying a domain first. Since this is becoming more and more important, now Google is also looking to hidden layers of content, ads etcetera and you want to test smartphone robot too, you can make use of a simple workaround.
Make sure you create a simple clean HTML file on a domain you already own. In my case, I use notprovided.eu which is verified in my Google Webmaster Tools account. Within this HTML you can easily add a iframe or embed element, containing the URL you want to test with the Fetch As Google function.
Within WMT you fetch the URL which includes the iframe or embed section. This will show the external website, parsed by the selected Google Bot. Make sure the file is not blocked by robots.txt from crawling, so add a noindex tag. If you want to fetch a HTTPS url, you also need to have a HTTPS domain verified in WMT to including such an URL by making use of iframe or embed codes.
Many articles have been written about content marketing in the past years, online marketers are sometimes completely obsessed about content marketing. Because of several projects I did during the past 12 months and all those people using “Content is King” whenever possible, I would like to discuss some of the factors content marketing has an impact on. These factors are not commonly considered by everyone involved in online marketing campaigns.
Due to the fast-paced development of Google, where there has been more focus on the quality of content and websites by developing algorithmic filters like Panda, many websites have produced enormous amounts of textual content on their domains. A lot of websites are still targeting specific queries with unique pages instead of targeting a specific group of people. In the company of the latest updates of Google’s algorithms, where Panda is one of them, some problems start to occur. From a technical point of view, more content is not always positively influencing the organic results.
Search engines like Yandex, Google and Baidu are rapidly changing these days. More and more, engineers try to imitate human behaviour and try to use user experience as a signal, compared to the old fashioned link graph way of ranking pages in their indices. Not only have we seen interesting developments like Yandex Islands and Google’s Knowledgde graph, but also social is starting to play a growing role in the technology behind search. Human search behaviour has changed considerably.
Together with the crowd, I’ll explore the world behind search engines and try to create an understanding of how search engines are working, how the future of search will be shaped and how one as an affiliate marketeer can make use of it.
During my flight from Washington to London yesterday night, I was thinking about who I am. What makes me different from the other 200+ passengers? Thanks to Google’s additional Knowledge cards, Yandex Islands any many more. Because search engines added all kinds of additional data in their search engine result pages during the last years, we get more and more “quantified” and we are getting more conscious about who or what we are, especially on the web. The whole “From strings to things” concept is being visualized for the average person by companies like Google. Currently, Google shares more facts about me (I’m sure they know a lot more since I used their services since 2000), than the average number of facts my friends will remember about me. How can you not like science? Continue reading
Building the biggest knowledge graph, in layman terms an information database, of all time. That is what Google is researching at this moment. New details were presented during the annual “Knowledge Discovery in Databases (KDD) conference” in New York two weeks ago. Behind the screens of corporations like Microsoft, Facebook, Amazon and IBM are similar systems being developed. It’s a race towards the first database which archives all human knowledge.