Captain's log, stardate d376.y39/AB
Quick entry today, after a few on the heavy side!
Sometimes, as an agency, you're forced to guesstimate the size of a project, in terms of number of pages, especially for migration projects. In most of these cases, the prospective client won't give you access to the code even if you have signed an NDA. Maybe they want to do it to replace their current provider, or they don't want to bother their internal technical team, so it's up to you to find out.
This kind of guesstimate is crucial in SEO-based projects, as you'll need as accurate a number as you can have in order to know where you will be putting your efforts later on. It's not the same to migrate a size with 1000 contents that one with 100000 pieces of unique content.
One quick trick is to take a look at the
sitemap.xml file(s) of the website.
For instance, let's suppose the popular tech portal TechCrunch comes to us and wants us to migrate their site. However, they don't give us access to the code nor database, and we need to submit a proposal for them with whatever information is publicly available.
For the client, SEO is extremely important but, at the same time, they might not have all the bandwidth available to help all the potential providers gather all the information, so the only instructions we receive are "I don't know how many contents I have, but they all have to be migrated". This happens more often than not, hence this blog post.
What we will do here is to open up this url: https://techcrunch.com/sitemap.xml (usually is always
/sitemap.xml) and see that in this project in particular, the sitemap is split into 2 sub-sitemaps. The first one contains 108 sitemap files, and the second one 184. This is a lot of content, but we're talking about one of the main sites for tech out there.
What we will do here, is count the number of items on each sitemap and we will have a number of how many contents are there on the site. If done properly, all sub-sitemap files have got the exact number of items, so you can do number of sub-sitemap files times the average of items on each file, and voilà!
Key idea: 🚨 Multiple sitemaps or sub-sitemaps mean LOTS OF CONTENTS 🚨
Actually, this answer is technically not 100% accurate: contents listed on sitemap files are only those that have got a URL. Unpublished or drafted contents will not appear here, but we can assume these won't be a significant number and negligible in the big scheme of things. You can always add a buffer to the guesstimate to deal with these later, or leave them for a subsequent phase once you will have taken control of the platform and you're more familiar with everything. Maybe they're not needed at all.
For the bonus points, if you open the
robots.txt file, you might be able to see disallowed contents for indexing robots, which also need to be taken into consideration in migrations because they need to be ported with the same level of access.
Robots.txt can also give other information as to routes of image or assets folders, so it's always worth taking a look into, if you're estimating a migration project.
Learn how we help our customers to reduce their time to market by prototyping on static pages, not on the real product.Read full article
MySQL and PostgreSQL are very similar but not exact. Take a look at this scenario that works with PostgreSQL but not with MySQL.Read full article
In this blog post, our CTO, Xavi, will show us how to query data from PostgreSQL to represent it in a time series graph.Read full article