The Importance of the Technical SEO
Technical SEO refers to the work that is done aside from content optimization. In technical SEO we usually focus on website crawling and indexing. Any website that has good technical SEO is crawled well and indexed fast. Technical SEO keeps an important role while ranking the website so it is must to know for every SEO’s and website owner to implement the Technical SEO activities completely for every website.
What are things included in Technical SEO?
Page Load Speed
This is very important part of technical SEO and most important activity to before starting the SEO for any website. Google has updated page speed algorithm recently. Page speed of any webpage is measured on basis of some metrics such as FCP and DCL.
FP:
First Paint (When a web page starts showing something at the very first time in a browser) or we can say when a browser starts rendering pixels.
FCP:
It stands for First content full paint. In this part, the browser renders the first bit of content.
FMP:
Is stands for First Meaningful Paint and the most important part of a webpage is rendered in this part.
TTI:
Time to interactive. It means the application or page have rendered fully and ready to reply the user’s input.
DCL:
It stands for DOM content loaded
Web page speed is divided into 3 parts
Fast:
If the median value of FCP and DCL is in top third then web page is considered fast.
Average:
If the median value of FCP and DCL is in Middle third then page speed is considered average.
Slow:
If the median value of FCP and DCL is in bottom third then page speed is considered slow.
Crawling
Crawling is also a very important part of technical SEO. To rank any website it is must that site should be crawled by search engine spiders without any issue. There are some errors that stops search engine spiders to crawl the particular webpage such as DNS errors, Server error, URL errors. If there is any page on website that is not available anymore, will be not found buy search engine robots. If you have used the wrong robots.txt file, it is confirmed those pages will not be visited by search engine robots. So finally keep your website neat and clean and error free so that search engine robots can crawl your site and index.
URL
In this part, you need to keep your page URL’s error free. It is recommended to use only one version of your URL. There are 4 types of different versions available of any URL such as
Xyz.com
www.xyz.com
https://xyz.com
https://www.xyz.com
These all are four different types of URL for search engines. You need to implement any one of them. It is recommended if you use a secure URL. The best version to work is
https://www.xyz.com
Robots File
- Robots.txt file has a very important role in SEO. Robots.txt file is used to block search engine spiders visiting a webpage. Robots.txt file should be written carefully, a wrong rotobs.txt file can create a big problem for you site. The syntax of a robots.txt file
- To Allow the whole website for crawling
Disallow:
- To Disallow robots not to crawl the whole website
User-agent:*
Disallow: /
- To Exclude Single Robot
User-agent: BadBot
Disallow: /
- To Allow Single robot – Put Following Code in the robots.txt File
User-agent: Google
Disallow:
User-agent: *
Disallow: /
1 Comment
Great article! That is the kind of info that should be shared across the net.
Shame on Google for not positioning this put
up upper! Come on over and talk over with my website . Thank you =)