Google Analytics Solutions offer free and enterprise analytics tools to live website, app, digital and offline data to realize customer insights.
Geotagging or Geo Tagging is that the process of adding geographical identification metadata to varied media like a geotagged photograph or video, websites, SMS messages, QR codes or RSS feeds and may be a sort of geospatial metadata.
Go to Google – type Google analytics – Open Google Analytics – Copy blog URL or website URL – then attend Google Analytics – Click on check in .
The Traffic source will report us the source of the device & monetarization of the web site Url, and therefore the location from the USA, UK or India.
It will give the report, the user visited the house page from Home Page – Services through products from the merchandise the user has existed. it’ll give us the user flow which page user has entered & which page user has existed.
Through this behavior to seek out the health condition of our website. The behavior flow & site speed, site search (the Google program , the Bing program , the yahoo search engine). The behavior is predicated on on-site content, keywords & sense console.
Using this setting you’ll add or delete Url. you’ll also create a replacement account for an equivalent (add blog & website additional)
Google Search Console
Google Search Console (previously Google Webmaster Tools) may be a no-charge web service by Google for webmasters. It allows webmasters to see indexing status and optimize visibility of their websites.
- attend Google
- Type Google Webmaster Tools
- Search – Open link
- Copy your blog URL or website URL
- Then attend webmaster tools
- Click on Add Property
- Paste your URL then click on Add
- Check your URL or blog/website you added
- Click on the crawl – then select sitemaps – add test site map – add site map XML on the URL it had been displaying then click on submit.
- Refresh the page
- Then check the small print .
The importance of a robots.txt file
You might be surprised to listen to that one small document , referred to as robots.txt, might be the downfall of your website. If you get the file wrong you’ll find yourself telling program robots to not crawl your site, meaning your webpages won’t appear within the search results. Therefore, it’s important that you simply understand the aim of a robots.txt enter SEO and find out how to see you’re using it correctly.
How does it (robots.txt file) work?
Before an enquiry engine crawls your site, it’ll check out your robots.txt file as instructions on where they’re allowed to crawl (visit) and index (save) on the program results.
Robots.txt files are useful
- If you would like search engines to ignore any duplicate pages on your website.
- If you don’t want search engines to index your internal search result pages.
- If you don’t want search engines to index certain areas of your website or an entire website.
- If you don’t want search engines to index certain files on your website (images, PDFs, etc.,)
- If you would like to inform search engines where your sitemap is found .
- User agent: Media partners – Google
- User – agent: *
- Disallow: / search
- Disallow: / *.gif$
- Disallow: / *.pdf$
- Disallow: /*.PDF$
- Disallow: /*.php$
- Disallow: /terms.html
- Disallow: /*search
- Disallow: /*_tests.php$
- Allow: /