The service includes tools that let webmasters • Submit and check a
sitemap. • Check the crawl rate, and view statistics about when
Googlebot accesses a particular site. • Receive alerts when Google encounters indexing, spam, or other issues on your site. • Show you which sites link to your website. • Write and check a
robots.txt file to help discover pages that are blocked in robots.txt accidentally. • List internal and external pages that link to the website. • Get a list of links which Googlebot had difficulty in crawling, including the error that Googlebot received when accessing the
URLs in question. • Set a preferred domain (e.g. prefer example.com over www.example.com or vice versa), which determines how the site URL is displayed in SERPs. • Highlight to
Google Search elements of structured data which are used to enrich search hit entries (released in December 2012 as Google Data Highlighter). • View site speed reports from the Chrome User Experience Report. • Page Experience Report including - Core Web Vitals, and HTTPS. • Receive notifications from Google for manual penalties. • Provide access to an
API to add, change and delete listings and list crawl errors. • Check the security issues if there are any with the website. (Hacked Site or Malware Attacks) • Add or remove the property owners and associates of the web property. ==See also==