Google uses several methods to find new pages on the internet. Which of the following is the main way Google discovers new pages to index?
Think about how Googlebot moves across the web.
Google primarily discovers new pages by crawling links found on pages it already knows. This process is called web crawling.
Which statement best describes the purpose of a sitemap.xml file in helping Google understand a website?
Think about what a sitemap lists.
A sitemap.xml file lists important URLs on a website to help search engines find and index them more efficiently.
Consider a page that Google can crawl but does not appear in search results. Which reason below best explains why Google might choose not to index it?
Think about instructions given to Google in the page's code.
A noindex meta tag tells Google explicitly not to include the page in its search index, even if it can crawl it.
When Google finds multiple pages with very similar or identical content, what does it usually do?
Think about how Google avoids showing the same content multiple times.
Google tries to pick a main version (canonical) of duplicate content to index and show, avoiding multiple copies in search results.
Consider a website that relies heavily on JavaScript to load its main content. How does this affect Google's ability to understand and index the page?
Think about how Google processes JavaScript compared to plain HTML.
Google can render JavaScript but sometimes delays indexing or misses content if JavaScript is complex or slow to load, affecting how well the page is understood.