At Google, we believe in open access to information. We believe that society works best when people have access to a wide variety of content from many sources. That’s why we do not remove content from search results, except in very limited circumstances.
Given how Search encompasses trillions of pages across the web, occasionally results may contain content that some find objectionable or offensive. This may especially happen if the language used in a search query matches closely with the language that appears within problematic content. It might also happen in situations where fairly little useful or reliable content has been published that aligns with a particular topic. Such problematic content does not reflect Google’s own opinions. However, our belief in open access to information means that we do not remove such content except in accordance with the specific policies or legal obligations that are detailed below.
Google processes billions of searches per day. In fact, every day, 15% of the searches we process are ones we’ve never seen before. Automation is how Google handles the immense scale of so many searches. Google uses automation to discover content from across the web and other sources. Automated systems -- our search algorithms -- are used to surface what seems to be the most useful or reliable content in response to particular queries. Automation also helps power our SafeSearch feature, allowing those who wish to use it to help prevent explicit content from appearing in search results.
Automation is also generally Google’s first line of defense in dealing with policy-violating content. Our systems are designed to prioritize what appears to be the most useful and helpful content on a given topic. Our systems are also designed not to surface content that violates our content policies.
No system is 100% perfect. If ours surfaces policy-violating content, we always look to resolve it by improving our automated systems. This allows us to better deal with both a particular issue that’s been detected and improve for related queries and other searches overall.
In some cases, we may also take manual action. This does not mean that Google uses human curation to rearrange the results on a page. Instead, humans are used to review cases where policy-violating content surfaces and take manual action to block this content, in the limited and well-defined situations that warrant this.
You can learn more about our policies that apply to Google Search here.