Google considers reducing web page filtering
Google is considering reducing web page frequency in an effort to save your computerThird, the resources
Image source: Unsplash
Google may reduce the frequency of web pages as it increases awareness of targeting clarity and targeting.
This topic was discussed by the Google Search Relationships team, consisting of John Mueller, Martin Split, and Gary Ellis.
Together, on the latest episode of the Search Without Record podcast, they discuss what you can expect from Google in 2022 and beyond.
Among the topics they talked about about crawling and indexing, SEO experts and website owners said they’ve seen little in the past year.
This will be Google’s main focus this year as it aims to make crawling more sustainable by saving computer resources.
Here’s what that means for your website and its performance in search results.
Crawl and Indexing Sustainability
Since the crawling and targeting of Googlebot is almost happening, you can’t think of it affecting the environment.
Ellis draws his attention to the problem when he says that computer use is not always stable:
“…using a computer on a regular basis and when you think about Bitcoin, for example, Bitcoin mining has a real area impact that you can really measure, especially if the electricity is coming from coal plants or other less sustainable plants.
We don’t have carbon, I don’t know, 2007 or something else, 2009, but that doesn’t mean we can never reduce our natural course. Crawling is one of the first things we can do to reduce fruit. “
Slightly hanging fruit, in this example, means unnecessary web crawling. Similar to showing web pages that do not have the latest updates.
Related: You can continue to search by creating these settings
How to fix content on a larger scale than fix screen errors, fix Google Webmaster Console error
How will Google make crawling more sustainable?
Ellis explains that web crawling can be sustained by limiting refreshing crawling.
There are two types of Googlebot Crawl: Crawl to find new content and crawl to update existing content.
Google is considering retrieving the crawl to refresh the content.
“Eyodwa one thing we do, and we may not need to do much, is crawl. Which means once we get the document and the URL, we go and crawl it, and then finally, we come back and visit that URL. That’s a recursive crawl.
And then every time we go back to that URL, that will always be the clarity of the update. “
He goes on to give an example of some web sites that allow a great deal of refinement for certain parts of the site but not others.
A site like the Wall Street Journal is constantly updating its homepage with new content, so it deserves a lot of clarity updates.
However, it is possible that the WSJ will not update its About page regularly, so Google does not need to keep updating updates on these types of pages.
“So you shouldn’t go back there often. A lot of the time, we can’t really measure that, and we definitely have room for improvement in search updates. Because sometimes it seems like a waste of time to click on the same URL over and over again.”
Sometimes we get to 404 pages, for example, for no apparent reason or for no apparent reason. And all of these things are really an improvement over our flaws. “
If Google lowers the update accuracy, which is 100% guaranteed, here’s the effect it can have on your website.
Related Topics: How can cybersecurity companies appeal to the tired digital world?
What does it mean to reduce crawl on your website?
There is a belief that a high level of visibility is a good sign of SEO, even if you don’t update your content as often as Google crawls.
This is a misconception, says Ellis, as the content cannot be measured better because it is more straightforward.
“A misconception that the more a page crawls, the more likely it will go up. Is this a misconception or is this true?”
“It’s a wrong idea.”
“Okay, so there’s no need to try to force something to crawl if it doesn’t really change. It wouldn’t rank better.”
Also, it’s not guaranteed that Google will downplay the visibility of the update, but that’s an opinion the team is considering.
More crawl doesn’t mean better rates.
In addition, the idea is to know the pages Need Refine clarity and pages that do not. This means that pages that you change frequently will continue to be refreshed and refreshed in search results.
Related Topics: Stop Turning Off Your Computer (Pros and Cons)