International SEO

SEO Success Case Study:

January 13, 2022
95 min
SEO Success Case Study:

The Importance of a Holistic Approach and Clear Communication to SEO Success

1.0 Summary of SEO Results

[Updated to include benefits from 2022 algorithm updates - scroll to section 12].

Digital PR and Local SEO worked together to improve the authority of the site and bring a significant increase in relevant traffic to within 3 to 6 months.

The general precept of this SEO project was:

“Every pixel, millisecond, byte, letter and user matters for SEO”
The project was lead by Koray Tuğberk GÜBÜR of Holisitic SEO who is the author of this case study.

Last 3 Months Comparison of the Encazip, taken on 6th February of 2021
Last 3 Months Comparison of the Encazip, taken on 6th February of 2021

This case study will demonstrate how clear lines of communication between software development and marketing teams can positively impact SEO results. The Encazip SEO Project achieved:

  • 155% Organic Traffic increase in 6 months
  • 110% Organic Traffic Increase in YoY Comparison for the same three months ( November to January)

To achieve this increase in Organic Traffic, the only targeted keywords were “Relevant and Possible Search Activities” with Semantic SEO Conception.

Comparison between August and January 2021 for Encazip
Comparison between August and January 2021 for Encazip

Organic Search View from Google Analytics with Average Session Duration
Organic Search View from Google Analytics with Average Session Duration

2.0 Encazip background & the SEO Project

Encazip is an energy consumption tariff comparison, modification, savings brokerage and affiliate company. is a Turkish based company, founded by localizing the know-how of the founders of

This is a YoY Comparison for the Encazip SEO Project’s GSC Data using DataBox
This is a YoY Comparison for the Encazip SEO Project’s GSC Data using DataBox

The idea of “” has been realized by a consortium under the leadership of Çağada KIRIM and backed by British investors, including Henry Mountbatten, The Earl of Medina. Electricity pricing and industrial or personal energy consumption is the primary target of the company. At the beginning of the SEO Project, the design, web page loading performance, branding, and content structure along with the education of the customer were looked at with a fresh perspective.

The “Everything matters in SEO” perspective was adopted by the customer.

We acknowledge the importance of SEO with no doubt, however, we would rather approach this subject to be one of the foundations of our corporate culture. I believe we have achieved the establishment of SEO culture in thanks to Koray and I am looking forward to being a part of more case studies in this very important subject.

Çağada Kırım, CEO,

2.1 Acknowledgements

If the Encazip’s team were not passionate about SEO, then it would not be possible to perform this case study at such a detailed level. Thanks and acknowledgements go to the whole team, especially Miss Yağmur Akyar, Mr. Erman Aydınlık, Mr. Nedim Taş, Mr. Oktay Kılınç and Mr. Can Sayan. CEO Mr. Çağada Kırım, who brought the whole team together and organized it for SEO was also a major factor in the success of the project.

Some parts of this SEO Case Study have been written by the Customer’s Team and will cover Technical SEO, Branding, Entitization, Content Marketing, Digital PR, and Web Page Loading Performance.

This screenshot has been taken on 9 June 2021 which is 7 days after the 2021 June Broad Core Algorithm Update of Google
This screenshot has been taken on 9 June 2021 which is 7 days after the 2021 June Broad Core Algorithm Update of Google

3.0 Page Speed Improvements: Every Millisecond Matters

According to Google’s RAIL model “Response”, “Animate”, “Idle”, “Load”, a developer only has 10 milliseconds to move a pixel. To acquire “60 Frames per seconds”, we have 16 Milliseconds, but also the browser needs 6 milliseconds to move a frame over another, this leaves us only 10 Milliseconds.

And, Pagespeed is actually a “health” and “trust” issue. If a frame doesn’t move or change itself in 16 milliseconds, users will notice some “motion bugs”. If it stays static for more than 100 milliseconds, it means that something is wrong. If it takes more than 100 milliseconds, users will start to “stress”. According to Google, a slow web page can create more stress than a fight for human-beings and it affects daily life negatively.

This is the “core of my statements” for any first meeting with my clients’ development teams.

RAIL performance market
RAIL Performance model

Below, you will see a quote from Nedim Taş who is the responsible developer for the front-end of Encazip.

The more important a brand’s image, the more important SEO and Code Performance are. SEO and Front-end Development can work together to create a better brand-image by taking users to the most usable and accessible website for their intents.

Nedim Taş

There is another “core lesson” here. I have many clients that have more than 15 members for their IT team, but yet, these crowded IT and Developer teams can’t create the same productivity and effectiveness as just a 2 person team. If you ask me what the difference is, I would say “SEO Passion and culture of the Company” and “growth hacking instinct”.

Page load time impact on bounce rates - Mach Metrics

Now, we can look at what we have done for improving web page loading performance of Encazip during the SEO Case Study and Project.

3.1 HTML Minification for Gaining 15 Kilobytes per Web Page

HTML Minification means deleting the HTML Inline Comments and Whitespaces of the HTML Document. This is one of the most essential page speed improvements but there are more benefits such as:.

  • Helps Search Engine crawlers explore the “link path” faster.
  • Helps HTML digestion for Search Engine Crawlers and their Indexing Systems.
  • Lets users’ devices build the Document Object Model faster.
  • It lessens the burden on the website’s server and lets users’ devices consume less bandwidth.
  • A complex and big HTML document might prevent Search Engine Crawlers from loading all of the document or evaluating it for ranking purposes.

For this last point, we can look at the old warning message from the previous version of Google Search Console”: “HTML Size is too large”. This warning was valid only for news sites. Google didn’t move this warning to the new Google Search Console, because I guess it didn’t want to share “weak points” or “deficits” of its indexing system. But, we still have the same warning from “Bing Webmaster Tools”.

“HTML is too Large Warning” from Old Google Search Console’s News Website Section:

“HTML is too Large Warning” from Old Google Search Console’s News Website Section:
“HTML is too Large Warning” from Old Google Search Console’s News Website Section

In the old Google Search Console, “extraction failed” error happened if the HTML was larger than 450 KB.

“HTML is too Large” warning from Microsoft Bing’s Webmaster Tools
“HTML is too Large” warning from Microsoft Bing’s Webmaster Tools

As a note, Microsoft Bing appears to be more open with information sharing about its algorithms and “desires” from webmasters than Google’s “BlackBox” attitude.

HTML Minification for Encazip was done in the first month of the project.  But further down the line, we had to can this work as we had some “server incapabilities” during the migration from .NET to .NET Core.

So, even the simplest Pagespeed Improvement has lots of aspects and value for SEO, also bear in mind that it might not be as simple as you may think!

3.2 CSS and JavaScript Refactoring

All Developers know that “refactoring a CSS File” is actually harder than writing a CSS File from scratch. Some developers call this situation “Code Spaghetti”. If you experience this, you should create a clean and efficient new CSS and JavaScript file for the same layout and functionality.

A screenshot from “CSS specificity calculation”
A screenshot from “CSS specificity calculation”

In this example, the “CSS and JavaScript Minification” took place at the same time with “CSS and JavaScript Refactoring” as there was already a plan for the website redesign.

(Editor’s Note: For more of this, see Koray’s article on “Advanced Page Speed Metrics”.  He covers what to know and what to do to achieve an efficient Rendering Tree).

Before the JavaScript and CSS Refactoring process, there were  more than 8 JS and CSS Files for only “page layout and functionality”. And, most of this code was not used during the web page loading process.

Test-driven development Methodology (TDD)

Test driven development (TDD)
Test driven development (TDD)

The total size of these web page assets was more than 550 KB.

At the end of this process, the development team decreased the total number of CSS and JS files to 3.

  1. Two of these 3 files were CSS and their total size was 14 KB.
  2. The JavaScript file that is being used for functionality was only 7 KB.

3.2.1 Why Have Two Different CSS Files?

Having the two different CSS Files lets Google cache and use just the necessary web page resources for its crawling routine. Googlebot and other Search Engine crawlers use “aggressive caching”, which means that even if you don’t cache something, Googlebot stores the necessary resources.

YoY Comparison for November, December and January
YoY Comparison for November, December and January

3.2.2 How does a Search Engine crawler know what to cache?

Thanks to aggressive caching of web page resources, a search engine uses less bandwidth from the website’s server, and makes fewer requests, thus the Search Engine Crawler can crawl a website more quickly.

To determine how to crawl a site, search engines use Artificial Intelligence. If a website uses all its “CSS and javascript files” site-wide, it means that the Search Engine will have to cache all of these resources.

Use long-lived caching

But, in this case only the necessary CSS File for the necessary web pages was needed, thus we can facilitate “fewer requests” for the search engine crawler and also help it use only the essential files.

In our example, only three different types of web page were used according to their functionality and importance:

  1. Home Page
  2. Product and Service Pages
  3. Blog Pages

Thanks to the “intelligence website design”, the code necessities had been united for the “product, service and blog pages”. Thus, fewer CSS files and code were used for managing the site.

  1. Encazip,  has “headfoot.css” which is only for the header and footer area of the website.
  2. A “homepage.css” file for just the homepage.
  3. “Subpage.css” for only the blog and product pages.

As you can see, Googlebot and other search engine crawlers can cache the “headfoot.css” easily because it affects the site-wide section of the web page. Also, “homepage.css” is only being used in the “homepage”, so for most of the crawling activity, search engine crawlers also can use the “subpage.css” file while saving itself from the “home page CSS codes”.

In short the effect of this strategy was:

  1. We decreased the CSS-JS File size from 550+ KB to only 22-25 KB per web page.
  2. We decreased the request count for CSS-JS files from 8+ to only 3 per web page.

And lastly, “CSS and JS Minification”, have been used. You can see again that “Every millisecond and byte really does matter”.

3.3 Text Compression on Server-side and Advantage of Brotli

Originally the company was using the “Gzip” algorithm for server-side compression. Many brands today are using “gzip” compression. Brotli was invented by Google. You can examine “Brotli’s Code” on Google’s Github Profile.

Brotli uses the LZ77 algorithm for lossless compression, for Brotli there is also a Google Working Group.  Brotli performs 36% better than Gzip!

And, in the fourth month of the SEO Project, the team started to use Brotli for the text compression on the server-side.

Compression technologies and file size
Compression technologies and file size


3.3.1 Why is Server-side Compression Important?

Here is a breakdown of the importance of Brotli usage.

  1. The hardest thing about loading and rendering a web page is transmitting the files from a server over the internet connection.
  2. If you check the Chrome DevTools Network Panel, you will see that for every web page resource, the longest part is the “requesting a file” and “downloading the file” over a network connection.
  3. Text-compression means compressing the files on the server-side and conveying these files to the requestor.
  4. Since the file sizes are being decreased by compression, conveying these resources over a network connection is simpler.
  5. Thus every web page resource will be loaded faster by the requestor which is the search engine crawler or the user.
  6. After loading the web page resources, the resources will be decompressed by the browser to use for rendering, parsing, and compiling the web page.

First byte improvement from Brotli compression
First byte improvement from Brotli compression

Basically, most of these articles include the “Time to First Byte Improvements”. By implementing Brotli, we have achieved an improved user experience, crawlability, crawl delay time and crawl efficiency of the web site.

3.4 HTTP 2.1 Usage Instead of HTTP 1.1

The essential difference between HTTP 1.1 and HTTP 2.1 is the request count capacity per round-trip for a requester to a server. What is the reason for this situation? HTTP 1.1 keeps every request and response in plain text format, while HTTP 2.1 keeps every request and response in binary. Thanks to the binary format, HTTP 2.1 can convey more web page resources over a single TCP Connection.

During the web page rendering, a requester can take only the first 6 resources from a server. If the first 6 requests don’t include all of the critical resources for the above the fold section, this means that the client will need a second round-trip for rendering the initial contact section of the web page.

A Schema for showing the HTTP 2.1’s Working Principle
A Schema for showing the HTTP 2.1’s Working Principle

HTTP 2.1 adoption was a critical change for Encazip’s SEO project. Since the critical resource count and size for above the fold section were already decreased, the “request round-trip need” was also decreased with the usage of HTTP 2.1.

Furthermore, Googlebot started to use HTTP 2.1 for its crawling purposes, thus our crawl efficiency improved.

3.4.1 HTTP 2.1 Server Push

HTTP 2 Server Push is the first reason for Encazip’s HTTP 2.1 migration from HTTP 1.1. After creating the “subpage.css” and “headfoot.css”, we used HTTP 2.1 Server Push for creating a faster initial contact with the users.

HTTP 2 Server Push lets a server “push a web page resource” to the requester even if the client doesn’t want it to. Thus, before the client makes the request from the server, the server will push the resource so that the connection and downloading process of the resource can happen faster.

Server Push’s working principle
Server Push’s working principle

For HTTP 2 Server Push, the team determined to include some resources from the above-the-fold section of the page, such as “logo, headfoot.css, subpage.css and main.js”.

Please bear in mind, HTTP 2 Server Push aslo has some side effects.

  • If you use HTTP 2 Server Push for too many resources, it will lose its efficacy
  • The main purpose of HTTP 2 Server Push is to only use it for certain and critical resources
  • The resources within the HTTP 2 Server Push can’t be cached
  • HTTP 2 Server Push can create a little bit more server overhead than usual

Because of only the final point, we ended up using HTTP 2 Server Push only for a short time.  Once we have made some back-end structure changes, Can Sayan plans to use it again.

Lack of a caching system or a strong server… Long queries that slow down the response times… All of these affect the User Experience and also SEO. Thus, we are racing against milliseconds.

Can Sayan, Backend Developer

3.5 Resource Loading Order and Prioritization

TCP Slow Start means that only the first 1460 bytes of an HTML Document can be read by the client. This is actually designed to protect servers. Thus, the most critical resources need to be at the top of the HTML document.

Our Resource Load Order is below with the HTML Tags and Browser Hints.

  1. <link rel=”preload” href=”/content/assets/image/promo/banner.avif” as=”image”>
  2. <link rel=”preload” href=”/content/assets/css/headfoot.css” as=”style”>
  3. <link rel=”preload” href=”/content/assets/font/NunitoVFBeta.woff2″ as=”font” crossorigin=”anonymous”>
  4. <link rel=”preload” href=”/content/assets/script/main.js” as=”script”>
  5. <link rel=”stylesheet” href=”/content/assets/css/headfoot.css”>
  6. <link rel=”stylesheet” href=”/content/assets/css/subpage.css?v=23578923562″>

You can see the translation of these resource load order’s logic below.

  1. “Banner.avif” is for the Largest Contentful Paint.
  2. “Headfoot.css” is for the First Contentful Paint
  3. “NunitoVFBeta.woff2” is for the “FOUT” and “FOIT” effect.
  4. “Main.js” is for the functionality of the web page.
  5. “Subpage.css” is for the general layout of the Product and Service web pages.

And, you can see the profile of resource load prioritization below.

Encazip Resource Load Order Prioritisation
Encazip Resource Load Order Prioritisation

3.5.1 Cross-Browser Compatibility for Preload Usage

In the image above you will see some “duplicate requests”, these have been purposely left in. This is because, until January 2021, Firefox didn’t let SEOs and developers or Holistic SEOs use “preload”. Thus if a user-agent included “Firefox”, it couldn’t use the “preload”, thus we have also put normal request links without preload.

Don’t worry, Google Chrome won’t request the same file twice!

Updated view of Preload browser compatibility profile
Updated view of Preload browser compatibility profile

3.5.2 What You Should Know About Preload

Preload does not work in Firefox, but lately, they have started to use Preload without implementing any kind of Firefox flag configuration. Here are other things you need to know while using preload.

  1. You can’t use preload for the resources that you have already “pushed with HTTP 2 Server Push” feature.
  2. If you try to use “preload” for everything, it won’t lose its meaning.
  3. Preload also “caches” the file for the browser so that returning clients can open the next pages faster.
  4. If you preload too many things, it can create a CPU bottleneck at the beginning of the web page loading process.
  5. A CPU bottleneck also can increase the Total Blocking Time, First Input Delay, and lastly “Time to Interactive”. Not an ideal scenario!

That’s why discussions with your development team are important. It can sound easy while saying “okay, we will just put ‘preload’ value to the ‘rel’ attribute”. But, it’s not that easy, everything needs to be examined repeatedly with a bad internet connection and mediocre mobile device.

Preload’s effect for resource loading
Preload’s effect for resource loading

3.5.3 Beyond Preload: Preconnect for Third-Party Trackers

Why didn’t we use DNS-Prefetch?  Simply put, DNS-Prefects is only for “DNS-Resolution” for the third-part resources’ server while “Preconnect” is for performing the “DNS-Resolution”, “TLS Negotiation” and the “TCP Handshake”. TLS Negotiation and TCP Handshake are essential processes for loading a resource from a server. Instead of DNS-Prefetch, we implemented “Preconnect” for third-party trackers for this reason.

Preconnect’s working principle
Preconnect’s working principle

Some of the “preconnect” requests for Encazip are below.

  • <link rel=”preconnect” href=””>
  • <link rel=”preconnect” href=””>
  • <link rel=”preconnect” href=””>
  • <link rel=”preconnect” href=””>

3.5.4 Loading CSS Files as Async

Loading CSS Files as “async” is an important but mostly ignored topic. Thus, it is recommended that you read the “Loading CSS Async” tutorial. To load a CSS File asynchronously, the browser needs to be manipulated with “media=all” and “media=print” attribute and value pairs. All of the CSS files are actually render-blocking, it means that while they are being loaded, a browser cannot also render the web page as with JavaScript files.  For JavaScript, we have the “async” attribute, but we don’t have anything for “CSS” files.

Thus, we have wanted to use “CSS Async loading” for Encazip, but at first, it wasn’t necessary. Because the total CSS File amount was only 16 KB per web page. It is so small that it couldn’t block the rendering.

CSS Async prevents blocking
CSS Async prevents blocking

ut, when we started to load more resources with the “preload” browser hint, CSS Files started to “block the rendering”. Thus, we wanted to re-plan it as we were only gaining 15-20 milliseconds per web page loading event.

But, we didn’t implement it in the end. Because, when we use the “CSS Async” feature for these CSS Files, it is creating a “flicker effect”. It means that the browser was rendering the web page without CSS first, then it was implementing the CSS Effects, this was creating a “turbulent page loading experience” or in other words, a flickering effect.

Just for gaining 15-20 Milliseconds, we didn’t want to cause such stress for the user. That’s why as an SEO and developer, you need to balance things while making a website more crawlable for search engines and usable for users.

So, at the moment, we have left this topic for further discussion, maybe we can use CSS Async only for one of the resources. But, I want you to remember that CSS Async is a feature that relies on “JavaScript” rendering, and this can also affect SEO. (Editor’s Note: I recommend you to check the section for “Image Place Holders”).

3.5.5 Deferring All of the Third Party non-Content Related Trackers

All of the third party and non-content relevant JavaScript files were deferred. A deferred JavaScript cannot be rendered until the “domInteractive” event.

And, there are two important things to consider when using the defer browser hint:

  1. If you use defer on the main JS file, you will probably not see its ‘initiator’ effects until it is being installed.
  2. If you use defer too much, you may cause a CPU bottleneck at the end of the page load.

Be careful whilst using “defer” to ensure that you do not block the user!

JavaScript loading methods and its effects
JavaScript loading methods and its effects

In Encazip, before the SEO Project’s launch, all of the non-important JavaScript files were being loaded before the important content relevant CSS and Javascript files.

We have changed the loading order for the resources, so the most important web page resources load first and have deferred the non-important ones.

N/B: Deferring third-party trackers can cause slightly different user tracking reports since they won’t be able to track the user from the first moment.

3.5.6 Using Async for Only the Necessary JavaScript File

As with using the “defer” attribute, using the “async” attribute is an important weapon at your disposal for creating the best possible user experience. In the case of Encazip, we have used the “async” feature for only the “Main.js” file since it was the only file that was focused on the “content” and “functionality”.

And, from the previous section, you can remember that you shouldn’t defer the most important and functional JS file.

With Puppeteer, you can block any resource so you can see which JavaScript file is affecting what for the above the fold section
With Puppeteer, you can block any resource so you can see which JavaScript file is affecting what for the above the fold section

3.5.7 Loading the Polyfill JS Only for Legacy Browsers

Polyfill is used for ensuring compatibility of these modern JavaScript methods and files with legacy browsers such as Internet Explorer. And, as an SEO, you always need to think about the user profile 100%. Since, there are millions of people who use Internet Explorer (I don’t know why, but they are using it!)

Since we don’t have “Dynamic Rendering” and “Dynamic Serving” technology at the moment for Encazip, Nedim Taş prepared another step to prevent loading Polyfill JS for modern browsers. If the browser is a modern browser, Polyfill JS will not be loaded with its content, but the request will still be performed. If the browser is a legacy browser, it will be loaded with its content.

Thus, for most of the users, we have saved them tens of KB.

From a Chrome session, you can see that “polyfill.min.js” is empty
From a Chrome session, you can see that “polyfill.min.js” is empty

3.6 Aggressive Image Optimization with SRCSet and AVIF Extension

First, let me explain what aggressive image optimization is? How is it different from regular image compression? There are four different aspects of image optimization; “pixels, extensions, resolutions, and EXIF data.”

Pixel optimization in terms of “image capping” is actually a new term. Image capping has been implemented by Medium and Twitter before to decrease the image size by 35% while decreasing the request latency by 32%. Image capping means “decreasing the pixel count of image 1x scale, in other words, 1×1 pixel per dot. Since 2010, “super retina” devices started to become more and more popular. “Super retina devices” mean that they include more than one pixel per dot on the screen and this gives a device a chance for showing more detailed images with higher pixels.

Colour and pixel differences based on devices. Above, you see “1 Pixel’s colour profile”
Colour and pixel differences based on devices. Above, you see “1 Pixel’s colour profile”

So, what is wrong with super-retina devices and 2x Scale or 3x Scale images?The human eye can’t actually see the details in 2x resolution or 3x resolution images.

  1. 2x resolution or 3x resolution images are bigger in terms of size.

For Pixel Optimization, you can use “different types of image pixel optimization algorithms”. Such as “NEAREST” or “BILINEAR” from PILLOW.

So, let me introduce you to other sections of aggressive image optimization for web page loading performance briefly.

Like “pixel optimization”, also “extensions are important”. This is common information now. But, most of the SEOs, Developers, or Holistic SEOs are not aware of “AVIF”. Most people also know WebP. But, I can say that WebP is already outdated and outranked by AVIF.

Justin Schmitz is the inventor of AVIF
Justin Schmitz is the inventor of AVIF

EXIF Data (Exchangeable Image File) is an important aspect of SEO. I won’t go deep into this aspect but you can watch the video of Matt Cutts from 2012.

As an aside: Google was sharing way too much information about their internal system before John Mueller. As an SEO, I can’t say that I like this change!

Exchangeable Image File includes the “light, camera, lens, geolocation, image title, description, ISO Number, Image Owner, and Licence information”. Some also call this IPTC Metadata (International Press Telecommunications Council).

For relevance, I recommend my clients to use IPTC Metadata with minimum dimensions. But for performance, you need to clean them.

We have chosen “600px” as the fixed-width since it is compatible with both mobile and desktop view screens

Resolution. For image optimization, unnecessarily big resolutions shouldn’t be used. If the web site is not from the News niche, you probably won’t need big resolutions.

To use the best possible image extension based on user-agent (browser) differences and the best possible resolution based on the device differences, we have used “srcset”.

Below, you will see an example.

srcset for images

Remember, we already “preloaded” the AVIF image, and now we are just telling the browser where to show it if it can do so. And, “<figure>” is being used for Semantic HTML as we will discover in the future sections of this SEO Case Study.

P.S: Do I really need to talk about “alt” tags? Or, Image URLs?

3.6.1 Intersection Observer for Image Lazy Loading

Intersection Observer is an API to load the images only if the image is close to the view screen or in the view screen. In the Intersection Observer, you can determine when or where to load and show an image. Basically, Intersection Observer API is the practical name of Lazy Loading. But, why didn’t we use the “load:lazy” attribute of Chrome? Or, why didn’t we use a third-party library for lazy loading?

  1. We didn’t use Chrome’s “loading” attribute and “lazy” value because it is not compatible with every browser. For cross-browser compatibility, we need to use Intersection Observer.
  2. We didn’t use the third party libraries for lazy loading, because it would also load unnecessary codes from another third-party domain. And, maintaining your own custom library is way much better than adding another dependency to your toolset.

Below, you will see what percentage of the users’ browsers support Intersection Observer API which is 91.98%.

Browser support for Intersection Observer API
Browser support for Intersection Observer API

On the other hand, only 69.39% percent of the browsers support “loading” attribute and “lazy” value for it.

Browser suport for lazy loading
Browser suport for lazy loading

I recommend checking out Mozilla’s How to Create an Intersection Observer API tutorial.

Thanks to the Intersection Observer API, we increased the size of the “main.js” file just a little bit, but we gained control of the lazy loading’s default behaviors without any other dependencies and, of course, we have improved the initial loading time more than 50% thanks to lazy-loading. This 50% improvement was measured from the following relevant page speed metrics: First Paint, First Contentful Paint, Largest Contentful Paint and Time to Interactive.

3.6.2 Image Placeholder for Better Speed Index and Largest Contentful Paint

Image placeholders are important for “visual progress” speed. To completely load the above-the-fold section of the web page in terms of “visual completeness”, placeholders provide a “smooth” and more “interactive” experience. Image placeholders are now being used in for improving “Speed Index” and “Largest Contentful Paint” timing.

But also, since image placeholder is a technology that relies on JavaScript, it also has some side-effects such as not showing the image for the not-rendering crawling schemas of Googlebot. You can see the effect of image placeholders due to their JavaScript based nature on Google’s SERPs.

And, this is what happens if a Search Engine doesn’t render your JS in a stable way.

Me: I'm using “placeholders” for better LCP and Speed Index.

Search Engine: I'm using the “placeholder” image instead of the actual image.

As you can see, since without the “rendering phase”, Google’s indexing engine can’t see the actual image, it doesn’t understand that the image is actually a placeholder. Thus, it shows this instead!

Thus, Google decided to not show the image after all because it was not in the “initial HTML”. Google thought that the image wasn’t important enough and also since they don’t render JavaScript every time, they couldn’t see the actual LCP Image continuously between crawl round-trips.

After a while, Google will show the image placeholder again, then the actual image and then it will remove it again… This will continue as a loop. And, you should think of this as a reason for “ranking fluctuation” also.

Ranking fluctuation will continue

After 15 days, Googlebot fixed it. But, if you know Google, “fresh data” is always more important than the “old data”. You should think of Google’s crawling behaviour as a “loop”.

You will see the “loop of” Search Engine’s Decision Tree
You will see the “loop of” Search Engine’s Decision Tree

3.6.3 Using Image Height and Width Attributes for Cumulative Layout Shift

Image height and width are important for Image SEO and Visual Search. But besides Image SEO, this is also an important aspect for User Experience and thus for SEO. I won’t dive deep into Cumulative Layout Shift here, but suffice to know that every unexpected “layout shift” or “moving web page component” is a cause for Cumulative Layout Shift for the user.

To prevent this situation there are certain rules:

  1. Give height and width values to images.
  2. Do not use dynamic content injection.
  3. Late uploaded web fonts.
  4. Network approval before DOM Loading.

In this context, we have given height and width values for images so that Cumulative Layout Shift can be decreased and Encazip can be ready for Google’s Page Experience Algorithm.

3.7 Web App Manifest Usage for Progressive Web Apps

Web App Manifest is the gateway for Progressive Web Applications. Simply put, Web App Manifest is a file that defines the website as an application and lets a device download the website to local storage with certain icons, shortcuts, colors, and definitions. Thanks to Web App Manifest, a website can be open without a browser like an app. That’s why it is called a Web App Manifest.

In Encazip, we have started to use Web App Manifest, also the “words” and “shortcuts” in the Web App Manifest can increase the user-retention while reinforcing your brand-entity to Google.

Web App Manifest
Web App Manifest

And, you will see that we have a “prompt pop-up” for installing Encazip as a local app. Below, you will see that we have the as a local app on my desktop screen.

Install as a local app
Install as a local app


3.8 Using Service Workers for Better HTML Payloads

Service Workers are another step for Progressive Web Applications. Thanks to service workers, a website can work offline. A service worker is actually local storage from the browser’s memory. A service worker can create a “cache within an array” and certain URLs are registered into this array. After these URLs and resources within them are being cached, the client doesn’t send requests to the server for these resources. And since they are in the local cache, website loading performance is improved for returning visitors.

Service Workers

With Service Workers, we have cached the most important resources for the “initial contact” with the user. But, to be honest, it’s not obvious what the limit of storage is for a service worker, so we have tried to use it carefully and sparingly.

3.9 Cleaning Unused Code from Third-party Trackers by Localizing

This section is actually debatable. In Technical SEO and Page Speed Improvements for creating the best possible time and cost balance for the SEO Project, I always focus on the most important points that will make the difference.

Generally third-party dependencies generally impact internet users in terms of data usage and page speed. Thus, “cleaning and localizing” the third-party trackers can decrease the page’s size enormously and also remove the need for connecting another outsourced service for the client.

But, Cleaning and Localizing the third-party resource has also some side-effects:

  1. If you localize a third-party tracker, you won’t get the updates automatically.
  2. Localized third-party dependencies might not work perfectly due to sloppy cleaning.
  3. If the marketing team wants to use another feature from the dependent script, the process might need to be repeated.

Below, you will see the positive effects:Removing the Single Point of Failure possibility.

  1. You will only use the necessary portions of the dependent script for lesser code.
  2. You won’t need to connect to another outsourced service to complete the web page loading.
  3. It is sustainable if the development team can make this a habit.

In Encazip, because of these side-effects, we didn’t implement this yet, but it is in the future scope of the project.  I’ve included it to show the lengths we will go to to get the best results and demonstrate the true “vision” and “perspective” of this case study.

3.10 Conflicting Document Type with Response Headers and HTML Files

Unfortunately the majority of SEOs do not care about the Response Headers – and they should!

Response headers and their messages are actually more important than the actual HTML. So, any kind of message in the Response Header should not conflict with the information within the tags from the HTML Document.

Content type within the Response Header
Content type within the Response Header

In our case, Encazip was using the “Windows-1258” encoding for the “content-type” response header while using the “UTF-8” for the HTML Document. And, this gives a mixed signal to the browser and indirectly to the Search Engine crawlers about the web page’s content type. To remove such a mixed-signal, we started to use only “UTF-8” within the HTML Document.

3.11 HTML Digestion and HTML Based Improvements

“HTML Digestion” is a term from the “Search Off the Record” Podcast Series which is created and published by Googlers, Danny Sullivan, Garry Illyes, Martin Splitt, and John Mueller. They also call this “HTML Normalization”. According to Google, the “actual HTML” and the “indexed HTML” are not the same. Googlebot and Caffeine Indexing System of Google are extracting the HTML Structure from the actual document with the signals they collect.

A headline from Barry Schawrtz for the specific topic
A headline from Barry Schawrtz for the specific topic

There is a simple quote below from Garry Illyes about “HTML Normalization”.

"If you have really broken HTML, then that’s kind of hard. So we push all the HTML through an HTML lexer. Again, search for the name. You can figure out what that is. But, basically, we normalize the HTML. And then, it’s much easier to process it. And then, there comes the hotstepper: h1, h2, h3, h4.

I know. All these header tags are also normalized through rendering. We try to understand the styling that was applied on the h tags, so we can determine the relative importance of the h tags compared to each other. Let’s see, what else we do there?

Do we also convert things, like PDFs or… Oh, yeah. Google Search can index many formats, not just text HTML, we can index PDFs, we can index spreadsheets, we can index Word document files, we can index… What else? Lotus files, for some reason".

Garry Illyes

You can listen to the Search off the Record Podcast Series, related episode.

Remember what happened during August, September and November in Google’s indexing system?  Everything went awry!  Google removed the Request Indexing function and it mixed the “canonicalized URLs”, etc…

Google’s confirmation for a series of bugs. Even the Google Search Console’s Coverage Report was not refreshed for days
Google’s confirmation for a series of bugs

Thus, having a simple, error-free, understandable HTML is helpful. Thus, while redesigning the website, we have also cleaned all of the HTML code errors from

P.S: HTML Code errors might make a browser work within “quirk mode” which is also harmful for web page loading performance even if it is just a small factor.

3.12 Semantic HTML Usage

Semantic HTML means that HTML tags can have a context and meaning within a hierarchy. It gives more hints and makes it easier to understand a web page for the Search Engine Crawlers. Also, Semantic HTML is useful for screen readers and web users who have disabilities.

A schema for Semantic HTML’s logic
A schema for Semantic HTML’s logic

In Encazip, at the beginning of the SEO Project, Semantic HTML was not used. But, with some certain rules we have used Semantic HTML. You can see some of the tags we prefer to use within the website.

  1. Header
  2. Footer
  3. Nav
  4. Main
  5. Headings
  6. Article
  7. Aside
  8. Section
  9. Ol and Li
  10. Picture
  11. Figure
  12. Quote
  13. Table
  14. Paragraph

Every “section” had at least and at the most only one “heading 2”. And every “visual transition” was also at the end of the section. In other words, the “visual design elements” of the website and the Semantic HTML are compatible with each other. This helps to “align the signals” unlike “bad and mixed signals”.

3.13 Decreasing the HTML DOM Size

The DOM Size is an important factor for “Reflow, Repaint Cost”. The Document Object Model is built from objects or nodes. Every additional node is an increase of +1 to the DOM Size. Google suggests having less than 1,500 Nodes in the Document Object Model. Because, having a large DOM Size makes it harder to layout, paint, or render processes for the browser.

You can see the DOM-Tree Analysis for’s Homepage
You can see the DOM-Tree Analysis for’s Homepage

n Encazip, we ended up with 570 nodes in the DOM. It is much better than Google’s suggested limit, but our main competitor has an average of 640 nodes. So, we are better, but not much better than our main competitor, at least for now.

3.14 Font File Count and Size Decreasing

Font file optimization is completely another discipline in page speed science. Thus, I will just give a simple and short summary here.

In Encazip there were more than 5 font files per web page. And most of these fonts were not even used for every web page or even if they were being used, they were only for a small portion of the web page.

I always recommend brands to use “less color” and “fewer fonts”. Because they are not really critical, but still costly for the users and crawlers of the search engines.

We use a single font file for now
We use a single font file for now

The first major issue is that all of the font files did not have the Woff2 file extension. This meant that their size was unnecessarily large. The total size of the font files was more than 200 KB per page.

  1. At the end of the day, we have decreased the font file count to one.
  2. We have decreased the font file size to 44 KB.
  3. We have gained 4 requests and an average of 150 KB per web page.

3.15 Using Font Variables

Font variables are one of the advanced page speed topics. Imagine that you are unifying the “bold”, “italic” and “regular” versions of a font into a single file. Thanks to font-variables we could use different font variations with only one request.

“Font-variation-setting” is for font-variable
“Font-variation-setting” is for font-variable

Thus, we have managed to stick to only one font and different styles. You can see “Font-variable” codes below from our CSS File. (Another thanks to Mr. Nedim Taş for this!).

3.16 FOUT and FOIT

Flash of Unstyled Text and Flash of Invisible Text are other important terms for web font optimization. FOUT and FOIT are also important for Cumulative Layout Shift and sometimes, Largest Contentful Paint if  the LCP is textual content. To prevent FOUT and FOIT situations, we have preloaded the font file while using “font-display:swap” CSS Feature within our CSS. Below, you can see the necessary code block.

Font-display is from our CSS Files
Font-display is from our CSS Files

3.17 Using Browser Side Caching for Static Resources

The browser-side cache is for the static resources of the web pages. If a resource on the web page doesn’t change frequently, it means that it can be stored in the browser’s cache. To perform this, “cache-control:max-age” and “Etag” or “Entity Tag” HTTP Header should be used.

Working principle of browser-side cache
Working principle of browser-side cache

n, we have used browser-side cache for some of the static resources, but some of the static resources’ browser caching is delayed due to some back-end infrastructure improvements.  So there’s some more incremental improvements to make here.

4.0 Structured Data Usage

Structured Data is one of the other important signals for the search engine. It shows the entities and their profile and connection with other entities to the search engine. Structured data can affect the relevance, SERP view and the web page’s main intent in the eyes of search engines.

Encazip’s Organization Structure Data Visualization
Encazip’s Organization Structure Data Visualization

In Encazip, the structured data had not been implemented correctly, so we adopted three different structured data types for

  1. Organization
  2. FAQ
  3. AggregateRating

Why did we use these types of structured data?

  1. Organization structured data was used for creating an entity reputation and definition for Encazip. Soon after, Google started to show Encazip’s social media profiles on the SERP.
  2. FAQ structured data has been used for the blog and service/product pages. In the future, we plan to also add more sections to the FAQ structured data within the guidelines.
  3. AggregateRating is for the business partners’ of and it was united with the Organization structured data. The main purpose here was to show the web page’s activity on the SERP with the reviews and stars.
  4. Images which are located in the Largest Contentful Paint Element have been added into the FAQ Structured Data for better web page elements, and layout functionality signals.

5.0 Website Accessibility: Every User Matters

Website accessibility is one of the most important things that are relevant to SEO, UX, and most importantly humanity. As an SEO, I must say that an accessible website is actually a human right. Thus, I believe that making websites accessible is one of the best sides of SEO. (And, as a color-blind person, I give extra-special attention to this area).

5.1 Using Accessible Rich Internet Applications for being a Better Brand

Encazip is a ‘mostly’ accessible website. I say “mostly” because, to be honest, learning and implementing “Accessible Rich Internet Applications” is not easy. But, we have implemented “role”, “aria-labelledby”, “aria-describedby” attributes with proper values.

Furthermore, I can also say that pages that are legible to a screen reader can be understood easier by a Search Engine, since it doesn’t leave anything to chance and connects every web page component to each other. (I recommend you to think about this, also having Semantic HTML in mind).

Lastly, we also cared about “light and color differences” between web page components.

PS: Do we really need to talk about alt tags, in 2021?

6.0 Website Redesign Process: Every Pixel Matters

For designing a website there are lots of dimensions. The layout of the web pages, component order of web pages, style of components, texts, images, links, categorization of pages and more are affecting rankings.

If a web page cannot satisfy the search intent, it cannot be ranked well by Search Engines. If a website’s layout is not understandable and requires “learning” by users and also Search Engine’s Quality Evaluation Algorithms, it can harm the SEO Performance.

Google can understand a website’s quality and expertise from its design, layout, or web page components. By just changing the design, I have overcome some SEO performance plateaus before, this also includes changes as minor as colour palettes. Google also has some patents about this, to show the detailed insights that Google might seek to extract from a website’s layout and design.  I have chosen only four Search Engine patents, one is from Microsoft, three are from Google.

6.1 Website Representation Vectors

vectors web representation
Vectors web representation

Website representation vectors cluster websites according to their layout and design quality along with expertise signals. According to their percentage similarity, Google labels sites as expert, practitioner or beginner by looking at texts, links, images, layout and a combination of this and more.

6.2 Read Time Calculation

Detection and utilisation of document reading speed
Detection and utilisation of document reading speed

Google might use “markers” to try to understand how a user can read a document, and how much time it would take them to find the right portion of the document for a specific information or query. It also tries to understand the language of the content and its layout for the users’ needs. This patent is from 2005, but it shows that at some point, Google’s Search Quality Team has cared about this in the middle of 2000-2010. And, we all know that the above-the-fold section and phrases and entities from the upper section of the content are more important than the middle and bottom sections.

6.3 Visual Segmentation of Web Pages based on Gaps and Text Blocks

Document segmentation based on visual gaps
Document segmentation based on visual gaps

Google can use visual gaps, text blocks, headings and some marks to understand the relationship of different blocks with each other. But, also if you leave too much gap between blocks it can affect the “completeness of the document” while increasing the scroll depth and also “read time”. So, having a “complete visual block” that follows another one within a hierarchy and harmony is important.

6.4 VIPS: a Vision-based Page Segmentation Algorithm:

Vision-based Page Segmentation Algorithm
Vision-based Page Segmentation Algorithm

Another patent, but this time it is from Microsoft.  As you can see above, the Vision-based Page Segmentation Algorithm uses the Document Object Model and also visual signals to analyze the relationship of different web page segments with each other. You will see some rules from ViPS:

  1. If the DOM node is not a text node and it has no valid children, then this node cannot be divided and will be cut.
  2. If the DOM node has only one valid child and the child is not a text node, then divide this node.
  3. If the DOM node is the root node of the sub-DOM tree (corresponding to the block), and there is only one sub-DOM tree corresponding to this block, divide this node.

Why do you think that I have shared these rules? Because, all of these are similar to what Google’s Lighthouse does for determining the Largest Contentful Paint. And, LCP “div” can also be used for understanding the actual purpose of the web page. Of course it is not a “directive”, it is just a hint, but that’s why LCP is important for search engines. It means that the initial contact section of a page with a fast LCP Score can satisfy the search intent faster.

I don’t go further in this section but know this, Google also has patents for page segmentation based on “function blocks and linguistic features”. It also checks the code blocks to understand which section is for what, whilst annotating the language style on these code blocks.

Classifying functions of web blocks based on linguistic features
Classifying functions of web blocks based on linguistic features

In Encazip, while designing the new website, the designers created a modern, useful web page layout and visual aesthetic for web users. During the design process, we also talked about the implications of mobile-only indexing, mobile-first indexing, search intent, visual consistency and Chris Goward’s “LIFT Model” for page layouts.

LIFT Model of Chris Goward
LIFT Model of Chris Goward

The LIFT Model is also important for me because it lets me optimize a web page based on the dominant search intent along with sub-intents. With a proper hierarchy, everything coalesces nicely with a clear signal. While designing the website, we also talked about DOM size, the need for CSS code, styling of heading elements and semantic HTML usage along with AIRA necessity. (Again, in this section, Mr. Nedim did an excellent job!).

This image shows Google’s Indexing System’s working style with “aligning ranking signals”
This image shows Google’s Indexing System’s working style with “aligning ranking signals”

So, web page layout is an important ranking factor.  Along with E-A-T, it affects the user experience and conversion rate, it is also directly related to the web page’s loading performance.  To be a good well-rounded SEO, one should be able to harmoniously manage these different aspects of an SEO project.

7.0 Kibana and ElasticSearch Usage for Log Analysis

At the beginning of the Encazip SEO Case Study, we didn’t perform any kind of log analysis and actually, we didn’t need it. But for the future stages of the SEO Project, we plan to use Kibana and ElasticSearch.

If you are looking for more on this, Jean Christopher has a great article on Search Engine Journal about how to use Kibana and ElasticSearch for SEO Log Analysis.

With this in mind we have started to prepare our log analysis environment. Of course, if you want you can read log files with a custom Python script or a paid service such as JetOctopus, or OnCrawl, it is up to you.

Authoritas also has a nice real-time log analysis tool in Alpha that uses a small JS snippet that you can insert into your website.  It detects bots and then sends the data to the Authoritas servers where it is then analysed in Kibana and Elastic Search.  It’s very fast to load (page impact 10-20 ms) as it loads over UDP rather than TCP/IP. It only works at the moment on sites running PHP.  As it’s loaded with the page, it won’t pick up any 5XX server errors, but it will help you track bots in real-time, find 3XX, 4XX issues and bad bots hitting your site.  If you have difficulty getting access to your server logs, then this could be a simple and easy step.

Real-time log analysis tool
Real-time log analysis tool

Analyse your website's logs
Analyse your website's logs


8.0 Branding, Digital PR, and Entitization of Every Mention Matters

Encazip is also a good example of being an entity. I won’t go deep about entities here, but there are four key differences between an entity and a phrase.

  1. Entity has a meaning, keyword has not.
  2. Entity does not have a sound, a keyword has.
  3. Entity have attributes, keywords not so.
  4. Entity is about understanding concepts, keywords are about matching the string.

Entitization means the process of giving a brand an actual meaning, a vision, attributes and connection with other concepts to a search engine. Being an entity will help improve your rankings. Google can evaluate a “source” beyond its own domain. To perform this, you need to implement entity-based Search Engine Optimization.

But, in the Encazip SEO Project, my general strategy failed. The best way to become an entity and get your entity ID is actually opening a profile page within a popular source for the Google Knowledge panel such as Wikipedia or Wiki Fandom.

Knowledge Panel Source Sites for Google
Knowledge Panel Source Sites for Google (Kalicube & Authoritas API data)

In the SEO Project, also “entitization” was not just about the “brand”. Also, “sponsor”, “founder” and “manager” should be entities. The reliability, news, mentions and relevance within the energy industry are also important for a search engine. Imagine that Google suggests a website on the SERP that the founder, owner or the manager of the site’s brand is actually a criminal. It wouldn’t be a reliable “source” right?

To create more E-A-T, we also used Mr. Çağada Kırım’s scientific articles and background on the energy field, we have opened some Wikipedia Pages for the Mountbatten Family’s members, because Encazip is also owned by Mountbatten Family members as partners.

But, using WikiPedia pages was not the best way to proceed for entitization. If you open a Wikipedia page for a brand, in a short time, you will get an entity ID. I have done this before for VavaCars, and you can see their entity ID and Knowledge Graph search result below.

Entity ID and Knowledge Graph search result
Entity ID and Knowledge Graph search result

The image above is a visual description of “how to query within Google’s Knowledge Graph for entities with Python”.  You can see the entity ID of VavaCars with the entity definition and link for the Wikipedia page that I created.

In, we did not manage to successfully create a Wikipedia page, so how did we solve the problem of becoming an entity?  Through increasing the volume of mentions, news stories and third-party definitive articles about the site.

Going deeper into this process is beyond the scope of this case study, but in doing so, we also increased the latent search demand for the brand’s site. It was also useful in increasing the SEO performance and organic rankings since it is a direct “ranking factor”.

If users search for you, Google will promote you on the SERP for the related concepts and terms. Look at the 22nd of December from Google Trends for the search trends related to Encazip:

Google Trends for Encazip
Google Trends for Encazip

It was 100. And, look how December 22nd affected the “Average Position” and also how it was a cornerstone for this SEO Case Study. After December 22nd, Google decided to use Encazip for broader queries with a more solid expertise, relevance and authority.

GSC data: Average Position
GSC data: Average Position

And, thanks to all this heavy branding and intensive news, mentions, search demand Google has recorded Encazip as an entity in its Knowledge Base. The screenshot below is from Google Trends. If you can see a search term as also a “topic” it means that they are an entity.

Google trends data for Encazip
Google trends data for Encazip

If you choose a topic in Google Trends, you will see all related search activity for the entity. And, during all the branding work, press releases and more, we always cared about the context. We always used the “” phrase with the most relevant and industry-centric concepts. Google calls this as “annotation text” within its patents. It means that the sentence’s sentiment and annotations will create a relevance between concepts.

Becoming an entity is not enough but it’s a good stat!  You should also create relevant annotations and connections between your brand-entity and the industry so that you can become an authority.

Encazip’s entity ID for Google’s Knowledge Base is “2F11cmtxkff9”. An entity’s ID can be seen within the URL of the Google Trends.

8.1 Social Media’s Effect for Becoming an Entity and Entity-based SEO

During the Encazip SEO Project, we also used social media actively. My general principle for Social Media is actually using “hashtags”, “images” and also mainstream social media accounts for giving constant activity signal to the search engines.

Social media signals
Social media signals

We know that Google Discovery acts in great parallel to Social Media activity, even without an official statement. We also know that Google indexes social media posts by dividing them into hashtags, videos, posts and images.

Google wants to see a brand’s social media accounts in the organization’s structured data. We also know that Google has placed these links in the knowledge panels of entities, and even placed special places on social media links in the “Update the Knowledge Panel” section.

Compare the SERP for two competitors’ social media searches
Compare the SERP for two competitors’ social media searches

We have some old explanations from Matt Cutts about “social media links” and how they try to interpret them for search quality, and from Google’s old changelogs, we know that they scrape social media accounts and posts to understand the web better. Between 2010-2015, social media activity was an important ranking factor, even back then there were “post services” as a black hat method.(Editor’s Note: Not that anyone we know used them of course 😉).

Below, you can see my general ten rules and suggestions for Encazip for social media activity

  1. Always, be more active than the competitors.
  2. Always, have more followers and connections than followers.
  3. Create new hashtags with long-tail keywords.
  4. Use hashtags within a hierarchy, such as “#brandname, #maintopic, #subtopic”.
  5. Always try to fetch the latest and most popular hashtags for every mainstream social media platform.
  6. Try to appear in Google’s Twitter, TikTok, or Instagram short video carousels.
  7. Have more indexed content on Google and Bing within the social media mainstream platforms than your competitors.
  8. Use original images with links to the main content.
  9. Syndicate the content distribution with social media platforms along with content-sharing platforms such as Quora, Reddit, Medium.
  10. Consolidate the ranking signals of the social media post and platforms with the brand entity’s main source which is the website.

The tenth point is actually the main purpose of my social media activity within all SEO Projects, and it can be acquired via links, mentions, image logos and entity-based connections.

During the SEO Case Study, was active on Instagram, Facebook, Linkedin, Quora, Reddit, Medium, Twitter, Facebook and YouTube with hierarchical and derived hashtags with keywords.

Instagram, Linkedin, Twitter, YouTube, Reddit, Quora, Medium and more…
Instagram, Linkedin, Twitter, YouTube, Reddit, Quora, Medium and more…


8.2 Local Search and GMB Listing’s Effect on Entities in the SERPs

As a holistic SEO, it’s not just the technical, coding side or content side that is important, the local search activity is as important as the social media arena. Google unifies every ranking and relevance signal along with quality signals from the different verticals of search and web.

In this context, we can clearly say that the Local Search quality of an entity is also an active SEO factor for web search results. Thus, Encazip has performed a “review marketing” campaign and with honest reviews that were requested from the customers.  The company’s social media posts with custom-designed images were being posted to the Google My Business posts.

Encazip Knowledge panel and SERP
You may see the review count, questions, GMB Posts and how Google relates with the most authoritative entities unlike the beginning phase of the SEO Project


After many positive reviews and lots of related questions which were answered by Encazip’s experts, started to be grouped with the biggest energy companies on the Google My Business panel within the “People also Search” feature. This was clearly a quality score increase and a relevance signal increase which was great.

In other words, from all search verticals and web platforms, we created a high activity level, with strong quality and authority signals while consolidating them for the search engines’  algorithms.

9.0 Protecting the Site Migration Route’s old domain was “”. It has been migrated to Encazip, but during the project, the old 301 redirects expired because the registration for the old domain expired!  (Fortunately, Mr. Cagada Kırım noticed this problem before we bought the old domain again and redirected it to

A screenshot from 2013 for the You can imagine the “historical data and relevance” for the industry
A screenshot from 2013 for the You can imagine the “historical data and relevance” for the industry

This is important to take advantage of the old domain’s brand authority and relevance for the queries that it has historical data. This section is also related to the “Uncertainty Principle of Search Engines”. It takes time to convince search engines’ algorithms, if you cancel a site migration, it can hurt your brand’s reliability, thus this is also another important step to get right.

10.0 Authoritative, Semantic SEO for Content Marketing: Every Letter Matters

In Encazip, I implemented the Semantic SEO principles. I used the “ontology” and “taxonomy” for all the relevant topical graphs under a logical hierarchy and structure. For content creation, we have educated the authors, taught them Natural Language Processing rules, terms, and their importance. In this process, I also should pinpoint the importance of educating the customer’s team.

If you don’t educate your customer, you will be over-exhausted and you will compromise the quality of the content. To prevent this unwanted situation, a holistic SEO should educate his customer. In this context, I recommend you to read some important Google Research Papers below:

  1. Translating Web Search Queries into Natural Language Questions
  2. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
  3. Siamese Multi-depth Transformer-based Hierarchical Encoder for Long-Form Document Matching

And, lastly, I recommend you to read anything Bill Slawski and Shaun Anderson write 😉
Lastly, if you are a true SEO nerd, then you can read my Topical Authority Lecture with 4 SEO Projects with a summary of Search Engine concepts and theories.

During the SEO Project, we have written, redesigned, repurposed, reformatted, and republished more than 60 articles. Also, we have started to create new content hubs by complying with Google’s entity taxonomies.

"Since the day we started working with Mr. Koray, his work has always had positive effects on
He also taught our team a lot about both SEO and technical SEO. Our work, with the guidance of Mr. Koray, has always ended well.
He contributed a lot to to become what it is today."

Yağmur Akyar

10.1 Different Contexts within Semantic Topical Graphs

An entity has different contexts. Google calls this Dynamic Organization of Content. A brand can be an authoritative source for an entity’s context. Such as ‘electricity prices’. But, there is also a connection between electricity production and electricity prices. Also, electricity production is connected to the electricity definition and science. Thus, for calculation, definition, production, consumption, science and scientists, Encazip started a comprehensive content production process based on semantic search features.

Dynamic Content Organization for the entity of electricity
Dynamic Content Organization for the entity of electricity

P.S: And, Do I need to talk about keyword gaps or etc… in 2021?

10.2 Image Search and Visual Content Creation

In images, there are two types of entity according to the Google patents, one is object entity, another one is the “attribution entity”. We also used “EXIF Data” and IPTC for image SEO. Images are designed as unique or chosen as unique. The brand’s logo is being used as a watermark. We have determined Search Engine Friendly URLs and alt tags for the images. I also specified how an image should be chosen for the article or a sub-section of an article. To determine this, I used Google’s Vision AI and other Search Engines’ image search features and tools

Relevance score
Relevance score

this section, I must also say that Microsoft Bing’s Advanced Image Search capacity helped me. Unlike Google, they index every image on a web page. And, they have a faster “snap and search” infrastructure for image search.

I recommend you read the documents below, to understand this section better.

  1. Ranking Image Search Results Using Machine Learning Models
  2. Facial Recognition with Social Network Aiding


10.3 Image Sitemap, Image Structured Data and Representative Images for Landing Pages

Google has announced its sitemap syntax in June 2005 and improved the sitemap understanding, syntax, and tag structure over time, for instance, in 2011 Google has announced that they can understand the hreflang from Sitemap Files. In this context, Image Sitemap files can have different tags such as “caption”, “geo_location”, “title”, and “license”. Image Sitemaps or images in the sitemap files are one of the useful communication surfaces with the search engine for helping its algorithms to understand the role, content, and meaning of the image for a web page.

In this context, all of the representative images and also Largest Contentful Paint Elements of the web pages are added into the sitemap files. In other words, a regular sitemap file has been turned into a complex sitemap with URLs and also images. Below, you can see a complex sitemap example for Encazip that includes images and URLs at the same time in the sitemap file.

As a second step for better communication with the search engine, the images on the web page have been added into the FAQ Structured Data. The object and subject entities within the images, texts, colors, and any visual communication element can strengthen the context of the content within the FAQ Structured Data. Thus, not just the first image, but all images are added into the FAQ Structured Data as JSON-LD. Below, you can see an example.

FAQ Schema Markup
FAQ Schema Markup

To differentiate these two image sections from each other, different subfolder names and paths are used. As a final step, to support web search via image search, and increase the quality, usability, and click satisfaction signals, the search engine’s overall selections for specific queries in terms of image search are used. In other words, if someone searches for a query and Google shows certain types of images, we have analyzed these images’ object and subject entities to use them within our featured images too.

Below, you can see an example. For the query “Electricity Prices”, you can use a “table with prices” as an image (I have used an HTML table which is more clear for a Search Engine) or you can add a “bill that shows the electricity prices”. As a result, from 60+ rank, the web page has ranked as 6th for image search.

Google SERP design changes
The image design and color choices also shows itself as an attention grabbing way in Google’s SERP Design.  Image N-Grams of the search engine have been used to understand the search engine’s perspective and content tuning for the Semantic SEO Optimization process

And, for topical authority analysis, it is not just about “text content”, it is about “all of the content”. Thus, every gain from every vertical of search, whether it is textual, visual or vocal is a contributor for winning the broad core algorithm updates and dominating a topic, network of search intent and queries from a certain type of context.’s Image Impression Change for the last 12 months’s Image Impression Change for the last 12 months

11.0 Importance of Clear Communication & Passion for SEO on the Customer-side

To be honest, Encazip is the most easy-going, problem-free SEO Case Study that I have ever performed until now. Because the customer’s team is very positive and passionate about SEO. I know that I have spent double or triple the energy that I spent for Encazip on other SEO Projects that produced less efficiency. But, what was the difference?

The difference is the mindset. A Polish proverb says that, “You can lead a horse to water, but you can’t make him drink”. And, SEO Project Management is on the same page with this proverb. That’s why I always try to be careful while choosing my customers. SEO is not just a “one-person job” anymore, it needs to pervade company culture.

In my opinion, the customer’s character, mindset and perspective on SEO are the main factors that govern an SEO Project’s success.

(As a confession: In the old days (before 1 August 2018, aka Medic update), I was a blackhat SEO, thus I didn’t need my customers to love or know SEO, but after a while, Google has fixed lots of gaps in its algorithm while changing my perspective too. That’s why I have learned to code, understand UX web design and much more!).

And, that’s why creating an SEO Case Study with an uneducated customer is like “Making the camel jump over a ditch!”, or with the Turkish version, “deveye hendek atlatmak”.

11.1 Importance of Educating the Customer in Advanced SEO Concepts

How can you educate the customer? If you want to talk about just simple and easy SEO terms, it won’t help you to create SEO success stories. That’s why my biggest priority for improving the customer’s comprehensiveness for SEO is paying attention to the “smallest details”.

That’s why the main headline of this SEO Case Study is that, “Every Pixel, Millisecond, Byte, Letter and User Matters for SEO”.

From the technical side, you should focus on “bytes” and “milliseconds” with the IT and Developer team, while focusing on “UX, Content, and Branding” with your marketing and editorial teams.

Our work with the Holistic SEO Approach was at a new level of difficulty for us. Every SEO meeting was like an education and ended with a to-do list that included a lot of hard-work.

Erman Aydınlık

12.0 Importance of Broad Core Algorithm Update Strategy for SEO

Broad Core Algorithm Updates are the algorithmic updates for the core features of Google Search Engine. Google announces its broad core algorithm updates officially with some extra details such as update rolling out time, and update rolling out finishing time. Before the Medic Update (1st August 2018 Google Update), all of the Broad Core Algorithm Updates are called “Phantom Updates”. Because these updates are not officially announced, the SEO Community calls them “Phantom Updates” while Google calls them “Quality Updates”. Since Broad Core Algorithm Updates affect the crawl budget, authority, and quality assignment of a source (domain) on the web in the eyes of Google Search Engine, having a solid Broad Core Algorithm Update Strategy will help an SEO to manage the SEO Project more effectively and time-efficient.

Google Search Liaison
Google Search Liaison

To use Broad Core Algorithm Updates as an SEO Strategy, I have written a concrete SEO Case Study with

During the SEO Case Study and Project, there were two Broad Core Algorithm Updates, one is the December 2020 Broad Core Algorithm Update, the latter one is the June 2021 Broad Core Algorithm Update. Encazip has won both of the Broad Core Algorithm Updates of Google, and in the next two sections, you will see how Google compares competing sources on the web with each other while deciding which one should be ranked for the determined topic, and niche.

12.1 Effects of December Broad Core Algorithm Update of Google on and its Competitors

When you put so much effort into planning and executing such a comprehensive SEO programme, then you need to ensure you have a variety of SEO tools and software at your disposal to help you coordinate teams and activity, and to manage and report on SEO performance.  I use a combination of tools including Authoritas and Ahrefs.

Here’s a few examples of how you can manage the SEO process for a project of this nature with Authoritas.

You can track queries and your performance for these queries along with the SERP Features and queries’ search intent with Authoritas.

Accurate keyword rank and intent tracking
Accurate keyword rank and intent tracking

With Authoritas’ Google Search Console Module, you can examine all of the queries and their traffic productivity as above.

Authoritas GSC module shows impact of SEO efforts
Authoritas GSC module shows impact of SEO efforts

With Authoritas, you can add tasks and also solve On-Page and Technical SEO problems:

Link relevance of internal and external link anchors


With Authoritas, examining the referring domains and their relevance and importance for the SEO Performance can be measured

Authoritas measures referring domains
Authoritas measures referring domains

With Authoritas, you can also find the best experts and authors for your industry, for your PR and Marketing campaigns.

Find influential authors for outreach
Find influential authors for outreach


With Authoritas, you can filter the branded queries and non-brand queries for CTR and keyword profile analysis:

Brand and non-brand filtering
Brand and non-brand filtering

And you cancreate a custom device-based CTR Model for estimating SEO visibility:

Custom CTR model

You can crawl a website with Authoritas and blend the Google Analytics, Google Search Console Data for the specific pages along with Technical SEO attributes. And you can generate FAQ questions for different queries, languages, Search Engines and regions. This is actually a unique feature for detailed semantic content marketing.

Discover frequently asked questions
Discover frequently asked questions


With Authoritas, you can also create SEO tasks and assign them to your team members.

SEO task management
SEO task management


Encazip was impacted by the December Broad Core Algorithm Update of Google. But, in this section, I will show a comparative analysis based on the Ahrefs data charts including Encazip and its competitors.

Its first competitor lost most of its traffic.

Chart showing loss of organic traffic
Chart showing loss of organic traffic - 1st competitor

Below, you will see the second competitor’s graphic.

Chart showing loss of organic traffic
Chart showing loss of organic traffic - 2nd competitor

And, this is the last year’s trend for Encazip.

Organic traffic growth chart
Organic traffic growth chart

Every pixel, millisecond, byte, letter and user is behind this difference!

12.2 Effects of June 2021 Broad Core Algorithm Update of Google on and Competitors

Google recently announced another Broad Core Algorithm Update on June 2nd 2021. Before the June Broad Core Algorithm Update was announced, Google was switching between sources in its SERPs and this was affecting the traffic of Encazip.

During these “source switching periods”, I tried to publish and update more content while supporting the site with press releases, social media, and accelerating the delayed improvements. Search engines always try to differentiate the noise from data, and when they try to gather meaningful data from the SERP, feeding them more positive trust, activity, and quality signals is helpful. In this context, you can check the effects of the June 2021 Broad Core Algorithm Update and its consistency with the December 2020 Broad Core Algorithm Update in terms of the direction of the decisions of the search engine.

The first competitor,’s traffic can be seen below. They increased the overall query count, but their traffic continues to decrease, it is an indicator that there is not enough contextual relevance between queries and the source.The first competitor,’s traffic can be seen below. They increased the overall query count, but their traffic continues to decrease, it is a signal for not enough contextual relevance between queries and the source.

Impact of broad core algorithm update
Impact of broad core algorithm update

The second competitor, also increased the overall query count, but the traffic continues to decrease, and you can see how these two main competitors’ graphics are too close to each other, it means that they have been clustered together by the Search Engine.

Impact of broad core algorithm update
Impact of broad core algorithm update

Below, you will see Encazip’s organic traffic change for the June 2021 Google Broad Core Algorithm Update which includes a 100% organic traffic increase.  (Editor’s Note: Don’t graphs like this make your day? 😉 )

Organic growth chart -
Organic growth chart -

A Broad Core Algorithm Update Strategy for every SEO Project should be improved and reinforced. Brands and Organizations might tend to forget the effect of Broad Core Algorithm Update’s strong effects, after two or three months. In this case, an SEO should make every member of the client remember how important it is, and how to create the trust, quality, and activity signals for the search engine.

12.3 Effects of July 2021 Broad Core Algorithm Update of Google on and Competitors

Google finished rolling out the Broad Core Algorithm Update on 13th of July. Also, another spam update that focuses on the affiliate links took place on the 27th of July. has tripled its organic traffic after the July 2021 Broad Core Algorithm Update. In other words, the consistent signals from the search engine became more obvious, and strongly reflected. Below you will see the,, and the’s change graphics during the 2021 July Broad Core Algorithm Update.’s organic performance change after the 2021 July Broad Core Algorithm Update:

SEMrush organic growth chart
SEMrush organic growth chart

Same change can be seen from Ahrefs too.

Ahrefs organic growth
Ahrefs organic growth’s organic performance change during the 2021 July Broad Core Algorithm update of Google can be seen with the negative impact as below.’s organic performance’s organic performance’s change can be seen below.

Gazelektrik organic performance
Gazelektrik organic performance

The last situation of’s organic performance, and positive changes thanks to the reliability of the brand can be seen below.

Encazip organic performance
Encazip organic performance

Because the SEO project has been approached holistically, the site has increased in SEO Visibility regardless of the type of Google update. With two broad core algorithm updates, two spam updates, page experience algorithm updates along with countless unconfirmed and unannounced updates, the search engine has favoured on the SERP time and time again, with higher confidence for click satisfaction due to the always-on multi-faceted SEO improvements we have been making and continue to make.

12.4 Effects of November 2021 Broad Core Algorithm Update on

The 2021 November Local Search Update, and many other changes, continue to affect’s SEO performance along with its competitors. Since August lots of things have changed in In this chapter, these SEO related changes and the ongoing search engine updates’ effects will be discussed.

Below,’s SEO Performance change during and after the November 2021 Broad Core Algorithm Update can be seen below.

Exciting organic growth trajectory has been affected positively by the November 2021 Broad Core Algorithm Update, the Content Spam Update and the Local Search Update. The website has reached the maximum query and organic search performance.

The changes that have been made over this period can be found below.

Website Migration to ReactJS and NextJS

During the last 3 months, has performed two different types of site migrations. A site migration can be performed in four different ways.

  1. Site Migration without URL Change.
  2. Site Migration with URL Change.
  3. Site Migration with Framework, Back-end Structure Change
  4. Site Migration with Design Change

Website Migration to ReactJS and NextJS represent a framework and back-end structure change. During the site migration, I determined the following basic terms and rules for the development and project management teams.

  1. Do not change the website structure, design or URL tree before a core algorithm update.
  2. Do not change the content, design and framework at the same time.
  3. Perform the migration during a “non-trending” season without risks.
  4. Be sure that image, text and link elements are visible on the web page even if the JS is not rendered.
  5. Be sure that the request count, size and request origins are fewer than before.
  6. Do not increase the size of DOM elements.
  7. Do not lose the previous improvements.
  8. Don’t lose structured data implementation.

After the July 2021 Broad Core Algorithm Update, I gave the go ahead for the website migration. Since there are less searches over the summer it was a safe-zone for a migration. During the website migration for the framework, we experienced the following mistakes and obstacles.

  • The Virtual DOM was not used despite the NextJS advantages.
  • The framework migration wasn’t performed on time due to technical problems.
  • The request size, and count were bigger than before.
  • Code-splitting gains were lost.
  • The image loading prioritization, and place-holders are lost.
  • The DOM Size was larger than before.

The biggest potential benefit of the NextJS and ReactJS migration is using the Virtual DOM. You can see how fast a Virtual DOM exercise is below.

And, this is from

From the Virtual DOM, I was able to open 4 different web pages in 6 seconds. It was only 1 for Another problem during the framework migration was that some of the Technical SEO earnings were lost. Thus, another Technical SEO sprint was started.

You can see how these mistakes and neglects affected the organic search performance.

Since the development team lost time for the framework migration, the URL Property migration was rushed. Based on those errors, the Core Web Vitals was negatively impacted for the website. You can see the change for Core Web Vitals as below.

Core Web Vitals reporting in GSC
Core Web Vitals reporting in GSC

Website Migration to the WWW URL Property from without WWW Version

During the URL Property Migration, the most important parts of the checklist are as below.

  1. Fix all of the internal Links, do not use the without “www” version.
  2. Fix all of the URLs within the sitemap.
  3. Fix all of the social media, GMB (GBP) links with the new version.
  4. Create a new GSC Account, and submit the new URLs.
  5. Remove the sitemap from the previous GSC Account.
  6. Have at least 45 days between the Framework and URL Migration.
  7. Do not include the legacy URLs, or old web server files.

During the URL Property Migration, the internal links were cleaned in an efficient way, but the legacy URLs are resurrected with 404 status codes. In other words, the old deleted URLs are put into the internal links, and Google starts to crawl, or even index some of those old URLs.

Crawl issues
Crawl issues

Some of those can be seen above. Because of these errors, Google couldn’t understand the canonical version of the website properly. It continued to keep both versions indexed. Below, you will find the “www” URL Property performance report.

GSC performance chart for www version
GSC performance chart for www version

As you see there is a sudden increase and then later a gradual increase. Google couldn’t find the canonical URLs or canonicalise the proper URL Property due to the framework and URL changes. Thus, 25% of the website has taken traffic from the root domain for nearly 2 months.

More GSC stats for root domain
More GSC stats for the root domain

To fix these types of problems, the “301” redirection might not be sufficient. Because, Google’s crawl hits might use the “previous content” for updating only some of the resources on the web page. If there are URLs with less traffic, and some external links, it might be harder for Google to see the canonical version. Since the traffic is less, Google might not update the content with new crawl hits, and as there are external links, the canonicalization might benefit the legacy URL Property.

After 3 months, some of the old URLs are still being used.

Old URLs are still indexed

And, we see that Google has indexed the old URL Property’s “sitemap” file. This might be because of deindexation of other URLs, and Google might choose a “left-out” URL to index since others have disappeared. It still shows there is a canonicalization problem for the website.

Some of the URLs are indexed with both versions despite the 301 redirect. For these situations, best practice is using the URL Hints with 100x Status Codes. In a Response Header, the canonical version of a content’s URL is shown below.

Link: <>; rel=”canonical”

By using the canonical in HTTP Headers, you can state an extra signal to Googlebot for the content’s new place. Adding multiple, consistent and meaningful aligned signals for a search engine is always beneficial to alert different algorithms to the same content. During the URL Property Migration, the external references from Digital PR, news and YouTube Channels have been used for further canonicalization.

Another change during the URL Property Migration was starting to use a CDN subdomain for the images.

After a while, Google started to drop the “non-www” version from the SERP. In this example, the external references, mentions, social media references, and external links from digital PR were a protective layer for the branding and canonicalization signals along with the date.

Due to the URL Property Change for the Image Resources, the Google Image Search Results were also affected.

Looking at both of the URLs below, you can see the “deindexation” and new “crawl pattern creation” needed by the search engine.×160/age-enerji.svg×160/age-enerji.svg

In the image you will see that Google is still looking for the old resource URL.

404 errors
404 errors

The effect of the image deindexation can be seen below.

Images de-indexed

As an example query, the “Jeotermal Enerji” had thousands of impressions from image search. While it is just a few, there are many queries in the same situation.

Thus, when the URL Property changes, the old image URLs should be redirected to the new image URLs. If, for image optimization, the URL extension needs to be changed, try to optimize the image format without a URL change.

After all of these differences and new improvements, here are the positive changes.

New Content Sprints and URL Count Restrictions

Most of these SEO related errors happened because the client changed the team members. Since the team changed, the SEO knowledge was lost. Thus, in SEO Consultancy, keeping the SEO culture alive in a company is a must. In this context, the URL Count Restriction is a new rule that I brought for the client, because the “press releases” are added to the website by creating tons of new URLs.

These new URLs didn’t provide any information about the brand, and it diluted ranking signals, and PageRank distribution by making the website bigger. Thus, I told the client to not add a new URL, or change a URL without agency approval. Besides these issues, five things were still worked for the website.

  • Content Sprint with Semantic SEO has been completed.
  • New Content Sprints were launched, especially for new industries, and localized search behaviors.
  • The Brand Reputation gets better with the new digital PR and press releases.
  • The trending searches and events in Turkey feed the website’s authority for the best ranking web pages.
  • The Social Media Activity, Signals and Subscribers increased along with the engagement rate.

Lastly, most of these bad things didn’t happen “too early”, before the Broad Core Algorithm Update. In other words, it didn’t change the website, because it didn’t create enough historical data. In this context, the Content Spam, Local Search and Broad Core Algorithm Updates during November were positive. And, most of these errors will be fixed or scheduled to be fixed in future updates.

Organic growth chart for Encazip Dec 2021
Organic growth chart for Encazip Dec 2021 latest SEO Performance graphic from the Ahrefs can be seen above.

Last 6 months KPIs
Last 6 months KPIs

The last 6 months growth comparison for can be seen above.

Competitors and Elektriksepeti continue to lose traffic.

During the November 2021 Broad Core Algorithm Update, the effect and prominence of Holistic SEO has improved. Gaining support from every SEO Practice will guarantee success despite organizational problems, or unexpected errors of individuals.

12.5 Effects of May 25th 2022, Broad Core Algorithm Update

The May 25th 2022 Google Broad Core Algorithm Update affected finance, insurance, credit, news and affiliate marketing websites heavily. The update is perceived as an affiliate marketing inhibitor, while it is decreasing the click distance between the product purchase and the SERP. The same “direct-actionable” SERP design affected the Aggregators for different types of services such as finance, insurance and credit. is affected from the May 25 2022, Broad Core Algorithm update slightly negatively since it is an aggregator and it lost most of the technical SEO improvements during the website migration.

You can see the organic search performance graphic for below.

Organic traffic impact of algorithm update
Organic traffic impact of algorithm update

The marked point demonstrates that the website started to lose queries and organic traffic. Below, you can see the AHrefs Organic Search performance graphic for

Organic traffic impact of algorithm update
Organic traffic impact of algorithm update

In a connected and correlative way, Ahrefs demonstrates the lost query count, and organic traffic after the 25th May 2022 Broad Core Algorithm Update.’ fared better than its competitors.

Gazelektrik has lost heavier traffic than during the May 2022 BCAU (Broad Core Algorithm Update).

Sepas has gained slightly better traffic, but it didn’t last. After the June 2022 Product Review Update, it started to lose traffic again. These types of short-term BCAU effects might signal that the Micro Core Updates and later Product Review-like updates are reversing the prioritisation of the web source.

Enerjiatlasi continued to lose traffic with May 2022 BCAU Update like other competitors.

However, EPDK (official government website) increased its traffic with a big jump along with the query count.

Conclusions for the 2022 May BCAU of Google

The summary of the conclusions for 2022 May BCAU of Google for are set out below.

  • has lost the technical SEO improvements during the website migration
  • The indexed URLs are changed over a long period of time
  • Google took nearly 3 months to remove all the “without WWW” versions of the website from the
  • The re-indexing, and 301 evaluation caused search engines to decrease the confidence for the indexed URLs, since their URL ID is different, and they are new on the index, and “processing the text”, comparing the redirection source to the redirection target takes longer. This is why most site migrations show positive effects in the long-term, when the search engine is ready to believe the migration, or that’s why the “Change of Address Tool” of GSC works with 6 months of timeline for waiting with a keep-alive 301 redirection
  • The May 2022 BCAU of Google targeted the affiliate marketers without expertise while giving more exposure to the actual service and product providers
  • and its competitors all lost some organic search traffic
  • May 2022 BCAU affected the websites according to their classification based on their site type, rather than individual site quality scores or a site’s authority
  • It means, even if you have improved your website, it would only keep the website neutral for traffic change or decrease the level of traffic loss
  • May 2022 BCAU of Google is more related to query to web source distance rather than web source and quality comparison
  • The results of the May 2022 BCAU shows connectedness with the Helpful Content Update, because it demonstrates a stricter connection between “who you are and why you should rank for the query”

Effects of Expanding the Brand Identity and Service Area

Connectedness of Topics and Propagation of Expertise from a Topic to Another

As an addition to the conclusions above, has expanded its identity and services further. It is important because, a brand might hinder its trustworthiness heavily if it doesn’t provide quality and unique services, information and conceptual expertise for all the verticals which it serves. has focused and proven its authority mainly for the electricity industry, now, it is also a project for Credit, Insurance, and even Natural Gas, etc. Thus, the newly published content causes a search engine to examine the questions below.

  • Did this website rank for these queries before?
  • Is this the first time that this web page appears for this query?
  • Did the web source define this concept before?
  • Did the web source explain new services and products within their brand identity?
  • What is the valid brand-related reason for business to dive into this vertical?
  • What is the historical data and user behaviors that overlap between the new and old query networks?
  • Are these new topics related to each other?

During this time, due to heavy investment pressure on the company, the articles and content on insurance, credit and other market verticals are not optimized for serving end-users. It is mainly created for publication frequency, indexation, historical data and higher chance of brand relevance for the industry, but not for competitive ranking. Thus this loses brand identity and connectedness between the topics might affect its situation in the May 2022 Broad Core Algorithm results and later.

12.6 Effects of September 2022 Broad Core Algorithm Update, Helpful Content Update and Fifth Product Review Update was the main winner from the 2022 September Broad Core Algorithm Update.

You can see the results of the organic search performance increase graphic for 2022 September for, below.

Impact of Sept 2022 Algorithm update
Impact of Sept 2022 Algorithm update

New top 3 rankings for 400 queries and 600 new queries between 4-10th place, along with 5,000 new organic queries in total, all helped return to the top of the industry again.

Impact of Sept 2022 Algorithm update
Impact of Sept 2022 Algorithm update

The 3 months comparison over Year of Year shows that the organic growth is over 650%. The quick summary of the organic traffic increase for the September 2022 Broad Core Algorithm is below.

  • 670% organic click increase which is equal to more than 600,000 organic clicks
  • 765% organic impression increase which is equal to an increase of impressions of more than 13,500,000
  • 26% organic average position increase which is equal to more than 2.9 organic positions 

The Last 28 days of organic search performance comparison year-over-year is below.

Impact of Sept 2022 Algorithm update -last 28 days
Impact of Sept 2022 Algorithm update -last 28 days

Three weeks later version of the from SEMrush is below.

Impact of Sept 2022 Algorithm update - 3 weeks later
Impact of Sept 2022 Algorithm update - 3 weeks later

The organic search increase here is over 175% and organic impression increase is over 245%.

To achieve this difference by dominating the niche while showing the brand for the new industries and fixing the technical SEO tasks as before, there is a tight publishing and semantic SEO research and effort. started to lose its traffic slightly.

Enerjiatlasi regained the positive ranking state by increasing its visibility slightly.

EPDK continued to positive ranking state with new gainings during the 2022 BCAU Core Algorithm Update.

Zorluenerji started to lose traffic as opposed to the results of the May 2022 BCAU of Google.

CKBogazicienstitu started to lose its traffic after September 2022 BCAU as a result of opposing effects of the May 2022 BCAU of Google.

Topical Consolidation for the New Sectors, and Initial Ranking

The topical consolidation is processed in the updated version of the Entity-oriented Search SEO Case Study. The topical consolidation is used for both of the projects, and The topical consolidation represents the increased topical relevance of web sources via using more central knowledge domain terms in the important information retrieval zones with semantically related entities. It helps for saving web sources from the irrelevant queries, while increasing the rankings for other relevant queries. Algorithmic hierarchy works with the output of an algorithm as input of another.

In other words, if a web source ranks for irrelevant queries to the web source identity, the irrelevant queries might make the web source less relevant to the core topic. For example, if a web source ranks for “electric consumption”, and “electric price fluctuations” mainly, the query like “electric vibes” represent a social context rather than economical, or scientific context. But, due to some common phrases, and phrase lists for index construction, the web sources might rank for these types of queries. And, these irrelevant queries might dissolve the overall relevance of the web source for the topic. has started its content production for the Credit and Insurance industry from scratch. And, the new technical SEO sprint is started by fixing the existing errors, while fixing the design problems and layout component order problems for macro context and search intent matching.

The initial results for the Insurance Sector initial rankings for are below. 

Explosive growth in insurance queries
Explosive growth in insurance queries

The Google Search Console graphic to show the initial rankings of the Insurance Industry is below. It signals the “How Google Ranks” SEO Case Study topics.

GSC reports explosive growth in insurance queries
GSC reports explosive growth in insurance queries

The Credit Industry subfolder’s organic search performance graphic for the from AHrefs is below.

Explosive growth in credit industry queries
Explosive growth in credit industry queries

The credit industry organic search graphic for is below.

GSC reports explosive growth in credit industry queries
GSC reports explosive growth in credit industry queries

Both industries create 10% of the daily clicks for the entire website for now. It means that the website started to unite different topics under its base identity which helps further rankings.

Summary of Conclusions for 2022 September BCAU Effects of Google

The summary of conclusions for 2022 September BCAU Effecs of Google is below.

  • The BCAU 2022 September Broad Core Algorithm Update of Google reversed some of the effects of the May 2022 Broad Core Algorithm Update between the aggregators, affiliates and actual service and product owners
  • The sources with higher brand authority, popularity and trust signals ranked higher even if they do not have a solid historical background. Because, despite being new for the insurance and credit industry, outranked many other sources thanks to the depth of information that is provided and high level of trust signals
  • The web sources are started to be judged with helpfulness of content and information literacy, quality and content advisories
  • Search engines have started to diversify the search results further with different types of SERP tests such as multi-answered featured snippets, things to know, popular products and other types of SERP features. started to rank for more featured snippets than its competitors by dominating a good deal of Information Extraction focused SERP features
  • The content publication frequency is kept higher than the competitors and it continues to help continuously gain new ranking queries 
  • An increase in brand mentions through digital PR and publicising the CEO of the company with high-authority sources helped even further
  • The authoritative brands increased their organic traffic while they lost traffic from some certain queries. The service-providers couldn’t rank better for the definitional queries, while they rank higher for their brand names, including comparative queries for certain regions. It signals that the brand search demand might make the source closer to its own brand-related topics while not leveraging their rankability for the non-brand related queries to the same degree
  • The Ranking State and Rankability of the web sources continue negatively or positively in the same direction until they get another ranking confidence from a core update. Thus, continues to rank higher thanks to historical data and continuous quality increase. The September 2022 Broad Core Algorithm Update helped to gain further queries and rankings thanks to its long-time positive ranking state, and constant quality increase which requires it to stay as an active source

The search engine Google’s quality and SERP serving tests, announcements and updates are explained in more detailed way in the “Entity-oriented Search SEO Case Study”.

Notes for Helpful Content Update and Product Review Update

Helpful Content Update and Fifth Product Review Update try to understand the first-person experience and expertise for different topics to rank real-world authorities higher on search results. Helpful Content Update focuses on Information Responsiveness to understand whether the information on the website is responsive to the query needs or not. The Fifth Product Review Update tries to understand whether the product review is responsive for the possible product seeker for a certain need. Both of the updates encapsulated the 2022 September Broad Core Algorithm Update. 

The focused on publication frequency rather than the individual article’s informativeness for a fast launch. It is advantageous because processing the text takes a long time. The headings of the articles and context vectors are optimised for higher relevance and responsiveness signals, but the actual information inside the articles, they are comprehensive and good for only the root pages, but not for the individual definitional concept pages.

Thus, in the coming months, the situation of organic search gains might slowly change. To prevent this situation, to support the gained rankings, a content revision and configuration campaign will be started. In this case, the historical data, positive ranking state, and search engine’s trust will be racing against the text processing capabilities of the search engine. To understand the effect of historical data and search engine trust, you can watch the related video SEO case study. serves product reviews too, reviewing brands, banks, or electric distributors are related to the product review update. And, these reviews need to be strengthened further. At the moment, they are above the thresholds. But, as I stated in the Quality Thresholds SEO Case Study, an understanding of quality depends on the competitors’ quality.

13.0 Last Thoughts on Holistic SEO and Case Studies

Most SEO Case Studies usually focus on just one perspective or angle.

Organic growth chart
Every SEO likes seeing organic growth like this

If a project focuses on only one vertical of Search or SEO, the scope of its findings will naturally be limited. SEO is even being impacted by weather changes, search behavior changes and the agenda of countries - so take a big picture view but pay attention to every important detail!

A search engine might not always evaluate different data dimensions and their relationship to each other as it should. That's why trying to understand the search engine's perspective and rules/guidelines is important, whilst keeping a strategic eye across all aspects of the SEO Project.

We're ready for SGE. Are you?

The rollout of SGE will create unprecedented risks to your hard earned organic traffic, as well as new opportunities to succeed.

You need to be ready.  The only question is, whether you want to be ready now or later?

SGE Keyword Rank Tracker Software -Authoritas