Over at Nype - https://npe.cm/ we had the same issue, we moved away from GitHub Pages and we're still recovering from it.
1st issue is basically that, due to how GitHub Pages are setup everything is stored in a limited amount of servers under a limited amount of IP addresses, so this leads to a quota threshold being met and Google pausing on crawling the servers at all.
- https://support.google.com/webmasters/thread/187613472/discovered-currently-not-indexed-error-on-github-pages-hosted-site
2nd issue (my guess) is that any problem with a site is "amplified" due to the slow crawling, and Google sort of abandons a site after too many problems occur. You can check the status of your crawl stats under Google Search Console -> Your Property -> Settings -> Crawling -> Crawl Stats (Open Report) -> "Between the chart and Crawl requests breakdown there is a status button for robots, DNS, Server connectivity", after a ~month after changing servers there is still "residue" of old issues and not all check-marks are green in our back panel .
3rd issue (my guess) is that in your case, you set up your redirections on the 404 page. GoogleBot should in theory handle JavaScript and detect the location change in JS, but overall when a bad URL loads GitHub sends the 404 server status code, and only later it loads the correct page. At Nype we played around with a combination of 404 page JavaScript redirects, mkdocs-redirects plugin to create a valid page with a redirect, so the server status code is at least 200, and some duplicate pages and link rel=canonical mangling, and the results were mediocre at best, likely due to the 1st issue and slow "reassessment" Later on, as the page has now it's own server, we moved on to the proper way of sending a 301 code when redirecting, but too soon to tell anything more.
4th issue (my guess) is that in your case, you have now a lot of pages with possibly low traffic / low amount of backlinks, so Google is reluctant in adding so many links at once
We have also used the currently trending script https://github.com/goenning/google-indexing-script it helped a bit, but due to the 2nd issue manual requests for indexing were better. I personally didn't do it as I'm not the property owner.
I hope at least some of the above helps you out but yeah after working a bit with Google "docs", A LOT of things are not said directly, and users have to guess too much imo.
GitHub Page Still Not Indexed by Google After a Month
GitHub page not indexed on Google YEARS later
Get Github Pages Site found in Google Search Results - Stack Overflow
My GitHub Pages site doesn't appear in google search index - Webmasters Stack Exchange
I created a GitHub Page about a month ago, but it still hasn’t been indexed by Google. I even added backlinks to it on some active websites, like LinkedIn, but that didn’t seem to help.
Is it possible that GitHub blocks Google crawlers for some pages? I’ve seen other people’s GitHub Pages appear in search results, but mine doesn’t. My GitHub repository also has not been indexed, although I have it for a couple of years...
Has anyone experienced something similar or have any tips on how to get it indexed?
Here are the links, just in case - pages: https://deividas-strole.github.io/
repo: https://github.com/Deividas-Strole
Appreciate any help!
You have to create a Google Search Console account and add your page, then typically you just drop a "marker" file in the root (Search Console generates this) so that Google can confirm you really own the page.
Google Search Console
Instructions
(Since the instructions are long and have many links to sub-steps, I'm only providing the link.)
Also, if you're going to use a registered domain name, set that up before you register the site for search.
(Edit: Technically you don't have to do this, sooner or later Google will find you... but this will give your content a much higher-quality score.)
Generally, google finds all website and index them. Sometimes, it's takes time to crawl the new website.
But, you can do this thing manually by following these steps:
- Go to Google Search Console
- Add the website as your property
- Then, verify your property that you're the owner of this.
Have you got any followable backlinks pointing to your profile from other sites?
Github is huge, and Google will presumably see most of its pages as low priority unless it has signals to suggest otherwise. If you only have internal links, it's unlikely Google's crawler will hit many of their pages on a frequent-enough basis to be able to find it by crawling Github alone.
I see the issue is the same on Bing and DuckDuckGo, so I think you're just not very findable.
If your GitHub profile is what you consider your homepage, then popping it in your profile on any other sites you frequent may be enough.
You have to have valuable back links to the repo. Use dev.to or medium to write shore articles about your repos. This is one of the quickest ways to get the repo indexed.
Hi everyone,
I created a very simple Github pages site using the instructions in the documentation. The repository for the site is very simple -- it basically just consists of an index.html file. However, the website is not appearing in any search engines even though it was published more than three months ago.
Any ideas why the website is not being indexed by Google?
Apologies if this is an entirely trivial question - I really appreciate any help!

