• Home
  • Get help
  • Ask a question
Last post 21 min ago
Posts last week 89
Average response time last week 30 min
All time posts 67739
All time tickets 10468
All time avg. posts per day 21

Helpdesk is open from Monday through Friday CET

Please create an (free) account to post any question in the support area.
Please check the development versions area. Look at the changelog, maybe your specific problem has been resolved already!
All tickets are private and they cannot be viewed by anyone. We have made public only a few tickets that we found helpful, after removing private information from them.

#8383 – 4SEO - issue getting started 403 error on homepage

Posted in ‘4SEO’
This is a public ticket. Everybody will be able to see its contents. Do not include usernames, passwords or any other sensitive information.
Tuesday, 14 December 2021 08:57 UTC
Lavers69

Hello Yannick

I have just installed 4seo on my site www.xxxx.com and I have fallen at the first hurdle. The site analysis is stuck at 0% and I am getting a 403 error on the homepage. I think that this might have something to do with a rule in the robots.txt file. From memory my host company put in a number of lines in the robots.txt a few years ago because I was getting a lot of bots visiting the site and eating up the resources. I used to use Jsitemap on this site to create a sitemap, but it is not playing nicely with a component I use called JomHoliday. So that is why I have switched to 4SEO (plus it makes sense given I have used sh404sef for years).

Site details & login

https://www.xxxx.com/administrator/

user: xxxx

pass: xxxx

Thanks

Iain Laverock

Tuesday, 14 December 2021 09:10 UTC
wb_weeblr

Hi

 I think that this might have something to do with a rule in the robots.txt file. 

robots.txt is entirely passive, it does not have any visible effect on the site and should not trigger a 403.

You did have some issues in your robots.txt, which I commented out:

but nothing that can actually trigger a 403.

A 403 is most likely caused by:

- your .htaccess file.

- a firewall rule on your server

Looked around a bit, what's the history of the site? you have some weird things showing:

Anyway, this is happening on the server: something is blocking 4SEO from crawling your page. Be sure to:

- review your .htaccess for anything blocking access from anywhere

- look at your hosting company control panel, and maybe talk to them, they often have some firewall rules enabled by default (seen on o2switch.fr for instance)

Best regards

Yannick Gaultier

weeblr.com / @weeblr

 

 

 
Tuesday, 14 December 2021 09:19 UTC
Lavers69

Thanks Yannick

Yes, I noticed that issue with the com_installer_type. I can't delete/uninstall it. I think it is a legacy from when I used the Jomres component. That was why I moved to JomHoliday, because Jomres became unworkable. I will contact my host and see if they can have a look at the .htaccess.

Thanks for your help

Iain

Tuesday, 14 December 2021 09:21 UTC
wb_weeblr

Hi

OK, I can see that you only disabled jomres, the files are still there, or at least showing in Joomla plugins and extensions list, which might explain why you cannot uninstall these installer things.

I will leave this ticket open in case you need to add something. It will automatically close in 2 weeks if no further comment is made.

Best regards

Yannick Gaultier

weeblr.com / @weeblr

 

 
Tuesday, 14 December 2021 11:16 UTC
Lavers69

Hi Yannick My host thought that they had got 4seo running ok, but it is still triggering a 403 error on the homepage. These are the rules at the top of the .htaccess file. Do you think that they could be preventing 4seo from running properly?

#redirect non-www to www

RewriteEngine On

RewriteCond %{HTTP_HOST} !^www\.

RewriteRule ^(.*)$ https://www.%{HTTP_HOST}/$1 [R=301,L]

#Block all bots except known crawlers

RewriteCond %{HTTP_USER_AGENT} ^$ [OR]

RewriteCond %{HTTP_USER_AGENT} (bot|crawl|robot|spider)

RewriteCond %{HTTP_USER_AGENT} !(uptimerobot|bing|Google|msn|MSR|Twitter|Yandex|Yahoo|Jetpack) [NC]

RewriteRule ^(.*)$ - [F,L]

 

#SetEnvIfNoCase User-agent guzzlehttp not-allowed=1

#Order Allow,Deny

#Allow from ALL

#Deny from env=not-allowed

 

Tuesday, 14 December 2021 13:24 UTC
wb_weeblr

Hi

Totally. 4SEO user agent string is: "4SEO - Internal crawler" so this would be blocked by RewriteCond %{HTTP_USER_AGENT} (bot|crawl|robot|spider)  

You need to change that or add 4SEO to the next option:

RewriteCond %{HTTP_USER_AGENT} !(4SEO|uptimerobot|bing|Google|msn|MSR|Twitter|Yandex|Yahoo|Jetpack) [NC]

 

 
Friday, 14 January 2022 05:34 UTC
system
This ticket has been automatically closed. All tickets which have been inactive for a long time are automatically closed. If you believe that this ticket was closed in error, please contact us.
This ticket is closed, therefore read-only. You can no longer reply to it. If you need to provide more information, please open a new ticket and mention this ticket's number.