Why Facebook’s crawler doesn’t create a cache?
-
Hi,
I see many entries in the access log file:2a03:2880:10ff:16::face:b00c - - [18/Sep/2020:13:39:26 +0200] "GET /some-page/ HTTP/1.1" 200 19153 "-" "facebookexternalhit/1.1 (+https://www.facebook.com/externalhit_uatext.php)"
Why LiteSpeed doesn’t create cache for “/some-page/”? When I type this address (a few minutes later) in the browser I have to wait a few seconds (cache is not working). When I refresh the page again (or I will run it on another computer), it appears immediately (cache works).
Another example – Slack’s bot:
3.89.134.152 - - [18/Sep/2020:15:01:05 +0200] "GET /site-test/ HTTP/1.1" 200 15155 "-" "Slackbot-LinkExpanding 1.0 (+https://api.slack.com/robots)"
Again, LiteSpeed doesn’t create a cache for this site. When I get a page using wget command:
wget https://mysite.com/site-test/
…again, LiteSpeed doesn’t create a cache for this site. I have checked all the options in the plugin and can’t find anything (“Do Not Cache User Agents” -> this field is empty).
Wordpress: 5.5.1
LiteSpeed Cache: 3.4.2
Viewing 7 replies - 1 through 7 (of 7 total)
Viewing 7 replies - 1 through 7 (of 7 total)
- The topic ‘Why Facebook’s crawler doesn’t create a cache?’ is closed to new replies.