Tag: Crawling

What is smart crawling?

Smart crawling allows the crawler to run more frequently, leveraging the XML sitemap(s) of your site. It uses the lastmod timestamps in the sitemap to detect if a page was updated since the last crawl. This allows the crawler to only re-crawl recently modified pages, saving time and resources when . . . Read more

Setting up time-scheduled crawling

It is possible to set a specific time schedule for the crawler to run at a specific time of day. Currently, it is not possible to set the time schedule for a crawler via MyCludo. To configure time-scheduled crawling, submit a support ticket, informing your timezone and at which time of . . . Read more

What is async crawling?

Async crawling is meant for websites where content is loaded asynchronously (AJAX-generated content). AJAX-generated content allows the web page and web browser to process data without having to reload the page. For example, if you hit a “Submit” button on the page, AJAX processes the information and updates the content . . . Read more

Can I set up smart crawling?

Currently, it’s not possible to enable smart crawling directly in MyCludo. To enable smart crawling, submit a support request, letting us know which crawler you would like to enable smart crawling for. A support agent will confirm that your site is eligible for smart crawling and inform you when it is . . . Read more

What are the crawlers’ user agent and IP addresses?

In some cases, the crawler may be blocked from indexing your website. To fix this, you may need to whitelist our IP address to allow the crawler to access the site. Our crawler’s user agent is: Our crawler’s user agent can be referred to simply as cludo. User-agent: cludoAllow: * Our . . . Read more

What is the maximum file size Cludo can index?

Cludo’s crawlers can index files up to 15 MB. Anything larger can be pushed directly via Cludo’s API. The extraction of files removes the size of images and other irrelevant information prior to looking at the file size. For reference, the raw text of the entire Bible is around 5MB.

Deleting a crawler

For security reasons, crawlers can only be deleted by Cludo staff. If you need to delete a crawler, please contact support and let us know the ID(s) of the crawler(s) you would like to delete.

What file types does Cludo index?

The indexability of a file is not defined by its extension (e.g. “.pdf”), but rather by the content type, as returned in the HTTP headers. In the list below, we have added extensions as examples. Supported file types

How many requests does the crawler make?

Our crawler will always attempt to make as many requests as possible, often requesting multiple pages per second, but the actual frequency of requests depends on the server response from the website. Some websites might also have a crawl delay set in their robots.txt, which can impact how many requests . . . Read more

Why is this page not indexed?

Once a crawler has crawled the defined domain(s), you may experience a specific page not being added to the search index. This will typically be due to one of the following reasons: Feel free to contact support if you have further questions on why a page was not indexed as . . . Read more