MENU

Crawling and authentication

Authentication - general

Crawling secured websites is supported, the configuration needed depends on the method which is being used to secure the website.  The LiSA search crawler can be configured to authenticate using basic/ntlm authentication methods.  To do so, simply input the username/password/domain under the "Crawler settings" in the search module.

Forms authentication

Forms authentication is a cookie/session based authentication scheme supported by the .NET framework.  This is the method used when you use LiSA to control access to various pages on the front-end.  This authentication scheme requires the user to log in via a form on the website.  The LiSA search crawler supports cookies/sessions so it can be configured to log in, prior to a crawl.

Setup

  1. Create a user for the crawler.  It should have the same permissions as the other users which have access to the website.
  2. Make sure you've got a login-page on the website.
  3. Test the user/login page by logging in using the user created in step 1.
  4. Configure the crawler to use the login page as a root page for the crawl.  Copy/paste the full URL of the login page and add the required query-string parameters to the url.  Example:  http://site-domain/loginpageurl/?action=login&username=usernamefromstep1&password=passwordfromstep1&returnurl=/
  5. Test the crawler by starting the crawler task.  If everything goes as planned, the crawler will start at the given login url and log in.  The returnUrl parameter redirects the crawler back to the front page after login and it should continue from there.

Search

Index

Mailing list

Get notified when new versions of LiSA are released!

Check the box next to the list you wish to register to.