views:

57

answers:

3

I'm creating a web crawling application that I want the 'average' user to be able to use. I'm concerned that a web crawling application is probably just too complex for most users though because users need to:

  1. Understand URL structure (domain, path, etc...).
  2. Understand crawling 'depth'.
  3. Understand file extensions and be able to setup 'filters' to narrow their crawling to achieve better performance (or they'll be frustrated with the program).
  4. Understand where URLs are found in pages (image srcs, links, plain text URLs, etc...).

What can I do to help users get quickly acquainted with my program? Or even better, what can I do so the program is intuitive enough that users just 'get it'? I know this seems pretty broad, but if you can confine your answers to web crawlers that should help. I've read up on general usability, ui design, etc... but I'm struggling with the domain I'm working in. Thanks.

+1  A: 

Some ideas:

  • Make an interactive user interface (e.g. a slider for depth, which shows a small picture of folders and subfolders opening as they move the slider)
  • Avoid clutter. Divide the settings into logical tabs.
  • Make video tutorials for the things you need to teach them.
Gert G
+1  A: 

Perhaps you could have a picture of "the web" showing two or three pages each from two or three websites. As the user selected where to find links (for example, images, plain text, links, etc), the parts of the page they selected would be highligted in the images.

Computerish
+1  A: 

Just because a web crawler is complex in implementation, doesn't mean it has to be complex to use. Only offer what is really necessary, use sensible defaults for the rest. That will get you 80% of use cases, and then rely on the other 20% being more willing for have a deeper understanding.

  1. Why should they have to understand this? Depends on the expected usage, but I would of assumed most uses where crawling a full website, so only the domain is needed.
  2. Gert G's suggestion of a slider with extending folder structure was a good one. This doesn't have to be dynamic with the site in question, just an illustration of what it means.
  3. Forget exposing file extensions, instead offer common types of file with icons, possibly even grouping them (e.g. all common image types, jpg, png, gif, go into one 'images' type). Only give raw file extension settings under an advanced config section, those that need it will understand it.
  4. I don't really see why they need to understand this? Surely that's a job for the crawler.
roryf
Thanks! I really like the idea of an simple/advanced options setup. I think this is the most far reaching idea. Setting sensible defaults that will generally find what the user is looking for and generally give good performance will be tricky, but I think it's the right thing to do. As far as file extensions: I already group them as you suggest, but I allow users to enter/remove extensions to the groups (e.g. maybe users want jpgs and not gifs). As for #4: This is a performance issue that can make the difference between waiting a few minutes or waiting a few hours for the crawler to finish.
Chad