I have a "jump page", which is dynamically generated with a big list of links that refers to a set of items I would like to index. I'm trying to point TREX to the jump page and have it follow through all the URLs on the page and index all the items. However, it seems to me that, TREX is only indexing the jump page and not following through to any of the links in that page. Is there a way to make that happen?
I have also tried to create a web address with a URL like http://www.mydomain.com/departments/sales/ (with no index.html or any file name), and that seem to be able to crawl a bunch of other files within that directory. What are the difference between the two scenario?
Currently, I'm doing that by creating a Web Address, and have the index point to the web address as a datasource. I've read other threads about creating Web Repository and/or HTTP Systems, will that make a difference?
Please help. Thanks.