• Comparison of Three Vertical Search Spiders

      Chau, Michael; Chen, Hsinchun (IEEE, 2003-05)
      Spiders are the software agents that search engines use to collect content for their databases. We investigated algorithms to improve the performance of vertical search engine spiders. The investigation addressed three approaches: a breadth-first graph-traversal algorithm with no heuristics to refine the search process, a best-first traversal algorithm that used a hyperlink-analysis heuristic, and a spreading-activation algorithm based on modeling the Web as a neural network.
    • Using Coplink to Analyze Criminal-Justice Data

      Hauck, Roslin V.; Atabakhsh, Homa; Ongvasith, Pichai; Gupta, Harsh; Chen, Hsinchun (IEEE, 2002-03)
      As information technologies and applications become more overwhelming and diverse, persistent information overload problems have become ever more urgent.1 Fallout from this trend has most affected government, specifically criminaljustice information systems. The explosive growth in the digital information maintained in the data repositories of federal, state, and local criminal-justice entities and the spiraling need for cross-agency access to that information have made utilizing it both increasingly urgent and increasingly difficult. The Coplink system applies a concept spaceâ a statistics-based, algorithmic technique that identifies relationships between suspects, victims, and other pertinent dataâ to accelerate criminal investigations and enhance law enforcement efforts.