System Summary and Timing Organization Name: Rutgers SCILS Interactive Track List of Run ID's: runit1, ruibsl Construction of Indices, Knowledge Bases, and other Data Structures Methods Used to build Data Structures - Controlled Vocabulary? : NO - Stemming Algorithm: PORTER - Morphological Analysis: NO - Term Weighting: DEFAULT INQUERY (SEE UMASS) - Phrase Discovery? : NO - Syntactic Parsing? : NO - Word Sense Disambiguation? : NO - Heuristic Associations (including short definition)? : NO - Spelling Checking (with manual correction)? : YES - Proper Noun Identification Algorithm? : NO - Tokenizer? : NO - Manually-Indexed Terms? : NO Statistics on Data Structures built from TREC Text - Inverted index - Run ID : ruint1, ruibsl SEE SYSTEM DESCRIPTION FOR GEORGIA TECH FOR THESE DATA - Clusters - N-grams, Suffix arrays, Signature Files - Knowledge Bases - Use of Manual Labor - Special Routing Structures - Other Data Structures built from TREC text Query construction Automatically Built Queries (Ad-Hoc) - Topic Fields Used: DESCRIPTION - Method used in Query Construction - Term Weighting (weights based on terms in topics)? : YES - Phrase Extraction from Topics? : NO - Syntactic Parsing of Topics? : NO - Word Sense Disambiguation? : NO - Proper Noun Identification Algorithm? : NO - Tokenizer? : NO - Heuristic Associations to Add Terms? : NO - Expansion of Queries using Previously-Constructed Data Structure? : - Structure Used: NO - Automatic Addition of Boolean Connectors or Proximity Operators? : NO Manually Constructed Queries (Ad-Hoc) - Type of Query Builder - Tools used to Build Query - Knowledge Base Browser? : - Structure Used: NO - Other Lexical Tools? : - Method used in Query Construction - Addition of Terms not Included in Topic? : Interactive Queries - Initial Query Built Automatically or Manually: MANUALLY - Type of Person doing Interaction - Domain Expert: NO, SEARCHERS WITH VARYING DEGREES OF EXPERIENCE - Average Time to do Complete Interaction - Clock Time from Initial Construction of Query to Completion of Final Query (in minutes): 28.32 - Average Number of Iterations: 9.46 - Average Number of Documents Examined per Iteration: FULL TEXT, 2.8 - Minimum Number of Iterations: 2 - Maximum Number of Iterations: 28 - What Determines the End of an Iteration: INVOKING THE "RUN QUERY" BUTTON, OR THE "EXIT" BUTTON - Methods used in Interaction - Automatic Term Reweighting from Relevant Documents? : YES - Automatic Query Expansion from Relevant Documents? : YES - All Terms in Relevant Documents added: NO - Only Top X Terms Added (what is X): FIVE IF ONE DOCUMENT IS RELEVANT, FIVE PLUS TWO FOR EACH SUBSEQUENT DOCUMENT IF MORE THAN ONE RELEVANT DOCUMENT - User Selected Terms Added: YES - Other Automatic Methods: NONE - Manual Methods - Using Individual Judgment (No Set Algorithm)? : YES - Following a Given Algorithm (Brief Description)? : NO Searching Machine Searching Methods - Other: PROBABILISTIC INFERENCE NET Factors in Ranking - Term Frequency? : YES - Inverse Document Frequency? : YES - Semantic Closeness? : NO - Position in Document? : NO - Syntactic Clues? : NO - Proximity of Terms? : YES - Information Theoretic Weights? : NO - Document Length? : YES - Percentage of Query Terms which match? : SORT OF - N-gram Frequency? : NO - Word Specificity? : NO - Word Sense Frequency? : NO - Cluster Distance? : NO Machine Information - Machine Type for TREC Experiment: SUN SPARCSTATION 5 - Was the Machine Dedicated or Shared: DEDICATED - Amount of Hard Disk Storage (in MB): 10GB - Amount of RAM (in MB): 64MB - Clock Rate of CPU (in MHz): 110MHZ System Comparisons - Amount of "Software Engineering" which went into the Development of the System: 2 PERSON-M0NTHS, FOR THE INTERFACE AND RELATED ASPECTS, ON TOP OF THE BASIC INQUERY SYSTEM