Search method melds results

By Chhavi Sachdev, Technology Research News

Search for the term ‘heartburn’ in a medical database and a typical search engine will spit out a long list of documents that contain the term along with a brief summary of each hit. Researchers at Columbia University have come up with an automated engine that promises to take the search a step further by comparing the returns.

The engine finds a set of documents, defines the query term in natural sentences using text pulled from those documents, and summarizes their hyperlinks, allowing the user to see both the information present in a set of documents and what is unique to each hit, according Min-Yen Kan, a computer science researcher at Columbia.

A summary from the heartburn search, for instance, would describe the causes, symptoms and prognosis of the malady, he said.

The query results also include details like relative length that differ among types of documents. A 10-page article about heart disease might be considered long, for instance, but a 100-page novel is rather short. The results also detail whether or not a document in the set has a diagnosis section, or which documents contain figures and tables.

In order to compare results, the system sets up a topic tree for every query term by comparing all the subheads in the documents. “Each document is viewed as a tree of topics and these trees are intersected to find commonalties and differences,” said Kan.

The definition, symptoms, and causes of most problems are nearly always signaled by distinct section headings within documents. Each of these branches could have five sub-branches.

“Once a search engine knows that the document is relevant to the search query... the system matches the query to a particular node of each document,” said Kan. The node might be the root node of the tree, as in the case of ‘heartburn treatments’, or somewhere in the middle, as in the case of ‘facts about heartburn,’ he said.

The system sets up a composite tree by comparing all the documents to get an idea of what an average document is like. Even if the heartburn documents don't have ‘prognosis’ as a topic, the system can infer the typical structure of patient information documents by looking at related documents that have the same structure, Kan said.

The software sorts references and subheads, focusing on those that occur more often in the query documents. When a section is not featured in another document and is too far down from the main branches, it might be discarded as irrelevant or too intricate, said Kan.

It generates natural language summaries of text based on the types of information that are really useful, Kan said. It extracts sentences or list items that occur across documents and checks morphological and subject-verb agreements; then it adds words to make clear sentences, he said.

The software also extracts descriptors such as the author’s name, media format, and categorization keywords if they are available as metadata in the documents.

The researchers can change the parameters of a search to trade speed for accuracy; the running time for the system to summarize five documents can vary from a few seconds to a few minutes, according to Kan.

The system is currently very accurate, he said, “but this is because we have been working only on developing the system using a closed set of example documents. As we expand the types of documents we will handle, more noise and error will be introduced, reducing the accuracy.”

The medical documents the Centrifuser project draws from are very similar in domain and genre. The software could easily be adapted to other sets of documents grouped by genre in achival repositories or intranets, Kan said.

Current document summarization techniques are not particularly satisfactory, said Justin Zobel, an assistant professor of computer science at RMIT University in Australia. “Developing a careful architecture for summarization, as in this work, is definitely a good way to proceed,” and including indicative summaries is also helpful, he said.

The Columbia engine is still in an early phase, however, Zobel added. “The system is not yet at the stage at which realistic user studies can be done.”

The researchers plan next to evaluate the current system and base future development according to users’ criticisms, Kan said. They also plan to make the topic-tree indexing software better. The system could be in use in 5 to 7 years.

Kan’s research colleagues were Kathleen R. McKeown and Judith L. Klavans. They presented the research at the 8th European Workshop on Natural Language Generation in Toulouse, France, July 2001. The research was funded by the Digital Libraries 2 Initiative of the National Science Foundation (NSF).

Timeline:  5-7 years
Funding:  Government
TRN Categories:  Natural Language Processing; Databases and Information Retrieval
Story Type:   News
Related Elements:  Technical paper, "Applying Natural Language Generation to Indicative Summarization," presented, held on 6 7 July, 2001. Posted on the Computing Research Repository on July 16, 2001 at http://xxx.lanl.gov/abs/cs.CL/0107019
http://www.cs.columbia.edu/~min/research/centrifuser/wnlg01/wnlg01.pdf




Advertisements:



January 9, 2002

Page One

Sunlight turns water to fuel

Search method melds results

VR tool keeps line of sight in hand

Laser pulse penetrates glass

Model tracks desert spread

News:

Research News Roundup
Research Watch blog

Features:
View from the High Ground Q&A
How It Works

RSS Feeds:
News  | Blog  | Books 



Ad links:
Buy an ad link

Advertisements:







Ad links: Clear History

Buy an ad link

 
Home     Archive     Resources    Feeds     Offline Publications     Glossary
TRN Finder     Research Dir.    Events Dir.      Researchers     Bookshelf
   Contribute      Under Development     T-shirts etc.     Classifieds
Forum    Comments    Feedback     About TRN


© Copyright Technology Research News, LLC 2000-2006. All rights reserved.