WO2001035282A2 - System and method for creating associations between digital data objects - Google Patents
System and method for creating associations between digital data objects Download PDFInfo
- Publication number
- WO2001035282A2 WO2001035282A2 PCT/US2000/041622 US0041622W WO0135282A2 WO 2001035282 A2 WO2001035282 A2 WO 2001035282A2 US 0041622 W US0041622 W US 0041622W WO 0135282 A2 WO0135282 A2 WO 0135282A2
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- metadata
- user
- objects
- multimedia
- associations
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/40—Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
- G06F16/48—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/40—Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
Definitions
- This invention relates generally to a system and method for organizing and interacting with digital data, and more particularly, to an automated, interactive system for creating and presenting associations between digital data objects in multimedia and information collections.
- Multimedia includes color images and black and white still images, video, audio, animation, text, and 2- and 3- dimensional graphic objects.
- conventional methods of managing physical media are not readily transferable to digital media.
- the conventional methods of managing physical media, such as pictures by storing the images in a single location, such as a photo album or a desk drawer, makes organizing, finding and presenting the physical media difficult.
- the physical media are difficult to index and search manually, and conventional methods for presenting physical media restrict users to static presentations that are very difficult to change. For example, pictures in a photo album are affixed to the album in a fixed order and changing the presentation requires the presenter to physically transfer the pictures to another spot in the album.
- the illustrated embodiment of the present invention provides a way of associating groups of digital multimedia data that share similar features.
- Multimedia objects in a multimedia stream that are input into a computer system using the illustrated embodiment share a great many features, including simple temporal and geographical attributes (i.e., a sequence of images taken at one time and place), as well as "high- level” features such as people and objects.
- Computer vision and other technologies can be used to detect such features in multimedia objects, and the present invention provides the mechanism needed to form associations among the multimedia objects that share features.
- Knowledge representation technology is then utilized to allow linkages among multimedia objects.
- the links are assigned an importance level weighted according to domain knowledge about identified multimedia objects. The weighting of the links allows coherent and meaningful presentations of related multimedia objects to a user.
- a method is practiced whereby digital multimedia data is input into a computer system and parsed into individual multimedia objects.
- the multimedia data is screened for recognizable features and the features are stored as metadata attached to an agent created for each digital multimedia object.
- the features contained in the metadata attached to agents for different multimedia objects are compared for associations and any associated multimedia object is noted.
- Knowledge representation technology is used in the association process to weight the linkages according to domain knowledge about different types of objects.
- An interactive component presents to the user for verification those decisions regarding features and linkages that are incomplete comparison matches.
- Digital multimedia objects are presented to a user by the system and contain linkages to all of the other associated objects recorded in the attached metadata.
- a method is practiced whereby digital information data is input into a computer system and parsed into individual information objects.
- the input digital information include streams of genealogy research data, forensic investigation data, industrial site survey data, intelligence, surveillance and reconnaisance data, museum collections data and medical data, etc..
- the information data is screened for recognizable features and the features are stored as metadata attached to an agent created for each digital information object.
- the features contained in the metadata attached to agents for different information objects are compared for associations and any associated information object is noted.
- Knowledge representation technology is used in the association process to weight the linkages according to domain knowledge about different types of objects.
- An interactive component presents to the user for verification decisions regarding features and linkages that are incomplete comparison matches.
- Digital information objects are presented to a user by the system and contain linkages to all of the other associated objects recorded in the attached metadata.
- Figure 1 is a block diagram of major modules employed in the embodiment of the invention
- Figure 2 is a block diagram of the EM- Agent and EMCON module of Figure 1 and interacting modules;
- FIG. 3 is a block diagram of the association engine module of Figure 2 and its related components
- Figure 4 is a flowchart of the association engine process
- Figure 5 is a flowchart of the multimedia input process.
- the illustrated embodiment of the present invention provides a method for creating associations between multimedia objects and presenting the associations to a user.
- the multimedia objects may be color images, black and white still images, video, audio, animation, text, and/or two and three dimensional graphic objects.
- Figure 1 depicts the major modules utilized in the illustrated embodiment of the present invention, including multimedia input 2.
- the multimedia input 2 may be entered into the storage medium of a computer which is using the illustrated embodiment in a number of different ways. For example, frame grabs may be used for digitizing analog video and 2D scanners may be used for digitizing text and visual images. Similarly 3D scanners may be used for digitizing 3D objects and high bandwidth cables that connect directly to digital cameras may be used to input the pictures taken by the digital cameras.
- the EM-Agents and EMCONS module 4 create an agent with attached metadata for each multimedia object input into the computer system and stored in memory.
- the metadata attached to the agent for a multimedia object contains information about the content of the multimeda object.
- the display processors module 6 converts display specifications received from the EM-Agents and EMCONS module 4 into display instructions for the multimedia displays module 8.
- the multimedia displays module 8 may be any one of numerous types of well known display devices used to make digital multimedia presentations to a user 10, including CRT monitors, flat panel LCDs, or head mounted displays.
- the user 10 interacts with the illustrated embodiment of the present invention through a human-machine interface module 12.
- the human-machine interface module 12 provides a mechanism for interacting with the computer system such as a keyboard, joystick, mouse, voice command, etc..
- the human-machine interface module parses messages received from the user 10 into tokens.
- the dialog manager module 14 receives the parsed messages from the user 10 as tokens from the human-machine interface module 12.
- the dialog manager 14 also interacts with a number of other modules including the blackboard module 16, the learning dialog module 18, the display manager module 22, the EM-Agents and EMCONS module 4, and the ontology module 20.
- the blackboard module 16 is primarily a storage area in which agent queries are stored in a priority order to be periodically retrieved by the dialog manager 14 for presentation to the user 10.
- An agent query may be questions to the user regarding the identification of a feature contained in a multimedia object or a question regarding the importance of such a feature.
- the blackboard module 16 also maintains a listing of identified features contained within the multimedia object data, a pointer to each feature and to all of the agents that share this feature, as well as one or more tags which indicate information such as whether the metadata for this feature has changed or whether new associations for this feature have been found.
- the ontology module 20 is a software module that stores information about the domain of discourse in a generalization hierarchy.
- the generalization hierarchy contains weighted preferences for different types of features contained in the multimedia objects being input into the computer system. For example, if the user 10 is interested in creating associations between features depicting types of transportation, but is primarily interested in airplanes, less interested in trains and not interested in automobiles, the weighted preferences assigned to those features will reflect the user's level of interest.
- the generalization hierarchy includes kind-of and part-of relationships between nodes in the hierarchy, such as a chair is a kind-of furniture and a hair is a part-of a mammal.
- Such hierarchical information may include a dictionary of commonly encountered concepts such as kinship relationships, commonly observed holidays, common events such as birthday parties, common objects such as automobiles and household appliances, and so forth.
- the agents created by the EM-Agents and EMCONS module 4 consult with ontology 20 to determine the relative location in the generalization hierarchy of an EMCON feature to be identified. Based on this information, an EM-Agent calculates a priority for an agent query posted on the blackboard module 16 for later presentment to the user 10.
- the dialog manager 14 retrieves agent queries from the blackboard module 16 and presents them to the learning dialog module 18.
- the learning dialog module 18 formats the agent query into instructions for the display manager module 22 on how to present the agent queries to the user 10.
- the agent queries may be spoken, graphical, or text based, and they may specify that certain multimedia objects be presented to the user 10.
- Replies to agent queries from the user are forwarded as tokens from the human- machine interface module 12 to the dialog manager 14 and to the learning dialog module 18.
- the learning dialog module 18 formats the input information from the user 10 and sends the information to EM-Agents and EMCONS module 4 as updates for the metadata attached to a particular agent.
- Figure 2 depicts a block diagram of the submodules making up the EM-Agents and EMCONS module 4 and the other modules that interact with the EM-Agents and EMCONS module .
- Incoming multimedia input 2 is processed by a registry module 24 and the raw data from the input multimedia stream is stored as multimedia objects in the media objects module 26.
- the multimedia object is referred to as an EMCON.
- the media objects module 26 may be any one of a number of well known storage device types such as a hard drive.
- the registry module 24 creates an agent for each multimedia object, each agent including metadata describing the contents of the multimedia object.
- the agent and associated metadata for a multimedia object are stored in the agent community module 30.
- Agents are programmed to seek out associations for features contained in their attached metadata and utilize the association engine module 28 to contact the other agents in the agent communtity 30 to identify associations. Any association found will be stored in the metadata attached to the agents for both multimedia objects so that both agents will contain a link to the other agent's multimedia object. The located feature will also be listed on the blackboard 16.
- the search for associations is an ongoing process that may occur automatically without a users input.
- the search for associations occurs as a background process.
- the search for associations between EMCON features occurs during idle clock cycles.
- the search for associations may also take place in a distributed computer network on available hosts.
- the search for associations between EMCON features may occur as the result of a user input command.
- association engine 28 and its interacting components are depicted in
- the association engine 28 passes an EMCON to feature finder module 34 to be segmented into features to be stored in the metadata associated with the agent assigned to the EMCON. After segmenting the EMCON into features, the feature finder module 34 passes the features to the training features module 35 to search for commonly occurring objects. Commonly occurring objects include objects such as people, faces, domestic and farm animals, appliances, equipment, vehicles and landscape. If a possible feature is found by the feature finder 34 the feature is tagged with a generic identity and a weight specifying the probability that the generic identify is correct. If no generic identifications can be found, the EM-Agent for the multimedia object prepares agent queries, prioritizes them and posts them on the blackboard 16.
- the EM-Agent 32 tries to uniquely identify each such feature.
- the features are sent to the association engine 28 and then to the feature recognizer 36.
- the feature recognizer 36 returns a unique identity of the feature together with a weight specifying the probability that the unique feature identity is correct.
- the weight is compared to a predefined parameter, and if the unique feature identity weight is above the threshold, the unique feature identities are formatted as metadata by the feature recognizer module 36 and returned to the EM-Agent 32 via the association engine 28.
- Uniquely recognized features are stored in training features module 35. As the number of training features increases, the feature identification process becomes more efficient.
- the association engine 28 is also used by EM-Agents 32 to perform metadata searches 29 on other EM-Agents in the agent community 30 for the purpose of forming associations between recognized features in the metadata describing the content of different multimedia objects.
- Figure 4 depicts a flow chart of the association engine process.
- An EM-Agent 32 will check to see if it has received notification of any new associations between features (step 38 ). If the EM-Agent 32 has received new associations it will compute the weight to be given to those associations using matched feature priorities (step 40 ). If there are more associations (step 42 ) the EM-Agent 32 will continue to compute the weight using matched feature priorities. If there are not more associations or if there were no associations when the EM-Agent 32 initially checked, the EM-Agent will select in priority order one of the features contained in its metadata and search for associations between the selected feature and the features contained in the metadata attached to other agents ( step 44 ). The selected feature is sent to the association engine 28 ( step 46 ).
- the association engine 28 checks to see if the feature is currently listed on the blackboard module 16 (step 48 ). If information about the requested feature is not currently listed on the blackboard, a query regarding the feature is placed in priority order on the blackboard 16 (step 50 ). If the requested feature is already listed on the blackboard 14, the information associated with EM-Agent 32 is merged with the metadata from the other listed EM-Agents ( step 52 ). A metadata search is then performed ( step 29 ) and the search results checked for new associations ( step 54 ). If an association is found the association weight is computed ( step 56 ). If all of the metadata has been searched ( step 58 ) the blackboard feature listing is updated ( step 60 ), otherwise another metadata search 29 ( step 29 ) on the remaining metadata is conducted. After updating the blackboard listing ( step 60 ) in the event of a completed metadata search, the new association metadata is returned to the EM-Agent 32 ( step 62 ).
- a flowchart of the multimedia input process used by the illustrated embodiment of the present invention is depicted in Figure 5.
- a multimedia object from a multimedia input stream 2 is examined to see if it is a new object (step 64 ). If the object is not a new object a message to that effect is displayed to the user (step 66 ).
- the multimedia input stream 2 is examined to see if there are any additional objects (step 68 ) and if there are not, control is returned to the dialog manager 14 ( step 70 ) which checks its own status ( step 71 ). If the dialog manager 14 has no pending queries to retrieve from the blackboard 16, the dialog manager quits ( step 72 ).
- the registry module 24 creates a new EM- Agent for the multimedia object (step 74), allocates memory and stores the EMCON object in the media objects module 26 ( step 76 ), and registers the EM-Agent and
- EMCON with the registry 24 (step 78). If the EMCON is part of a hierarchy, such as a single frame grab out of a sequence of video, it is segmented into constituent EMCONs (step 80 ) and a new agent creation and registration process is carried out for each EMCON (step 82 ). Each individual EMCON is examined after registration for the presence of generic identifiable features (step 84 ). If no generic features are identifed, an agent query is prepared ( step 86 ). The agent query is forwarded to the dialog manager 14 ( step 70) for posting on the blackboard 16 for later presentment to a user 10.
- the weight of the identification is computed (step 88 ) and the EM-Agent' s metadata is updated if the weight of the identification exceeds a threshold parameter ( step 90 ).
- the EMCON is searched for unique features (step 92 ). If unique features are not found, an agent query is prepared ( step 86 ). If a unique feature is found, and the feature weight exceeds a threshold parameter (step 96 ), the feature is stored as a training feature ( step 98 ) in the training features module 35, and the metadata for the EM-Agent is updated ( step 100 ). Thereafter, control is returned to the dialog manager module 14 ( step 70 ).
Abstract
Description
Claims
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
AU27477/01A AU2747701A (en) | 1999-10-26 | 2000-10-26 | System and method for creating associations between digital data objects |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US16143199P | 1999-10-26 | 1999-10-26 | |
US60/161,431 | 1999-10-26 |
Publications (2)
Publication Number | Publication Date |
---|---|
WO2001035282A2 true WO2001035282A2 (en) | 2001-05-17 |
WO2001035282A3 WO2001035282A3 (en) | 2002-11-07 |
Family
ID=22581146
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2000/041622 WO2001035282A2 (en) | 1999-10-26 | 2000-10-26 | System and method for creating associations between digital data objects |
Country Status (2)
Country | Link |
---|---|
AU (1) | AU2747701A (en) |
WO (1) | WO2001035282A2 (en) |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5734893A (en) * | 1995-09-28 | 1998-03-31 | Ibm Corporation | Progressive content-based retrieval of image and video with adaptive and iterative refinement |
US5930783A (en) * | 1997-02-21 | 1999-07-27 | Nec Usa, Inc. | Semantic and cognition based image retrieval |
-
2000
- 2000-10-26 AU AU27477/01A patent/AU2747701A/en not_active Abandoned
- 2000-10-26 WO PCT/US2000/041622 patent/WO2001035282A2/en active Application Filing
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5734893A (en) * | 1995-09-28 | 1998-03-31 | Ibm Corporation | Progressive content-based retrieval of image and video with adaptive and iterative refinement |
US5930783A (en) * | 1997-02-21 | 1999-07-27 | Nec Usa, Inc. | Semantic and cognition based image retrieval |
Non-Patent Citations (1)
Title |
---|
AGOSTI M ET AL: "Design and implementation of a tool for the automatic construction of hypertexts for information retrieval" INFORMATION PROCESSING & MANAGEMENT, ELSEVIER, BARKING, GB, vol. 32, no. 4, 1 July 1996 (1996-07-01), pages 459-476, XP004007247 ISSN: 0306-4573 * |
Also Published As
Publication number | Publication date |
---|---|
AU2747701A (en) | 2001-06-06 |
WO2001035282A3 (en) | 2002-11-07 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10942982B2 (en) | Employing organizational context within a collaborative tagging system | |
US11861150B2 (en) | Methods and apparatus for managing and exchanging information using information objects | |
US9977827B2 (en) | System and methods of automatic query generation | |
US8180767B2 (en) | Inferred relationships from user tagged content | |
CA2677464C (en) | Providing unique views of data based on changes or rules | |
US8583592B2 (en) | System and methods of searching data sources | |
US9281963B2 (en) | Method and system for email search | |
US8122023B2 (en) | Data access using multilevel selectors and contextual assistance | |
US20020049689A1 (en) | Systems and methods for visual optimal ordered knowledge learning structures | |
US20080189336A1 (en) | Creating and managing digital media content using contacts and relational information | |
US20110307791A1 (en) | Social Networking Application for Knowledge Sharing and Management | |
US20090119572A1 (en) | Systems and methods for finding information resources | |
US20080244429A1 (en) | System and method of presenting search results | |
US20080243786A1 (en) | System and method of goal-oriented searching | |
US20080243787A1 (en) | System and method of presenting search results | |
US20130091162A1 (en) | Data Access Using Multilevel Selectors and Contextual Assistance | |
CN102597994A (en) | User-defined profile tags, rules, and recommendations for portal | |
US20090049078A1 (en) | Autofoldering process in content management | |
US9185147B1 (en) | System and methods for remote collaborative intelligence analysis | |
US20030131317A1 (en) | Method and system for organizing non-document specifications | |
KR101821832B1 (en) | Information management | |
US8103660B2 (en) | Computer method and system for contextual management and awareness of persistent queries and results | |
US9256672B2 (en) | Relevance content searching for knowledge bases | |
WO2001035282A2 (en) | System and method for creating associations between digital data objects | |
Zhu et al. | New query refinement and semantics integrated image retrieval system with semiautomatic annotation scheme |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A2 Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BY BZ CA CH CN CR CU CZ DE DK DM DZ EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT TZ UA UG UZ VN YU ZA ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A2 Designated state(s): GH GM KE LS MW MZ SD SL SZ TZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
REG | Reference to national code |
Ref country code: DE Ref legal event code: 8642 |
|
AK | Designated states |
Kind code of ref document: A3 Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BY BZ CA CH CN CR CU CZ DE DK DM DZ EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT TZ UA UG UZ VN YU ZA ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A3 Designated state(s): GH GM KE LS MW MZ SD SL SZ TZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG |
|
122 | Ep: pct application non-entry in european phase | ||
NENP | Non-entry into the national phase |
Ref country code: JP |