US6408293B1 - Interactive framework for understanding user's perception of multimedia data - Google Patents
Interactive framework for understanding user's perception of multimedia data Download PDFInfo
- Publication number
- US6408293B1 US6408293B1 US09/328,968 US32896899A US6408293B1 US 6408293 B1 US6408293 B1 US 6408293B1 US 32896899 A US32896899 A US 32896899A US 6408293 B1 US6408293 B1 US 6408293B1
- Authority
- US
- United States
- Prior art keywords
- user
- perception
- query
- query object
- database
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Lifetime
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/58—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/583—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
- G06F16/5838—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using colour
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10—TECHNICAL SUBJECTS COVERED BY FORMER USPC
- Y10S—TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10S707/00—Data processing: database and file management or data structures
- Y10S707/99931—Database or file accessing
- Y10S707/99933—Query processing, i.e. searching
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10—TECHNICAL SUBJECTS COVERED BY FORMER USPC
- Y10S—TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10S707/00—Data processing: database and file management or data structures
- Y10S707/99931—Database or file accessing
- Y10S707/99933—Query processing, i.e. searching
- Y10S707/99935—Query augmenting and refining, e.g. inexact access
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10—TECHNICAL SUBJECTS COVERED BY FORMER USPC
- Y10S—TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10S707/00—Data processing: database and file management or data structures
- Y10S707/99941—Database schema or data structure
- Y10S707/99943—Generating database or data structure, e.g. via user interface
Definitions
- the present invention generally relates to retrieval of multimedia data (images, video and audio) from a database and, more particularly, to a system which understands the user's perception from the query object(s) itself via user interaction, thereby increasing the relevance of the data retrieved from the database, and subsequently increasing the speed of retrieval of the objects of interest.
- CBIR content-based image retrieval
- Performance of such a computer-centric retrieval paradigm is not satisfactory (i.e., number of images irrelevant to the user is large), essentially due to the gap between high-level concepts (i.e., user's actual intention) and low-level visual features, and the inherent subjectivity of human perception.
- the retrieval process is interactive between the computer and the user. Based on the initial query image, the computer returns a set of similar images from the database. The user assigns relevance to the retrieved images (from highly relevant to irrelevant). The computer tries to correlate the user's perception of the image in terms of the low-level features, typically by employing some machine learning techniques. It then performs the retrieval process again. This interactive process is repeated until the user finds the image of interest.
- the query can consist of any of the following: an image, an image set, image(s) derived from a video sequence, a video sequence or an audio clip.
- the query is adjusted using the feedback given by the user about the relevance of previously extracted part(s) from the query object itself, such that the adjusted query is a better approximation to the user's perception. Since a single query object is utilized in the system according to the invention, high-performance learning techniques, which are often computationally intensive, can be employed for this intra-query object learning of user's perception.
- the refined query can be subsequently used using prior-art techniques for inter-query object relevance feedback where data is retrieved from the database based on parameters learned by intra-query object feedback mechanism, and the user provides feedback by ranking the retrieved data in order of their relevance to her/him.
- inter-query object learning of user's perception is expedited by utilizing the learned parameters in the intra-query object relevance feedback.
- the methodology of the invention allows for building refined queries based on part(s) of the query object rather than the entire object itself, thereby reducing irrelevant data being retrieved from the database.
- the present invention allows the user to synthesize/modify the query object(s) starting from a textual query or from a set of dictionary objects; e.g., using a “drag-and-drop” approach. The user's action during the synthesis and modification process is further used to learn his or her perception of the image.
- FIG. 1 is a block diagram illustrating apparatus on which the process according to the invention may be implemented
- FIG. 2 is a block diagram illustrating another apparatus on which the process according to the invention may be implemented
- FIG. 3 is a flow diagram illustrating the logic of the process for the complete multimedia retrieval system with user feedback and learning according to the invention
- FIG. 4 is a flow diagram illustrating in more detail the first learning loop, i.e., intra-query object learning, of the process shown in FIG. 3;
- FIGS. 5A and 5B are illustrations of an image showing segmentation of the image in the practice of the invention.
- FIG. 6 is a flow diagram illustrating query object(s) synthesis from a textual query through user interaction using a pre-processed (i.e., pre-compiled) object set (i.e., dictionary of objects); and
- FIG. 7 is a flow diagram illustrating an alternative embodiment of FIG. 6, where a Graphic User Interface (GUI) is used for query object synthesis, instead of a textual query.
- GUI Graphic User Interface
- FIG. 1 A user 10 interacts via a monitor 11 and an input device 12 , such as a keyboard and/or mouse, with a processing element 13 .
- the processing element may be part of a computer system such as a personal computer or workstation, for example, which is programmed to perform the process according to the invention.
- the computer system typically includes memory and mass storage, such as a hard disk, as well as other supporting peripherals.
- the processing element 13 accesses multimedia database 14 , typically stored on the mass storage of the computer system, in response to queries by the user 10 .
- the computer system illustrated in FIG. 1 is a stand alone computer system, but the invention may be implemented on a networked computer system, an example of which is illustrated in FIG. 2 .
- a client computer system 20 accesses a server 21 via a network 22 .
- the network 22 may be a Local Area Network (LAN) or Wide Area Network (WAN), for example.
- the network 22 may be the Internet.
- the server 21 accesses a multimedia database 23 in response to queries received from the client computer system 20 .
- the database 23 is typically stored on a mass storage system using magnetic or optical media.
- the user starts the search process with the help of query object(s) in block 4000 .
- the query object(s) may either be selected from a small database of query objects or may be supplied by the user. Alternatively, if the query object does not exist in the database and if the user does not have a representative query object, the user may then synthesize such a query using a pre-compiled dictionary as described later.
- the user's query is processed by an intra-query object processing block 1000 .
- This processing block includes an intra-query object learning function 301 which receives user feedback 302 , and is illustrated in more detail in FIG. 4 .
- the output of the intra-object processing block 1000 is used by the computer to perform database retrieval in function block 303 .
- the retrieved objects are displayed to the user who is prompted to indicate whether he or she is satisfied with the retrieval or not.
- the user's response to the prompt is evaluated in decision block 304 . If the user is satisfied, the present query stops and the user picks, i.e., downloads or saves, from the set of data returned by the computer in block 305 . Otherwise, the process enters the second learning loop, the inter-query object processing block 2000 .
- the inter-query object processing block 2000 includes an interquery object learning function 306 which receives user feedback 307 , just like the intra-query object processing block 1000 .
- This second learning loop i.e., inter-query object processing, is prior art and hence is not described in more detail.
- the intra-query object processing performed in processing block 1000 consists of (a) specification of part(s) of interest by the user (i.e., user's interest is only in part of the query image or the user's interest is only in the vocals of an audio clip), (b) tentative over-segmentation of the query object (or part of it) using low-level features like color, texture, shape, audio frequency, motion, etc. (It could also include for example separation of audio signal into high and low pitch components.
- Inter-query object processing comprises of (a) searching the database for similar query object(s) using learned parameters from intra-query object processing, (b) user's ranking of the retrieved media as relevant and irrelevant images, and (c) understanding user's perception to improve the retrieval efficiency for the next iteration of database retrieval.
- the intra-query object processing of processing block 1000 is shown in more detail in FIG. 4 .
- the first step in the process is media object segmentation performed in function block 401 .
- Multimedia segmentation could imply image segmentation using low-level features like color, texture, etc. for images or audio segmentation into signals of various frequencies based on the pitch, or determination of video segmentation using motion vectors from the frames of a video clip.
- the segmentation results are displayed to the user in function block 402 with a prompt to indicate whether or not the user is satisfied with the segmentation results, at least in the region of interest.
- the user's input is evaluated in decision block 403 , and if this input indicates that the user is not satisfied with the segmentation, then the user is prompted for further input, such as merge, de-merge, validate, re-order, audio mix, etc.
- This input is indicated by input block 404 and is used to learn the user's perception in function block 405 .
- a re-segmentation of the media object is performed in function block 406 .
- the block 406 could also modify the media object or a part of the object to better learn the user's perception, e.g., re-coloring an image, modifying the pitch to assess the user's sensitivity to such changes.
- results of re-segmentation or modification are again displayed to the user at function block 402 and, again, the user is prompted to indicate his or her satisfaction.
- User's response is used to learn the perception; e.g., if color was changed and the user did not respond negatively, then the user cares more about texture and not color.
- the query object is defined by a set of segments, each segment for an image being a homogeneous region of the image, in terms of gray-level value, RGB value, texture and/or a combination of these.
- a segment is characterized by a feature vector x, a shape description d, and a similarity measure for an image or video clip.
- the query object is characterized by a set of segments, their topological and chronological relationship G. The segmentation results are shown to the user. If the user is interested only in the partial query image, he or she selects areas of interest by enclosing them in a bounded polygon, for example.
- a subset of segments surrounding the user's areas of interest are shown to the user.
- segmentation is performed only in the area(s) of interest.
- the user is now allowed to merge two or more segments, e.g., by clicking on the edge between them or by clicking points inside the segments.
- This enables the computer to capture the users notion of similarity, topological and chronological relationship between different parts, and the user's area of interest in the query object.
- the user is also given an option of splitting a segment into two or more based on his perception, for example, by clicking on representative points in the segment.
- the computer now learns the model parameters and the similarity measures which would lead to merging or splitting of the selected segments.
- the re-grouped object is again shown to the user. This interactive process is repeated until the user is satisfied.
- the present invention allows the user to create a composite object consisting essentially of two (or more) dissimilar segments, e.g., a flower and its stem.
- the two segments which are dissimilar are retained as separate segments and are added to query object set.
- the topological arrangements and features of individual segments along with learned similarity measures are used in subsequent database search.
- the similarity measures are adaptive quantities, which are adjusted with user's feedback. The measure could be different for different segments. Multimedia objects which contain the similar segments or a subset of segments and satisfy certain topological arrangement (if specified) are retrieved during inter-query object processing.
- the segmentation results might consist of an image frame along with motion vectors, e.g., a clip of a moving car might be shown as a single car with translational vector and rotational vectors for the tires.
- the user may then select the car and the corresponding motion vector to say that he or she is interested in a moving car.
- the computer might separate the vocals from the instrumentals based on pitch or percussion from a stringed instrument. The user may then select the desired sound effects.
- the image is segmented into regions a, b, c, d, e and f.
- Segments a, b, c and d differ in color but are similar in texture.
- Segment a differs from segments e and f in texture.
- the computer learns that user cares more about texture than color.
- the computer then groups the regions a, b, c and d together to make the query object. If now the user requests the search, the shape and the corresponding feature vector of only this segment abcd shown in FIG. 5B is used for searching the database.
- the system according to the invention enables the user to create an approximate query object from a textual query using a pre-compiled dictionary of objects.
- the drag and drop approach is a good candidate for query object synthesis.
- Textual phrases can also be used for specifying topological relationship of dictionary objects.
- the dictionary is large, the query can be formulated textually in terms of the dictionary objects and their topological relationships.
- the dictionary is augmented in run-time by letting the user add an object to the dictionary, every time he or she creates one in intra-object processing phase.
- the present invention is different from prior art techniques since the object in the present invention consists of the feature vector describing the region and a similarity measure f which will be used to measure the similarity of this segment to others.
- two segments with more or less the same feature vector could have different similarity measures, e.g., a homogeneous region made up of bricks in the shape of a cottage could either have a measure that says the shape matching is to be used to find cottages made up of either bricks or concrete or wood, or instead the measure could denote that the brick texture is to be used to find any brick region whether on the floor or wall, i.e., any shape.
- This is a different embodiment of this invention with the same objective of understanding the user's perception of the multimedia query object, achieved with synthesis and/or modification through user interaction.
- FIG. 6 Query-object synthesis using a pre-processed (i.e., pre-compiled) object set (i.e., dictionary) of objects as described above is shown in FIG. 6, which describes in detail the block 4000 of FIG. 3 .
- the process could begin with a textual query, as shown in input block 601 , that is first parsed in function block 602 and then categories are accessed from an “annotated” pre-processed object set, or dictionary, in function block 603 .
- the object is then synthesized with user feedback for adding and/or deleting components or specifying the topological relationships between objects. This is performed in processing block 3000 , the output of which is input to processing block 1000 in FIG. 3 .
- the object is synthesized in function block 604 , and the synthesized query object is displayed in block 605 with a prompt for the user to input his or her satisfaction.
- the user input is evaluated in decision block 606 , and if the user is not satisfied, the user is prompted to provide additional user input. This input is indicated at input block 607 and, once the input is entered, the process loops back to function block 603 .
- An alternative to this “object synthesis” approach could be just using a pre-processed object set (dictionary of objects); i.e., there is no annotation in the object set and, hence, the user may not browse the categories.
- FIG. 7 where the query is input at input block 701 , and in response to the query, the pre-processed object set is accessed in function block 702 . At this point the process enters processing block 3000 , as in FIG. 6 . This is a simpler version as compared to that illustrated in FIG. 6 . Consequently, the user may just drag-and-drop objects from the object set, but there is no textual query.
- the present invention is different and more user-friendly from the prior art reported by T. P. Minka and R. W. Picard, “Interactive learning using a society of models”, Proc. IEEE CVPR , 1996, where the user indicates a region of interest in the image itself by marking a number of pixels.
- This system generalizes by selecting and, if necessary, combining a number of pre-compiled grouping of highly specialized and context-dependent features from a “society of models” using a restricted set of rules.
Landscapes
- Engineering & Computer Science (AREA)
- Library & Information Science (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Image Analysis (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Processing Or Creating Images (AREA)
Abstract
Description
Claims (9)
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US09/328,968 US6408293B1 (en) | 1999-06-09 | 1999-06-09 | Interactive framework for understanding user's perception of multimedia data |
CA002307264A CA2307264C (en) | 1999-06-09 | 2000-05-01 | An interactive framework for understanding user's perception of multimedia data |
SG200002914A SG91283A1 (en) | 1999-06-09 | 2000-05-30 | An interactive framework for understanding user's perception of multimedia data |
JP2000169919A JP2001043246A (en) | 1999-06-09 | 2000-06-07 | Interactive framework for comprehending recognition of user to multimedia data |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US09/328,968 US6408293B1 (en) | 1999-06-09 | 1999-06-09 | Interactive framework for understanding user's perception of multimedia data |
Publications (1)
Publication Number | Publication Date |
---|---|
US6408293B1 true US6408293B1 (en) | 2002-06-18 |
Family
ID=23283263
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US09/328,968 Expired - Lifetime US6408293B1 (en) | 1999-06-09 | 1999-06-09 | Interactive framework for understanding user's perception of multimedia data |
Country Status (4)
Country | Link |
---|---|
US (1) | US6408293B1 (en) |
JP (1) | JP2001043246A (en) |
CA (1) | CA2307264C (en) |
SG (1) | SG91283A1 (en) |
Cited By (66)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020136468A1 (en) * | 2001-03-20 | 2002-09-26 | Hung-Ming Sun | Method for interactive image retrieval based on user-specified regions |
US20020154695A1 (en) * | 2001-04-20 | 2002-10-24 | Cornog Katherine H. | Correcting motion vector maps for image processing |
US20030065661A1 (en) * | 2001-04-02 | 2003-04-03 | Chang Edward Y. | Maximizing expected generalization for learning complex query concepts |
WO2004019752A2 (en) * | 2002-08-30 | 2004-03-11 | Rheinische Friedrich-Wilhelms-Uni Versität Bonn | Method and device for the generation of intersensory perception-associations |
US20040073543A1 (en) * | 2002-10-14 | 2004-04-15 | Samsung Electronics Co., Ltd. | Image retrieval method and apparatus using iterative matching |
US20040177138A1 (en) * | 2003-03-05 | 2004-09-09 | Mathias Salle | Method and system for processing user feedback received from a user of a website |
US20040176990A1 (en) * | 2003-03-05 | 2004-09-09 | Vacante Robert C. | Method and system enabling the trading of a futures contract for the purchase of goods or services |
US20040176992A1 (en) * | 2003-03-05 | 2004-09-09 | Cipriano Santos | Method and system for evaluating performance of a website using a customer segment agent to interact with the website according to a behavior model |
US20040176991A1 (en) * | 2003-03-05 | 2004-09-09 | Mckennan Carol | System, method and apparatus using biometrics to communicate dissatisfaction via stress level |
US20040176998A1 (en) * | 2003-03-05 | 2004-09-09 | Evan Kirshenbaum | Method and system for selling an item over a computer network |
US20040177092A1 (en) * | 2003-03-05 | 2004-09-09 | Mathias Salle | Apparatus and method for theorem certification without disclosing documents that lead to the theorem |
US6795818B1 (en) * | 1999-07-05 | 2004-09-21 | Lg Electronics Inc. | Method of searching multimedia data |
US20040199584A1 (en) * | 2003-03-05 | 2004-10-07 | Evan Kirshenbaum | Method and system for customized configuration of an appearance of a website for a user |
US20040267740A1 (en) * | 2000-10-30 | 2004-12-30 | Microsoft Corporation | Image retrieval systems and methods with semantic and feature based relevance feedback |
US20050010553A1 (en) * | 2000-10-30 | 2005-01-13 | Microsoft Corporation | Semi-automatic annotation of multimedia objects |
US6859802B1 (en) * | 1999-09-13 | 2005-02-22 | Microsoft Corporation | Image retrieval based on relevance feedback |
US6915292B2 (en) * | 1999-12-02 | 2005-07-05 | Lg Electronics Inc. | Method for updating multimedia feature information |
US20050162515A1 (en) * | 2000-10-24 | 2005-07-28 | Objectvideo, Inc. | Video surveillance system |
US20050270283A1 (en) * | 2003-07-16 | 2005-12-08 | Plut William J | Methods for spatial-based power savings |
US20050270265A1 (en) * | 2003-07-16 | 2005-12-08 | Plut William J | LCD plateau power conservation |
US20060001660A1 (en) * | 2003-07-16 | 2006-01-05 | Plut William J | Color preservation for spatially varying power conservation |
US20060001659A1 (en) * | 2003-07-16 | 2006-01-05 | Plut William J | Window information preservation for spatially varying power conservation |
US20060001658A1 (en) * | 2003-07-16 | 2006-01-05 | Plut William J | Edge preservation for spatially varying power conservation |
US20060020906A1 (en) * | 2003-07-16 | 2006-01-26 | Plut William J | Graphics preservation for spatially varying display device power conversation |
US20060236893A1 (en) * | 2005-04-22 | 2006-10-26 | Xerox Corporation | Photoreceptors |
US20060250525A1 (en) * | 2005-05-04 | 2006-11-09 | Plut William J | White-based power savings |
US20070002035A1 (en) * | 2003-07-16 | 2007-01-04 | Plut William J | Background plateau manipulation for display device power conservation |
US20070219983A1 (en) * | 2006-03-14 | 2007-09-20 | Fish Robert D | Methods and apparatus for facilitating context searching |
US20080148179A1 (en) * | 2006-12-18 | 2008-06-19 | Microsoft Corporation | Displaying relatedness of media items |
US20090106233A1 (en) * | 2007-08-24 | 2009-04-23 | Veenstra John W | Query engine interpreter and prioritization engine |
US20090254539A1 (en) * | 2008-04-03 | 2009-10-08 | Microsoft Corporation | User Intention Modeling For Interactive Image Retrieval |
US20100026736A1 (en) * | 2005-05-04 | 2010-02-04 | Plut William J | Luminance suppression power conservation |
US20110035406A1 (en) * | 2009-08-07 | 2011-02-10 | David Petrou | User Interface for Presenting Search Results for Multiple Regions of a Visual Query |
US20110038512A1 (en) * | 2009-08-07 | 2011-02-17 | David Petrou | Facial Recognition with Social Network Aiding |
US20110125735A1 (en) * | 2009-08-07 | 2011-05-26 | David Petrou | Architecture for responding to a visual query |
US20110131235A1 (en) * | 2009-12-02 | 2011-06-02 | David Petrou | Actionable Search Results for Street View Visual Queries |
US20110129153A1 (en) * | 2009-12-02 | 2011-06-02 | David Petrou | Identifying Matching Canonical Documents in Response to a Visual Query |
US20110128288A1 (en) * | 2009-12-02 | 2011-06-02 | David Petrou | Region of Interest Selector for Visual Queries |
US20110131241A1 (en) * | 2009-12-02 | 2011-06-02 | David Petrou | Actionable Search Results for Visual Queries |
US20110179385A1 (en) * | 2008-09-24 | 2011-07-21 | Wencheng Li | Content classification utilizing a reduced description palette to simplify content analysis |
US20130101209A1 (en) * | 2010-10-29 | 2013-04-25 | Peking University | Method and system for extraction and association of object of interest in video |
TWI396105B (en) * | 2009-07-21 | 2013-05-11 | Univ Nat Taiwan | Digital data processing method for personalized information retrieval and computer readable storage medium and information retrieval system thereof |
US8463045B2 (en) | 2010-11-10 | 2013-06-11 | Microsoft Corporation | Hierarchical sparse representation for image retrieval |
US8564661B2 (en) | 2000-10-24 | 2013-10-22 | Objectvideo, Inc. | Video analytic rule detection system and method |
EP2685420A1 (en) * | 2011-03-11 | 2014-01-15 | Omron Corporation | Image processing device, image processing method and control program |
WO2014009751A1 (en) * | 2012-07-13 | 2014-01-16 | Deepmind Technologies Limited | Method and apparatus for image searching |
US8711217B2 (en) | 2000-10-24 | 2014-04-29 | Objectvideo, Inc. | Video surveillance system employing video primitives |
US8805079B2 (en) | 2009-12-02 | 2014-08-12 | Google Inc. | Identifying matching canonical documents in response to a visual query and in accordance with geographic information |
US8811742B2 (en) | 2009-12-02 | 2014-08-19 | Google Inc. | Identifying matching canonical documents consistent with visual query structural information |
US8935246B2 (en) | 2012-08-08 | 2015-01-13 | Google Inc. | Identifying textual terms in response to a visual query |
EP2894574A1 (en) * | 2014-01-08 | 2015-07-15 | Lg Electronics Inc. | Mobile terminal and controlling method thereof |
US9176986B2 (en) | 2009-12-02 | 2015-11-03 | Google Inc. | Generating a combination of a visual query and matching canonical document |
US9317533B2 (en) | 2010-11-02 | 2016-04-19 | Microsoft Technology Licensing, Inc. | Adaptive image retrieval database |
WO2016109056A1 (en) * | 2014-12-31 | 2016-07-07 | Ebay, Inc. | Dynamic product placement based on perceived value |
US9720937B2 (en) | 2008-12-22 | 2017-08-01 | Koninklijke Philips N.V. | Relevance feedback on a segment of a data object |
US9852156B2 (en) | 2009-12-03 | 2017-12-26 | Google Inc. | Hybrid use of location sensor data and visual query to return local listings for visual query |
US9881229B2 (en) * | 2005-09-30 | 2018-01-30 | Facebook, Inc. | Apparatus, method and program for image search |
US9892606B2 (en) | 2001-11-15 | 2018-02-13 | Avigilon Fortress Corporation | Video surveillance system employing video primitives |
US10482340B2 (en) * | 2016-12-06 | 2019-11-19 | Samsung Electronics Co., Ltd. | System and method for object recognition and ranging by deformation of projected shapes in a multimodal vision and sensing system for autonomous devices |
US10572520B2 (en) | 2012-07-31 | 2020-02-25 | Veveo, Inc. | Disambiguating user intent in conversational interaction system for large corpus information retrieval |
US10592575B2 (en) | 2012-07-20 | 2020-03-17 | Veveo, Inc. | Method of and system for inferring user intent in search input in a conversational interaction system |
US10728351B2 (en) | 2015-01-30 | 2020-07-28 | Rovi Guides, Inc. | Systems and methods for resolving ambiguous terms in social chatter based on a user profile |
US10896184B2 (en) | 2013-05-10 | 2021-01-19 | Veveo, Inc. | Method and system for capturing and exploiting user intent in a conversational interaction based information retrieval system |
US10978094B2 (en) | 2013-05-07 | 2021-04-13 | Veveo, Inc. | Method of and system for real time feedback in an incremental speech input interface |
CN114040227A (en) * | 2021-10-22 | 2022-02-11 | 游密科技(深圳)有限公司 | Audio and video playing control method and device, electronic equipment and storage medium |
US11423074B2 (en) | 2014-12-23 | 2022-08-23 | Rovi Guides, Inc. | Systems and methods for determining whether a negation statement applies to a current or past query |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR100902010B1 (en) | 2007-06-18 | 2009-06-12 | 서강대학교산학협력단 | Content-based multimedia retrieval method with associative feedback |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5802361A (en) * | 1994-09-30 | 1998-09-01 | Apple Computer, Inc. | Method and system for searching graphic images and videos |
US5893095A (en) * | 1996-03-29 | 1999-04-06 | Virage, Inc. | Similarity engine for content-based retrieval of images |
US5899999A (en) * | 1996-10-16 | 1999-05-04 | Microsoft Corporation | Iterative convolution filter particularly suited for use in an image classification and retrieval system |
US5913205A (en) * | 1996-03-29 | 1999-06-15 | Virage, Inc. | Query optimization for visual information retrieval system |
US6181818B1 (en) * | 1994-11-15 | 2001-01-30 | Canon Kabushiki Kaisha | Image retrieval method and apparatus |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0788450B1 (en) * | 1994-11-03 | 2001-10-04 | Amcor Packaging (New Zealand) Limited | Filling means for containers |
-
1999
- 1999-06-09 US US09/328,968 patent/US6408293B1/en not_active Expired - Lifetime
-
2000
- 2000-05-01 CA CA002307264A patent/CA2307264C/en not_active Expired - Fee Related
- 2000-05-30 SG SG200002914A patent/SG91283A1/en unknown
- 2000-06-07 JP JP2000169919A patent/JP2001043246A/en active Pending
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5802361A (en) * | 1994-09-30 | 1998-09-01 | Apple Computer, Inc. | Method and system for searching graphic images and videos |
US6181818B1 (en) * | 1994-11-15 | 2001-01-30 | Canon Kabushiki Kaisha | Image retrieval method and apparatus |
US6246804B1 (en) * | 1994-11-15 | 2001-06-12 | Canon Kabushiki Kaisha | Image retrieval method and apparatus using a compound image formed from a plurality of detected regions |
US5893095A (en) * | 1996-03-29 | 1999-04-06 | Virage, Inc. | Similarity engine for content-based retrieval of images |
US5913205A (en) * | 1996-03-29 | 1999-06-15 | Virage, Inc. | Query optimization for visual information retrieval system |
US5899999A (en) * | 1996-10-16 | 1999-05-04 | Microsoft Corporation | Iterative convolution filter particularly suited for use in an image classification and retrieval system |
Non-Patent Citations (7)
Title |
---|
C. Carson et al., "Region-Based Image Querying", CVPR '97 Workshop on Content-Based Access of Image and Video Libraries, Puerto Rico, Jun. 1997, pp. 1-8. |
C. Nastar et al., "Efficient Query Refinement for Image Retrieval", Proc. IEEE CVPR, 1998. |
J.R. Smith and S.F. Chung, "Visual Seek: A fully automated content-based image query system", Proc. ACM Multimedia 96, 1996. |
T.P. Minka and R.W. Picard, "Interactive Learning with a "Society of Models'", Proc. IEEE CVPR, 1996, pp. 447-452. |
T.P. Minka and R.W. Picard, "Interactive Learning with a ‘Society of Models’", Proc. IEEE CVPR, 1996, pp. 447-452. |
W.Y. Ma and B.S. Manjunath, "NeTra: A Toolbox for Navigating Large Image Databases", Proc. IEEE Int. Conf. On Image Processing, 1997, pp. 568-571. |
Y. Rui et al., "Relevance Feedback: A Power Tool for Interactive Content-Based Image Retrieval", IEEE Trans. Circuits and Systems for Video Technology, Special Issue on Interactive Multimedia Systems for the Internet, 1998, pp. 1-13. |
Cited By (143)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6795818B1 (en) * | 1999-07-05 | 2004-09-21 | Lg Electronics Inc. | Method of searching multimedia data |
US7493340B2 (en) | 1999-09-13 | 2009-02-17 | Microsoft Corporation | Image retrieval based on relevance feedback |
US7613686B2 (en) | 1999-09-13 | 2009-11-03 | Microsoft Corporation | Image retrieval based on relevance feedback |
US20050065929A1 (en) * | 1999-09-13 | 2005-03-24 | Microsoft Corporation | Image retrieval based on relevance feedback |
US6859802B1 (en) * | 1999-09-13 | 2005-02-22 | Microsoft Corporation | Image retrieval based on relevance feedback |
US6915292B2 (en) * | 1999-12-02 | 2005-07-05 | Lg Electronics Inc. | Method for updating multimedia feature information |
US8711217B2 (en) | 2000-10-24 | 2014-04-29 | Objectvideo, Inc. | Video surveillance system employing video primitives |
US9378632B2 (en) | 2000-10-24 | 2016-06-28 | Avigilon Fortress Corporation | Video surveillance system employing video primitives |
US20050162515A1 (en) * | 2000-10-24 | 2005-07-28 | Objectvideo, Inc. | Video surveillance system |
US8564661B2 (en) | 2000-10-24 | 2013-10-22 | Objectvideo, Inc. | Video analytic rule detection system and method |
US10026285B2 (en) | 2000-10-24 | 2018-07-17 | Avigilon Fortress Corporation | Video surveillance system employing video primitives |
US10645350B2 (en) | 2000-10-24 | 2020-05-05 | Avigilon Fortress Corporation | Video analytic rule detection system and method |
US10347101B2 (en) | 2000-10-24 | 2019-07-09 | Avigilon Fortress Corporation | Video surveillance system employing video primitives |
US7627556B2 (en) | 2000-10-30 | 2009-12-01 | Microsoft Corporation | Semi-automatic annotation of multimedia objects |
US7349895B2 (en) | 2000-10-30 | 2008-03-25 | Microsoft Corporation | Semi-automatic annotation of multimedia objects |
US20040267740A1 (en) * | 2000-10-30 | 2004-12-30 | Microsoft Corporation | Image retrieval systems and methods with semantic and feature based relevance feedback |
US20050010553A1 (en) * | 2000-10-30 | 2005-01-13 | Microsoft Corporation | Semi-automatic annotation of multimedia objects |
US7499916B2 (en) | 2000-10-30 | 2009-03-03 | Microsoft Corporation | Image retrieval systems and methods with semantic and feature based relevance feedback |
US20050055344A1 (en) * | 2000-10-30 | 2005-03-10 | Microsoft Corporation | Image retrieval systems and methods with semantic and feature based relevance feedback |
US7529732B2 (en) | 2000-10-30 | 2009-05-05 | Microsoft Corporation | Image retrieval systems and methods with semantic and feature based relevance feedback |
US20050114325A1 (en) * | 2000-10-30 | 2005-05-26 | Microsoft Corporation | Semi-automatic annotation of multimedia objects |
US6970860B1 (en) * | 2000-10-30 | 2005-11-29 | Microsoft Corporation | Semi-automatic annotation of multimedia objects |
US20020136468A1 (en) * | 2001-03-20 | 2002-09-26 | Hung-Ming Sun | Method for interactive image retrieval based on user-specified regions |
US7158970B2 (en) | 2001-04-02 | 2007-01-02 | Vima Technologies, Inc. | Maximizing expected generalization for learning complex query concepts |
US6976016B2 (en) * | 2001-04-02 | 2005-12-13 | Vima Technologies, Inc. | Maximizing expected generalization for learning complex query concepts |
US20030065661A1 (en) * | 2001-04-02 | 2003-04-03 | Chang Edward Y. | Maximizing expected generalization for learning complex query concepts |
US7043058B2 (en) * | 2001-04-20 | 2006-05-09 | Avid Technology, Inc. | Correcting motion vector maps for image processing |
US20020154695A1 (en) * | 2001-04-20 | 2002-10-24 | Cornog Katherine H. | Correcting motion vector maps for image processing |
US9892606B2 (en) | 2001-11-15 | 2018-02-13 | Avigilon Fortress Corporation | Video surveillance system employing video primitives |
WO2004019752A2 (en) * | 2002-08-30 | 2004-03-11 | Rheinische Friedrich-Wilhelms-Uni Versität Bonn | Method and device for the generation of intersensory perception-associations |
WO2004019752A3 (en) * | 2002-08-30 | 2004-07-15 | Univ Bonn | Method and device for the generation of intersensory perception-associations |
EP1398012A1 (en) * | 2002-08-30 | 2004-03-17 | Friedrich-Wilhelms-Universität Bonn Rheinische | Methods and apparatus for generating intersensory perceptions of associations |
US20040073543A1 (en) * | 2002-10-14 | 2004-04-15 | Samsung Electronics Co., Ltd. | Image retrieval method and apparatus using iterative matching |
US7298931B2 (en) * | 2002-10-14 | 2007-11-20 | Samsung Electronics Co., Ltd. | Image retrieval method and apparatus using iterative matching |
US20040176990A1 (en) * | 2003-03-05 | 2004-09-09 | Vacante Robert C. | Method and system enabling the trading of a futures contract for the purchase of goods or services |
US20040199584A1 (en) * | 2003-03-05 | 2004-10-07 | Evan Kirshenbaum | Method and system for customized configuration of an appearance of a website for a user |
US8069075B2 (en) | 2003-03-05 | 2011-11-29 | Hewlett-Packard Development Company, L.P. | Method and system for evaluating performance of a website using a customer segment agent to interact with the website according to a behavior model |
US20040177138A1 (en) * | 2003-03-05 | 2004-09-09 | Mathias Salle | Method and system for processing user feedback received from a user of a website |
US20040177092A1 (en) * | 2003-03-05 | 2004-09-09 | Mathias Salle | Apparatus and method for theorem certification without disclosing documents that lead to the theorem |
US20040176998A1 (en) * | 2003-03-05 | 2004-09-09 | Evan Kirshenbaum | Method and system for selling an item over a computer network |
US20040176991A1 (en) * | 2003-03-05 | 2004-09-09 | Mckennan Carol | System, method and apparatus using biometrics to communicate dissatisfaction via stress level |
US20040176992A1 (en) * | 2003-03-05 | 2004-09-09 | Cipriano Santos | Method and system for evaluating performance of a website using a customer segment agent to interact with the website according to a behavior model |
US7583260B2 (en) | 2003-07-16 | 2009-09-01 | Honeywood Technologies, Llc | Color preservation for spatially varying power conservation |
US20100079595A1 (en) * | 2003-07-16 | 2010-04-01 | Plut William J | Televisions with reduced power consumption |
US20070002035A1 (en) * | 2003-07-16 | 2007-01-04 | Plut William J | Background plateau manipulation for display device power conservation |
US7580031B2 (en) | 2003-07-16 | 2009-08-25 | Honeywood Technologies, Llc | Histogram and spatial-based power savings |
US20050270283A1 (en) * | 2003-07-16 | 2005-12-08 | Plut William J | Methods for spatial-based power savings |
US8203551B2 (en) | 2003-07-16 | 2012-06-19 | Samsung Electronics Co., Ltd | Televisions with reduced power consumption |
US7602388B2 (en) | 2003-07-16 | 2009-10-13 | Honeywood Technologies, Llc | Edge preservation for spatially varying power conservation |
US20060020906A1 (en) * | 2003-07-16 | 2006-01-26 | Plut William J | Graphics preservation for spatially varying display device power conversation |
US20060001658A1 (en) * | 2003-07-16 | 2006-01-05 | Plut William J | Edge preservation for spatially varying power conservation |
US7629971B2 (en) | 2003-07-16 | 2009-12-08 | Honeywood Technologies, Llc | Methods for spatial-based power savings |
US9715846B2 (en) | 2003-07-16 | 2017-07-25 | Samsung Electronics Co., Ltd. | Background plateau manipulation for display device power conservation |
US20060001659A1 (en) * | 2003-07-16 | 2006-01-05 | Plut William J | Window information preservation for spatially varying power conservation |
US7663597B2 (en) | 2003-07-16 | 2010-02-16 | Honeywood Technologies, Llc | LCD plateau power conservation |
US9953553B2 (en) | 2003-07-16 | 2018-04-24 | Samsung Electronics Co., Ltd. | Background plateau manipulation for display device power conservation |
US7714831B2 (en) | 2003-07-16 | 2010-05-11 | Honeywood Technologies, Llc | Background plateau manipulation for display device power conservation |
US20100141635A1 (en) * | 2003-07-16 | 2010-06-10 | Plut William J | Lcd plateau power conservation |
US20100149197A1 (en) * | 2003-07-16 | 2010-06-17 | Plut William J | Spatial based power savings for lcd televisions |
US20060001660A1 (en) * | 2003-07-16 | 2006-01-05 | Plut William J | Color preservation for spatially varying power conservation |
US7786988B2 (en) | 2003-07-16 | 2010-08-31 | Honeywood Technologies, Llc | Window information preservation for spatially varying power conservation |
US20100220090A1 (en) * | 2003-07-16 | 2010-09-02 | Plut William J | Background plateau manipulation for display device power conservation |
US9135884B2 (en) | 2003-07-16 | 2015-09-15 | Samsung Electronics Co., Ltd. | LCD plateau power conservation |
US8912999B2 (en) | 2003-07-16 | 2014-12-16 | Samsung Electronics Co., Ltd. | Background plateau manipulation for display device power conservation |
US20050275651A1 (en) * | 2003-07-16 | 2005-12-15 | Plut William J | Histogram and spatial-based power savings |
US20050270265A1 (en) * | 2003-07-16 | 2005-12-08 | Plut William J | LCD plateau power conservation |
US8207934B2 (en) | 2003-07-16 | 2012-06-26 | Samsung Electronics Co., Ltd | Spatial based power savings for LCD televisions |
US20060236893A1 (en) * | 2005-04-22 | 2006-10-26 | Xerox Corporation | Photoreceptors |
US9659544B2 (en) | 2005-05-04 | 2017-05-23 | Samsung Electronics Co., Ltd. | Luminance suppression power conservation |
US10140945B2 (en) | 2005-05-04 | 2018-11-27 | Samsung Electronics Co., Ltd. | Luminance suppression power conservation |
US20100026735A1 (en) * | 2005-05-04 | 2010-02-04 | Plut William J | Luminance suppression power conservation |
US11145270B2 (en) | 2005-05-04 | 2021-10-12 | Samsung Electronics Co., Ltd. | Luminance suppression power conservation |
US20100026736A1 (en) * | 2005-05-04 | 2010-02-04 | Plut William J | Luminance suppression power conservation |
US20100277648A1 (en) * | 2005-05-04 | 2010-11-04 | Plut William J | White-based power savings |
US10685620B2 (en) | 2005-05-04 | 2020-06-16 | Samsung Electronics Co., Ltd. | Luminance suppression power conservation |
US20060250525A1 (en) * | 2005-05-04 | 2006-11-09 | Plut William J | White-based power savings |
US9785215B2 (en) | 2005-05-04 | 2017-10-10 | Samsung Electronics Co., Ltd. | White-based power savings |
US7760210B2 (en) | 2005-05-04 | 2010-07-20 | Honeywood Technologies, Llc | White-based power savings |
US9881229B2 (en) * | 2005-09-30 | 2018-01-30 | Facebook, Inc. | Apparatus, method and program for image search |
US10810454B2 (en) | 2005-09-30 | 2020-10-20 | Facebook, Inc. | Apparatus, method and program for image search |
US20070219983A1 (en) * | 2006-03-14 | 2007-09-20 | Fish Robert D | Methods and apparatus for facilitating context searching |
US9767184B2 (en) * | 2006-03-14 | 2017-09-19 | Robert D. Fish | Methods and apparatus for facilitating context searching |
US8458606B2 (en) | 2006-12-18 | 2013-06-04 | Microsoft Corporation | Displaying relatedness of media items |
US20080148179A1 (en) * | 2006-12-18 | 2008-06-19 | Microsoft Corporation | Displaying relatedness of media items |
US20090106233A1 (en) * | 2007-08-24 | 2009-04-23 | Veenstra John W | Query engine interpreter and prioritization engine |
US8190604B2 (en) | 2008-04-03 | 2012-05-29 | Microsoft Corporation | User intention modeling for interactive image retrieval |
US20090254539A1 (en) * | 2008-04-03 | 2009-10-08 | Microsoft Corporation | User Intention Modeling For Interactive Image Retrieval |
US20110179385A1 (en) * | 2008-09-24 | 2011-07-21 | Wencheng Li | Content classification utilizing a reduced description palette to simplify content analysis |
US9720937B2 (en) | 2008-12-22 | 2017-08-01 | Koninklijke Philips N.V. | Relevance feedback on a segment of a data object |
TWI396105B (en) * | 2009-07-21 | 2013-05-11 | Univ Nat Taiwan | Digital data processing method for personalized information retrieval and computer readable storage medium and information retrieval system thereof |
US20110038512A1 (en) * | 2009-08-07 | 2011-02-17 | David Petrou | Facial Recognition with Social Network Aiding |
US20110035406A1 (en) * | 2009-08-07 | 2011-02-10 | David Petrou | User Interface for Presenting Search Results for Multiple Regions of a Visual Query |
US9087059B2 (en) | 2009-08-07 | 2015-07-21 | Google Inc. | User interface for presenting search results for multiple regions of a visual query |
US10031927B2 (en) | 2009-08-07 | 2018-07-24 | Google Llc | Facial recognition with social network aiding |
US9135277B2 (en) * | 2009-08-07 | 2015-09-15 | Google Inc. | Architecture for responding to a visual query |
US20110125735A1 (en) * | 2009-08-07 | 2011-05-26 | David Petrou | Architecture for responding to a visual query |
US10534808B2 (en) | 2009-08-07 | 2020-01-14 | Google Llc | Architecture for responding to visual query |
US8670597B2 (en) | 2009-08-07 | 2014-03-11 | Google Inc. | Facial recognition with social network aiding |
US9208177B2 (en) | 2009-08-07 | 2015-12-08 | Google Inc. | Facial recognition with social network aiding |
US10515114B2 (en) | 2009-08-07 | 2019-12-24 | Google Llc | Facial recognition with social network aiding |
US8811742B2 (en) | 2009-12-02 | 2014-08-19 | Google Inc. | Identifying matching canonical documents consistent with visual query structural information |
US8977639B2 (en) | 2009-12-02 | 2015-03-10 | Google Inc. | Actionable search results for visual queries |
US20110128288A1 (en) * | 2009-12-02 | 2011-06-02 | David Petrou | Region of Interest Selector for Visual Queries |
US9405772B2 (en) | 2009-12-02 | 2016-08-02 | Google Inc. | Actionable search results for street view visual queries |
US8805079B2 (en) | 2009-12-02 | 2014-08-12 | Google Inc. | Identifying matching canonical documents in response to a visual query and in accordance with geographic information |
US9183224B2 (en) | 2009-12-02 | 2015-11-10 | Google Inc. | Identifying matching canonical documents in response to a visual query |
US9176986B2 (en) | 2009-12-02 | 2015-11-03 | Google Inc. | Generating a combination of a visual query and matching canonical document |
US9087235B2 (en) | 2009-12-02 | 2015-07-21 | Google Inc. | Identifying matching canonical documents consistent with visual query structural information |
US20110129153A1 (en) * | 2009-12-02 | 2011-06-02 | David Petrou | Identifying Matching Canonical Documents in Response to a Visual Query |
US20110131235A1 (en) * | 2009-12-02 | 2011-06-02 | David Petrou | Actionable Search Results for Street View Visual Queries |
US20110131241A1 (en) * | 2009-12-02 | 2011-06-02 | David Petrou | Actionable Search Results for Visual Queries |
US9852156B2 (en) | 2009-12-03 | 2017-12-26 | Google Inc. | Hybrid use of location sensor data and visual query to return local listings for visual query |
US10346463B2 (en) | 2009-12-03 | 2019-07-09 | Google Llc | Hybrid use of location sensor data and visual query to return local listings for visual query |
US20130101209A1 (en) * | 2010-10-29 | 2013-04-25 | Peking University | Method and system for extraction and association of object of interest in video |
US9317533B2 (en) | 2010-11-02 | 2016-04-19 | Microsoft Technology Licensing, Inc. | Adaptive image retrieval database |
US8463045B2 (en) | 2010-11-10 | 2013-06-11 | Microsoft Corporation | Hierarchical sparse representation for image retrieval |
EP2685420A1 (en) * | 2011-03-11 | 2014-01-15 | Omron Corporation | Image processing device, image processing method and control program |
US20140112598A1 (en) * | 2011-03-11 | 2014-04-24 | Omron Corporation | Image processing device, image processing method and control program |
EP2685420A4 (en) * | 2011-03-11 | 2014-09-17 | Omron Tateisi Electronics Co | Image processing device, image processing method and control program |
WO2014009751A1 (en) * | 2012-07-13 | 2014-01-16 | Deepmind Technologies Limited | Method and apparatus for image searching |
US12032643B2 (en) | 2012-07-20 | 2024-07-09 | Veveo, Inc. | Method of and system for inferring user intent in search input in a conversational interaction system |
US11436296B2 (en) | 2012-07-20 | 2022-09-06 | Veveo, Inc. | Method of and system for inferring user intent in search input in a conversational interaction system |
US10592575B2 (en) | 2012-07-20 | 2020-03-17 | Veveo, Inc. | Method of and system for inferring user intent in search input in a conversational interaction system |
US10572520B2 (en) | 2012-07-31 | 2020-02-25 | Veveo, Inc. | Disambiguating user intent in conversational interaction system for large corpus information retrieval |
US11847151B2 (en) | 2012-07-31 | 2023-12-19 | Veveo, Inc. | Disambiguating user intent in conversational interaction system for large corpus information retrieval |
US11093538B2 (en) | 2012-07-31 | 2021-08-17 | Veveo, Inc. | Disambiguating user intent in conversational interaction system for large corpus information retrieval |
US12169514B2 (en) | 2012-07-31 | 2024-12-17 | Adeia Guides Inc. | Methods and systems for supplementing media assets during fast-access playback operations |
US8935246B2 (en) | 2012-08-08 | 2015-01-13 | Google Inc. | Identifying textual terms in response to a visual query |
US9372920B2 (en) | 2012-08-08 | 2016-06-21 | Google Inc. | Identifying textual terms in response to a visual query |
US10978094B2 (en) | 2013-05-07 | 2021-04-13 | Veveo, Inc. | Method of and system for real time feedback in an incremental speech input interface |
US12169496B2 (en) | 2013-05-10 | 2024-12-17 | Adeia Guides Inc. | Method and system for capturing and exploiting user intent in a conversational interaction based information retrieval system |
US10896184B2 (en) | 2013-05-10 | 2021-01-19 | Veveo, Inc. | Method and system for capturing and exploiting user intent in a conversational interaction based information retrieval system |
EP2894574A1 (en) * | 2014-01-08 | 2015-07-15 | Lg Electronics Inc. | Mobile terminal and controlling method thereof |
US9778825B2 (en) | 2014-01-08 | 2017-10-03 | Lg Electronics Inc. | Mobile terminal and controlling method thereof |
EP3522030A1 (en) * | 2014-01-08 | 2019-08-07 | LG Electronics Inc. | Mobile terminal and controlling method thereof |
US11423074B2 (en) | 2014-12-23 | 2022-08-23 | Rovi Guides, Inc. | Systems and methods for determining whether a negation statement applies to a current or past query |
WO2016109056A1 (en) * | 2014-12-31 | 2016-07-07 | Ebay, Inc. | Dynamic product placement based on perceived value |
US11843676B2 (en) | 2015-01-30 | 2023-12-12 | Rovi Guides, Inc. | Systems and methods for resolving ambiguous terms based on user input |
US11811889B2 (en) | 2015-01-30 | 2023-11-07 | Rovi Guides, Inc. | Systems and methods for resolving ambiguous terms based on media asset schedule |
US11991257B2 (en) | 2015-01-30 | 2024-05-21 | Rovi Guides, Inc. | Systems and methods for resolving ambiguous terms based on media asset chronology |
US10728351B2 (en) | 2015-01-30 | 2020-07-28 | Rovi Guides, Inc. | Systems and methods for resolving ambiguous terms in social chatter based on a user profile |
US10482340B2 (en) * | 2016-12-06 | 2019-11-19 | Samsung Electronics Co., Ltd. | System and method for object recognition and ranging by deformation of projected shapes in a multimodal vision and sensing system for autonomous devices |
CN114040227A (en) * | 2021-10-22 | 2022-02-11 | 游密科技(深圳)有限公司 | Audio and video playing control method and device, electronic equipment and storage medium |
Also Published As
Publication number | Publication date |
---|---|
CA2307264A1 (en) | 2000-12-09 |
JP2001043246A (en) | 2001-02-16 |
CA2307264C (en) | 2007-12-04 |
SG91283A1 (en) | 2002-09-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US6408293B1 (en) | Interactive framework for understanding user's perception of multimedia data | |
CN1094219C (en) | Multimedia Database Retrieval System | |
US6606623B1 (en) | Method and apparatus for content-based image retrieval with learning function | |
JP4540970B2 (en) | Information retrieval apparatus and method | |
US6445834B1 (en) | Modular image query system | |
US7478091B2 (en) | System and method for measuring image similarity based on semantic meaning | |
CA2505779C (en) | Searching images | |
US20020026449A1 (en) | Method of content driven browsing in multimedia databases | |
US6522782B2 (en) | Image and text searching techniques | |
US20030123737A1 (en) | Perceptual method for browsing, searching, querying and visualizing collections of digital images | |
JP4363792B2 (en) | Information retrieval system and method | |
EP1202187A2 (en) | Image retrieval system and methods with semantic and feature based relevance feedback | |
Gupta | Visual information retrieval technology: A virage perspective | |
CN100410928C (en) | Surface Search Support Devices | |
CN1102897A (en) | Improved database retrieval facility for user interface | |
JP2002366575A (en) | Image search method and apparatus | |
GB2395807A (en) | Information retrieval | |
WO2002099703A2 (en) | Modular intelligent multimedia analysis system | |
US6522780B1 (en) | Indexing of images and/or text | |
AU2004218694A1 (en) | Systems and methods for indexing and retrieving images | |
JP2000222418A (en) | Database search method and apparatus | |
Chen et al. | Large-scale Emperor Digital Library and semantics-sensitive region-based retrieval | |
JP4295510B2 (en) | Multimedia information classification and arrangement device and classification and arrangement method | |
King et al. | A feature-based image retrieval database for the fashion, textile, and clothing industry in Hong Kong | |
Borowski et al. | Structuring the visual content of digital libraries using CBIR systems |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: INTERNATIONAL BUSINESS MACHINES CORPORATION, NEW Y Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:AGGARWAL, GAURAV;DUBEY, PRADEEP K.;GHOSAL, SUGATA;AND OTHERS;REEL/FRAME:010174/0432 Effective date: 19990802 |
|
AS | Assignment |
Owner name: INTERNATIONAL BUSINESS MACHINES CORPORATION, NEW Y Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KURTZBERG, JEROME M.;LEVANONI, MENACHEM;REEL/FRAME:010278/0418 Effective date: 19990614 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
AS | Assignment |
Owner name: IPG HEALTHCARE 501 LIMITED, UNITED KINGDOM Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:INTERNATIONAL BUSINESS MACHINES CORPORATION;REEL/FRAME:020083/0864 Effective date: 20070926 |
|
FPAY | Fee payment |
Year of fee payment: 8 |
|
AS | Assignment |
Owner name: PENDRAGON NETWORKS LLC, WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:IPG HEALTHCARE 501 LIMITED;REEL/FRAME:028594/0204 Effective date: 20120410 |
|
AS | Assignment |
Owner name: PHONENICIA INNOVATIONS LLC; SUBSIDIARY OF PENDRELL Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:PENDRAGON NETWORKS LLC;REEL/FRAME:029305/0427 Effective date: 20121031 |
|
FEPP | Fee payment procedure |
Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
FPAY | Fee payment |
Year of fee payment: 12 |
|
AS | Assignment |
Owner name: UNILOC LUXEMBOURG S.A., LUXEMBOURG Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:PHOENICIA INNOVATIONS LLC;REEL/FRAME:045338/0667 Effective date: 20180131 |
|
AS | Assignment |
Owner name: UNILOC 2017 LLC, DELAWARE Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:UNILOC LUXEMBOURG S.A.;REEL/FRAME:046532/0088 Effective date: 20180503 |