US9218063B2 - Sessionless pointing user interface - Google Patents
Sessionless pointing user interface Download PDFInfo
- Publication number
- US9218063B2 US9218063B2 US13/592,352 US201213592352A US9218063B2 US 9218063 B2 US9218063 B2 US 9218063B2 US 201213592352 A US201213592352 A US 201213592352A US 9218063 B2 US9218063 B2 US 9218063B2
- Authority
- US
- United States
- Prior art keywords
- computer
- gesture
- user
- hand
- maps
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/013—Eye tracking input arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/0304—Detection arrangements using opto-electronic means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
- G06F3/167—Audio in a user interface, e.g. using voice commands for navigating, audio feedback
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
- G06T7/75—Determining position or orientation of objects or cameras using feature-based methods involving models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/20—Movements or behaviour, e.g. gesture recognition
- G06V40/28—Recognition of hand or arm movements, e.g. recognition of deaf sign language
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30196—Human being; Person
Definitions
- This invention relates generally to user interfaces for computerized systems, and specifically to user interfaces that are based on three-dimensional sensing.
- tactile interface devices include the computer keyboard, mouse and joystick.
- Touch screens detect the presence and location of a touch by a finger or other object within the display area.
- Infrared remote controls are widely used, and “wearable” hardware devices have been developed, as well, for purposes of remote control.
- Computer interfaces based on three-dimensional (3D) sensing of parts of the user's body have also been proposed.
- 3D three-dimensional
- PCT International Publication WO 03/071410 whose disclosure is incorporated herein by reference, describes a gesture recognition system using depth-perceptive sensors.
- a 3D sensor provides position information, which is used to identify gestures created by a body part of interest.
- the gestures are recognized based on a shape of a body part and its position and orientation over an interval.
- the gesture is classified for determining an input into a related electronic device.
- U.S. Pat. No. 7,348,963 whose disclosure is incorporated herein by reference, describes an interactive video display system, in which a display screen displays a visual image, and a camera captures 3D information regarding an object in an interactive area located in front of the display screen.
- a computer system directs the display screen to change the visual image in response to changes in the object.
- Three-dimensional human interface systems may identify not only the user's hands, but also other parts of the body, including the head, torso and limbs.
- U.S. Patent Application Publication 2010/0034457 whose disclosure is incorporated herein by reference, describes a method for modeling humanoid forms from depth maps.
- the depth map is segmented so as to find a contour of the body.
- the contour is processed in order to identify a torso and one or more limbs of the subject.
- An input is generated to control an application program running on a computer by analyzing a disposition of at least one of the identified limbs in the depth map.
- a method including receiving, by a computer, a sequence of three-dimensional maps containing at least a hand of a user of the computer, identifying, in the maps, a device coupled to the computer, analyzing the maps to detect a gesture performed by the user toward the device, and actuating the device responsively to the gesture.
- an apparatus including a three-dimensional sensing device, and a computer configured to receive from the three-dimensional sensing device a sequence of three-dimensional maps containing at least a hand of a user of the computer, to identify, in the maps, an entity coupled to the computer, to analyze the maps to detect a gesture performed by the user toward the entity, and to actuate the entity responsively to the gesture.
- a computer software product including a non-transitory computer-readable medium, in which program instructions are stored, which instructions, when read by a computer, cause the computer to receive a sequence of three-dimensional maps containing at least a hand of a user of the computer, to identify, in the maps, a device coupled to the computer, to analyze the maps to detect a gesture performed by the user toward the device, and to actuate the device responsively to the gesture.
- FIG. 1 is a schematic, pictorial illustration of a computer system executing a sessionless pointing user interface (SPUI), in accordance with an embodiment of the present invention
- FIG. 2 is a flow diagram that schematically illustrates a method of using a gesture to actuate a device coupled to the computer, in accordance with an embodiment of the present invention
- FIG. 3 is a schematic pictorial illustration of a first scene comprising a user interacting with the SPUI, in accordance with an embodiment of the present invention.
- FIG. 4 is a schematic pictorial illustration of a second scene comprising multiple users interacting with the SPUI, in accordance with an embodiment of the present invention.
- the user When using a tactile input device such as a mouse, the user typically manipulates the physical device in a two-dimensional plane comprising a horizontal X-axis and a vertical Y-axis.
- a non-tactile three-dimensional (3D) user interface also referred to herein as a 3D user interface
- the user may perform gestures in mid-air, and perform the gestures from different positions within a field of view of a 3D sensor coupled to the interface.
- Embodiments of the present invention provide methods and systems for a user to actuate a device (e.g., a television or a lighting fixture) via gestures described hereinbelow.
- a device e.g., a television or a lighting fixture
- the device is driven by a computer executing a sessionless pointing user interface (SPUI) that enables sessionless control of the device, i.e., wherein no session is initiated by the user.
- SPUI sessionless pointing user interface
- the user can turn on a light by pointing at the light.
- the computer can be configured to reduce false positives (i.e., the user may point at the light inadvertently during a conversation) by actuating the device in response to a gesture and a vocal command.
- the computer can turn on the light in response to the user saying “Light” while pointing at the light.
- FIG. 1 is a schematic, pictorial illustration of a sessionless pointing user interface (SPUI) 20 for operation by a user 22 of a computer 26 , in accordance with an embodiment of the present invention.
- the SPUI is based on a 3D sensing device 24 coupled to the computer, which captures 3D scene information of a scene that includes the body (or at least a body part, such as hands 29 ) of the user.
- Device 24 or a separate camera may also capture video images of the scene.
- the information captured by device 24 is processed by computer 26 , which drives a display 28 accordingly.
- Computer 26 executing SPUI 20 , processes data generated by device 24 in order to reconstruct a 3D map of user 22 .
- the term “3D map” refers to a set of 3D coordinates measured with reference to a fixed set of axes based on device 24 .
- the 3D coordinates represent the surface of a given object, in this case the user's body.
- device 24 projects a pattern of spots onto the object and captures an image of the projected pattern.
- Computer 26 then computes the 3D coordinates of points on the surface of the user's body by triangulation, based on transverse shifts of the spots in the pattern.
- interface 20 may use other methods of 3D mapping, using single or multiple cameras or other types of sensors, as are known in the art.
- sensing device 24 may include a microphone (not shown) configured to convey audio signals generated in response to speech or sound from user 22 to computer 26 .
- Computer 26 can be configured to process the audio signals, thereby enabling the computer to respond to vocal commands from user 22 , in addition to physical gestures performed by the user.
- Computer 26 typically comprises a general-purpose computer processor, which is programmed in software to carry out the functions described hereinbelow.
- the software may be downloaded to the processor in electronic form, over a network, for example, or it may alternatively be provided on tangible media, such as optical, magnetic, or electronic memory media.
- some or all of the functions of the image processor may be implemented in dedicated hardware, such as a custom or semi-custom integrated circuit or a programmable digital signal processor (DSP).
- DSP programmable digital signal processor
- computer 26 is shown in FIG. 1 , by way of example, as a separate unit from sensing device 24 , some or all of the processing functions of the computer may be performed by suitable dedicated circuitry within the housing of the sensing device or otherwise associated with the sensing device.
- processing functions may be carried out by a suitable processor that is integrated with display 28 (in a television set, for example) or with any other suitable sort of computerized device, such as a game console or media player.
- the sensing functions of device 24 may likewise be integrated into the computer or other computerized apparatus that is to be controlled by the sensor output.
- computer 26 extracts 3D connected components corresponding to the parts of the body from the depth data generated by device 24 .
- Techniques that may be used for this purpose are described, for example, in U.S. patent application Ser. No. 12/854,187, filed Aug. 11, 2010, whose disclosure is incorporated herein by reference.
- the computer analyzes these extracted components in order to reconstruct a “skeleton” of the user's body, as described in the above-mentioned U.S. Patent Application Publication 2010/0034457, or in U.S. patent application Ser. No. 12/854,188, filed Aug. 11, 2010, whose disclosure is also incorporated herein by reference.
- other techniques may be used to identify certain parts of the user's body, and there is no need for the entire body to be visible to device 24 or for the skeleton to be reconstructed, in whole or even in part.
- computer 26 can assume a position of a body part such as a tip of finger 30 , even though the body part (e.g., the fingertip) may not be detected by the depth map due to issues such as minimal object size and reduced resolution at greater distances from device 24 .
- computer 26 can auto-complete a body part based on an expected shape of the human part from an earlier detection of the body part, or from tracking the body part along several (previously) received depth maps.
- the information generated by computer 26 as a result of this skeleton reconstruction includes the location and direction of the user's head, as well as of the arms, torso, and possibly legs, hands and other features, as well. Changes in these features from frame to frame (i.e. depth maps) or in postures of the user can provide an indication of gestures and other motions made by the user. User posture, gestures and other motions may provide a control input for user interaction with interface 20 . These body motions may be combined with other interaction modalities that are sensed by device 24 , including user eye movements, as described above, as well as voice commands and other sounds. Interface 20 thus enables user 22 to perform various remote control functions and to interact with applications, interfaces, video programs, images, games and other multimedia content appearing on display 28 .
- FIG. 2 is a flow diagram that schematically illustrates a method of actuating display 28
- FIG. 3 is a schematic pictorial illustration of a first scene 40 comprising user 22 interacting with computer 26 , in accordance with an embodiment of the present invention.
- 3D sensing device 24 and computer 26 are both integrated into display 28 .
- a receive step 30 computer 26 receives, from sensing device 24 , a sequence of 3D maps containing at least hand 29 , and in an identification step 32 , the computer identifies in the maps the hand and at least one device or entity coupled to, and driven by (i.e., controlled by), the computer.
- the device comprises display 28 that incorporates sensing device 24 . Therefore, computer 26 can interpret a gesture toward sensing device 24 as a gesture toward display 28 .
- computer 26 may control multiple devices in proximity to user 22 .
- devices that can be driven by computer 26 include, but are not limited to lamp fixtures, ventilation (i.e., heating/cooling) units, ceiling fans and electronic entertainment systems.
- the user can identify the location of the devices during an initialization step.
- computer 26 can identify devices in proximity to the sensing device (e.g., via Bluetooth or another communication protocol known in the art), present interactive items on display 28 corresponding to each of the identified devices, and direct the user to first point at a given one of the interactive items and then point to the corresponding device.
- computer 26 can initialize sessionless pointing user interface (SPUI) 20 by defining a pyramid shaped region 42 within a field of view of 3D sensor 24 .
- pyramid shaped region 42 is rectangular and comprises an apex (i.e., the narrow tip) 56 that meets user 22 , and a base (i.e., the wide end) 58 that encompasses display 28 .
- computer 26 positions apex 56 at a head 44 or an eyeball 46 of user 22 .
- a detect step 34 computer 26 detects, in the sequence of 3D maps, a gesture directed toward display 28 .
- gestures that user 22 can direct toward display 28 include, but are not limited to a grab gesture and pointing gestures.
- the grab gesture is described in U.S. patent application Ser. No. 13/423,314, filed on Mar. 19, 2012, whose disclosure is incorporated herein by reference, and comprises user 22 closing at least some fingers of hand 29 .
- Pointing gestures are described in PCT International Application PCT/IB2012/050577, filed Feb. 9, 2012, whose disclosure is incorporated herein by reference, and include a point-select, a point-touch, and a point-hold gesture.
- a point-select For example, to perform the point-touch gesture, user 22 points hand 29 toward display 28 , stops or slows down the hand, and then pulls the hand back toward the user.
- computer 26 may be configured to define conditions for identifying a pointing gesture when analyzing the 3D maps. Examples of conditions include:
- computer 26 may be configured to detect gestures performed by user 22 with fingers of hand 29 , such as an index finger 52 .
- Gestures performed by user 22 with hand 29 are referred to herein as hand gestures, and gestures performed with the fingers of the hand are referred to herein as finger gestures.
- finger gestures can help reduce fatigue while interacting with computer 26 .
- user 22 can keep the hand resting on a lap (i.e., while sitting), and perform gestures with the fingers of hand 29 .
- computer 26 can be configured at either a higher resolution (also referred to herein as finger resolution) to detect the position (i.e., location and orientation) of individual fingers of hand 29 , or a lower resolution (also referred to herein as hand resolution) to detect the position of hand 29 . Additionally or alternatively, computer 26 may be configured to detect hand gestures and finger gestures at shorter distances between user 22 and 3D sensing device 24 , and to detect hand gestures at longer distances between the user and the 3D sensing device. When configured for finger resolution, computer can respond to finger gestures such as a finger pointing gesture or a grab gesture.
- computer 26 actuates the device to which the user is pointing, and the method ends.
- computer 26 can actuate (i.e., turn on) display 28 in response to a pointing gesture performed by user 22 .
- computer 26 can present multiple interactive items (not shown) on display 28 , identify a given interactive item to which the user is pointing, and actuate the identified interactive item.
- computer 26 may present advertisements for multiple restaurants on display 28 , and present detailed information on a given restaurant upon detecting a gesture directed toward the advertisement of the given restaurant.
- selecting a given interactive item comprises executing a software application associated with the given interactive item.
- the given interactive item is associated with a media item (e.g., a music track or a movie), and selecting a given interactive item comprises playing a media file associated with the given interactive item.
- computer 26 can calculate a pointing geometry for any number of users 22 interacting with the computer.
- the computer typically responds to a given user 22 performing a pointing gesture within pyramid shaped region 42 . Therefore, if there is more than one user interacting with computer 26 , then the computer can define a separate pyramid shaped region 42 for each of the users.
- the computer can be configured to respond to pointing gestures performed by user 22 both inside and outside pyramid shaped region 42 .
- FIG. 4 is a schematic pictorial illustration of a second scene 60 (also referred to herein as room 60 ) comprising multiple users 22 interacting with computer 26 , in accordance with an embodiment of the present invention.
- users 22 may be differentiated by appending a letter to the identifying numeral, so that the users comprise users 22 A and 22 B.
- computer 26 can be configured to control display 28 , lighting fixtures 64 and 66 , and air vents 68 and 70 .
- a given user 22 can individually control multiple devices configured to be driven by computer 26 (i.e., in addition to display 28 ).
- computer 26 can turn on (or off) a given one of the lighting fixtures in response to detecting, in the 3D maps, a pointing gesture directed toward the given lighting fixture.
- computer 26 can open and close a given one of the air vents in response to detecting, in the 3D maps, a gesture directed toward the given air vent.
- a gesture directed toward the given air vent To open the given air vent, user 22 can perform a release gesture toward the given air vent.
- the release gesture described in U.S. patent application Ser. No. 13/423,314, filed on Mar. 19, 2012, whose disclosure is incorporated herein by reference, comprises user relaxing hand 29 so as to open the hand from a closed or folded state.
- a grab gesture as described supra
- While the configuration of scene 60 includes a single 3D sensing device 24 , other configurations may include multiple 3D sensors, and are considered to be within the spirit and scope of the present invention.
- additional sensing devices 24 may be positioned in room 60 , thereby enabling computer 26 to monitor the entire room (i.e., no blind spots) for gestures performed by the users.
- the devices e.g., the lighting fixtures and the air vents
- a given user 22 in addition to controlling multiple devices, can perform a pointing gesture to select one or more additional users in the user's vicinity. For example, user 22 A can perform a first pointing gesture to select an icon on display 28 to start a multiplayer game, and then perform a subsequent pointing gesture directed toward user 22 B user, thereby indicating user 22 B as an additional participant in the game.
- computer 26 can be configured to have the actuated device communicate with a further device in response to a subsequent gesture directed toward the further device.
- user 22 A wants to transfer a digital photograph 72 from a smartphone 62 to a photo album application 74 that computer 26 is presenting on display 28 .
- Computer 26 can update photo album 74 upon detecting, in the 3D maps, user 22 A performing an initial pointing gesture directed toward smartphone 62 and a subsequent pointing gesture directed toward display 28 .
- computer 26 presents a web page on display 28 , and user 22 B wants to copy the uniform resource locator (URL) of the web page to a web browser application executing on a laptop computer 74 .
- Computer 26 can copy the URL to the web browser application upon detecting, in the 3D maps, user 22 B performing an initial pointing gesture directed toward display 28 and a subsequent pointing gesture directed toward laptop computer 74 .
- URL uniform resource locator
- Computer 26 can present an air conditioning control menu (e.g., with icons for controlling settings such as temperature and fan speed) on display 28 in response to detecting, in the 3D maps, user 22 A performing an initial pointing gesture directed toward one of the air vents and a subsequent pointing gesture directed toward display 28 .
- an air conditioning control menu e.g., with icons for controlling settings such as temperature and fan speed
- computer 26 can copy the URL of the web page presented on display 28 to laptop computer 74 in response to detecting, in the 3D maps, user 22 B performing a grab gesture directed toward display 28 and subsequent performing a release gesture directed toward laptop computer 74 .
- light fixture 64 is positioned in proximity to air vent 68
- light fixture 66 is positioned in proximity to air vent 70 .
- the computer may actuate a different device responsively to the gesture. For example, if user 22 B intends to turn on light 66 but performs a pointing gesture toward air vent 70 , computer 26 may turn on air vent 70 in response to the pointing gesture.
- computer 26 be configured to actuate a given device in response to a combination of a vocal (i.e., audio) command and a gesture.
- a vocal command i.e., audio
- the user can say “Light” while pointing in proximity to lighting fixture 66 .
- Combining vocal command processing with gesture recognition can enhance the accuracy of computer 26 controlling multiple devices in proximity to each other.
- user 22 performs a pointing gesture without intending to actuate a device controlled by computer 26 .
- user 22 A can point toward user 22 B during a conversation.
- computer 26 can be configured to determine the intent of user 22 by detecting (and tracking), in the 3D maps, a gaze of the user, and actuating a given device in response to the gaze and a gesture directed toward the given device.
- tracking the gaze of the user comprises detecting an orientation of head 44 and/or eyeball 46 .
- computer 26 can be configured to respond to a pointing gesture directed towards display 28 only if the 3D maps indicate that the user's gaze is directed toward the display. Identifying a gaze direction of user 22 is described in PCT International Application PCT/IB2012/050577, filed Feb. 9, 2012, whose disclosure is incorporated herein by reference.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Multimedia (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Psychiatry (AREA)
- Social Psychology (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
Description
-
- Defining an
interaction region 48 within pyramid shapedregion 42, and identifying a pointing gesture upon the 3Dmaps indicating user 22positioning hand 29 withinregion 48 and moving the hand toward the display. - Defining an angle threshold (e.g., 90 degrees) for
elbow 50, and identifying a pointing gesture upon the 3Dmaps indicating user 22 extendinghand 29 toward the display and extendingelbow 50 at anangle 54 greater than or equal to the angle threshold. - Defining a minimum time period (e.g., 200 milliseconds), and identifying a pointing gesture upon the 3D
maps indicating user 22 pausinghand 29 for the minimum time period after extendinghand 29 toward the display. Requiring a minimum time period enablescomputer 26 to mimic the natural behavior of an individual pointing at an object.
- Defining an
Claims (10)
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/592,352 US9218063B2 (en) | 2011-08-24 | 2012-08-23 | Sessionless pointing user interface |
US14/919,751 US10088909B2 (en) | 2011-08-24 | 2015-10-22 | Sessionless pointing user interface |
US16/106,028 US10642371B2 (en) | 2011-08-24 | 2018-08-21 | Sessionless pointing user interface |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201161526692P | 2011-08-24 | 2011-08-24 | |
US13/592,352 US9218063B2 (en) | 2011-08-24 | 2012-08-23 | Sessionless pointing user interface |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/919,751 Continuation US10088909B2 (en) | 2011-08-24 | 2015-10-22 | Sessionless pointing user interface |
Publications (2)
Publication Number | Publication Date |
---|---|
US20130055120A1 US20130055120A1 (en) | 2013-02-28 |
US9218063B2 true US9218063B2 (en) | 2015-12-22 |
Family
ID=47745504
Family Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/592,352 Active US9218063B2 (en) | 2011-08-24 | 2012-08-23 | Sessionless pointing user interface |
US14/919,751 Active 2033-01-12 US10088909B2 (en) | 2011-08-24 | 2015-10-22 | Sessionless pointing user interface |
US16/106,028 Active US10642371B2 (en) | 2011-08-24 | 2018-08-21 | Sessionless pointing user interface |
Family Applications After (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/919,751 Active 2033-01-12 US10088909B2 (en) | 2011-08-24 | 2015-10-22 | Sessionless pointing user interface |
US16/106,028 Active US10642371B2 (en) | 2011-08-24 | 2018-08-21 | Sessionless pointing user interface |
Country Status (1)
Country | Link |
---|---|
US (3) | US9218063B2 (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140370470A1 (en) * | 2013-06-13 | 2014-12-18 | Gary And Mary West Health Institute | Systems, apparatus and methods for delivery and augmentation of behavior modification therapy and teaching |
US20160144867A1 (en) * | 2014-11-20 | 2016-05-26 | Toyota Motor Engineering & Manufacturing North America, Inc. | Autonomous vehicle detection of and response to traffic officer presence |
USD800586S1 (en) * | 2015-03-25 | 2017-10-24 | Muv Interactive Ltd. | Wearable sensing device |
US9983684B2 (en) | 2016-11-02 | 2018-05-29 | Microsoft Technology Licensing, Llc | Virtual affordance display at virtual target |
US10528794B2 (en) * | 2017-06-05 | 2020-01-07 | Motorola Solutions, Inc. | System and method for tailoring an electronic digital assistant inquiry response as a function of previously detected user ingestion of related video information |
US10671065B2 (en) * | 2013-10-25 | 2020-06-02 | Ioannis Micros | Optically assisted landing and takeoff of drones |
Families Citing this family (51)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9035876B2 (en) | 2008-01-14 | 2015-05-19 | Apple Inc. | Three-dimensional user interface session control |
US8933876B2 (en) | 2010-12-13 | 2015-01-13 | Apple Inc. | Three dimensional user interface session control |
US8639020B1 (en) | 2010-06-16 | 2014-01-28 | Intel Corporation | Method and system for modeling subjects from a depth map |
US9201501B2 (en) | 2010-07-20 | 2015-12-01 | Apple Inc. | Adaptive projector |
JP5791131B2 (en) | 2010-07-20 | 2015-10-07 | アップル インコーポレイテッド | Interactive reality extension for natural interactions |
US8959013B2 (en) | 2010-09-27 | 2015-02-17 | Apple Inc. | Virtual keyboard for a non-tactile three dimensional user interface |
US8872762B2 (en) | 2010-12-08 | 2014-10-28 | Primesense Ltd. | Three dimensional user interface cursor control |
WO2012107892A2 (en) | 2011-02-09 | 2012-08-16 | Primesense Ltd. | Gaze detection in a 3d mapping environment |
US9857868B2 (en) | 2011-03-19 | 2018-01-02 | The Board Of Trustees Of The Leland Stanford Junior University | Method and system for ergonomic touch-free interface |
US20120257035A1 (en) * | 2011-04-08 | 2012-10-11 | Sony Computer Entertainment Inc. | Systems and methods for providing feedback by tracking user gaze and gestures |
US8840466B2 (en) | 2011-04-25 | 2014-09-23 | Aquifi, Inc. | Method and system to create three-dimensional mapping in a two-dimensional game |
US11048333B2 (en) | 2011-06-23 | 2021-06-29 | Intel Corporation | System and method for close-range movement tracking |
JP6074170B2 (en) | 2011-06-23 | 2017-02-01 | インテル・コーポレーション | Short range motion tracking system and method |
US8881051B2 (en) | 2011-07-05 | 2014-11-04 | Primesense Ltd | Zoom-based gesture user interface |
US9459758B2 (en) | 2011-07-05 | 2016-10-04 | Apple Inc. | Gesture-based interface with enhanced features |
US9377865B2 (en) | 2011-07-05 | 2016-06-28 | Apple Inc. | Zoom-based gesture user interface |
US9030498B2 (en) | 2011-08-15 | 2015-05-12 | Apple Inc. | Combining explicit select gestures and timeclick in a non-tactile three dimensional user interface |
US9122311B2 (en) | 2011-08-24 | 2015-09-01 | Apple Inc. | Visual feedback for tactile and non-tactile user interfaces |
US9218063B2 (en) | 2011-08-24 | 2015-12-22 | Apple Inc. | Sessionless pointing user interface |
US8854433B1 (en) | 2012-02-03 | 2014-10-07 | Aquifi, Inc. | Method and system enabling natural user interface gestures with an electronic system |
JP6044079B2 (en) * | 2012-02-06 | 2016-12-14 | ソニー株式会社 | Information processing apparatus, information processing method, and program |
US9229534B2 (en) | 2012-02-28 | 2016-01-05 | Apple Inc. | Asymmetric mapping for tactile and non-tactile user interfaces |
US9671869B2 (en) * | 2012-03-13 | 2017-06-06 | Eyesight Mobile Technologies Ltd. | Systems and methods of direct pointing detection for interaction with a digital device |
US11169611B2 (en) | 2012-03-26 | 2021-11-09 | Apple Inc. | Enhanced virtual touchpad |
US9477303B2 (en) | 2012-04-09 | 2016-10-25 | Intel Corporation | System and method for combining three-dimensional tracking with a three-dimensional display for a user interface |
US9111135B2 (en) | 2012-06-25 | 2015-08-18 | Aquifi, Inc. | Systems and methods for tracking human hands using parts based template matching using corresponding pixels in bounded regions of a sequence of frames that are a specified distance interval from a reference camera |
US8934675B2 (en) | 2012-06-25 | 2015-01-13 | Aquifi, Inc. | Systems and methods for tracking human hands by performing parts based template matching using images from multiple viewpoints |
US8836768B1 (en) | 2012-09-04 | 2014-09-16 | Aquifi, Inc. | Method and system enabling natural user interface gestures with user wearable glasses |
US10241639B2 (en) * | 2013-01-15 | 2019-03-26 | Leap Motion, Inc. | Dynamic user interactions for display control and manipulation of display objects |
US9092665B2 (en) | 2013-01-30 | 2015-07-28 | Aquifi, Inc | Systems and methods for initializing motion tracking of human hands |
US9129155B2 (en) | 2013-01-30 | 2015-09-08 | Aquifi, Inc. | Systems and methods for initializing motion tracking of human hands using template matching within bounded regions determined using a depth map |
US20140244267A1 (en) * | 2013-02-26 | 2014-08-28 | Avaya Inc. | Integration of user orientation into a voice command system |
US20140258942A1 (en) * | 2013-03-05 | 2014-09-11 | Intel Corporation | Interaction of multiple perceptual sensing inputs |
US9298266B2 (en) | 2013-04-02 | 2016-03-29 | Aquifi, Inc. | Systems and methods for implementing three-dimensional (3D) gesture based graphical user interfaces (GUI) that incorporate gesture reactive interface objects |
US9798388B1 (en) | 2013-07-31 | 2017-10-24 | Aquifi, Inc. | Vibrotactile system to augment 3D input systems |
US10168873B1 (en) | 2013-10-29 | 2019-01-01 | Leap Motion, Inc. | Virtual interactions for machine control |
US9996797B1 (en) | 2013-10-31 | 2018-06-12 | Leap Motion, Inc. | Interactions with virtual objects for machine control |
US9507417B2 (en) | 2014-01-07 | 2016-11-29 | Aquifi, Inc. | Systems and methods for implementing head tracking based graphical user interfaces (GUI) that incorporate gesture reactive interface objects |
US9804737B2 (en) | 2014-01-27 | 2017-10-31 | Groupon, Inc. | Learning user interface |
US9619105B1 (en) | 2014-01-30 | 2017-04-11 | Aquifi, Inc. | Systems and methods for gesture based interaction with viewpoint dependent user interfaces |
IL232391A0 (en) * | 2014-04-30 | 2014-08-31 | Pointgrab Ltd | Method and system for controlling fixtures |
US10429923B1 (en) | 2015-02-13 | 2019-10-01 | Ultrahaptics IP Two Limited | Interaction engine for creating a realistic experience in virtual reality/augmented reality environments |
US9696795B2 (en) | 2015-02-13 | 2017-07-04 | Leap Motion, Inc. | Systems and methods of creating a realistic grab experience in virtual reality/augmented reality environments |
CN107436678B (en) * | 2016-05-27 | 2020-05-19 | 富泰华工业(深圳)有限公司 | Gesture control system and method |
DE112017003398T5 (en) * | 2016-07-05 | 2019-03-21 | Sony Corporation | INFORMATION PROCESSING DEVICE, INFORMATION PROCESSING PROCESS AND PROGRAM |
KR101974086B1 (en) * | 2016-09-30 | 2019-05-02 | 삼성디스플레이 주식회사 | Display module |
CN108089708A (en) * | 2017-12-22 | 2018-05-29 | 西安交通大学 | A suspended gesture interaction method for improving gesture fatigue |
JP7017689B2 (en) * | 2017-12-29 | 2022-02-09 | 富士通株式会社 | Information processing equipment, information processing system and information processing method |
US20200125175A1 (en) * | 2018-10-17 | 2020-04-23 | WiSilica Inc. | System using location, video-processing, and voice as user interface for controlling devices |
CN109872283A (en) * | 2019-01-18 | 2019-06-11 | 维沃移动通信有限公司 | An image processing method and mobile terminal |
US20200393156A1 (en) * | 2019-06-12 | 2020-12-17 | Alibaba Group Holding Limited | Temperature adjustment feedback system in response to user input |
Citations (256)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4550250A (en) | 1983-11-14 | 1985-10-29 | Hei, Inc. | Cordless digital graphics input device |
US4789921A (en) | 1987-02-20 | 1988-12-06 | Minnesota Mining And Manufacturing Company | Cone shaped Fresnel reflector |
US4988981A (en) | 1987-03-17 | 1991-01-29 | Vpl Research, Inc. | Computer data entry and manipulation apparatus and method |
US5264836A (en) * | 1991-01-15 | 1993-11-23 | Apple Computer, Inc. | Three dimensional cursor |
US5495576A (en) | 1993-01-11 | 1996-02-27 | Ritchey; Kurtis J. | Panoramic image based virtual reality/telepresence audio-visual system and method |
US5588139A (en) | 1990-06-07 | 1996-12-24 | Vpl Research, Inc. | Method and system for generating objects for a multi-person virtual world using data flow networks |
US5594469A (en) | 1995-02-21 | 1997-01-14 | Mitsubishi Electric Information Technology Center America Inc. | Hand gesture machine control system |
US5846134A (en) | 1995-07-14 | 1998-12-08 | Latypov; Nurakhmed Nurislamovich | Method and apparatus for immersion of a user into virtual reality |
US5852672A (en) | 1995-07-10 | 1998-12-22 | The Regents Of The University Of California | Image system for three dimensional, 360 DEGREE, time sequence surface mapping of moving objects |
US5862256A (en) | 1996-06-14 | 1999-01-19 | International Business Machines Corporation | Distinguishing gestures from handwriting in a pen based computer by size discrimination |
US5864635A (en) | 1996-06-14 | 1999-01-26 | International Business Machines Corporation | Distinguishing gestures from handwriting in a pen based computer by stroke analysis |
US5870196A (en) | 1995-10-16 | 1999-02-09 | European Community | Optical three-dimensional profilometry method based on processing SPECKLE images in partially coherent light, and interferometer implementing such a method |
US5917937A (en) | 1997-04-15 | 1999-06-29 | Microsoft Corporation | Method for performing stereo matching to recover depths, colors and opacities of surface elements |
WO1999035633A2 (en) | 1998-01-06 | 1999-07-15 | The Video Mouse Group | Human motion following computer mouse and game controller |
US5973700A (en) * | 1992-09-16 | 1999-10-26 | Eastman Kodak Company | Method and apparatus for optimizing the resolution of images which have an apparent depth |
US6002808A (en) | 1996-07-26 | 1999-12-14 | Mitsubishi Electric Information Technology Center America, Inc. | Hand gesture control system |
US6005548A (en) | 1996-08-14 | 1999-12-21 | Latypov; Nurakhmed Nurislamovich | Method for tracking and displaying user's spatial position and orientation, a method for representing virtual reality for a user, and systems of embodiment of such methods |
US6064387A (en) | 1998-01-23 | 2000-05-16 | Dell, Usa, L.P. | Animated cursor and icon for computers |
US6072494A (en) | 1997-10-15 | 2000-06-06 | Electric Planet, Inc. | Method and apparatus for real-time gesture recognition |
US6084979A (en) | 1996-06-20 | 2000-07-04 | Carnegie Mellon University | Method for creating virtual reality |
US6111580A (en) * | 1995-09-13 | 2000-08-29 | Kabushiki Kaisha Toshiba | Apparatus and method for controlling an electronic device with user action |
US6191773B1 (en) | 1995-04-28 | 2001-02-20 | Matsushita Electric Industrial Co., Ltd. | Interface apparatus |
US6215890B1 (en) * | 1997-09-26 | 2001-04-10 | Matsushita Electric Industrial Co., Ltd. | Hand gesture recognizing device |
US6229541B1 (en) | 1999-09-03 | 2001-05-08 | Isurftv | Use of templates for cost-effective secure linking of video stream objects |
US6243054B1 (en) | 1998-07-01 | 2001-06-05 | Deluca Michael | Stereoscopic user interface method and apparatus |
US6252988B1 (en) | 1998-07-09 | 2001-06-26 | Lucent Technologies Inc. | Method and apparatus for character recognition using stop words |
US6262740B1 (en) | 1997-08-01 | 2001-07-17 | Terarecon, Inc. | Method for rendering sections of a volume data set |
US6345111B1 (en) | 1997-02-28 | 2002-02-05 | Kabushiki Kaisha Toshiba | Multi-modal interface apparatus and method |
US6345893B2 (en) * | 1998-06-15 | 2002-02-12 | Vega Vista, Inc. | Ergonomic systems and methods for operating computers |
US20020057383A1 (en) | 1998-10-13 | 2002-05-16 | Ryuichi Iwamura | Motion sensing interface |
US20020071607A1 (en) | 2000-10-31 | 2002-06-13 | Akinori Kawamura | Apparatus, method, and program for handwriting recognition |
US6452584B1 (en) | 1997-04-23 | 2002-09-17 | Modern Cartoon, Ltd. | System for data management based on hand gestures |
US6456262B1 (en) | 2000-05-09 | 2002-09-24 | Intel Corporation | Microdisplay with eye gaze detection |
US20020158873A1 (en) | 2001-01-26 | 2002-10-31 | Todd Williamson | Real-time virtual viewpoint in simulated reality environment |
US6507353B1 (en) | 1999-12-10 | 2003-01-14 | Godot Huard | Influencing virtual actors in an interactive environment |
US6512838B1 (en) | 1999-09-22 | 2003-01-28 | Canesta, Inc. | Methods for enhancing performance and data acquired from three-dimensional image systems |
US6519363B1 (en) | 1999-01-13 | 2003-02-11 | International Business Machines Corporation | Method and system for automatically segmenting and recognizing handwritten Chinese characters |
US20030057972A1 (en) | 1999-07-26 | 2003-03-27 | Paul Pfaff | Voltage testing and measurement |
US20030088463A1 (en) | 1999-10-21 | 2003-05-08 | Steven Fischman | System and method for group advertisement optimization |
US20030156756A1 (en) | 2002-02-15 | 2003-08-21 | Gokturk Salih Burak | Gesture recognition system using depth perceptive sensors |
US20030185444A1 (en) | 2002-01-10 | 2003-10-02 | Tadashi Honda | Handwriting information processing apparatus, handwriting information processing method, and storage medium having program stored therein for handwriting information processing |
US20030227453A1 (en) | 2002-04-09 | 2003-12-11 | Klaus-Peter Beier | Method, system and computer program product for automatically creating an animated 3-D scenario from human position and path data |
US20030235341A1 (en) | 2002-04-11 | 2003-12-25 | Gokturk Salih Burak | Subject segmentation and tracking using 3D sensing technology for video compression in multimedia applications |
US6681031B2 (en) | 1998-08-10 | 2004-01-20 | Cybernet Systems Corporation | Gesture-controlled interfaces for self-service machines and other applications |
US6686921B1 (en) | 2000-08-01 | 2004-02-03 | International Business Machines Corporation | Method and apparatus for acquiring a set of consistent image maps to represent the color of the surface of an object |
US6690370B2 (en) | 1995-06-07 | 2004-02-10 | Geovector Corp. | Vision system computer modeling apparatus including interaction with real scenes with respect to perspective and spatial relationship as measured in real-time |
US20040046744A1 (en) | 1999-11-04 | 2004-03-11 | Canesta, Inc. | Method and apparatus for entering data using a virtual input device |
US6741251B2 (en) | 2001-08-16 | 2004-05-25 | Hewlett-Packard Development Company, L.P. | Method and apparatus for varying focus in a scene |
US20040104935A1 (en) | 2001-01-26 | 2004-06-03 | Todd Williamson | Virtual reality immersion system |
US20040135744A1 (en) | 2001-08-10 | 2004-07-15 | Oliver Bimber | Virtual showcases |
US20040155962A1 (en) | 2003-02-11 | 2004-08-12 | Marks Richard L. | Method and apparatus for real time motion capture |
US20040174770A1 (en) | 2002-11-27 | 2004-09-09 | Rees Frank L. | Gauss-Rees parametric ultrawideband system |
US6791540B1 (en) | 1999-06-11 | 2004-09-14 | Canon Kabushiki Kaisha | Image processing apparatus |
US20040184640A1 (en) | 2003-03-17 | 2004-09-23 | Samsung Electronics Co., Ltd. | Spatial motion recognition system and method using a virtual handwriting plane |
US20040183775A1 (en) | 2002-12-13 | 2004-09-23 | Reactrix Systems | Interactive directed light/sound system |
US20040184659A1 (en) | 2003-03-17 | 2004-09-23 | Samsung Electronics Co., Ltd. | Handwriting trajectory recognition system and method |
US20040193413A1 (en) * | 2003-03-25 | 2004-09-30 | Wilson Andrew D. | Architecture for controlling a computer using hand gestures |
US6803928B2 (en) | 2000-06-06 | 2004-10-12 | Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. | Extended virtual table: an optical extension for table-like projection systems |
US20040222977A1 (en) | 2003-05-05 | 2004-11-11 | Bear Eric Gould | Notification lights, locations and rules for a computer system |
WO2004107272A1 (en) | 2003-05-29 | 2004-12-09 | Sony Computer Entertainment Inc. | System and method for providing a real-time three-dimensional interactive environment |
US20040258314A1 (en) | 2003-06-23 | 2004-12-23 | Shoestring Research, Llc | Region segmentation and characterization systems and methods for augmented reality |
WO2005003948A1 (en) | 2003-07-08 | 2005-01-13 | Sony Computer Entertainment Inc. | Control system and control method |
US6853935B2 (en) | 2000-11-30 | 2005-02-08 | Canon Kabushiki Kaisha | Information processing apparatus, mixed reality presentation apparatus, method thereof, and storage medium |
US20050031166A1 (en) | 2003-05-29 | 2005-02-10 | Kikuo Fujimura | Visual tracking using depth data |
US6857746B2 (en) | 2002-07-01 | 2005-02-22 | Io2 Technology, Llc | Method and system for free-space imaging display and interface |
US20050089194A1 (en) | 2003-10-24 | 2005-04-28 | Matthew Bell | Method and system for processing captured image information in an interactive video display system |
US20050088407A1 (en) | 2003-10-24 | 2005-04-28 | Matthew Bell | Method and system for managing an interactive video display system |
US20050110964A1 (en) | 2002-05-28 | 2005-05-26 | Matthew Bell | Interactive video window display system |
US20050122308A1 (en) | 2002-05-28 | 2005-06-09 | Matthew Bell | Self-contained interactive video display system |
US20050162381A1 (en) | 2002-05-28 | 2005-07-28 | Matthew Bell | Self-contained interactive video display system |
US20050190972A1 (en) | 2004-02-11 | 2005-09-01 | Thomas Graham A. | System and method for position determination |
US6951515B2 (en) | 1999-06-11 | 2005-10-04 | Canon Kabushiki Kaisha | Game apparatus for mixed reality space, image processing method thereof, and program storage medium |
WO2005094958A1 (en) | 2004-03-23 | 2005-10-13 | Harmonix Music Systems, Inc. | Method and apparatus for controlling a three-dimensional character in a three-dimensional gaming environment |
US20050254726A1 (en) | 2004-02-25 | 2005-11-17 | The University Of North Carolina At Chapel Hill | Methods, systems, and computer program products for imperceptibly embedding structured light patterns in projected color images for display on planar and non-planar surfaces |
US20050265583A1 (en) | 1999-03-08 | 2005-12-01 | Vulcan Patents Llc | Three dimensional object pose estimation which employs dense depth information |
US6977654B2 (en) | 2002-10-30 | 2005-12-20 | Iviz, Inc. | Data visualization with animated speedometer dial charts |
US20060010400A1 (en) | 2004-06-28 | 2006-01-12 | Microsoft Corporation | Recognizing gestures and using gestures for interacting with software applications |
US20060020905A1 (en) | 2004-07-20 | 2006-01-26 | Hillcrest Communications, Inc. | Graphical cursor navigation methods |
US7023436B2 (en) | 2000-04-19 | 2006-04-04 | Sony Corporation | Three-dimensional model processing device, three-dimensional model processing method, program providing medium |
US20060092138A1 (en) | 2004-10-29 | 2006-05-04 | Microsoft Corporation | Systems and methods for interacting with a computer through handwriting to a screen |
US7042440B2 (en) | 1997-08-22 | 2006-05-09 | Pryor Timothy R | Man machine interfaces and applications |
US7042442B1 (en) | 2000-06-27 | 2006-05-09 | International Business Machines Corporation | Virtual invisible keyboard |
US20060110008A1 (en) * | 2003-11-14 | 2006-05-25 | Roel Vertegaal | Method and apparatus for calibration-free eye tracking |
US20060115155A1 (en) | 2000-11-10 | 2006-06-01 | Microsoft Corporation | Implicit page breaks for digitally represented handwriting |
US20060139314A1 (en) | 2002-05-28 | 2006-06-29 | Matthew Bell | Interactive video display system |
US20060149737A1 (en) | 2002-09-28 | 2006-07-06 | Yonggang Du | Method for accessing and additional content with parental control |
US20060159344A1 (en) | 2002-12-26 | 2006-07-20 | Xiaoling Shao | Method and system for three-dimensional handwriting recognition |
US20060187196A1 (en) | 2005-02-08 | 2006-08-24 | Underkoffler John S | System and method for gesture based control system |
US20060239670A1 (en) | 2005-04-04 | 2006-10-26 | Dixon Cleveland | Explicit raytracing for gimbal-based gazepoint trackers |
US20060248475A1 (en) | 2002-09-09 | 2006-11-02 | Thomas Abrahamsson | Graphical user interface system |
US7151530B2 (en) | 2002-08-20 | 2006-12-19 | Canesta, Inc. | System and method for determining an input selected by a user through a virtual interface |
US7170492B2 (en) | 2002-05-28 | 2007-01-30 | Reactrix Systems, Inc. | Interactive video display system |
US20070060336A1 (en) | 2003-09-15 | 2007-03-15 | Sony Computer Entertainment Inc. | Methods and systems for enabling depth and direction detection when interfacing with a computer program |
US20070078552A1 (en) | 2006-01-13 | 2007-04-05 | Outland Research, Llc | Gaze-based power conservation for portable media players |
WO2007043036A1 (en) | 2005-10-11 | 2007-04-19 | Prime Sense Ltd. | Method and system for object reconstruction |
US7227526B2 (en) * | 2000-07-24 | 2007-06-05 | Gesturetek, Inc. | Video-based image control system |
US20070130547A1 (en) | 2005-12-01 | 2007-06-07 | Navisense, Llc | Method and system for touchless user interface control |
US20070154116A1 (en) | 2005-12-30 | 2007-07-05 | Kelvin Shieh | Video-based handwriting input method and apparatus |
WO2007078639A1 (en) | 2005-12-12 | 2007-07-12 | Sony Computer Entertainment Inc. | Methods and systems for enabling direction detection when interfacing with a computer program |
US7257237B1 (en) | 2003-03-07 | 2007-08-14 | Sandia Corporation | Real time markerless motion tracking using linked kinematic chains |
US7259747B2 (en) | 2001-06-05 | 2007-08-21 | Reactrix Systems, Inc. | Interactive video display system |
US7264554B2 (en) | 2005-01-26 | 2007-09-04 | Bentley Kinetics, Inc. | Method and system for athletic motion analysis and instruction |
WO2007105205A2 (en) | 2006-03-14 | 2007-09-20 | Prime Sense Ltd. | Three-dimensional sensing using speckle patterns |
US20070230789A1 (en) | 2006-04-03 | 2007-10-04 | Inventec Appliances Corp. | Method of controlling an electronic device by handwriting |
US7289645B2 (en) * | 2002-10-25 | 2007-10-30 | Mitsubishi Fuso Truck And Bus Corporation | Hand pattern switch device |
US7289227B2 (en) | 2004-10-01 | 2007-10-30 | Nomos Corporation | System and tracker for tracking an object, and related methods |
US7295697B1 (en) | 1999-12-06 | 2007-11-13 | Canon Kabushiki Kaisha | Depth information measurement apparatus and mixed reality presentation system |
WO2007132451A2 (en) | 2006-05-11 | 2007-11-22 | Prime Sense Ltd. | Modeling of humanoid forms from depth maps |
US7301648B2 (en) | 2000-01-28 | 2007-11-27 | Intersense, Inc. | Self-referenced tracking |
US7302099B2 (en) | 2003-11-10 | 2007-11-27 | Microsoft Corporation | Stroke segmentation for template-based cursive handwriting recognition |
WO2007135376A2 (en) | 2006-05-24 | 2007-11-29 | Sony Computer Entertainment Europe Ltd | Control of data processing using cumulative inter-frame motion |
US20070285554A1 (en) | 2005-10-31 | 2007-12-13 | Dor Givon | Apparatus method and system for imaging |
US7333113B2 (en) | 2003-03-13 | 2008-02-19 | Sony Corporation | Mobile motion capture cameras |
US7340399B2 (en) * | 2000-01-27 | 2008-03-04 | Siemens Ag | System and method for eye tracking controlled speech processing |
US20080059915A1 (en) | 2006-09-05 | 2008-03-06 | Marc Boillot | Method and Apparatus for Touchless Control of a Device |
US20080062123A1 (en) | 2001-06-05 | 2008-03-13 | Reactrix Systems, Inc. | Interactive video display system using strobed light |
US7358972B2 (en) | 2003-05-01 | 2008-04-15 | Sony Corporation | System and method for capturing facial and body motion |
US20080094371A1 (en) | 2006-09-06 | 2008-04-24 | Scott Forstall | Deletion Gestures on a Portable Multifunction Device |
US7370883B2 (en) | 2002-06-03 | 2008-05-13 | Intelligent Mechatronic Systems, Inc. | Three dimensional occupant position sensor |
US20080123940A1 (en) | 2006-11-29 | 2008-05-29 | The Mitre Corporation | Cursive character handwriting recognition system and method |
US20080170776A1 (en) | 2007-01-12 | 2008-07-17 | Albertson Jacob C | Controlling resource access based on user gesturing in a 3d captured image stream of the user |
US7428542B1 (en) | 2005-05-31 | 2008-09-23 | Reactrix Systems, Inc. | Method and system for combining nodes into a mega-node |
US7427996B2 (en) | 2002-10-16 | 2008-09-23 | Canon Kabushiki Kaisha | Image processing apparatus and image processing method |
US20080236902A1 (en) | 2007-03-28 | 2008-10-02 | Oki Data Corporation | Handwriting input system |
WO2008120217A2 (en) | 2007-04-02 | 2008-10-09 | Prime Sense Ltd. | Depth mapping using projected patterns |
US20080256494A1 (en) | 2007-04-16 | 2008-10-16 | Greenfield Mfg Co Inc | Touchless hand gesture device controller |
US20080252596A1 (en) | 2007-04-10 | 2008-10-16 | Matthew Bell | Display Using a Three-Dimensional vision System |
US20080260250A1 (en) | 2001-04-09 | 2008-10-23 | I.C. + Technologies Ltd. | Apparatus and methods for hand motion detection and handwriting recognition generally |
US20080281583A1 (en) | 2007-05-07 | 2008-11-13 | Biap , Inc. | Context-dependent prediction and learning with a universal re-entrant predictive text input software component |
US20080287189A1 (en) | 2007-05-09 | 2008-11-20 | Nintendo Of America Inc. | System and method for using accelerometer outputs to control an object rotating on a display |
US7474256B2 (en) | 2003-08-21 | 2009-01-06 | Sharp Kabushiki Kaisha | Position detecting system, and transmitting and receiving apparatuses for the position detecting system |
US20090009593A1 (en) | 2006-11-29 | 2009-01-08 | F.Poszat Hu, Llc | Three dimensional projection display |
US20090027335A1 (en) | 2005-08-22 | 2009-01-29 | Qinzhong Ye | Free-Space Pointing and Handwriting |
US20090027337A1 (en) | 2007-07-27 | 2009-01-29 | Gesturetek, Inc. | Enhanced camera-based input |
US20090033623A1 (en) | 2007-08-01 | 2009-02-05 | Ming-Yen Lin | Three-dimensional virtual input and simulation apparatus |
US20090040215A1 (en) | 2007-08-10 | 2009-02-12 | Nitin Afzulpurkar | Interpreting Sign Language Gestures |
US20090077504A1 (en) | 2007-09-14 | 2009-03-19 | Matthew Bell | Processing of Gesture-Based User Interactions |
US20090073117A1 (en) | 2007-09-19 | 2009-03-19 | Shingo Tsurumi | Image Processing Apparatus and Method, and Program Therefor |
US7508377B2 (en) | 2004-03-05 | 2009-03-24 | Nokia Corporation | Control and a control arrangement |
US20090078473A1 (en) | 2007-09-26 | 2009-03-26 | Digital Pen Systems | Handwriting Capture For Determining Absolute Position Within A Form Layout Using Pen Position Triangulation |
US20090083622A1 (en) | 2007-09-26 | 2009-03-26 | Inventec Appliances Corp. | Handwriting recording apparatus |
US20090083122A1 (en) | 2007-09-26 | 2009-03-26 | Robert Lee Angell | Method and apparatus for identifying customer behavioral types from a continuous video stream for use in optimizing loss leader merchandizing |
US7526120B2 (en) | 2002-09-11 | 2009-04-28 | Canesta, Inc. | System and method for providing intelligent airbag deployment |
US20090183125A1 (en) | 2008-01-14 | 2009-07-16 | Prime Sense Ltd. | Three-dimensional user interface |
US20090195392A1 (en) | 2008-01-31 | 2009-08-06 | Gary Zalewski | Laugh detector and system and method for tracking an emotional response to a media presentation |
US7573480B2 (en) | 2003-05-01 | 2009-08-11 | Sony Corporation | System and method for capturing facial and body motion |
US20090228841A1 (en) | 2008-03-04 | 2009-09-10 | Gesture Tek, Inc. | Enhanced Gesture-Based Image Manipulation |
US7590941B2 (en) | 2003-10-09 | 2009-09-15 | Hewlett-Packard Development Company, L.P. | Communication and collaboration system using rich media environments |
US20090256817A1 (en) | 2008-02-28 | 2009-10-15 | New York University | Method and apparatus for providing input to a processor, and a sensor pad |
US20090284542A1 (en) * | 2001-06-12 | 2009-11-19 | Noregin Assets N.V., L.L.C. | Lens-defined adjustment of displays |
US20090297028A1 (en) | 2008-05-30 | 2009-12-03 | De Haan Ido Gert | Method and device for handwriting detection |
US20100002936A1 (en) | 2003-09-26 | 2010-01-07 | Khomo Malome T | Spatial character recognition technique and chirographic text character reader |
US20100007717A1 (en) | 2008-07-09 | 2010-01-14 | Prime Sense Ltd | Integrated processor for 3d mapping |
US20100036717A1 (en) | 2004-12-29 | 2010-02-11 | Bernard Trest | Dynamic Information System |
US20100053151A1 (en) | 2008-09-02 | 2010-03-04 | Samsung Electronics Co., Ltd | In-line mediation for manipulating three-dimensional content on a display device |
US20100053304A1 (en) | 2006-02-08 | 2010-03-04 | Oblong Industries, Inc. | Control System for Navigating a Principal Dimension of a Data Space |
US20100071965A1 (en) | 2008-09-23 | 2010-03-25 | Panasonic Corporation | System and method for grab and drop gesture recognition |
US7688998B2 (en) | 2004-02-11 | 2010-03-30 | Brainlab Ag | Adjustable marker arrangement |
US20100083189A1 (en) * | 2008-09-30 | 2010-04-01 | Robert Michael Arlein | Method and apparatus for spatial context based coordination of information among multiple devices |
US7696876B2 (en) | 2001-12-20 | 2010-04-13 | Calypso Medical Technologies, Inc. | System for spatially adjustable excitation of leadless miniature marker |
US20100103106A1 (en) | 2007-07-11 | 2010-04-29 | Hsien-Hsiang Chui | Intelligent robotic interface input device |
US7724250B2 (en) | 2002-12-19 | 2010-05-25 | Sony Corporation | Apparatus, method, and program for processing information |
US20100149096A1 (en) | 2008-12-17 | 2010-06-17 | Migos Charles J | Network management using interaction with display surface |
US20100164897A1 (en) | 2007-06-28 | 2010-07-01 | Panasonic Corporation | Virtual keypad systems and methods |
US20100177933A1 (en) | 2005-08-19 | 2010-07-15 | Koninklijke Philips Electronics N.V. | System and method of analyzing the movement of a user |
US7762665B2 (en) | 2003-03-21 | 2010-07-27 | Queen's University At Kingston | Method and apparatus for communication between humans and devices |
US20100199228A1 (en) | 2009-01-30 | 2010-08-05 | Microsoft Corporation | Gesture Keyboarding |
US20100199231A1 (en) * | 2009-01-30 | 2010-08-05 | Microsoft Corporation | Predictive determination |
US7774155B2 (en) | 2006-03-10 | 2010-08-10 | Nintendo Co., Ltd. | Accelerometer-based controller |
US20100229125A1 (en) | 2009-03-09 | 2010-09-09 | Samsung Electronics Co., Ltd. | Display apparatus for providing a user menu, and method for providing user interface (ui) applicable thereto |
US20100234094A1 (en) * | 2007-11-09 | 2010-09-16 | Wms Gaming Inc. | Interaction with 3d space in a gaming system |
US20100235034A1 (en) * | 2009-03-16 | 2010-09-16 | The Boeing Company | Method, Apparatus And Computer Program Product For Recognizing A Gesture |
US20100235786A1 (en) | 2009-03-13 | 2010-09-16 | Primesense Ltd. | Enhanced 3d interfacing for remote devices |
US7821541B2 (en) | 2002-04-05 | 2010-10-26 | Bruno Delean | Remote control apparatus using gesture recognition |
US7834847B2 (en) | 2005-12-01 | 2010-11-16 | Navisense | Method and system for activating a touchless control |
US7840031B2 (en) | 2007-01-12 | 2010-11-23 | International Business Machines Corporation | Tracking a range of body movement based on 3D captured image streams of a user |
US20100295781A1 (en) | 2009-05-22 | 2010-11-25 | Rachid Alameh | Electronic Device with Sensing Assembly and Method for Interpreting Consecutive Gestures |
US7844914B2 (en) | 2004-07-30 | 2010-11-30 | Apple Inc. | Activating virtual keys of a touch-screen virtual keyboard |
US20110006978A1 (en) | 2009-07-10 | 2011-01-13 | Yuan Xiaoru | Image manipulation based on tracked eye movement |
US20110007035A1 (en) | 2007-08-19 | 2011-01-13 | Saar Shai | Finger-worn devices and related methods of use |
US20110018795A1 (en) | 2009-07-27 | 2011-01-27 | Samsung Electronics Co., Ltd. | Method and apparatus for controlling electronic device using user interaction |
US20110029918A1 (en) | 2009-07-29 | 2011-02-03 | Samsung Electronics Co., Ltd. | Apparatus and method for navigation in digital object using gaze information of user |
US20110052006A1 (en) | 2009-08-13 | 2011-03-03 | Primesense Ltd. | Extraction of skeletons from 3d maps |
US20110081072A1 (en) | 2008-06-13 | 2011-04-07 | Techno Dream 21 Co., Ltd. | Image processing device, image processing method, and program |
US7925549B2 (en) | 2004-09-17 | 2011-04-12 | Accenture Global Services Limited | Personalized marketing architecture |
US20110087970A1 (en) | 2009-10-14 | 2011-04-14 | At&T Mobility Ii Llc | Systems, apparatus, methods and computer-readable storage media for facilitating integrated messaging, contacts and social media for a selected entity |
US20110144543A1 (en) * | 2009-05-27 | 2011-06-16 | Takashi Tsuzuki | Behavior recognition apparatus |
US20110141053A1 (en) | 2009-12-14 | 2011-06-16 | Synaptics Incorporated | System and method for measuring individual force in multi-object sensing |
US20110164141A1 (en) | 2008-07-21 | 2011-07-07 | Marius Tico | Electronic Device Directional Audio-Video Capture |
US20110164032A1 (en) * | 2010-01-07 | 2011-07-07 | Prime Sense Ltd. | Three-Dimensional User Interface |
US20110193939A1 (en) | 2010-02-09 | 2011-08-11 | Microsoft Corporation | Physical interaction zone for gesture-based user interfaces |
US20110205186A1 (en) | 2009-12-04 | 2011-08-25 | John David Newton | Imaging Methods and Systems for Position Detection |
US20110211754A1 (en) | 2010-03-01 | 2011-09-01 | Primesense Ltd. | Tracking body parts by combined color image and depth processing |
US8018579B1 (en) | 2005-10-21 | 2011-09-13 | Apple Inc. | Three-dimensional imaging and display system |
US20110225536A1 (en) | 2010-02-28 | 2011-09-15 | Osterhout Group, Inc. | Sliding keyboard input control in an augmented reality eyepiece |
US20110242102A1 (en) * | 2010-03-30 | 2011-10-06 | Harman Becker Automotive Systems Gmbh | Vehicle user interface unit for a vehicle electronic device |
US20110248914A1 (en) | 2010-04-11 | 2011-10-13 | Sherr Alan B | System and Method for Virtual Touch Typing |
US20110254765A1 (en) | 2010-04-18 | 2011-10-20 | Primesense Ltd. | Remote text input using handwriting |
US20110254798A1 (en) | 2009-12-18 | 2011-10-20 | Peter S Adamson | Techniques for recognizing a series of touches with varying intensity or angle of descending on a touch panel interface |
US20110261058A1 (en) | 2010-04-23 | 2011-10-27 | Tong Luo | Method for user input from the back panel of a handheld computerized device |
US20110260965A1 (en) | 2010-04-22 | 2011-10-27 | Electronics And Telecommunications Research Institute | Apparatus and method of user interface for manipulating multimedia contents in vehicle |
US20110279397A1 (en) | 2009-01-26 | 2011-11-17 | Zrro Technologies (2009) Ltd. | Device and method for monitoring the object's behavior |
US20110291926A1 (en) | 2002-02-15 | 2011-12-01 | Canesta, Inc. | Gesture recognition system using depth perceptive sensors |
US20110296353A1 (en) * | 2009-05-29 | 2011-12-01 | Canesta, Inc. | Method and system implementing user-centric gesture control |
US20110293137A1 (en) | 2010-05-31 | 2011-12-01 | Primesense Ltd. | Analysis of three-dimensional scenes |
US20110292036A1 (en) | 2010-05-31 | 2011-12-01 | Primesense Ltd. | Depth sensor with application interface |
US20110310010A1 (en) | 2010-06-17 | 2011-12-22 | Primesense Ltd. | Gesture based user interface |
US20120001875A1 (en) | 2010-06-29 | 2012-01-05 | Qualcomm Incorporated | Touchless sensing and gesture recognition using continuous wave ultrasound signals |
WO2012011044A1 (en) | 2010-07-20 | 2012-01-26 | Primesense Ltd. | Interactive reality augmentation for natural interaction |
US20120035934A1 (en) | 2010-08-06 | 2012-02-09 | Dynavox Systems Llc | Speech generation device with a projected display and optical inputs |
US20120038550A1 (en) * | 2010-08-13 | 2012-02-16 | Net Power And Light, Inc. | System architecture and methods for distributed multi-sensor gesture processing |
WO2012020380A1 (en) | 2010-08-11 | 2012-02-16 | Primesense Ltd. | Scanning projectors and image capture modules for 3d mapping |
US20120078614A1 (en) | 2010-09-27 | 2012-03-29 | Primesense Ltd. | Virtual keyboard for a non-tactile three dimensional user interface |
US8154781B2 (en) | 2006-10-26 | 2012-04-10 | Seereal Technologies S.A. | Compact holographic display device |
US20120117514A1 (en) * | 2010-11-04 | 2012-05-10 | Microsoft Corporation | Three-Dimensional User Interaction |
US20120124516A1 (en) * | 2010-11-12 | 2012-05-17 | At&T Intellectual Property I, L.P. | Electronic Device Control Based on Gestures |
US8183977B2 (en) * | 2009-02-27 | 2012-05-22 | Seiko Epson Corporation | System of controlling device in response to gesture |
US8194921B2 (en) * | 2008-06-27 | 2012-06-05 | Nokia Corporation | Method, appartaus and computer program product for providing gesture analysis |
US8214098B2 (en) * | 2008-02-28 | 2012-07-03 | The Boeing Company | System and method for controlling swarm of remote unmanned vehicles through human gestures |
US20120169583A1 (en) | 2011-01-05 | 2012-07-05 | Primesense Ltd. | Scene profiles for non-tactile user interfaces |
US20120173067A1 (en) * | 2010-12-30 | 2012-07-05 | GM Global Technology Operations LLC | Graphical vehicle command system for autonomous vehicles on full windshield head-up display |
US8218211B2 (en) | 2007-05-16 | 2012-07-10 | Seereal Technologies S.A. | Holographic display with a variable beam deflection |
US20120200494A1 (en) | 2009-10-13 | 2012-08-09 | Haim Perski | Computer vision gesture based control of a device |
US20120202569A1 (en) | 2009-01-13 | 2012-08-09 | Primesense Ltd. | Three-Dimensional User Interface for Game Applications |
US20120204133A1 (en) | 2009-01-13 | 2012-08-09 | Primesense Ltd. | Gesture-Based User Interface |
WO2012107892A2 (en) | 2011-02-09 | 2012-08-16 | Primesense Ltd. | Gaze detection in a 3d mapping environment |
US20120223882A1 (en) | 2010-12-08 | 2012-09-06 | Primesense Ltd. | Three Dimensional User Interface Cursor Control |
US20120249416A1 (en) * | 2011-03-29 | 2012-10-04 | Giuliano Maciocci | Modular mobile connected pico projectors for a local multi-user collaboration |
US20120268369A1 (en) * | 2011-04-19 | 2012-10-25 | Microsoft Corporation | Depth Camera-Based Relative Gesture Detection |
US20120275680A1 (en) | 2008-02-12 | 2012-11-01 | Canon Kabushiki Kaisha | X-ray image processing apparatus, x-ray image processing method, program, and storage medium |
US20120295661A1 (en) | 2011-05-16 | 2012-11-22 | Yongsin Kim | Electronic device |
US20120309535A1 (en) | 2011-05-31 | 2012-12-06 | Microsoft Corporation | Action selection gesturing |
US20120313848A1 (en) | 2010-12-13 | 2012-12-13 | Primesense Ltd. | Three Dimensional User Interface Session Control |
US20120320080A1 (en) | 2011-06-14 | 2012-12-20 | Microsoft Corporation | Motion based virtual object navigation |
US20130002801A1 (en) * | 2011-06-28 | 2013-01-03 | Mock Wayne E | Adjusting Volume of a Videoconference Using Touch-Based Gestures |
US20130014052A1 (en) | 2011-07-05 | 2013-01-10 | Primesense Ltd. | Zoom-based gesture user interface |
US20130044053A1 (en) | 2011-08-15 | 2013-02-21 | Primesense Ltd. | Combining Explicit Select Gestures And Timeclick In A Non-Tactile Three Dimensional User Interface |
US20130055150A1 (en) | 2011-08-24 | 2013-02-28 | Primesense Ltd. | Visual feedback for tactile and non-tactile user interfaces |
US20130055120A1 (en) | 2011-08-24 | 2013-02-28 | Primesense Ltd. | Sessionless pointing user interface |
US20130058565A1 (en) | 2002-02-15 | 2013-03-07 | Microsoft Corporation | Gesture recognition system using depth perceptive sensors |
US8396252B2 (en) | 2010-05-20 | 2013-03-12 | Edge 3 Technologies | Systems and related methods for three dimensional gesture recognition in vehicles |
US8416276B2 (en) | 2006-10-26 | 2013-04-09 | Seereal Technologies S.A. | Mobile telephony system comprising holographic display |
US20130106692A1 (en) | 2010-07-20 | 2013-05-02 | Primesense Ltd. | Adaptive Projector |
US8448083B1 (en) | 2004-04-16 | 2013-05-21 | Apple Inc. | Gesture control of multimedia editing applications |
US8446459B2 (en) | 2008-06-17 | 2013-05-21 | Huawei Device Co., Ltd. | Video communication method, device, and system |
US20130155070A1 (en) | 2010-04-23 | 2013-06-20 | Tong Luo | Method for user input from alternative touchpads of a handheld computerized device |
US20130207920A1 (en) | 2010-08-20 | 2013-08-15 | Eric McCann | Hand and finger registration for control applications |
US8514221B2 (en) | 2010-01-05 | 2013-08-20 | Apple Inc. | Working with 3D objects |
US8514251B2 (en) * | 2008-06-23 | 2013-08-20 | Qualcomm Incorporated | Enhanced character input using recognized gestures |
US8625882B2 (en) | 2010-05-31 | 2014-01-07 | Sony Corporation | User interface with three dimensional user input |
US20140108930A1 (en) | 2012-10-12 | 2014-04-17 | Sling Media Inc. | Methods and apparatus for three-dimensional graphical user interfaces |
US20140237432A1 (en) | 2011-09-15 | 2014-08-21 | Koninklijke Philips Electronics N.V. | Gesture-based user-interface with user-feedback |
US8996173B2 (en) | 2010-09-21 | 2015-03-31 | Intuitive Surgical Operations, Inc. | Method and apparatus for hand gesture control in a minimally invasive surgical system |
US9030529B2 (en) | 2011-04-14 | 2015-05-12 | Industrial Technology Research Institute | Depth image acquiring device, system and method |
US9075441B2 (en) | 2006-02-08 | 2015-07-07 | Oblong Industries, Inc. | Gesture based control using three-dimensional information extracted over an extended depth of field |
US9104271B1 (en) | 2011-06-03 | 2015-08-11 | Richard Adams | Gloved human-machine interface |
Family Cites Families (33)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
FR2806183B1 (en) | 1999-12-01 | 2006-09-01 | Cartesis S A | DEVICE AND METHOD FOR INSTANT CONSOLIDATION, ENRICHMENT AND "REPORTING" OR BACKGROUND OF INFORMATION IN A MULTIDIMENSIONAL DATABASE |
US6853375B2 (en) | 2000-05-10 | 2005-02-08 | Cognos Incorporated | Method for preemptive screen rendering |
US7263668B1 (en) | 2000-11-09 | 2007-08-28 | International Business Machines Corporation | Display interface to a computer controlled display system with variable comprehensiveness levels of menu items dependent upon size of variable display screen available for menu item display |
WO2003032125A2 (en) | 2001-10-11 | 2003-04-17 | Visualsciences, Llc | System, method, and computer program product for processing and visualization of information |
US8905834B2 (en) | 2007-11-09 | 2014-12-09 | Igt | Transparent card display |
EP1987484A4 (en) * | 2006-02-16 | 2011-05-25 | Hillcrest Lab Inc | Systems and methods for placing advertisements |
WO2007096893A2 (en) | 2006-02-27 | 2007-08-30 | Prime Sense Ltd. | Range mapping using speckle decorrelation |
US8139029B2 (en) | 2006-03-08 | 2012-03-20 | Navisense | Method and device for three-dimensional sensing |
US8583772B2 (en) | 2008-08-14 | 2013-11-12 | International Business Machines Corporation | Dynamically configurable session agent |
US9058307B2 (en) | 2007-01-26 | 2015-06-16 | Microsoft Technology Licensing, Llc | Presentation generation using scorecard elements |
WO2009042579A1 (en) | 2007-09-24 | 2009-04-02 | Gesturetek, Inc. | Enhanced interface for voice and video communications |
US8223121B2 (en) | 2008-10-20 | 2012-07-17 | Sensor Platforms, Inc. | Host system and method for determining an attitude of a device undergoing dynamic acceleration |
US20160026265A1 (en) | 2008-10-20 | 2016-01-28 | Kevin A. Shaw | System and Method for Determining an Attitude of a Device Undergoing Dynamic Acceleration |
JP5361349B2 (en) * | 2008-11-28 | 2013-12-04 | 任天堂株式会社 | Information processing apparatus, computer program, information processing system, and information processing method |
US8738118B2 (en) * | 2009-05-20 | 2014-05-27 | Sotera Wireless, Inc. | Cable system for generating signals for detecting motion and measuring vital signs |
US8487871B2 (en) * | 2009-06-01 | 2013-07-16 | Microsoft Corporation | Virtual desktop coordinate transformation |
US8977972B2 (en) * | 2009-12-31 | 2015-03-10 | Intel Corporation | Using multi-modal input to control multiple objects on a display |
US8756532B2 (en) * | 2010-01-21 | 2014-06-17 | Cisco Technology, Inc. | Using a gesture to transfer an object across multiple multi-touch devices |
US20110289455A1 (en) * | 2010-05-18 | 2011-11-24 | Microsoft Corporation | Gestures And Gesture Recognition For Manipulating A User-Interface |
US8856688B2 (en) | 2010-10-11 | 2014-10-07 | Facebook, Inc. | Pinch gesture to navigate application layers |
US20120159395A1 (en) | 2010-12-20 | 2012-06-21 | Microsoft Corporation | Application-launching interface for multiple modes |
US8689123B2 (en) | 2010-12-23 | 2014-04-01 | Microsoft Corporation | Application reporting in an application-selectable user interface |
US8612874B2 (en) | 2010-12-23 | 2013-12-17 | Microsoft Corporation | Presenting an application change through a tile |
US9519418B2 (en) | 2011-01-18 | 2016-12-13 | Nokia Technologies Oy | Method and apparatus for providing a multi-stage device transition mechanism initiated based on a touch gesture |
US9104239B2 (en) * | 2011-03-09 | 2015-08-11 | Lg Electronics Inc. | Display device and method for controlling gesture functions using different depth ranges |
US8840466B2 (en) | 2011-04-25 | 2014-09-23 | Aquifi, Inc. | Method and system to create three-dimensional mapping in a two-dimensional game |
US9104307B2 (en) | 2011-05-27 | 2015-08-11 | Microsoft Technology Licensing, Llc | Multi-application environment |
US9158445B2 (en) | 2011-05-27 | 2015-10-13 | Microsoft Technology Licensing, Llc | Managing an immersive interface in a multi-application immersive environment |
US9104440B2 (en) | 2011-05-27 | 2015-08-11 | Microsoft Technology Licensing, Llc | Multi-application environment |
US20130057587A1 (en) | 2011-09-01 | 2013-03-07 | Microsoft Corporation | Arranging tiles |
US8854433B1 (en) | 2012-02-03 | 2014-10-07 | Aquifi, Inc. | Method and system enabling natural user interface gestures with an electronic system |
US9075460B2 (en) | 2012-08-10 | 2015-07-07 | Blackberry Limited | Method of momentum based zoom of content on an electronic device |
US9619105B1 (en) | 2014-01-30 | 2017-04-11 | Aquifi, Inc. | Systems and methods for gesture based interaction with viewpoint dependent user interfaces |
-
2012
- 2012-08-23 US US13/592,352 patent/US9218063B2/en active Active
-
2015
- 2015-10-22 US US14/919,751 patent/US10088909B2/en active Active
-
2018
- 2018-08-21 US US16/106,028 patent/US10642371B2/en active Active
Patent Citations (284)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4550250A (en) | 1983-11-14 | 1985-10-29 | Hei, Inc. | Cordless digital graphics input device |
US4789921A (en) | 1987-02-20 | 1988-12-06 | Minnesota Mining And Manufacturing Company | Cone shaped Fresnel reflector |
US4988981B1 (en) | 1987-03-17 | 1999-05-18 | Vpl Newco Inc | Computer data entry and manipulation apparatus and method |
US4988981A (en) | 1987-03-17 | 1991-01-29 | Vpl Research, Inc. | Computer data entry and manipulation apparatus and method |
US5588139A (en) | 1990-06-07 | 1996-12-24 | Vpl Research, Inc. | Method and system for generating objects for a multi-person virtual world using data flow networks |
US5264836A (en) * | 1991-01-15 | 1993-11-23 | Apple Computer, Inc. | Three dimensional cursor |
US5973700A (en) * | 1992-09-16 | 1999-10-26 | Eastman Kodak Company | Method and apparatus for optimizing the resolution of images which have an apparent depth |
US5495576A (en) | 1993-01-11 | 1996-02-27 | Ritchey; Kurtis J. | Panoramic image based virtual reality/telepresence audio-visual system and method |
US5594469A (en) | 1995-02-21 | 1997-01-14 | Mitsubishi Electric Information Technology Center America Inc. | Hand gesture machine control system |
US6191773B1 (en) | 1995-04-28 | 2001-02-20 | Matsushita Electric Industrial Co., Ltd. | Interface apparatus |
US6690370B2 (en) | 1995-06-07 | 2004-02-10 | Geovector Corp. | Vision system computer modeling apparatus including interaction with real scenes with respect to perspective and spatial relationship as measured in real-time |
US5852672A (en) | 1995-07-10 | 1998-12-22 | The Regents Of The University Of California | Image system for three dimensional, 360 DEGREE, time sequence surface mapping of moving objects |
US5846134A (en) | 1995-07-14 | 1998-12-08 | Latypov; Nurakhmed Nurislamovich | Method and apparatus for immersion of a user into virtual reality |
US6111580A (en) * | 1995-09-13 | 2000-08-29 | Kabushiki Kaisha Toshiba | Apparatus and method for controlling an electronic device with user action |
US5870196A (en) | 1995-10-16 | 1999-02-09 | European Community | Optical three-dimensional profilometry method based on processing SPECKLE images in partially coherent light, and interferometer implementing such a method |
US5862256A (en) | 1996-06-14 | 1999-01-19 | International Business Machines Corporation | Distinguishing gestures from handwriting in a pen based computer by size discrimination |
US5864635A (en) | 1996-06-14 | 1999-01-26 | International Business Machines Corporation | Distinguishing gestures from handwriting in a pen based computer by stroke analysis |
US6084979A (en) | 1996-06-20 | 2000-07-04 | Carnegie Mellon University | Method for creating virtual reality |
US6002808A (en) | 1996-07-26 | 1999-12-14 | Mitsubishi Electric Information Technology Center America, Inc. | Hand gesture control system |
US6005548A (en) | 1996-08-14 | 1999-12-21 | Latypov; Nurakhmed Nurislamovich | Method for tracking and displaying user's spatial position and orientation, a method for representing virtual reality for a user, and systems of embodiment of such methods |
US6345111B1 (en) | 1997-02-28 | 2002-02-05 | Kabushiki Kaisha Toshiba | Multi-modal interface apparatus and method |
US5917937A (en) | 1997-04-15 | 1999-06-29 | Microsoft Corporation | Method for performing stereo matching to recover depths, colors and opacities of surface elements |
US6452584B1 (en) | 1997-04-23 | 2002-09-17 | Modern Cartoon, Ltd. | System for data management based on hand gestures |
US6262740B1 (en) | 1997-08-01 | 2001-07-17 | Terarecon, Inc. | Method for rendering sections of a volume data set |
US7042440B2 (en) | 1997-08-22 | 2006-05-09 | Pryor Timothy R | Man machine interfaces and applications |
US8405604B2 (en) | 1997-08-22 | 2013-03-26 | Motion Games, Llc | Advanced video gaming methods for education and play using camera based inputs |
US6215890B1 (en) * | 1997-09-26 | 2001-04-10 | Matsushita Electric Industrial Co., Ltd. | Hand gesture recognizing device |
US6072494A (en) | 1997-10-15 | 2000-06-06 | Electric Planet, Inc. | Method and apparatus for real-time gesture recognition |
US6256033B1 (en) | 1997-10-15 | 2001-07-03 | Electric Planet | Method and apparatus for real-time gesture recognition |
WO1999035633A2 (en) | 1998-01-06 | 1999-07-15 | The Video Mouse Group | Human motion following computer mouse and game controller |
US6064387A (en) | 1998-01-23 | 2000-05-16 | Dell, Usa, L.P. | Animated cursor and icon for computers |
US6345893B2 (en) * | 1998-06-15 | 2002-02-12 | Vega Vista, Inc. | Ergonomic systems and methods for operating computers |
US6559813B1 (en) | 1998-07-01 | 2003-05-06 | Deluca Michael | Selective real image obstruction in a virtual reality display apparatus and method |
US6243054B1 (en) | 1998-07-01 | 2001-06-05 | Deluca Michael | Stereoscopic user interface method and apparatus |
US6252988B1 (en) | 1998-07-09 | 2001-06-26 | Lucent Technologies Inc. | Method and apparatus for character recognition using stop words |
US6681031B2 (en) | 1998-08-10 | 2004-01-20 | Cybernet Systems Corporation | Gesture-controlled interfaces for self-service machines and other applications |
US20020057383A1 (en) | 1998-10-13 | 2002-05-16 | Ryuichi Iwamura | Motion sensing interface |
US6519363B1 (en) | 1999-01-13 | 2003-02-11 | International Business Machines Corporation | Method and system for automatically segmenting and recognizing handwritten Chinese characters |
US20050265583A1 (en) | 1999-03-08 | 2005-12-01 | Vulcan Patents Llc | Three dimensional object pose estimation which employs dense depth information |
US7003134B1 (en) | 1999-03-08 | 2006-02-21 | Vulcan Patents Llc | Three dimensional object pose estimation which employs dense depth information |
US6951515B2 (en) | 1999-06-11 | 2005-10-04 | Canon Kabushiki Kaisha | Game apparatus for mixed reality space, image processing method thereof, and program storage medium |
US6791540B1 (en) | 1999-06-11 | 2004-09-14 | Canon Kabushiki Kaisha | Image processing apparatus |
US20030057972A1 (en) | 1999-07-26 | 2003-03-27 | Paul Pfaff | Voltage testing and measurement |
US6229541B1 (en) | 1999-09-03 | 2001-05-08 | Isurftv | Use of templates for cost-effective secure linking of video stream objects |
US20030063775A1 (en) | 1999-09-22 | 2003-04-03 | Canesta, Inc. | Methods for enhancing performance and data acquired from three-dimensional image systems |
US6512838B1 (en) | 1999-09-22 | 2003-01-28 | Canesta, Inc. | Methods for enhancing performance and data acquired from three-dimensional image systems |
US20030088463A1 (en) | 1999-10-21 | 2003-05-08 | Steven Fischman | System and method for group advertisement optimization |
US20040046744A1 (en) | 1999-11-04 | 2004-03-11 | Canesta, Inc. | Method and apparatus for entering data using a virtual input device |
US7295697B1 (en) | 1999-12-06 | 2007-11-13 | Canon Kabushiki Kaisha | Depth information measurement apparatus and mixed reality presentation system |
US6507353B1 (en) | 1999-12-10 | 2003-01-14 | Godot Huard | Influencing virtual actors in an interactive environment |
US7340399B2 (en) * | 2000-01-27 | 2008-03-04 | Siemens Ag | System and method for eye tracking controlled speech processing |
US7301648B2 (en) | 2000-01-28 | 2007-11-27 | Intersense, Inc. | Self-referenced tracking |
US7023436B2 (en) | 2000-04-19 | 2006-04-04 | Sony Corporation | Three-dimensional model processing device, three-dimensional model processing method, program providing medium |
US6456262B1 (en) | 2000-05-09 | 2002-09-24 | Intel Corporation | Microdisplay with eye gaze detection |
US6803928B2 (en) | 2000-06-06 | 2004-10-12 | Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. | Extended virtual table: an optical extension for table-like projection systems |
US7042442B1 (en) | 2000-06-27 | 2006-05-09 | International Business Machines Corporation | Virtual invisible keyboard |
US20080030460A1 (en) | 2000-07-24 | 2008-02-07 | Gesturetek, Inc. | Video-based image control system |
US7227526B2 (en) * | 2000-07-24 | 2007-06-05 | Gesturetek, Inc. | Video-based image control system |
US6686921B1 (en) | 2000-08-01 | 2004-02-03 | International Business Machines Corporation | Method and apparatus for acquiring a set of consistent image maps to represent the color of the surface of an object |
US20020071607A1 (en) | 2000-10-31 | 2002-06-13 | Akinori Kawamura | Apparatus, method, and program for handwriting recognition |
US7013046B2 (en) | 2000-10-31 | 2006-03-14 | Kabushiki Kaisha Toshiba | Apparatus, method, and program for handwriting recognition |
US20060115155A1 (en) | 2000-11-10 | 2006-06-01 | Microsoft Corporation | Implicit page breaks for digitally represented handwriting |
US6853935B2 (en) | 2000-11-30 | 2005-02-08 | Canon Kabushiki Kaisha | Information processing apparatus, mixed reality presentation apparatus, method thereof, and storage medium |
US20040104935A1 (en) | 2001-01-26 | 2004-06-03 | Todd Williamson | Virtual reality immersion system |
US20020158873A1 (en) | 2001-01-26 | 2002-10-31 | Todd Williamson | Real-time virtual viewpoint in simulated reality environment |
US20080260250A1 (en) | 2001-04-09 | 2008-10-23 | I.C. + Technologies Ltd. | Apparatus and methods for hand motion detection and handwriting recognition generally |
US20080062123A1 (en) | 2001-06-05 | 2008-03-13 | Reactrix Systems, Inc. | Interactive video display system using strobed light |
US7259747B2 (en) | 2001-06-05 | 2007-08-21 | Reactrix Systems, Inc. | Interactive video display system |
US20090284542A1 (en) * | 2001-06-12 | 2009-11-19 | Noregin Assets N.V., L.L.C. | Lens-defined adjustment of displays |
US20040135744A1 (en) | 2001-08-10 | 2004-07-15 | Oliver Bimber | Virtual showcases |
US6741251B2 (en) | 2001-08-16 | 2004-05-25 | Hewlett-Packard Development Company, L.P. | Method and apparatus for varying focus in a scene |
US7696876B2 (en) | 2001-12-20 | 2010-04-13 | Calypso Medical Technologies, Inc. | System for spatially adjustable excitation of leadless miniature marker |
US7215815B2 (en) | 2002-01-10 | 2007-05-08 | Ricoh Company, Ltd. | Handwriting information processing apparatus, handwriting information processing method, and storage medium having program stored therein for handwriting information processing |
US20030185444A1 (en) | 2002-01-10 | 2003-10-02 | Tadashi Honda | Handwriting information processing apparatus, handwriting information processing method, and storage medium having program stored therein for handwriting information processing |
US20030156756A1 (en) | 2002-02-15 | 2003-08-21 | Gokturk Salih Burak | Gesture recognition system using depth perceptive sensors |
US20130058565A1 (en) | 2002-02-15 | 2013-03-07 | Microsoft Corporation | Gesture recognition system using depth perceptive sensors |
WO2003071410A2 (en) | 2002-02-15 | 2003-08-28 | Canesta, Inc. | Gesture recognition system using depth perceptive sensors |
US20110291926A1 (en) | 2002-02-15 | 2011-12-01 | Canesta, Inc. | Gesture recognition system using depth perceptive sensors |
US7340077B2 (en) | 2002-02-15 | 2008-03-04 | Canesta, Inc. | Gesture recognition system using depth perceptive sensors |
US7821541B2 (en) | 2002-04-05 | 2010-10-26 | Bruno Delean | Remote control apparatus using gesture recognition |
US20030227453A1 (en) | 2002-04-09 | 2003-12-11 | Klaus-Peter Beier | Method, system and computer program product for automatically creating an animated 3-D scenario from human position and path data |
US20030235341A1 (en) | 2002-04-11 | 2003-12-25 | Gokturk Salih Burak | Subject segmentation and tracking using 3D sensing technology for video compression in multimedia applications |
US20050122308A1 (en) | 2002-05-28 | 2005-06-09 | Matthew Bell | Self-contained interactive video display system |
US20050110964A1 (en) | 2002-05-28 | 2005-05-26 | Matthew Bell | Interactive video window display system |
US7170492B2 (en) | 2002-05-28 | 2007-01-30 | Reactrix Systems, Inc. | Interactive video display system |
US20060139314A1 (en) | 2002-05-28 | 2006-06-29 | Matthew Bell | Interactive video display system |
US20080150913A1 (en) | 2002-05-28 | 2008-06-26 | Matthew Bell | Computer vision based touch screen |
US20050162381A1 (en) | 2002-05-28 | 2005-07-28 | Matthew Bell | Self-contained interactive video display system |
US7348963B2 (en) | 2002-05-28 | 2008-03-25 | Reactrix Systems, Inc. | Interactive video display system |
US20080150890A1 (en) | 2002-05-28 | 2008-06-26 | Matthew Bell | Interactive Video Window |
US7370883B2 (en) | 2002-06-03 | 2008-05-13 | Intelligent Mechatronic Systems, Inc. | Three dimensional occupant position sensor |
US6857746B2 (en) | 2002-07-01 | 2005-02-22 | Io2 Technology, Llc | Method and system for free-space imaging display and interface |
US7151530B2 (en) | 2002-08-20 | 2006-12-19 | Canesta, Inc. | System and method for determining an input selected by a user through a virtual interface |
US20060248475A1 (en) | 2002-09-09 | 2006-11-02 | Thomas Abrahamsson | Graphical user interface system |
US7526120B2 (en) | 2002-09-11 | 2009-04-28 | Canesta, Inc. | System and method for providing intelligent airbag deployment |
US20060149737A1 (en) | 2002-09-28 | 2006-07-06 | Yonggang Du | Method for accessing and additional content with parental control |
US7427996B2 (en) | 2002-10-16 | 2008-09-23 | Canon Kabushiki Kaisha | Image processing apparatus and image processing method |
US7289645B2 (en) * | 2002-10-25 | 2007-10-30 | Mitsubishi Fuso Truck And Bus Corporation | Hand pattern switch device |
US6977654B2 (en) | 2002-10-30 | 2005-12-20 | Iviz, Inc. | Data visualization with animated speedometer dial charts |
US20040174770A1 (en) | 2002-11-27 | 2004-09-09 | Rees Frank L. | Gauss-Rees parametric ultrawideband system |
US7576727B2 (en) | 2002-12-13 | 2009-08-18 | Matthew Bell | Interactive directed light/sound system |
US20040183775A1 (en) | 2002-12-13 | 2004-09-23 | Reactrix Systems | Interactive directed light/sound system |
US7724250B2 (en) | 2002-12-19 | 2010-05-25 | Sony Corporation | Apparatus, method, and program for processing information |
US20060159344A1 (en) | 2002-12-26 | 2006-07-20 | Xiaoling Shao | Method and system for three-dimensional handwriting recognition |
US20040155962A1 (en) | 2003-02-11 | 2004-08-12 | Marks Richard L. | Method and apparatus for real time motion capture |
US7257237B1 (en) | 2003-03-07 | 2007-08-14 | Sandia Corporation | Real time markerless motion tracking using linked kinematic chains |
US7333113B2 (en) | 2003-03-13 | 2008-02-19 | Sony Corporation | Mobile motion capture cameras |
US7812842B2 (en) | 2003-03-13 | 2010-10-12 | Sony Corporation | Mobile motion capture cameras |
US7580572B2 (en) | 2003-03-17 | 2009-08-25 | Samsung Electronics Co., Ltd. | Spatial motion recognition system and method using a virtual handwriting plane |
US20040184659A1 (en) | 2003-03-17 | 2004-09-23 | Samsung Electronics Co., Ltd. | Handwriting trajectory recognition system and method |
US20040184640A1 (en) | 2003-03-17 | 2004-09-23 | Samsung Electronics Co., Ltd. | Spatial motion recognition system and method using a virtual handwriting plane |
US7762665B2 (en) | 2003-03-21 | 2010-07-27 | Queen's University At Kingston | Method and apparatus for communication between humans and devices |
US20040193413A1 (en) * | 2003-03-25 | 2004-09-30 | Wilson Andrew D. | Architecture for controlling a computer using hand gestures |
US7358972B2 (en) | 2003-05-01 | 2008-04-15 | Sony Corporation | System and method for capturing facial and body motion |
US7573480B2 (en) | 2003-05-01 | 2009-08-11 | Sony Corporation | System and method for capturing facial and body motion |
US20040222977A1 (en) | 2003-05-05 | 2004-11-11 | Bear Eric Gould | Notification lights, locations and rules for a computer system |
WO2004107272A1 (en) | 2003-05-29 | 2004-12-09 | Sony Computer Entertainment Inc. | System and method for providing a real-time three-dimensional interactive environment |
US20050031166A1 (en) | 2003-05-29 | 2005-02-10 | Kikuo Fujimura | Visual tracking using depth data |
US20040258314A1 (en) | 2003-06-23 | 2004-12-23 | Shoestring Research, Llc | Region segmentation and characterization systems and methods for augmented reality |
WO2005003948A1 (en) | 2003-07-08 | 2005-01-13 | Sony Computer Entertainment Inc. | Control system and control method |
US7474256B2 (en) | 2003-08-21 | 2009-01-06 | Sharp Kabushiki Kaisha | Position detecting system, and transmitting and receiving apparatuses for the position detecting system |
US20070060336A1 (en) | 2003-09-15 | 2007-03-15 | Sony Computer Entertainment Inc. | Methods and systems for enabling depth and direction detection when interfacing with a computer program |
US20100002936A1 (en) | 2003-09-26 | 2010-01-07 | Khomo Malome T | Spatial character recognition technique and chirographic text character reader |
US7590941B2 (en) | 2003-10-09 | 2009-09-15 | Hewlett-Packard Development Company, L.P. | Communication and collaboration system using rich media environments |
US7536032B2 (en) | 2003-10-24 | 2009-05-19 | Reactrix Systems, Inc. | Method and system for processing captured image information in an interactive video display system |
US20050089194A1 (en) | 2003-10-24 | 2005-04-28 | Matthew Bell | Method and system for processing captured image information in an interactive video display system |
US20050088407A1 (en) | 2003-10-24 | 2005-04-28 | Matthew Bell | Method and system for managing an interactive video display system |
US7302099B2 (en) | 2003-11-10 | 2007-11-27 | Microsoft Corporation | Stroke segmentation for template-based cursive handwriting recognition |
US20060110008A1 (en) * | 2003-11-14 | 2006-05-25 | Roel Vertegaal | Method and apparatus for calibration-free eye tracking |
US20050190972A1 (en) | 2004-02-11 | 2005-09-01 | Thomas Graham A. | System and method for position determination |
US7688998B2 (en) | 2004-02-11 | 2010-03-30 | Brainlab Ag | Adjustable marker arrangement |
US20050254726A1 (en) | 2004-02-25 | 2005-11-17 | The University Of North Carolina At Chapel Hill | Methods, systems, and computer program products for imperceptibly embedding structured light patterns in projected color images for display on planar and non-planar surfaces |
US7508377B2 (en) | 2004-03-05 | 2009-03-24 | Nokia Corporation | Control and a control arrangement |
WO2005094958A1 (en) | 2004-03-23 | 2005-10-13 | Harmonix Music Systems, Inc. | Method and apparatus for controlling a three-dimensional character in a three-dimensional gaming environment |
US8448083B1 (en) | 2004-04-16 | 2013-05-21 | Apple Inc. | Gesture control of multimedia editing applications |
US20060010400A1 (en) | 2004-06-28 | 2006-01-12 | Microsoft Corporation | Recognizing gestures and using gestures for interacting with software applications |
US20060020905A1 (en) | 2004-07-20 | 2006-01-26 | Hillcrest Communications, Inc. | Graphical cursor navigation methods |
US7844914B2 (en) | 2004-07-30 | 2010-11-30 | Apple Inc. | Activating virtual keys of a touch-screen virtual keyboard |
US7925549B2 (en) | 2004-09-17 | 2011-04-12 | Accenture Global Services Limited | Personalized marketing architecture |
US7289227B2 (en) | 2004-10-01 | 2007-10-30 | Nomos Corporation | System and tracker for tracking an object, and related methods |
US20060092138A1 (en) | 2004-10-29 | 2006-05-04 | Microsoft Corporation | Systems and methods for interacting with a computer through handwriting to a screen |
US20100036717A1 (en) | 2004-12-29 | 2010-02-11 | Bernard Trest | Dynamic Information System |
US7264554B2 (en) | 2005-01-26 | 2007-09-04 | Bentley Kinetics, Inc. | Method and system for athletic motion analysis and instruction |
US20060187196A1 (en) | 2005-02-08 | 2006-08-24 | Underkoffler John S | System and method for gesture based control system |
US20060239670A1 (en) | 2005-04-04 | 2006-10-26 | Dixon Cleveland | Explicit raytracing for gimbal-based gazepoint trackers |
US7428542B1 (en) | 2005-05-31 | 2008-09-23 | Reactrix Systems, Inc. | Method and system for combining nodes into a mega-node |
US20100177933A1 (en) | 2005-08-19 | 2010-07-15 | Koninklijke Philips Electronics N.V. | System and method of analyzing the movement of a user |
US20090027335A1 (en) | 2005-08-22 | 2009-01-29 | Qinzhong Ye | Free-Space Pointing and Handwriting |
WO2007043036A1 (en) | 2005-10-11 | 2007-04-19 | Prime Sense Ltd. | Method and system for object reconstruction |
US20090096783A1 (en) * | 2005-10-11 | 2009-04-16 | Alexander Shpunt | Three-dimensional sensing using speckle patterns |
US8018579B1 (en) | 2005-10-21 | 2011-09-13 | Apple Inc. | Three-dimensional imaging and display system |
US8462199B2 (en) | 2005-10-31 | 2013-06-11 | Extreme Reality Ltd. | Apparatus method and system for imaging |
US20070285554A1 (en) | 2005-10-31 | 2007-12-13 | Dor Givon | Apparatus method and system for imaging |
US20070130547A1 (en) | 2005-12-01 | 2007-06-07 | Navisense, Llc | Method and system for touchless user interface control |
US7834847B2 (en) | 2005-12-01 | 2010-11-16 | Navisense | Method and system for activating a touchless control |
WO2007078639A1 (en) | 2005-12-12 | 2007-07-12 | Sony Computer Entertainment Inc. | Methods and systems for enabling direction detection when interfacing with a computer program |
US20070154116A1 (en) | 2005-12-30 | 2007-07-05 | Kelvin Shieh | Video-based handwriting input method and apparatus |
US20070078552A1 (en) | 2006-01-13 | 2007-04-05 | Outland Research, Llc | Gaze-based power conservation for portable media players |
US9075441B2 (en) | 2006-02-08 | 2015-07-07 | Oblong Industries, Inc. | Gesture based control using three-dimensional information extracted over an extended depth of field |
US20100053304A1 (en) | 2006-02-08 | 2010-03-04 | Oblong Industries, Inc. | Control System for Navigating a Principal Dimension of a Data Space |
US7774155B2 (en) | 2006-03-10 | 2010-08-10 | Nintendo Co., Ltd. | Accelerometer-based controller |
WO2007105205A2 (en) | 2006-03-14 | 2007-09-20 | Prime Sense Ltd. | Three-dimensional sensing using speckle patterns |
US20070230789A1 (en) | 2006-04-03 | 2007-10-04 | Inventec Appliances Corp. | Method of controlling an electronic device by handwriting |
US20100034457A1 (en) | 2006-05-11 | 2010-02-11 | Tamir Berliner | Modeling of humanoid forms from depth maps |
WO2007132451A2 (en) | 2006-05-11 | 2007-11-22 | Prime Sense Ltd. | Modeling of humanoid forms from depth maps |
WO2007135376A2 (en) | 2006-05-24 | 2007-11-29 | Sony Computer Entertainment Europe Ltd | Control of data processing using cumulative inter-frame motion |
US20080059915A1 (en) | 2006-09-05 | 2008-03-06 | Marc Boillot | Method and Apparatus for Touchless Control of a Device |
US20080094371A1 (en) | 2006-09-06 | 2008-04-24 | Scott Forstall | Deletion Gestures on a Portable Multifunction Device |
US8154781B2 (en) | 2006-10-26 | 2012-04-10 | Seereal Technologies S.A. | Compact holographic display device |
US8416276B2 (en) | 2006-10-26 | 2013-04-09 | Seereal Technologies S.A. | Mobile telephony system comprising holographic display |
US20090009593A1 (en) | 2006-11-29 | 2009-01-08 | F.Poszat Hu, Llc | Three dimensional projection display |
US20080123940A1 (en) | 2006-11-29 | 2008-05-29 | The Mitre Corporation | Cursive character handwriting recognition system and method |
US7840031B2 (en) | 2007-01-12 | 2010-11-23 | International Business Machines Corporation | Tracking a range of body movement based on 3D captured image streams of a user |
US7971156B2 (en) | 2007-01-12 | 2011-06-28 | International Business Machines Corporation | Controlling resource access based on user gesturing in a 3D captured image stream of the user |
US20080170776A1 (en) | 2007-01-12 | 2008-07-17 | Albertson Jacob C | Controlling resource access based on user gesturing in a 3d captured image stream of the user |
US20080236902A1 (en) | 2007-03-28 | 2008-10-02 | Oki Data Corporation | Handwriting input system |
WO2008120217A2 (en) | 2007-04-02 | 2008-10-09 | Prime Sense Ltd. | Depth mapping using projected patterns |
US20080252596A1 (en) | 2007-04-10 | 2008-10-16 | Matthew Bell | Display Using a Three-Dimensional vision System |
US20080256494A1 (en) | 2007-04-16 | 2008-10-16 | Greenfield Mfg Co Inc | Touchless hand gesture device controller |
US20080281583A1 (en) | 2007-05-07 | 2008-11-13 | Biap , Inc. | Context-dependent prediction and learning with a universal re-entrant predictive text input software component |
US20080287189A1 (en) | 2007-05-09 | 2008-11-20 | Nintendo Of America Inc. | System and method for using accelerometer outputs to control an object rotating on a display |
US8218211B2 (en) | 2007-05-16 | 2012-07-10 | Seereal Technologies S.A. | Holographic display with a variable beam deflection |
US20100164897A1 (en) | 2007-06-28 | 2010-07-01 | Panasonic Corporation | Virtual keypad systems and methods |
US20100103106A1 (en) | 2007-07-11 | 2010-04-29 | Hsien-Hsiang Chui | Intelligent robotic interface input device |
US20090027337A1 (en) | 2007-07-27 | 2009-01-29 | Gesturetek, Inc. | Enhanced camera-based input |
US20090031240A1 (en) | 2007-07-27 | 2009-01-29 | Gesturetek, Inc. | Item selection using enhanced control |
US20090033623A1 (en) | 2007-08-01 | 2009-02-05 | Ming-Yen Lin | Three-dimensional virtual input and simulation apparatus |
US8368647B2 (en) | 2007-08-01 | 2013-02-05 | Ming-Yen Lin | Three-dimensional virtual input and simulation apparatus |
US20090040215A1 (en) | 2007-08-10 | 2009-02-12 | Nitin Afzulpurkar | Interpreting Sign Language Gestures |
US20110007035A1 (en) | 2007-08-19 | 2011-01-13 | Saar Shai | Finger-worn devices and related methods of use |
US20090077504A1 (en) | 2007-09-14 | 2009-03-19 | Matthew Bell | Processing of Gesture-Based User Interactions |
US20090073117A1 (en) | 2007-09-19 | 2009-03-19 | Shingo Tsurumi | Image Processing Apparatus and Method, and Program Therefor |
US20090083122A1 (en) | 2007-09-26 | 2009-03-26 | Robert Lee Angell | Method and apparatus for identifying customer behavioral types from a continuous video stream for use in optimizing loss leader merchandizing |
US20090083622A1 (en) | 2007-09-26 | 2009-03-26 | Inventec Appliances Corp. | Handwriting recording apparatus |
US20090078473A1 (en) | 2007-09-26 | 2009-03-26 | Digital Pen Systems | Handwriting Capture For Determining Absolute Position Within A Form Layout Using Pen Position Triangulation |
US20100234094A1 (en) * | 2007-11-09 | 2010-09-16 | Wms Gaming Inc. | Interaction with 3d space in a gaming system |
US8166421B2 (en) | 2008-01-14 | 2012-04-24 | Primesense Ltd. | Three-dimensional user interface |
US20090183125A1 (en) | 2008-01-14 | 2009-07-16 | Prime Sense Ltd. | Three-dimensional user interface |
US20090195392A1 (en) | 2008-01-31 | 2009-08-06 | Gary Zalewski | Laugh detector and system and method for tracking an emotional response to a media presentation |
US20120275680A1 (en) | 2008-02-12 | 2012-11-01 | Canon Kabushiki Kaisha | X-ray image processing apparatus, x-ray image processing method, program, and storage medium |
US20090256817A1 (en) | 2008-02-28 | 2009-10-15 | New York University | Method and apparatus for providing input to a processor, and a sensor pad |
US8214098B2 (en) * | 2008-02-28 | 2012-07-03 | The Boeing Company | System and method for controlling swarm of remote unmanned vehicles through human gestures |
US20090228841A1 (en) | 2008-03-04 | 2009-09-10 | Gesture Tek, Inc. | Enhanced Gesture-Based Image Manipulation |
US20090297028A1 (en) | 2008-05-30 | 2009-12-03 | De Haan Ido Gert | Method and device for handwriting detection |
US20110081072A1 (en) | 2008-06-13 | 2011-04-07 | Techno Dream 21 Co., Ltd. | Image processing device, image processing method, and program |
US8446459B2 (en) | 2008-06-17 | 2013-05-21 | Huawei Device Co., Ltd. | Video communication method, device, and system |
US8514251B2 (en) * | 2008-06-23 | 2013-08-20 | Qualcomm Incorporated | Enhanced character input using recognized gestures |
US8194921B2 (en) * | 2008-06-27 | 2012-06-05 | Nokia Corporation | Method, appartaus and computer program product for providing gesture analysis |
US20100007717A1 (en) | 2008-07-09 | 2010-01-14 | Prime Sense Ltd | Integrated processor for 3d mapping |
US20110164141A1 (en) | 2008-07-21 | 2011-07-07 | Marius Tico | Electronic Device Directional Audio-Video Capture |
US20100053151A1 (en) | 2008-09-02 | 2010-03-04 | Samsung Electronics Co., Ltd | In-line mediation for manipulating three-dimensional content on a display device |
US20100071965A1 (en) | 2008-09-23 | 2010-03-25 | Panasonic Corporation | System and method for grab and drop gesture recognition |
US20100083189A1 (en) * | 2008-09-30 | 2010-04-01 | Robert Michael Arlein | Method and apparatus for spatial context based coordination of information among multiple devices |
US20100149096A1 (en) | 2008-12-17 | 2010-06-17 | Migos Charles J | Network management using interaction with display surface |
US20120204133A1 (en) | 2009-01-13 | 2012-08-09 | Primesense Ltd. | Gesture-Based User Interface |
US20120202569A1 (en) | 2009-01-13 | 2012-08-09 | Primesense Ltd. | Three-Dimensional User Interface for Game Applications |
US20110279397A1 (en) | 2009-01-26 | 2011-11-17 | Zrro Technologies (2009) Ltd. | Device and method for monitoring the object's behavior |
US20100199228A1 (en) | 2009-01-30 | 2010-08-05 | Microsoft Corporation | Gesture Keyboarding |
US20100199231A1 (en) * | 2009-01-30 | 2010-08-05 | Microsoft Corporation | Predictive determination |
US8183977B2 (en) * | 2009-02-27 | 2012-05-22 | Seiko Epson Corporation | System of controlling device in response to gesture |
US20100229125A1 (en) | 2009-03-09 | 2010-09-09 | Samsung Electronics Co., Ltd. | Display apparatus for providing a user menu, and method for providing user interface (ui) applicable thereto |
US20100235786A1 (en) | 2009-03-13 | 2010-09-16 | Primesense Ltd. | Enhanced 3d interfacing for remote devices |
US20100235034A1 (en) * | 2009-03-16 | 2010-09-16 | The Boeing Company | Method, Apparatus And Computer Program Product For Recognizing A Gesture |
US20100295781A1 (en) | 2009-05-22 | 2010-11-25 | Rachid Alameh | Electronic Device with Sensing Assembly and Method for Interpreting Consecutive Gestures |
US20110144543A1 (en) * | 2009-05-27 | 2011-06-16 | Takashi Tsuzuki | Behavior recognition apparatus |
US20110296353A1 (en) * | 2009-05-29 | 2011-12-01 | Canesta, Inc. | Method and system implementing user-centric gesture control |
US20110006978A1 (en) | 2009-07-10 | 2011-01-13 | Yuan Xiaoru | Image manipulation based on tracked eye movement |
US20110018795A1 (en) | 2009-07-27 | 2011-01-27 | Samsung Electronics Co., Ltd. | Method and apparatus for controlling electronic device using user interaction |
US20110029918A1 (en) | 2009-07-29 | 2011-02-03 | Samsung Electronics Co., Ltd. | Apparatus and method for navigation in digital object using gaze information of user |
US20110052006A1 (en) | 2009-08-13 | 2011-03-03 | Primesense Ltd. | Extraction of skeletons from 3d maps |
US20120200494A1 (en) | 2009-10-13 | 2012-08-09 | Haim Perski | Computer vision gesture based control of a device |
US20110087970A1 (en) | 2009-10-14 | 2011-04-14 | At&T Mobility Ii Llc | Systems, apparatus, methods and computer-readable storage media for facilitating integrated messaging, contacts and social media for a selected entity |
US20110205186A1 (en) | 2009-12-04 | 2011-08-25 | John David Newton | Imaging Methods and Systems for Position Detection |
US20110141053A1 (en) | 2009-12-14 | 2011-06-16 | Synaptics Incorporated | System and method for measuring individual force in multi-object sensing |
US20110254798A1 (en) | 2009-12-18 | 2011-10-20 | Peter S Adamson | Techniques for recognizing a series of touches with varying intensity or angle of descending on a touch panel interface |
US8514221B2 (en) | 2010-01-05 | 2013-08-20 | Apple Inc. | Working with 3D objects |
US20110164032A1 (en) * | 2010-01-07 | 2011-07-07 | Prime Sense Ltd. | Three-Dimensional User Interface |
US20110193939A1 (en) | 2010-02-09 | 2011-08-11 | Microsoft Corporation | Physical interaction zone for gesture-based user interfaces |
US20110231757A1 (en) | 2010-02-28 | 2011-09-22 | Osterhout Group, Inc. | Tactile control in an augmented reality eyepiece |
US20110227820A1 (en) | 2010-02-28 | 2011-09-22 | Osterhout Group, Inc. | Lock virtual keyboard position in an augmented reality eyepiece |
US20110225536A1 (en) | 2010-02-28 | 2011-09-15 | Osterhout Group, Inc. | Sliding keyboard input control in an augmented reality eyepiece |
US20110211754A1 (en) | 2010-03-01 | 2011-09-01 | Primesense Ltd. | Tracking body parts by combined color image and depth processing |
US20110242102A1 (en) * | 2010-03-30 | 2011-10-06 | Harman Becker Automotive Systems Gmbh | Vehicle user interface unit for a vehicle electronic device |
US20110248914A1 (en) | 2010-04-11 | 2011-10-13 | Sherr Alan B | System and Method for Virtual Touch Typing |
US20110254765A1 (en) | 2010-04-18 | 2011-10-20 | Primesense Ltd. | Remote text input using handwriting |
US20110260965A1 (en) | 2010-04-22 | 2011-10-27 | Electronics And Telecommunications Research Institute | Apparatus and method of user interface for manipulating multimedia contents in vehicle |
US20130155070A1 (en) | 2010-04-23 | 2013-06-20 | Tong Luo | Method for user input from alternative touchpads of a handheld computerized device |
US20110261058A1 (en) | 2010-04-23 | 2011-10-27 | Tong Luo | Method for user input from the back panel of a handheld computerized device |
US8396252B2 (en) | 2010-05-20 | 2013-03-12 | Edge 3 Technologies | Systems and related methods for three dimensional gesture recognition in vehicles |
US8625882B2 (en) | 2010-05-31 | 2014-01-07 | Sony Corporation | User interface with three dimensional user input |
US20110293137A1 (en) | 2010-05-31 | 2011-12-01 | Primesense Ltd. | Analysis of three-dimensional scenes |
US20110292036A1 (en) | 2010-05-31 | 2011-12-01 | Primesense Ltd. | Depth sensor with application interface |
US20110310010A1 (en) | 2010-06-17 | 2011-12-22 | Primesense Ltd. | Gesture based user interface |
US20120001875A1 (en) | 2010-06-29 | 2012-01-05 | Qualcomm Incorporated | Touchless sensing and gesture recognition using continuous wave ultrasound signals |
US20130106692A1 (en) | 2010-07-20 | 2013-05-02 | Primesense Ltd. | Adaptive Projector |
US20130107021A1 (en) | 2010-07-20 | 2013-05-02 | Primesense Ltd. | Interactive Reality Augmentation for Natural Interaction |
WO2012011044A1 (en) | 2010-07-20 | 2012-01-26 | Primesense Ltd. | Interactive reality augmentation for natural interaction |
US20120035934A1 (en) | 2010-08-06 | 2012-02-09 | Dynavox Systems Llc | Speech generation device with a projected display and optical inputs |
WO2012020380A1 (en) | 2010-08-11 | 2012-02-16 | Primesense Ltd. | Scanning projectors and image capture modules for 3d mapping |
US20120038550A1 (en) * | 2010-08-13 | 2012-02-16 | Net Power And Light, Inc. | System architecture and methods for distributed multi-sensor gesture processing |
US20130207920A1 (en) | 2010-08-20 | 2013-08-15 | Eric McCann | Hand and finger registration for control applications |
US8996173B2 (en) | 2010-09-21 | 2015-03-31 | Intuitive Surgical Operations, Inc. | Method and apparatus for hand gesture control in a minimally invasive surgical system |
US20120078614A1 (en) | 2010-09-27 | 2012-03-29 | Primesense Ltd. | Virtual keyboard for a non-tactile three dimensional user interface |
US20120117514A1 (en) * | 2010-11-04 | 2012-05-10 | Microsoft Corporation | Three-Dimensional User Interaction |
US20120124516A1 (en) * | 2010-11-12 | 2012-05-17 | At&T Intellectual Property I, L.P. | Electronic Device Control Based on Gestures |
US20120223882A1 (en) | 2010-12-08 | 2012-09-06 | Primesense Ltd. | Three Dimensional User Interface Cursor Control |
US20120313848A1 (en) | 2010-12-13 | 2012-12-13 | Primesense Ltd. | Three Dimensional User Interface Session Control |
US20120173067A1 (en) * | 2010-12-30 | 2012-07-05 | GM Global Technology Operations LLC | Graphical vehicle command system for autonomous vehicles on full windshield head-up display |
US20120169583A1 (en) | 2011-01-05 | 2012-07-05 | Primesense Ltd. | Scene profiles for non-tactile user interfaces |
WO2012107892A2 (en) | 2011-02-09 | 2012-08-16 | Primesense Ltd. | Gaze detection in a 3d mapping environment |
US20120249416A1 (en) * | 2011-03-29 | 2012-10-04 | Giuliano Maciocci | Modular mobile connected pico projectors for a local multi-user collaboration |
US9030529B2 (en) | 2011-04-14 | 2015-05-12 | Industrial Technology Research Institute | Depth image acquiring device, system and method |
US20120268369A1 (en) * | 2011-04-19 | 2012-10-25 | Microsoft Corporation | Depth Camera-Based Relative Gesture Detection |
US20120295661A1 (en) | 2011-05-16 | 2012-11-22 | Yongsin Kim | Electronic device |
US20120309535A1 (en) | 2011-05-31 | 2012-12-06 | Microsoft Corporation | Action selection gesturing |
US9104271B1 (en) | 2011-06-03 | 2015-08-11 | Richard Adams | Gloved human-machine interface |
US20120320080A1 (en) | 2011-06-14 | 2012-12-20 | Microsoft Corporation | Motion based virtual object navigation |
US20130002801A1 (en) * | 2011-06-28 | 2013-01-03 | Mock Wayne E | Adjusting Volume of a Videoconference Using Touch-Based Gestures |
US20130014052A1 (en) | 2011-07-05 | 2013-01-10 | Primesense Ltd. | Zoom-based gesture user interface |
US20130044053A1 (en) | 2011-08-15 | 2013-02-21 | Primesense Ltd. | Combining Explicit Select Gestures And Timeclick In A Non-Tactile Three Dimensional User Interface |
US20130055120A1 (en) | 2011-08-24 | 2013-02-28 | Primesense Ltd. | Sessionless pointing user interface |
US20130055150A1 (en) | 2011-08-24 | 2013-02-28 | Primesense Ltd. | Visual feedback for tactile and non-tactile user interfaces |
US20140237432A1 (en) | 2011-09-15 | 2014-08-21 | Koninklijke Philips Electronics N.V. | Gesture-based user-interface with user-feedback |
US20140108930A1 (en) | 2012-10-12 | 2014-04-17 | Sling Media Inc. | Methods and apparatus for three-dimensional graphical user interfaces |
Non-Patent Citations (157)
Title |
---|
"Gesture-Aware Remote controls: Guidelines and Interaction Techniques", Gilles Bailly et al., 2011, Proceeding ICMI '11 Proceedings of the 13th international conference on multimodal interfaces. * |
Agrawala et al., "The two-user Responsive Workbench :support for collaboration through individual views of a shared space", Proceedings on the 24th conference on computer graphics and interactive techniques (SIGGRAPH 97), Los Angeles, USA, pp. 327-332 , Aug. 3-8, 1997. |
Allard et al., "Marker-less Real Time 3D modeling for Virtual Reality", Immersive Projection Technology, Iowa State University, year 2004. |
Arm Ltd., "AMBA Specification: AHB", Version 2, pp. 35-92, year 1999. |
Asada et al., "Determining Surface Orientation by Projecting a Stripe Pattern", IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI), vol. 10, No. 5, pp. 749-754, Sep. 1988. |
Ascension Technology Corporation, "Flock of Birds: Real-Time Motion Tracking", 2008. |
Avidan et al., "Trajectory triangulation: 3D reconstruction of moving points from amonocular image sequence", IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI), vol. 22, No. 4, pp. 348-3537, Apr. 2000. |
Azuma et al., "Recent Advances in Augmented Reality", IEEE Computer Graphics and Applications, vol. 21, issue 6, pp. 34-47, Nov. 2001. |
Berenson et al., U.S. Appl. No. 13/904,050 "Zoom-based gesture user interface", filed May 29, 2013. |
Berenson et al., U.S. Appl. No. 13/904,052 "Gesture-based interface with enhanced features", filed May 29, 2013. |
Besl, P., "Active, Optical Range Imaging Sensors", Machine vision and applications, vol. 1, pp. 127-152, year 1988. |
Bevilacqua et al., "People Tracking Using a Time-Of-Flight Depth Sensor", Proceedings of the IEEE International Conference on Video and Signal Based Surveillance, Sydney, Australia, Nov. 22-24, 2006. |
Bleiwess et al., "Fusing Time-of-Flight Depth and Color for Real-Time Segmentation and Tracking", Dyn3D 2009, Lecture Notes in Computer Science 5742, pp. 58-69, Jena, Germany, Sep. 9, 2009. |
Bleiwess et al., "Markerless Motion Capture Using a Single Depth Sensor", SIGGRAPH Asia 2009, Yokohama, Japan, Dec. 16-19, 2009. |
Bradski, G., "Computer Vision Face Tracking for Use in a Perceptual User Interface", Intel Technology Journal, vol. 2, issue 2 (2nd Quarter 2008). |
Breen et al., "Interactive Occlusion and Collision of Real and Virtual Objects in Augmented Reality", Technical Report ECRC-95-02, ECRC, Munich, Germany, 22 pages, year 1995. |
Bruckstein, A., "On shape from shading", Computer Vision, Graphics & Image Processing, vol. 44, pp. 139-154, year 1988. |
Burdea et al., "A Distributed Virtual Environment with Dextrous Force Feedback", Proceedings of Interface to Real and Virtual Worlds Conference, pp. 255-265, Mar. 1992. |
Bychkov et al., U.S. Appl. No. 13/849,514 "Gaze-enhanced Virtual Touchscreen", filed Mar. 24, 2013. |
Chai et al., "Parallel Projections for Stereo Reconstruction", Proceedings of IEEE Conference on Computer Vision and Pattern Recognition ,vol. 2, pp. 493-500, year 2000. |
Cheng et al., "Articulated Human Body Pose Inference from Voxel Data Using a Kinematically Constrained Gaussian Mixture Model", CVPR EHuM2: 2nd Workshop on Evaluation of Articulated Human Motion and Pose Estimation, 2007. |
Comaniciu et al., "Kernel-Based Object Tracking", IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 25, No. 5, pp. 564-577, May 2003. |
Commission Regulation (EC) No. 1275/2008, Official Journal of the European Union, Dec. 17, 2008. |
Dainty, J.C., "Introduction", Laser Speckle and Related Phenomena, pp. 1-7, Springer-Verlag, Berlin Heidelberg, 1975. |
Dekker, L., "Building Symbolic Information for 3D Human Body Modeling from Range Data", Proceedings of the Second International Conference on 3D Digital Imaging and Modeling, IEEE computer Society, pp. 388-397, 1999. |
Evers et al,. "Image-based Interactive rendering with view dependent geometry", Computer Graphics Forum, (Eurographics '03), vol. 22, No. 3, pp. 573-582, year 2003. |
Evers et al., "Image-based rendering of complex scenes from multi-camera rig", IEEE Proceedings on Vision, Image and Signal Processing, vol. 152, No. 4, pp. 470-480, Aug. 5, 2005. |
Fua et al., "Human Shape and Motion Recovery Using Animation Models", 19th Congress, International Society for Photogrammetry and Remote Sensing, Amsterdam, The Netherlands, Jul. 2000. |
Galor, M., U.S. Appl. No. 13/778,172 "Asymmetric Mapping in Tactile and Non-Tactile User Interfaces", filed Feb. 27, 2013. |
Gargallo et al., "Bayesian 3D Modeling from Images Using Multiple Depth Maps", Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05), vol. 2, pp. 885-891, Jun. 20-25, 2005. |
Gesturetec Inc., "Gesture Control Solutions for Consumer Devices", Canada, 2009. |
Gobbetti et al., "VB2: an Architecture for Interaction in Synthetic Worlds", Proceedings of the 6th Annual ACM Symposium on User Interface Software and Technology (UIST'93), pp. 167-178, Nov. 3-5, 1993. |
Gokturk et al., "A Time-Of-Flight Depth Sensor-System Description, Issues and Solutions", Proceedings of the 2004 Conference on Computer Vision and Patter Recognition Workshop (CVPRW'04), vol. 3, pp. 35, Jun. 27-Jul. 2, 2004. |
Goodman, J.W., "Statistical Properties of Laser Speckle Patterns", Laser Speckle and Related Phenomena, pp. 9-75, Springer-Verlag, Berlin Heidelberg, 1975. |
Gordon et al., "The use of Dense Stereo Range Date in Augmented Reality", Proceedings of the 1st International Symposium on Mixed and Augmented Reality (ISMAR), Darmstadt, Germany, pp. 1-10, Sep. 30-Oct. 1, 2002. |
Grammalidis et al., "3-D Human Body Tracking from Depth Images Using Analysis by Synthesis", Proceedings of the IEEE International Conference on Image Processing (ICIP2001), pp. 185-188, Greece, Oct. 7-10, 2001. |
Grest et al., "Single View Motion Tracking by Depth and Silhouette Information", SCIA 2007-Scandinavian Conference on Image Analysis, Lecture Notes in Computer Science 4522, pp. 719-729, Aalborg, Denmark, Jun. 10-14, 2007. |
Guendelman et al., U.S. Appl. No. 13/849,514 "Enhanced Virtual Touchpad", filed Mar. 24, 2013. |
Haritaoglu et al., "Ghost 3d: Detecting Body Posture and Parts Using Stereo", Proceedings of the IEEE Workshop on Motion and Video Computing (Motion'02), pp. 175-180, Orlando, USA, Dec. 5-6, 2002. |
Haritaoglu et al., "W4S : A real-time system for detecting and tracking people in 2<1/2>D", ECCV 98-5th European conference on computer vision, vol. 1407, pp. 877-892, Freiburg , Germany, Jun. 2-6, 1998. |
Haritaoglu et al., "W4S : A real-time system for detecting and tracking people in 2D", ECCV 98-5th European conference on computer vision, vol. 1407, pp. 877-892, Freiburg , Germany, Jun. 2-6, 1998. |
Harman et al., "Rapid 2D-to 3D conversion", Proceedings of SPIE Conference on Stereoscopic Displays and Virtual Reality Systems, vol. 4660, pp. 78-86, Jan. 21-23, 2002. |
Hart, D., U.S. Appl. No. 09/616,606 "Method and System for High Resolution , Ultra Fast 3-D Imaging" , filed Jul. 14, 2000. |
Harville, M., "Stereo Person Tracking with Short and Long Term Plan-View Appearance Models of Shape and Color", Proceedings of the IEEE International Conference on Advanced Video and Signal-Based Surveillance (AVSSS-2005), pp. 522-527, Como, Italy, Sep. 15-16, 2005. |
Hilliges et al, "Interactions in the air: adding further depth to interactive tabletops", Proceedings of the 22nd annual ACM symposium on User interface software and technology, ACM, pp. 139-148, Oct. 2009. |
Hoff et al., "Analysis of head pose accuracy in augmented reality", IEEE Transactions on Visualization and Computer Graphics, vol. 6, No. 4, pp. 319-334, Oct.-Dec. 2000. |
Holte et al., "Gesture Recognition using a Range Camera", Technical Report CVMT-07-01 ISSN 1601-3646, Feb. 2007. |
Holte, M., "Fusion of Range and Intensity Information for View Invariant Gesture Recognition", IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW '08), pp. 1-7, Anchorage, USA, Jun. 23-28, 2008. |
Horn et al., "Toward optimal structured light patterns", Proceedings of International Conference on Recent Advances in 3D Digital Imaging and Modeling, pp. 28-37, Ottawa, Canada, May 1997. |
Horn, B., "Height and gradient from shading", International Journal of Computer Vision, vol. 5, No. 1, pp. 37-76, Aug. 1990. |
Howe et al., "Bayesian Reconstruction of 3D Human Motion from Single-Camera Video", Advanced in Neural Information Processing Systems, vol. 12, pp. 820-826, USA 1999. |
International Application PCT/IB2010/051055 Search Report dated Sep. 1, 2010. |
International Application PCT/IB2011/053192 Search Report dated Dec. 6, 2011. |
International Application PCT/IB2012/050577 Search Report dated Aug. 6, 2012. |
International Application PCT/IB2013/052332 Search Report dated Aug. 26, 2013. |
International Application PCT/IL2006/000335 Preliminary Report on Patentability dated Apr. 24, 2008. |
International Application PCT/IL2007/000306 Search Report dated Oct. 2, 2008. |
International Application PCT/IL2007/000574 Search Report dated Sep. 10, 2008. |
Kaewtrakulpong et al., "An Improved Adaptive Background Mixture Model for Real-Time Tracking with Shadow Detection", Proceedings of the 2nd European Workshop on Advanced Video Based Surveillance Systems (AVBS'01), Kingston, UK, Sep. 2001. |
Kauff et al., "Depth map creation and image-based rendering for advanced 3DTV Services Providing Interoperability and Scalability", Signal Processing: Image Communication, vol. 22, No. 2, pp. 217-234, year 2007. |
Kimmel et al., "Analyzing and synthesizing images by evolving curves with the Osher-Sethian method", International Journal of Computer Vision, vol. 24, No. 1, pp. 37-56, year 1997. |
Kolb et al., "ToF-Sensors: New Dimensions for Realism and Interactivity", Proceedings of the IEEE Conference on Computer Vision and Patter Recognition Workshops, pp. 1-6, Anchorage, USA, Jun. 23-28, 2008. |
Kolsch et al., "Fast 2D Hand Tracking with Flocks of Features and Multi-Cue Integration", IEEE Workshop on Real-Time Time Vision for Human Computer Interaction (at CVPR'04), Washington, USA, Jun. 27-Jul. 2, 2004. |
Koninckx et al., "Efficient, Active 3D Acquisition, based on a Pattern-Specific Snake", Luc Van Gool (Editor), (DAGM 2002), Lecture Notes in Computer Science 2449, pp. 557-565, Springer 2002. |
Koutek, M., "Scientific Visualization in Virtual Reality: Interaction Techniques and Application Development", PhD Thesis, Delft University of Technology, 264 pages, Jan. 2003. |
Krumm et al., "Multi-Camera Multi-Person Tracking for EasyLiving", 3rd IEEE International Workshop on Visual Surveillance, Dublin, Ireland, Jul. 1, 2000. |
La Viola, J. Jr., "Whole-Hand and Speech Input in Virtual Environments", Computer Science Department, Florida Atlantic University, USA, 1996. |
Leclerc et al., "The direct computation of height from shading", The Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), pp. 552-558, USA, Jun. 1991. |
Leens et al., "Combining Color, Depth, and Motion for Video Segmentation", ICVS 2009-7th International Conference on Computer Vision Systems, Liege, Belgium Oct. 13-15, 2009. |
Li et al., "Real-Time 3D Motion Tracking with Known Geometric Models", Real-Time Imaging Journal, vol. 5, pp. 167-187, Academic Press 1999. |
MacCormick et al., "Partitioned Sampling, Articulated Objects, and Interface-Quality Hand Tracking", ECCV '00-Proceedings of the 6th European Conference on Computer Vision-Part II , pp. 3-19, Dublin, Ireland, Jun. 26-Jul. 1, 2000. |
Malassiotis et al., "Real-Time Hand Posture Recognition Using Range Data", Image and Vision Computing, vol. 26, No. 7, pp. 1027-1037, Jul. 2, 2008. |
Manning et al., "Foundations of Statistical Natural Language Processing", chapters 6,7,9 and 12, MIT Press 1999. |
Martell, C., "Form: An Experiment in the Annotation of the Kinematics of Gesture", Dissertation, Computer and Information Science, University of Pennsylvania, 2005. |
Mendlovic et al., "Composite harmonic filters for scale, projection and shift invariant pattern recognition", Applied Optics Journal, vol. 34, No. 2, Jan. 10, 1995. |
Microvision Inc., "PicoP® Display Engine-How it Works", 1996-2012. |
Miller, R., "Kinect for XBox 360 Review", Engadget, Nov. 4, 2010. |
Morano et al., "Structured Light Using Pseudorandom Codes", IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 20, issue 3, pp. 322-327, Mar. 1998. |
Munoz-Salinas et al., "People Detection and Tracking Using Stereo Vision and Color", Image and Vision Computing, vol. 25, No. 6, pp. 995-1007, Jun. 1, 2007. |
Nakamura et al, "Occlusion detectable stereo-occlusion patterns in camera matrix", Proceedings of the 1996 Conference on Computer Vision and Pattern Recognition (CVPR '96), pp. 371-378, Jun. 1996. |
Nam et al., "Recognition of Hand Gestures with 3D, Nonlinear Arm Movements", Pattern Recognition Letters, vol. 18, No. 1, pp. 105-113, Elsevier Science B.V. 1997. |
Nanda et al., "Visual Tracking Using Depth Data", Proceedings of the 2004 Conference on Computer Vision and Patter Recognition Workshop, vol. 3, Washington, USA Jun. 27-Jul. 2, 2004. |
Nesbat, S., "A System for Fast, Full-Text Entry for Small Electronic Devices", Proceedings of the 5th International Conference on Multimodal Interfaces, ICMI 2003, Vancouver, Nov. 5-7, 2003. |
Noveron, "Madison video eyewear", year 2012. |
Ohta et al., "Share-Z: Client/Server Depth Sensing for See-Through Head-Mounted Displays", Presence: Teleoperators and Virtual Environments, vol. 11, No. 2, pp. 176-188, Apr. 2002. |
Poupyrev et al., "The go-go interaction technique: non-liner mapping for direct manipulation in VR", Proceedings of the 9th annual ACM Symposium on User interface software and technology (UIST '96), Washington, USA, pp. 79-80, Nov. 6-8, 1996. |
Prime Sense Inc., "Prime Sensor(TM) NITE 1.1 Framework Programmer's Guide", Version 1.2, year 2009. |
Prime Sense Inc., "Prime Sensor™ NITE 1.1 Framework Programmer's Guide", Version 1.2, year 2009. |
Primesense Corporation, "PrimeSensor NITE 1.1", USA, year 2010. |
Primesense Corporation, "PrimeSensor Reference Design 1.08", USA, year 2010. |
Primesense, "Natural Interaction", YouTube Presentation, Jun. 9, 2010 http://www.youtube.com/watch?v=TzLKsex43zl~. |
Primesense, "Natural Interaction", YouTube Presentation, Jun. 9, 2010 http://www.youtube.com/watch?v=TzLKsex43zl˜. |
Scharstein et al., "High-Accuracy Stereo Depth Maps Using Structured Light", IEEE Conference on Computer Vision and Patter Recognition, vol. 1, pp. 195-2002, Madison, USA, Jun. 2003. |
Scharstein, D., "Stereo vision for view synthesis", Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, pp. 852-858, year 1996. |
Schmalstieg et al., "The Studierstube Augmented Reality Project", Presence: Teleoperators and Virtual Environments, vol. 11, No. 1, pp. 33-54, Feb. 2002. |
Segen et al., "Human-computer interaction using gesture recognition and 3D hand tracking", ICIP 98, Proceedings of the IEEE International Conference on Image Processing, vol. 3, pp. 188-192, Oct. 4-7, 1998. |
Segen et al., "Shadow gestures: 3D hand pose estimation using a single camera", Proceedings of IEEE International Conference on Computer Vision and Pattern Recognition, pp. 479-485, Fort Collins, USA, 1999. |
Shi et al., "Good Features to Track", IEEE Conference on Computer Vision and Pattern Recognition, pp. 593-600, Seattle, USA, Jun. 21-23, 1994. |
Siddiqui et al., "Robust Real-Time Upper Body Limb Detection and Tracking", Proceedings of the 4th ACM International Workshop on Video Surveillance and Sensor Networks, Santa Barbara, USA, Oct. 27, 2006. |
Slinger et al, "Computer-Generated Holography as a Generic Display Technology", IEEE Computer, vol. 28, Issue 8, pp. 46-53, Aug. 2005. |
Softkinetic S.A., IISU(TM)-3D Gesture Recognition Platform for Developers of 3D Applications, Belgium, Brussels, 2007 - 2010. |
Softkinetic S.A., IISU™-3D Gesture Recognition Platform for Developers of 3D Applications, Belgium, Brussels, 2007 - 2010. |
Sudderth et al., "Visual Hand Tracking Using Nonparametric Belief Propagation", IEEE Workshop on Generative Model Based Vision at CVPR'04, Washington, USA, Jun. 27-Jul. 2, 2004. |
Sun et al., "SRP Based Natural Interaction Between Real and Virtual Worlds in Augmented Reality", Proceedings of the International Conference on Cyberworlds (CW'08), pp. 117-124, Sep. 22-24, 2008. |
Tobii Technology, "The World Leader in Eye Tracking and Gaze Interaction", Mar. 2012. |
Tsap, L. "Gesture-Tracking in Real Time with Dynamic Regional Range Computation", Real-Time Imaging, vol. 8, issue 2, pp. 115-126, Apr. 2002. |
U.S. Appl. No. 12/314,210 Office Action dated Jan. 10, 2014. |
U.S. Appl. No. 12/352,622 Official Action dated Mar. 31, 2011. |
U.S. Appl. No. 12/352,622 Official Action dated Sep. 30, 2011. |
U.S. Appl. No. 12/683,452 Office Action dated Jan. 22, 2014. |
U.S. Appl. No. 12/683,452 Office Action dated Jul. 16, 2014. |
U.S. Appl. No. 12/683,452 Office Action dated Jun. 7, 2013. |
U.S. Appl. No. 12/683,452 Office Action dated Nov. 21, 2014. |
U.S. Appl. No. 12/683,452 Official Action dated Sep. 7, 2012. |
U.S. Appl. No. 12/721,582 Office Action dated Apr. 17, 2014. |
U.S. Appl. No. 12/762,336 Official Action dated May 15, 2012. |
U.S. Appl. No. 13/161,508 Office Action dated Apr. 10, 2013. |
U.S. Appl. No. 13/161,508 Office Action dated Aug. 18, 2015. |
U.S. Appl. No. 13/161,508 Office Action dated Dec. 23, 2014. |
U.S. Appl. No. 13/161,508 Office Action dated Sep. 9, 2013. |
U.S. Appl. No. 13/244,490 Office Action dated Dec. 6, 2013. |
U.S. Appl. No. 13/314,207 Office Action dated Apr. 3, 2014. |
U.S. Appl. No. 13/314,207 Office Action dated Aug. 5, 2013. |
U.S. Appl. No. 13/314,210 Office Action dated Jul. 19, 2013. |
U.S. Appl. No. 13/314,210, filed Dec. 8, 2011. |
U.S. Appl. No. 13/423,314 Advisory Action dated Jun. 26, 2014. |
U.S. Appl. No. 13/423,314 Office Action dated Dec. 4, 2013. |
U.S. Appl. No. 13/423,314 Office Action dated Jul. 31, 2014. |
U.S. Appl. No. 13/423,322 Office Action dated Apr. 7, 2014. |
U.S. Appl. No. 13/423,322 Office Action dated Nov. 1, 2013. |
U.S. Appl. No. 13/541,786 Office Action dated Feb. 13, 2014. |
U.S. Appl. No. 13/541,786, filed Jul. 5, 2012. |
U.S. Appl. No. 13/584,831 Office Action dated Jul. 8, 2014. |
U.S. Appl. No. 13/584,831 Office Action dated Mar. 20, 2014. |
U.S. Appl. No. 13/584,831, filed Aug. 14, 2012. |
U.S. Appl. No. 13/592,369, filed Aug. 23, 2012. |
U.S. Appl. No. 13/904,052 Office Action dated Oct. 14, 2015. |
U.S. Appl. No. 14/055,997 Office Action dated May 28, 2014. |
U.S. Appl. No. 14/055,997 Office Action dated Nov. 21, 2014. |
U.S. Appl. No. 61/525,771, filed Aug. 21, 2011. |
U.S. Appl. No. 61/538,970, filed Sep. 26, 2011. |
U.S. Appl. No. 61/603,949, filed Feb. 28, 2012. |
U.S. Appl. No. 61/615,403, filed Mar. 26, 2012. |
U.S. Appl. No. 61/652,899, filed May 30, 2012. |
U.S. Appl. No. 61/663,638, filed Jun. 25, 2012. |
Vogler et al., "ASL recognition based on a coupling between HMMs and 3D motion analysis", Proceedings of IEEE International Conference on Computer Vision, pp. 363-369, Mumbai, India, 1998. |
Wexelblat et al., "Virtual Reality Applications and Explorations", Academic Press Inc., San Diego, USA, 262 pages, year 1993. |
Winkelbach et al., "Shape from Single Stripe Pattern Illumination", Luc Van Gool (Editor), (DAGM 2002), Lecture Notes in Computer Science 2449, p. 240-247, Springer 2002. |
Xu et al., "A Multi-Cue-Based Human Body Tracking System", Proceedings of the 5ths International Conference on Computer Vision Systems (ICVS 2007), Germany, Mar. 21-24, 2007. |
Xu et al., "Human Detecting Using Depth and Gray Images", Proceedings of the IEE Conference on Advanced Video and Signal Based Surveillance (AVSS'03), Miami, USA, Jul. 21-22, 2003. |
Yilmaz et al., "Object Tracking: A Survey", ACM Computing Surveys, vol. 38, No. 4, article 13, Dec. 2006. |
Zhang et al., "Height recovery from intensity gradients", IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), pp. 508-513, Jun. 21-23, 1994. |
Zhang et al., "Rapid Shape Acquisition Using Color Structured Light and Multi-Pass Dynamic Programming", 1st International Symposium on 3D Data Processing Visualization and Transmission (3DPVT), Italy, Jul. 2002. |
Zhang et al., "Shape from intensity gradient", IEEE Transactions on Systems, Man and Cybernetics-Part A: Systems and Humans, vol. 29, No. 3, pp. 318-325, May 1999. |
Zhu et al., "Controlled Human Pose Estimation From Depth Image Streams", IEEE Conference on Computer Vision and Patter Recognition Workshops, pp. 1-8, Anchorage, USA, Jun. 23-27, 2008. |
Zhu et al., "Generation and Error Characterization of Pararell-Perspective Stereo Mosaics from Real Video", In-Video Registration, Springer, US, chapter 4,pp. 72-105, year 2003. |
Zigelman et al., "Texture mapping using surface flattening via multi-dimensional scaling", IEEE Transactions on Visualization and Computer Graphics, vol. 8, No. 2, pp. 198-207, Apr. 2002. |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140370470A1 (en) * | 2013-06-13 | 2014-12-18 | Gary And Mary West Health Institute | Systems, apparatus and methods for delivery and augmentation of behavior modification therapy and teaching |
US10474793B2 (en) * | 2013-06-13 | 2019-11-12 | Northeastern University | Systems, apparatus and methods for delivery and augmentation of behavior modification therapy and teaching |
US10671065B2 (en) * | 2013-10-25 | 2020-06-02 | Ioannis Micros | Optically assisted landing and takeoff of drones |
US20160144867A1 (en) * | 2014-11-20 | 2016-05-26 | Toyota Motor Engineering & Manufacturing North America, Inc. | Autonomous vehicle detection of and response to traffic officer presence |
US9586585B2 (en) * | 2014-11-20 | 2017-03-07 | Toyota Motor Engineering & Manufacturing North America, Inc. | Autonomous vehicle detection of and response to traffic officer presence |
USD800586S1 (en) * | 2015-03-25 | 2017-10-24 | Muv Interactive Ltd. | Wearable sensing device |
US9983684B2 (en) | 2016-11-02 | 2018-05-29 | Microsoft Technology Licensing, Llc | Virtual affordance display at virtual target |
US10528794B2 (en) * | 2017-06-05 | 2020-01-07 | Motorola Solutions, Inc. | System and method for tailoring an electronic digital assistant inquiry response as a function of previously detected user ingestion of related video information |
Also Published As
Publication number | Publication date |
---|---|
US20180356898A1 (en) | 2018-12-13 |
US10642371B2 (en) | 2020-05-05 |
US20130055120A1 (en) | 2013-02-28 |
US20160041623A1 (en) | 2016-02-11 |
US10088909B2 (en) | 2018-10-02 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10642371B2 (en) | Sessionless pointing user interface | |
US20220164032A1 (en) | Enhanced Virtual Touchpad | |
US11262840B2 (en) | Gaze detection in a 3D mapping environment | |
US12032746B2 (en) | Systems and methods of creating a realistic displacement of a virtual object in virtual reality/augmented reality environments | |
US9122311B2 (en) | Visual feedback for tactile and non-tactile user interfaces | |
US20240319867A1 (en) | Systems, methods and user interfaces for object tracing | |
AU2015252151B2 (en) | Enhanced virtual touchpad and touchscreen | |
CN116204060A (en) | Gesture-based movement and manipulation of a mouse pointer |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: PRIMESENSE LTD., ISRAEL Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:GALOR, MICHA;POKRASS, JONATHAN;HOFFNUNG, AMIR;AND OTHERS;SIGNING DATES FROM 20120820 TO 20120821;REEL/FRAME:028832/0938 |
|
FEPP | Fee payment procedure |
Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
AS | Assignment |
Owner name: APPLE INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:PRIMESENSE LTD.;REEL/FRAME:034293/0092 Effective date: 20140828 |
|
AS | Assignment |
Owner name: APPLE INC., CALIFORNIA Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE APPLICATION # 13840451 AND REPLACE IT WITH CORRECT APPLICATION # 13810451 PREVIOUSLY RECORDED ON REEL 034293 FRAME 0092. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT;ASSIGNOR:PRIMESENSE LTD.;REEL/FRAME:035624/0091 Effective date: 20140828 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |