US20160255333A1 - Generating Images from Light Fields Utilizing Virtual Viewpoints - Google Patents
Generating Images from Light Fields Utilizing Virtual Viewpoints Download PDFInfo
- Publication number
- US20160255333A1 US20160255333A1 US14/876,024 US201514876024A US2016255333A1 US 20160255333 A1 US20160255333 A1 US 20160255333A1 US 201514876024 A US201514876024 A US 201514876024A US 2016255333 A1 US2016255333 A1 US 2016255333A1
- Authority
- US
- United States
- Prior art keywords
- light field
- image data
- virtual
- viewpoint
- field image
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/271—Image signal generators wherein the generated image signals comprise depth maps or disparity maps
-
- H04N13/0271—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/18—Image warping, e.g. rearranging pixels individually
-
- G06T7/0075—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
- G06T7/55—Depth or shape recovery from multiple images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
- G06T7/55—Depth or shape recovery from multiple images
- G06T7/557—Depth or shape recovery from multiple images from light fields, e.g. from plenoptic cameras
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
- G06T7/55—Depth or shape recovery from multiple images
- G06T7/593—Depth or shape recovery from multiple images from stereo images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
- G06T7/55—Depth or shape recovery from multiple images
- G06T7/593—Depth or shape recovery from multiple images from stereo images
- G06T7/596—Depth or shape recovery from multiple images from stereo images from three or more stereo images
-
- H04N13/0242—
-
- H04N13/0282—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/111—Transformation of image signals corresponding to virtual viewpoints, e.g. spatial image interpolation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/111—Transformation of image signals corresponding to virtual viewpoints, e.g. spatial image interpolation
- H04N13/117—Transformation of image signals corresponding to virtual viewpoints, e.g. spatial image interpolation the virtual viewpoint locations being selected by the viewers or determined by viewer tracking
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/207—Image signal generators using stereoscopic image cameras using a single 2D image sensor
- H04N13/232—Image signal generators using stereoscopic image cameras using a single 2D image sensor using fly-eye lenses, e.g. arrangements of circular lenses
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/243—Image signal generators using stereoscopic image cameras using three or more 2D image sensors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/282—Image signal generators for generating image signals corresponding to three or more geometrical viewpoints, e.g. multi-view systems
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/10—Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from different wavelengths
- H04N23/13—Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from different wavelengths with multiple sensors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/95—Computational photography systems, e.g. light-field imaging systems
-
- H04N9/09—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10004—Still image; Photographic image
- G06T2207/10012—Stereo images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10028—Range image; Depth image; 3D point clouds
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10052—Images from lightfield camera
Definitions
- the present invention relates to systems and methods for generating images from light field image data and more specifically to systems and methods for generating images from light field image data using virtual viewpoints.
- Imaging devices such as cameras, can be used to capture images of portions of the electromagnetic spectrum, such as the visible light spectrum, incident upon an image sensor.
- the term light is generically used to cover radiation across the entire electromagnetic spectrum.
- light enters through an opening (aperture) at one end of the imaging device and is directed to an image sensor by one or more optical elements such as lenses.
- the image sensor includes pixels or sensor elements that generate signals upon receiving light via the optical element.
- Commonly used image sensors include charge-coupled device (CCDs) sensors and complementary metal-oxide semiconductor (CMOS) sensors.
- CCDs charge-coupled device
- CMOS complementary metal-oxide semiconductor
- Image sensors are devices capable of converting an optical image into a digital signal.
- Image sensors utilized in digital cameras are made up of an array of pixels; the number of pixels determines the megapixel rating of the image sensor. For example, an image sensor having a width ⁇ height of 2272 ⁇ 1704 pixels would have an actual pixel count of 3,871,488 pixels and would be considered a 4 megapixel image sensor.
- Each pixel in an image sensor is capable of capturing light and converting the captured light into electrical signals.
- a Bayer filter is often placed over the image sensor, filtering the incoming light into its red, blue, and green (RGB) components which are then captured by the image sensor.
- RGB red, blue, and green
- image capture utilizes a single image sensor, to capture individual images, one at a time.
- a digital camera typically combines both an image sensor and processing capabilities. When the digital camera takes a photograph, the data captured by the image sensor is provided to the processor by the image sensor. Processors are able to control aspects of a captured image by changing image capture parameters of the sensor elements or groups of sensor elements used to capture the image.
- the ISO/IEC 10918-1 standard more commonly referred to as the JPEG standard after the Joint Photographic Experts Group that developed the standard, establishes a standard process for digital compression and coding of still images.
- the JPEG standard specifies a codec for compressing an image into a bitstream and for decompressing the bitstream back into an image.
- a system configured to synthesize images using captured light field image data includes a processor and a memory connected to the processor and configured to store captured light field image data and an image manipulation application, wherein the captured light field image data includes image data, pixel position data, and a depth map, wherein the depth map includes depth information for one or more pixels in the image data, and wherein the image manipulation application configures the processor to obtain captured light field image data, determine a virtual viewpoint for the captured light field image data based on the pixel position data and the depth map for the captured light field image data, where the virtual viewpoint includes a virtual location and virtual depth information, compute a virtual depth map based on the captured light field image data and the virtual viewpoint, and generate an image from the perspective of the virtual viewpoint based on the captured light field image data and the virtual depth map, where the generated image includes a plurality of pixels selected from the image data based on the pixel position data and the virtual
- the virtual viewpoint corresponds to a focal plane in an array camera utilized to create the captured light field image data.
- the captured light field image data further includes a reference viewpoint within the captured light field image data and the virtual viewpoint is a separate viewpoint within the captured light field image data from the reference viewpoint.
- the captured light field image data was captured by an array camera having an imager array including a plurality of imagers and the reference viewpoint corresponds to the viewpoint of a first imager within the imager array in the array camera.
- the virtual viewpoint corresponds to the viewpoint of a second imager within the imager array, where the second imager is separate from the first imager.
- the virtual viewpoint is a viewpoint that does not correspond to the viewpoint of any of the imagers within the imager array.
- the virtual viewpoint is selected from a position selected from the group consisting of in front of the imager array and behind the imager array.
- the image manipulation application further configures the processor to generate an image from the perspective of the virtual viewpoint by projecting pixels from the captured light field image data based on the pixel position data and the depth map, where the projected pixels are described in the image data and the depth map.
- the captured light field image data further includes occluded pixel information describing pixels not visible from a reference viewpoint of the captured light field image data and the projected pixels include at least one occluded pixel in the occluded pixel information that is visible from the perspective of the virtual viewpoint.
- At least one projected pixel in the generated image is not described in the image data, the pixel position data, and the depth map and the image manipulation application further configures the processor to generate the at least one projected pixel by resampling the image data, the pixel position data, and the depth map.
- a pinhole camera model is utilized to project pixels within the generated image based on light rays projecting from the virtual viewpoint, where each projected pixels is associated with at least one of the projected light rays.
- projected pixel depth information is determined for at least one pixel in the generated image based on the depth map, the virtual viewpoint, and the light rays associated with the projected pixel.
- the depth information for a projected pixel is based on minimizing the variance for the projected pixel across the image data within the captured light field image data.
- the image manipulation application further configures the processor to combine projected pixels having the same location within the generated image.
- the pixels are combined based on the weighted average of the pixels, where the weighted average of the pixels is the inverse of the distance from the imager from which the projected pixel originated to the virtual viewpoint.
- system further includes an input device configured to obtain input data indicative of a position within the captured light field image data.
- the input device is a touchscreen interface.
- the input device is a sensor configured to obtain spatial location information.
- the input device is a camera configured to obtain input data selected from the group consisting of head tracking data and gaze tracking data.
- the virtual viewpoint is selected based on the input data.
- the image manipulation application further configures the processor to obtain a first input data indicative of a first position within the captured light field image data, determine a first virtual viewpoint based on the first input data, generate a first image from the perspective of the first virtual viewpoint, obtain a second input indicative of a second position within the captured light field image data, where the second position is separate from the first position, determine a second virtual viewpoint based on the second input data, generate at least one intermediate virtual viewpoint by interpolating between the first virtual viewpoint and the second virtual viewpoint, generate at least one intermediate image based on the generated at least one intermediate virtual viewpoints, where each intermediate image is from the perspective of an intermediate virtual viewpoint, and generate a second image from the perspective of the second virtual viewpoint.
- the image is generated utilizing a super-resolution process.
- Still another embodiment of the invention includes a process for generating an image from a virtual viewpoint, including obtaining captured light field image data using an image manipulation device, where the captured light field image data includes image data, pixel position data, and a depth map and where the depth map includes depth information for one or more pixels in the image data, determining a virtual viewpoint for the captured light field image data based on the pixel position data and the depth map for the captured light field image data using the image manipulation device, where the virtual viewpoint includes a virtual location and virtual depth information, computing a virtual depth map based on the captured light field image data and the virtual viewpoint using the image manipulation device, and generating an image from the perspective of the virtual viewpoint based on the captured light field image data and the virtual depth map using the image manipulation device, where the generated image includes a plurality of pixels selected from the image data based on the pixel position data and the virtual depth map.
- FIG. 1A conceptually illustrates an array camera including a 5 ⁇ 5 imager array connected with a processor in accordance with an embodiment of the invention.
- FIG. 1B conceptually illustrates a 5 ⁇ 5 array camera module in accordance with an embodiment of the invention.
- FIG. 1C conceptually illustrates a color filter pattern for a 4 ⁇ 4 array camera module in accordance with an embodiment of the invention.
- FIG. 2 is a diagram conceptually illustrating a device capable of processing light field images in accordance with an embodiment of the invention.
- FIG. 3A is a diagram conceptually illustrating virtual viewpoints for a given light field in accordance with an embodiment of the invention.
- FIG. 3B is a diagram conceptually illustrating a light field image rendered from a virtual viewpoint for a given light field in accordance with an embodiment of the invention.
- FIG. 3C is a diagram conceptually illustrating a light field image rendered from a second virtual viewpoint for a given light field in accordance with an embodiment of the invention.
- FIG. 4 is a flow chart conceptually illustrating a process for generating a light field image from a virtual viewpoint in a light field in accordance with an embodiment of the invention.
- FIG. 4B is a flow chart conceptually illustrating a process for generating a light field image from a virtual viewpoint using projected light rays in accordance with an embodiment of the invention.
- FIG. 5 is a flow chart conceptually illustrating a process for reprojecting light rays in relation to a virtual viewpoint in accordance with an embodiment of the invention.
- FIG. 6 is a flow chart conceptually illustrating a process for computing a depth map for a virtual viewpoint in accordance with an embodiment of the invention.
- FIG. 7 is a flow chart conceptually illustrating a process for projecting pixels to form a light field image corresponding to a virtual viewpoint in accordance with an embodiment of the invention.
- FIG. 8 is a flow chart conceptually illustrating a process for interactively generating light field images from virtual viewpoints in accordance with an embodiment of the invention.
- a light field is often defined as a 4D function characterizing the light from all direction at all points in a scene and can be interpreted as a two-dimensional (2D) collection of 2D images of a scene.
- Array cameras such as those described in U.S. patent application Ser. No. 12/935,504 entitled “Capturing and Processing of Images using Monolithic Camera Array with Heterogeneous Imagers” to Venkataraman et al., can be utilized to capture light fields.
- super-resolution processes such as those described in U.S. patent application Ser. No.
- the terms high or higher resolution and low or lower resolution are used here in a relative sense and not to indicate the specific resolutions of the images captured by the array camera.
- the disclosures of U.S. patent application Ser. No. 12/935,504 and U.S. patent application Ser. No. 12/967,807 are hereby incorporated by reference in their entirety.
- a file containing an image synthesized from light field image data and metadata derived from the light field image data can be referred to as a light field image file.
- the encoded image in a light field image file is typically synthesized using a super resolution process from a number of lower resolution images.
- the light field image file can also include metadata describing the synthesized image derived from the light field image data that enables post processing of the synthesized image.
- a light field image file is created by encoding an image synthesized from light field image data and combining the encoded image with a depth map derived from the light field image data.
- the encoded image is synthesized from a reference viewpoint and the metadata includes information concerning pixels in the light field image that are occluded from the reference viewpoint.
- the metadata can also include additional information including (but not limited to) auxiliary maps such as confidence maps, edge maps, occluded pixel information, and missing pixel maps that can be utilized during post processing of the encoded image to improve the quality of an image rendered using the light field image data file.
- a device configured to generate an image rendered using the information within the light field image file
- rendering devices can render new images using the information within the file without the need to perform super resolution processing on the original light field image data. In this way, the amount of data transmitted to the rendering device and the computational complexity of rendering an image is reduced.
- rendering devices are configured to perform processes including (but not limited to) refocusing the encoded image based upon a focal plane specified by the user, synthesizing an image from a different viewpoint, and generating a stereo pair of images.
- file formats may be utilized to store light field image files in accordance with embodiments of the invention.
- One such file format is the JPEG-DX extension to ISO/IEC 10918-1 described in U.S. Provisional Patent Application No. 61/540,188 entitled “JPEG-DX: A Backwards-compatible, Dynamic Focus Extension to JPEG”, to Venkataraman et al., filed Sep. 28, 2011, the entirety of which is incorporated by reference.
- An image is captured from the viewpoint of a camera.
- array cameras capture at least one image of a scene from a reference viewpoint and the reference viewpoint is used to synthesize an image.
- the reference viewpoint corresponds to the viewpoint of one focal plane or a reference focal plane in an imager array.
- the term virtual viewpoint describes a viewpoint used to synthesize an image that may or may not correspond to one of the viewpoints of the focal planes in the imager array that capture the light field image data. Given an array of images taken using an array camera, an image from the perspective of a virtual viewpoint in a light field can be synthesized.
- the viewpoint from which the image is synthesized can be the viewpoint of a focal plane in an imager array or from an interpolated position, such as a point between focal planes, in front of the imager array, or behind the imager array.
- images are synthesized from virtual viewpoints.
- the (virtual) viewpoint used to synthesize the image incorporates a rotation and/or translation of one or more of the optical axes of the viewpoint.
- devices capable of displaying images synthesized from light field image data include one or more input devices, such as cameras configured to allow for gaze and/or motion tracking and/or sensors configured to determine the orientation of the device in space.
- These devices can utilize input received from the input devices to select one or more virtual viewpoints within the light field image data and synthesize images from the virtual viewpoints.
- the devices allow for the exploration and interaction with the light field image data by dynamically synthesizing multiple images from multiple virtual viewpoints, providing a variety of views of the scene described in the light field image data.
- light fields contain additional information captured by an array camera, such as location data and/or data computed from information captured by an array camera.
- the synthesized image includes a depth map as metadata that can describe the distance to objects in a scene.
- a variety of user interfaces can be utilized to enable interaction with and exploration of light fields. Systems and methods for capturing light fields and synthesizing images from virtual viewpoints in accordance with embodiments of the invention are described below.
- FIG. 1 An array camera including an imager array in accordance with an embodiment of the invention is illustrated in FIG. 1 .
- the array camera 100 includes an imager array 102 configured to communicate with a processor 108 .
- the processor receives image data generated by the imager array.
- the imager array 102 includes multiple focal planes 104 . Imager arrays including multiple focal planes are discussed in U.S. patent application Ser. No. 13/106,797 entitled “Architectures for System on Chip Array Cameras,” to Pain et al., the disclosure of which is incorporated by reference in its entirety.
- the array camera module 150 includes an imager array 152 including an array of focal planes 154 along with a corresponding optic array 156 including an array of lens stacks 158 .
- each lens stack 158 creates an optical channel that forms an image of the scene on an array of light sensitive pixels within a corresponding focal plane 154 .
- Each pairing of a lens stack 158 and focal plane 154 forms a single camera 104 within the camera module.
- Each pixel within a focal plane 154 of a camera 104 generates image data that can be sent from the camera 104 to the processor 108 .
- the lens stack within each optical channel is configured so that pixels of each focal plane 158 sample the same object space or region within the scene.
- the lens stacks are configured so that the pixels that sample the same object space do so with sub-pixel offsets to provide sampling diversity that can be utilized to recover increased resolution through the use of super-resolution processes.
- sampling diversity refers to the fact that the images from different viewpoints sample the same object in the scene but with slight sub-pixel offsets.
- the focal planes are configured in a 5 ⁇ 5 array. In other embodiments, any of a variety of array configurations can be utilized including linear arrays.
- Each focal plane 154 of the imager array is capable of capturing an image of the scene.
- the sensor elements utilized in the focal planes can be individual light sensing elements such as, but not limited to, traditional CIS (CMOS Image Sensor) pixels, CCD (charge-coupled device) pixels, high dynamic range sensor elements, multispectral sensor elements and/or any other structure configured to generate an electrical signal indicative of light incident on the structure.
- the sensor elements of each focal plane have similar physical properties and receive light via the same optical channel and color filter (where present). In other embodiments, the sensor elements have different characteristics and, in many instances, the characteristics of the sensor elements are related to the color filter applied to each sensor element.
- color filters in individual cameras can be used to pattern the camera module with ⁇ filter groups as further discussed in U.S. Provisional patent application Ser. No. 13/875,248 entitled “Camera Modules Patterned with pi FilterGroups” filed May 1, 2013, the disclosure of which is incorporated by reference herein in its entirety.
- Any of a variety of color filter configurations can be utilized including the configuration in FIG. 1C including eight Green cameras, four Blue cameras, and four Red cameras, where the cameras are distributed around the center of the camera.
- the cameras can be used to capture data with respect to different colors, or a specific portion of the spectrum.
- color filters in many embodiments of the invention are included in the lens stack.
- a Green color camera can include a lens stack with a Green light filter that allows Green light to pass through the optical channel.
- the pixels in each focal plane are the same and the light information captured by the pixels is differentiated by the color filters in the corresponding lens stack for each filter plane.
- camera modules including it filter groups can be implemented in a variety of ways including (but not limited to) by applying color filters to the pixels of the focal planes of the camera module similar to the manner in which color filters are applied to the pixels of a conventional color camera.
- at least one of the cameras in the camera module can include uniform color filters applied to the pixels in its focal plane.
- a Bayer filter pattern is applied to the pixels of one of the cameras in a camera module.
- camera modules are constructed in which color filters are utilized in both the lens stacks and on the pixels of the imager array.
- the processor 108 is configured to take the image data captured by the sensor and synthesize high resolution images.
- the captured light field image data includes color information and depth information.
- the image is synthesized from a reference viewpoint, typically that of a reference focal plane 104 within the imager array 102 .
- an array of images is created using the image data captured by the focal planes in the imager array and can be considered to be a number of images of the scene captured from different viewpoints.
- the processor 108 is configured to take the image data captured by the sensor and create a light field containing the captured image data.
- the array of images is utilized by the processor 108 to synthesize high resolution images.
- the processor 108 is able to synthesize an image from a virtual viewpoint.
- a virtual viewpoint is any viewpoint which is not the reference viewpoint.
- the virtual viewpoint corresponds to a viewpoint of one of the focal planes 104 in the imager array 102 .
- the processor is able to synthesize an image from a virtual viewpoint, which does not correspond to any focal plane 104 in the imager array 102 .
- FIGS. 1A, 1B , and 1 C Although a specific array camera architecture is illustrated in FIGS. 1A, 1B , and 1 C, alternative architectures can also be utilized in accordance with embodiments of the invention. Systems and methods for synthesizing images from virtual viewpoints in accordance with embodiments of the invention are discussed below.
- FIG. 2 A diagram of a device capable of synthesizing and manipulating images synthesized from light field image data in accordance with an embodiment of the invention is illustrated in FIG. 2 .
- the device 200 contains a processor 210 capable of being configured via software (e.g. an image synthesis application) to load and manipulate images synthesized from light field image data.
- the image synthesis application configures the processor to synthesize images from virtual viewpoints based on the light field image data utilizing processes similar to those described below.
- the processor 210 is connected to a sensory array 220 capable of capturing light fields.
- An array camera includes the processor 210 and the imager array 220 .
- the processor is connected to a display 212 capable of displaying 2D images.
- the display 212 is capable of displaying 3D images.
- the display displays 3D images utilizing a stereo pair of images.
- the stereo pairs of images can include images synthesized from captured light field image data from different virtual viewpoints.
- the processor 210 is connected to a user interface 214 , such as a touchscreen interface, a pointing device, and/or a keyboard.
- the user interface 214 can be a camera or array camera capable of tracking user movements. As is described in more detail below, the user interface 214 can utilize input received from the input devices to select one or more virtual viewpoints within the light field image data and synthesize images from the virtual viewpoints.
- the processor 210 is connected to a storage device 216 .
- the storage device 216 is capable of storing light fields and light field image files and delivering those images to the processor 210 for manipulation.
- the processor 210 is connected to a network interface 218 capable of communication via a network. The network communication involves receiving and transmitting light fields and light field image files, where the light fields and light field image files can be stored in storage device 216 (if present) or loaded directly into the processor 210 .
- any device capable of synthesizing images from different viewpoints using light field image data and/or generating images from different viewpoints using image data contained within a light field image file can be used in accordance with embodiments of the invention.
- Methods for synthesizing images from a virtual viewpoint in a light field in accordance with embodiments of the invention are discussed below.
- an image can be synthesized from a reference viewpoint or from a virtual viewpoint in a captured light field.
- Images synthesized from a light field from a reference viewpoint and from virtual viewpoints in accordance with embodiments of the invention are illustrated in FIGS. 3A, 3B, and 3C .
- the image 310 captures a scene containing a person 317 , a box 318 , and a tree 319 .
- the image 310 has been captured from the reference viewpoint of a light field camera (not illustrated).
- Virtual viewpoints V 1 312 and V 2 314 are shown, where the dashed lines projection from virtual viewpoints V 1 312 and V 2 314 conceptually illustrate the sight lines from the virtual viewpoints.
- FIG. 3B An image synthesized from virtual viewpoint V 1 312 in accordance with an embodiment of the invention is illustrated in FIG. 3B .
- the image 320 illustrates the same person 317 , box 318 , and tree 319 as in light field image 310 .
- the person 317 appears to be located in front of box 318 and the tree 319 appears closer to box 318 than from the reference viewpoint of the light field camera.
- FIG. 3C An image synthesized from virtual viewpoint V 2 314 in accordance with embodiments of the invention is illustrated in FIG. 3C .
- the image 330 illustrates the same person 317 , box 318 , and tree 319 as in the image 310 and the image 320 .
- the person 317 appears to be located closer to the box 318 and further to the right of the box 318 than in the reference viewpoint or virtual viewpoint V 1 312 and the tree 319 appears closer to box 318 than from the reference viewpoint of the light field camera, yet further behind the box 318 than in the light field image synthesized using virtual viewpoint V 1 312 .
- FIGS. 3A, 3B , and 3 C Although a specific illustration of images synthesized from a variety of viewpoints in a captured light field is described above and illustrated in FIGS. 3A, 3B , and 3 C, the above is only an example and any variety of light fields and virtual viewpoints can be utilized in accordance with embodiments of the invention. Methods for synthesizing images from virtual viewpoints in light fields in accordance with embodiments of the invention are described below.
- An image can be synthesized from a variety of viewpoints within a captured light field.
- a process for capturing a light field and using the light field to synthesize an image from a virtual viewpoint in accordance with an embodiment of the invention is illustrated in FIG. 4 .
- the process 400 includes selecting ( 410 ) a virtual viewpoint in a light field.
- the displacement is calculated ( 412 ) for each image in the light field to the virtual viewpoint.
- the depth map is computed ( 414 ) for the virtual viewpoint and an image is synthesized ( 416 ) from the virtual viewpoint.
- a light field includes an array of images.
- the virtual viewpoint selected ( 410 ) corresponds to a camera (i.e. the viewpoint of an imager within an imager array) which captured one of the images in the array of images. In a number of embodiments, the virtual viewpoint selected ( 410 ) does not correspond to the viewpoints of any of the imagers within the imager array that captured the light field.
- synthesizing ( 416 ) an image utilizes color and depth information from the array of images. In several embodiments, synthesizing ( 416 ) an image results in a color image being created. A variety of processes can be utilized to synthesize ( 416 ) an image from a light field, including those disclosed in U.S. patent application Ser. No. 12/967,807 incorporated by reference above.
- images may be synthesized from light field image data.
- many devices lack the capability to process light fields directly. These devices can utilize light field image files in order to render images from virtual viewpoints without needing the capability to directly process a light field and/or to synthesize images from virtual viewpoints using super resolution processes.
- a process for rendering an image from a virtual viewpoint within a light field using the pixels, depth information, and information concerning occluded pixels contained within a light field image file is illustrated in FIG. 4B .
- the process 450 includes selecting ( 460 ) a virtual viewpoint within the light field. Light rays are projected ( 462 ). In a number of embodiments, a depth map is calculated ( 464 ).
- the depth map can be utilized in a variety of ways, including solid object detection in the synthesized image and/or determination of three-dimensional information for synthesizing a three-dimensional image.
- the image is generated ( 466 ).
- virtual viewpoints are selected ( 460 ) utilizing processes similar to those described below.
- the light rays are projected ( 462 ) from the virtual viewpoint selected ( 460 ) to the pixels and occluded pixels contained within the light field image file.
- some occluded pixels can be visible from the virtual viewpoint, but not from the reference viewpoint.
- some of the pixels visible from the reference viewpoint can be occluded pixels relative to the virtual viewpoint.
- the visible and occluded pixels are determined by computing the geometry of the pixels using depth information.
- a rendered image may not contain a pixel corresponding to a light ray projected from a virtual viewpoint (i.e. missing pixels).
- missing pixels can be estimated using probabilistic methods including, but not limited to, expectation-maximization, Markov networks, or texture synthesis methods such as inpainting. Cues such as color, depth, and saliency can be used to infer the missing pixels as appropriate to the requirements of specific applications in accordance with embodiments of the invention.
- Images can be generated ( 466 ) utilizing a variety of processes, including synthesizing images utilizing a super-resolution process (such as the processes described above) and rendering an image utilizing projected pixel information and depth information (such as the processes described below) in accordance with many embodiments of the invention.
- images 320 and 330 can be images rendered from the perspective of virtual viewpoints V 1 312 and V 2 314 using the pixels, depth information, and information concerning occluded pixels contained in a light field image file.
- a specific process for rendering an image based upon a light field image file is described above, a number of processes can be utilized for rendering images from light field image data utilizing virtual viewpoints in accordance with embodiments of the invention.
- Much of the discussion that follows references synthesizing images from virtual viewpoints.
- similar processes can be utilized in rendering an image from a light field image file in accordance with embodiments of the invention. Methods for projecting light rays and creating images from virtual viewpoints using captured light fields in accordance with embodiments of the invention are described below.
- the displacement of pixels and/or objects between the different viewpoints is calculated.
- Calculating the displacement of an object in a captured scene involves projecting light rays from a virtual viewpoint to the object.
- a process for projecting light rays in accordance with an embodiment of the invention is illustrated in FIG. 5 .
- the process 500 includes determining ( 510 ) the location of the camera. Displacements are determined ( 512 ) for each pixel in an object in the captured scene. The depth for each pixel in the object is determined ( 514 ). The pixels are mapped ( 516 ) to the appropriate location.
- the location of the camera is determined ( 510 ) based on a virtual viewpoint and a pinhole camera model.
- a pinhole camera model enables determining ( 512 ) displacements of one or more pixels utilizing the formula:
- ⁇ t x ,t y ,t z ⁇ are the displacements from camera C ij to the virtual viewpoint, and z is the distance from the camera C 1j to the object.
- the location of the camera is determined ( 510 ) utilizing a rotation matrix utilizing the formula:
- R is a 3 ⁇ 3 rotation matrix
- T is a 3 ⁇ 1 translation vector
- p is a three-dimensional point in space with respect to one camera
- the coordinates are the coordinates of p in the virtual camera, and the subscripts indicate the ith coordinate of vector v. In many embodiments, the coordinates are normalized for focal length and pixel size of the imager array capturing the light field image data.
- the depth for each pixel is determined ( 514 ) based on the location determined ( 510 ) of the camera.
- pixels are mapped ( 516 ) based on the determined ( 510 ) camera location and the determined ( 514 ) pixel depth information. Processes that can be utilized to determine ( 514 ) pixel depths and map ( 516 ) pixels in accordance with embodiments of the invention are described below.
- synthesizing an image from a virtual viewpoint using a light field having an array of images involves computing the depth for each image in the array of images from the virtual viewpoint.
- a process for computing the depth for each image in an array of images from a virtual viewpoint in accordance with an embodiment of the invention is illustrated in FIG. 6 .
- the process 600 involves shifting ( 610 ) images into a stack of images for a particular depth level.
- the variance of each pixel in the image stack is computed ( 612 ).
- Steps ( 610 ) and ( 612 ) are repeated until each depth level in the light field has been considered ( 614 ).
- the depth level is determined ( 616 ).
- determining ( 616 ) the depth level involves minimizing the variance for a particular pixel across the image stack.
- images are shifted to different positions using bilinear resampling, although a variety of resampling techniques may be utilized in accordance with embodiments of the invention.
- synthesizing a light field from a virtual viewpoint from a light field having an array of images involves projecting each pixel in the array of images based on the virtual viewpoint.
- a process for projecting pixels based on a virtual viewpoint in accordance with an embodiment of the invention is illustrated in FIG. 7 .
- the process 700 includes determining ( 710 ) the depth for each pixel using the virtual viewpoint. Pixels are projected ( 712 ) based on the virtual viewpoint. In many embodiments, if multiple pixels project to the same location, the pixels are combined ( 714 ).
- processes similar to those described above with regard to FIG. 6 are utilized to determine ( 710 ) the depth for each pixel.
- the process described above with regard to FIG. 5 is utilized to project ( 712 ) pixels based on the virtual viewpoint.
- the combination ( 714 ) of pixels utilizes the weighted average of the pixels.
- the weighted average of the pixels is the inverse of the distance from the camera from which the pixel originated to the virtual viewpoint.
- the projected ( 712 ) pixels are utilized to render an image from the virtual viewpoint.
- the projected ( 712 ) pixels are utilized to determine occlusion for one or more pixels based upon the projection ( 712 ) of the pixels from the virtual viewpoint.
- the projected ( 712 ) pixels are utilized to synthesize stereoscopic images (or video) from a captured light field image.
- the synthesis of stereoscopic images from virtual viewpoints includes synthesizing two or more images from different virtual viewpoints.
- two images are synthesized from two virtual viewpoints differing in their x-coordinates.
- the baseline e.g. the distance between optical centers of the two synthesized images
- These images can be rendered with parallel or converging optical axes as appropriate to the requirements of specific applications in accordance with embodiments of the invention.
- stereo images can be used for many applications, such as gaming, augmented reality using captured images, and vivid depth presentation of images on 3D displays. Stereo images can also be consumed on 2D displays, as the rapid alternation of left and right stereo views also gives a vivid depth sensation. Additionally, images synthesized from multiple virtual viewpoints are utilized to improve the depth map associated with the captured light field image data by providing additional information regarding the depth of particular pixels within the captured scene and/or by providing information regarding occluded pixels.
- the process 800 includes receiving ( 810 ) input.
- the input is received ( 810 ) using a touchscreen interface.
- the point selected on the touchscreen interface is used as the virtual viewpoint for the image.
- An image is synthesized ( 814 ) from a light field or rendered using a light field image file based upon the selected virtual viewpoint.
- a number of images are interpolated ( 812 ) between the original viewpoint and the selected virtual viewpoint.
- interpolating ( 812 ) the image involves smoothly transitioning from the original viewpoint to the virtual viewpoint.
- smoothly transitioning from the original viewpoint to the virtual viewpoint involves generating a number of images corresponding to a set of viewpoints interpolated ( 812 ) between the original viewpoint and the virtual viewpoint.
- interpolating ( 812 ) between viewpoints allows for a high degree of user interaction.
- devices capable of displaying and manipulating light fields and images rendered using light field image files include head and/or gaze tracking capabilities.
- the process 800 allows for users to view a variety of virtual viewpoints within a light field corresponding to where the user is looking.
- users may step into or out of a light field by generating images using virtual viewpoints in front of or behind the reference viewpoint; the detection of which can be performed using the head and/or gaze tracking capabilities of the device.
- the head and/or gaze tracking may be performed on a plurality of images simultaneously in order to provide a 3D experience.
- devices capable of manipulating light fields and images rendered using light field image files may include the capability of detecting the location of the device in space.
- These spatial location capabilities can be utilized using a variety of sensors as appropriate to the requirements of specific applications in accordance with embodiments of the invention, including, but not limited to, accelerometers, gyroscopes, and altimeters.
- Devices with spatial location determination abilities allow users to tilt and pan around the light field by generating images using virtual viewpoints determined using the spatial location ability of the device.
- devices capable of manipulating light fields and/or images rendered using light field image files contain several of the capabilities described above; these devices can combine the above described techniques in a variety of ways to allow users to determine virtual viewpoints for synthesizing light field images and/or images rendered using light field image files. Specific processes for interacting with a light field and/or an image rendered using a light field image file are described above; however, a variety of processes not described above can be utilized in accordance with embodiments of the invention.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Computing Systems (AREA)
- Image Processing (AREA)
- Studio Devices (AREA)
- Processing Or Creating Images (AREA)
- Image Generation (AREA)
Abstract
Description
- This application is a continuation of U.S. patent application Ser. No. 14/042,275, filed on Sep. 30, 2013, that claims the benefit of U.S. Provisional Patent Application No. 61/707,691, filed on Sep. 28, 2012, the disclosures of which are hereby incorporated by reference in their entirety.
- The present invention relates to systems and methods for generating images from light field image data and more specifically to systems and methods for generating images from light field image data using virtual viewpoints.
- Imaging devices, such as cameras, can be used to capture images of portions of the electromagnetic spectrum, such as the visible light spectrum, incident upon an image sensor. For ease of discussion, the term light is generically used to cover radiation across the entire electromagnetic spectrum. In a typical imaging device, light enters through an opening (aperture) at one end of the imaging device and is directed to an image sensor by one or more optical elements such as lenses. The image sensor includes pixels or sensor elements that generate signals upon receiving light via the optical element. Commonly used image sensors include charge-coupled device (CCDs) sensors and complementary metal-oxide semiconductor (CMOS) sensors.
- Image sensors are devices capable of converting an optical image into a digital signal. Image sensors utilized in digital cameras are made up of an array of pixels; the number of pixels determines the megapixel rating of the image sensor. For example, an image sensor having a width×height of 2272×1704 pixels would have an actual pixel count of 3,871,488 pixels and would be considered a 4 megapixel image sensor. Each pixel in an image sensor is capable of capturing light and converting the captured light into electrical signals. In order to separate the colors of light and capture a color image, a Bayer filter is often placed over the image sensor, filtering the incoming light into its red, blue, and green (RGB) components which are then captured by the image sensor. The RGB signal captured by the image sensor plus Bayer filter can then be processed and a color image can be created.
- Generally, image capture utilizes a single image sensor, to capture individual images, one at a time. A digital camera typically combines both an image sensor and processing capabilities. When the digital camera takes a photograph, the data captured by the image sensor is provided to the processor by the image sensor. Processors are able to control aspects of a captured image by changing image capture parameters of the sensor elements or groups of sensor elements used to capture the image.
- The ISO/IEC 10918-1 standard, more commonly referred to as the JPEG standard after the Joint Photographic Experts Group that developed the standard, establishes a standard process for digital compression and coding of still images. The JPEG standard specifies a codec for compressing an image into a bitstream and for decompressing the bitstream back into an image.
- Systems and methods for the synthesis of light field images from virtual viewpoints in accordance with embodiments of the invention are disclosed. In one embodiment of the invention, a system configured to synthesize images using captured light field image data includes a processor and a memory connected to the processor and configured to store captured light field image data and an image manipulation application, wherein the captured light field image data includes image data, pixel position data, and a depth map, wherein the depth map includes depth information for one or more pixels in the image data, and wherein the image manipulation application configures the processor to obtain captured light field image data, determine a virtual viewpoint for the captured light field image data based on the pixel position data and the depth map for the captured light field image data, where the virtual viewpoint includes a virtual location and virtual depth information, compute a virtual depth map based on the captured light field image data and the virtual viewpoint, and generate an image from the perspective of the virtual viewpoint based on the captured light field image data and the virtual depth map, where the generated image includes a plurality of pixels selected from the image data based on the pixel position data and the virtual depth map.
- In another embodiment of the invention, the virtual viewpoint corresponds to a focal plane in an array camera utilized to create the captured light field image data.
- In an additional embodiment of the invention, the captured light field image data further includes a reference viewpoint within the captured light field image data and the virtual viewpoint is a separate viewpoint within the captured light field image data from the reference viewpoint.
- In yet another additional embodiment of the invention, the captured light field image data was captured by an array camera having an imager array including a plurality of imagers and the reference viewpoint corresponds to the viewpoint of a first imager within the imager array in the array camera.
- In still another additional embodiment of the invention, the virtual viewpoint corresponds to the viewpoint of a second imager within the imager array, where the second imager is separate from the first imager.
- In yet still another additional embodiment of the invention, the virtual viewpoint is a viewpoint that does not correspond to the viewpoint of any of the imagers within the imager array.
- In yet another embodiment of the invention, the virtual viewpoint is selected from a position selected from the group consisting of in front of the imager array and behind the imager array.
- In still another embodiment of the invention, the image manipulation application further configures the processor to generate an image from the perspective of the virtual viewpoint by projecting pixels from the captured light field image data based on the pixel position data and the depth map, where the projected pixels are described in the image data and the depth map.
- In yet still another embodiment of the invention, the captured light field image data further includes occluded pixel information describing pixels not visible from a reference viewpoint of the captured light field image data and the projected pixels include at least one occluded pixel in the occluded pixel information that is visible from the perspective of the virtual viewpoint.
- In yet another additional embodiment of the invention, at least one projected pixel in the generated image is not described in the image data, the pixel position data, and the depth map and the image manipulation application further configures the processor to generate the at least one projected pixel by resampling the image data, the pixel position data, and the depth map.
- In still another additional embodiment of the invention, a pinhole camera model is utilized to project pixels within the generated image based on light rays projecting from the virtual viewpoint, where each projected pixels is associated with at least one of the projected light rays.
- In yet still another additional embodiment of the invention, projected pixel depth information is determined for at least one pixel in the generated image based on the depth map, the virtual viewpoint, and the light rays associated with the projected pixel.
- In yet another embodiment of the invention, the depth information for a projected pixel is based on minimizing the variance for the projected pixel across the image data within the captured light field image data.
- In still another embodiment of the invention, the image manipulation application further configures the processor to combine projected pixels having the same location within the generated image.
- In yet still another embodiment of the invention, the pixels are combined based on the weighted average of the pixels, where the weighted average of the pixels is the inverse of the distance from the imager from which the projected pixel originated to the virtual viewpoint.
- In yet another additional embodiment of the invention, the system further includes an input device configured to obtain input data indicative of a position within the captured light field image data.
- In still another additional embodiment of the invention, the input device is a touchscreen interface.
- In yet still another additional embodiment of the invention, the input device is a sensor configured to obtain spatial location information.
- In yet another embodiment of the invention, the input device is a camera configured to obtain input data selected from the group consisting of head tracking data and gaze tracking data.
- In still another embodiment of the invention, the virtual viewpoint is selected based on the input data.
- In yet still another embodiment of the invention, the image manipulation application further configures the processor to obtain a first input data indicative of a first position within the captured light field image data, determine a first virtual viewpoint based on the first input data, generate a first image from the perspective of the first virtual viewpoint, obtain a second input indicative of a second position within the captured light field image data, where the second position is separate from the first position, determine a second virtual viewpoint based on the second input data, generate at least one intermediate virtual viewpoint by interpolating between the first virtual viewpoint and the second virtual viewpoint, generate at least one intermediate image based on the generated at least one intermediate virtual viewpoints, where each intermediate image is from the perspective of an intermediate virtual viewpoint, and generate a second image from the perspective of the second virtual viewpoint.
- In yet another additional embodiment of the invention, the image is generated utilizing a super-resolution process.
- Still another embodiment of the invention includes a process for generating an image from a virtual viewpoint, including obtaining captured light field image data using an image manipulation device, where the captured light field image data includes image data, pixel position data, and a depth map and where the depth map includes depth information for one or more pixels in the image data, determining a virtual viewpoint for the captured light field image data based on the pixel position data and the depth map for the captured light field image data using the image manipulation device, where the virtual viewpoint includes a virtual location and virtual depth information, computing a virtual depth map based on the captured light field image data and the virtual viewpoint using the image manipulation device, and generating an image from the perspective of the virtual viewpoint based on the captured light field image data and the virtual depth map using the image manipulation device, where the generated image includes a plurality of pixels selected from the image data based on the pixel position data and the virtual depth map.
-
FIG. 1A conceptually illustrates an array camera including a 5×5 imager array connected with a processor in accordance with an embodiment of the invention. -
FIG. 1B conceptually illustrates a 5×5 array camera module in accordance with an embodiment of the invention. -
FIG. 1C conceptually illustrates a color filter pattern for a 4×4 array camera module in accordance with an embodiment of the invention. -
FIG. 2 is a diagram conceptually illustrating a device capable of processing light field images in accordance with an embodiment of the invention. -
FIG. 3A is a diagram conceptually illustrating virtual viewpoints for a given light field in accordance with an embodiment of the invention. -
FIG. 3B is a diagram conceptually illustrating a light field image rendered from a virtual viewpoint for a given light field in accordance with an embodiment of the invention. -
FIG. 3C is a diagram conceptually illustrating a light field image rendered from a second virtual viewpoint for a given light field in accordance with an embodiment of the invention. -
FIG. 4 is a flow chart conceptually illustrating a process for generating a light field image from a virtual viewpoint in a light field in accordance with an embodiment of the invention. -
FIG. 4B is a flow chart conceptually illustrating a process for generating a light field image from a virtual viewpoint using projected light rays in accordance with an embodiment of the invention. -
FIG. 5 is a flow chart conceptually illustrating a process for reprojecting light rays in relation to a virtual viewpoint in accordance with an embodiment of the invention. -
FIG. 6 is a flow chart conceptually illustrating a process for computing a depth map for a virtual viewpoint in accordance with an embodiment of the invention. -
FIG. 7 is a flow chart conceptually illustrating a process for projecting pixels to form a light field image corresponding to a virtual viewpoint in accordance with an embodiment of the invention. -
FIG. 8 is a flow chart conceptually illustrating a process for interactively generating light field images from virtual viewpoints in accordance with an embodiment of the invention. - Turning now to the drawings, systems and methods for generating images from light field image data using virtual viewpoints in accordance with embodiments of the invention are illustrated. A light field is often defined as a 4D function characterizing the light from all direction at all points in a scene and can be interpreted as a two-dimensional (2D) collection of 2D images of a scene. Array cameras, such as those described in U.S. patent application Ser. No. 12/935,504 entitled “Capturing and Processing of Images using Monolithic Camera Array with Heterogeneous Imagers” to Venkataraman et al., can be utilized to capture light fields. In a number of embodiments, super-resolution processes such as those described in U.S. patent application Ser. No. 12/967,807 entitled “Systems and Methods for Synthesizing High Resolution Images Using Super-Resolution Processes” to Lelescu et al., are utilized to synthesize a higher resolution 2D image or a stereo pair of higher resolution 2D images from the lower resolution images in the light field captured by an array camera. The terms high or higher resolution and low or lower resolution are used here in a relative sense and not to indicate the specific resolutions of the images captured by the array camera. The disclosures of U.S. patent application Ser. No. 12/935,504 and U.S. patent application Ser. No. 12/967,807 are hereby incorporated by reference in their entirety.
- A file containing an image synthesized from light field image data and metadata derived from the light field image data can be referred to as a light field image file. The encoded image in a light field image file is typically synthesized using a super resolution process from a number of lower resolution images. The light field image file can also include metadata describing the synthesized image derived from the light field image data that enables post processing of the synthesized image. In many embodiments, a light field image file is created by encoding an image synthesized from light field image data and combining the encoded image with a depth map derived from the light field image data. In several embodiments, the encoded image is synthesized from a reference viewpoint and the metadata includes information concerning pixels in the light field image that are occluded from the reference viewpoint. In a number of embodiments, the metadata can also include additional information including (but not limited to) auxiliary maps such as confidence maps, edge maps, occluded pixel information, and missing pixel maps that can be utilized during post processing of the encoded image to improve the quality of an image rendered using the light field image data file. By transmitting a light field image file including an encoded image, and metadata describing the encoded image, a rendering device (i.e. a device configured to generate an image rendered using the information within the light field image file) can render new images using the information within the file without the need to perform super resolution processing on the original light field image data. In this way, the amount of data transmitted to the rendering device and the computational complexity of rendering an image is reduced. In several embodiments, rendering devices are configured to perform processes including (but not limited to) refocusing the encoded image based upon a focal plane specified by the user, synthesizing an image from a different viewpoint, and generating a stereo pair of images. A variety of file formats may be utilized to store light field image files in accordance with embodiments of the invention. One such file format is the JPEG-DX extension to ISO/IEC 10918-1 described in U.S. Provisional Patent Application No. 61/540,188 entitled “JPEG-DX: A Backwards-compatible, Dynamic Focus Extension to JPEG”, to Venkataraman et al., filed Sep. 28, 2011, the entirety of which is incorporated by reference.
- An image is captured from the viewpoint of a camera. In many embodiments, array cameras capture at least one image of a scene from a reference viewpoint and the reference viewpoint is used to synthesize an image. Stated another way, the reference viewpoint corresponds to the viewpoint of one focal plane or a reference focal plane in an imager array. The term virtual viewpoint describes a viewpoint used to synthesize an image that may or may not correspond to one of the viewpoints of the focal planes in the imager array that capture the light field image data. Given an array of images taken using an array camera, an image from the perspective of a virtual viewpoint in a light field can be synthesized. In several embodiments, the viewpoint from which the image is synthesized can be the viewpoint of a focal plane in an imager array or from an interpolated position, such as a point between focal planes, in front of the imager array, or behind the imager array. In many embodiments of the invention, images are synthesized from virtual viewpoints. In a number of embodiments, the (virtual) viewpoint used to synthesize the image incorporates a rotation and/or translation of one or more of the optical axes of the viewpoint. In a variety of embodiments, devices capable of displaying images synthesized from light field image data include one or more input devices, such as cameras configured to allow for gaze and/or motion tracking and/or sensors configured to determine the orientation of the device in space. These devices can utilize input received from the input devices to select one or more virtual viewpoints within the light field image data and synthesize images from the virtual viewpoints. In this way, the devices allow for the exploration and interaction with the light field image data by dynamically synthesizing multiple images from multiple virtual viewpoints, providing a variety of views of the scene described in the light field image data.
- Using the 2D images in a captured light field, spatial X/Y luma/chroma values, and/or depth information for each pixel in each image captured by the array camera can be computed. In many embodiments, light fields contain additional information captured by an array camera, such as location data and/or data computed from information captured by an array camera. In many embodiments, the synthesized image includes a depth map as metadata that can describe the distance to objects in a scene. By utilizing depth information associated with each pixel in a light field, the light field can be utilized in an interactive manner in a variety of ways not possible with an image taken using a traditional camera. One way in which light fields can be utilized is the detection of solid objects; the solid objects are utilized in methods for synthesizing light fields in accordance with embodiments of the invention. Systems and methods for detecting objects in a light field are disclosed in U.S. patent application Ser. No. 12/967,807, the entirety of which is incorporated by reference above.
- A variety of user interfaces can be utilized to enable interaction with and exploration of light fields. Systems and methods for capturing light fields and synthesizing images from virtual viewpoints in accordance with embodiments of the invention are described below.
- Array cameras in accordance with embodiments of the invention are configured so that the array camera software can capture light fields and synthesize images from virtual viewpoints within the light field. An array camera including an imager array in accordance with an embodiment of the invention is illustrated in
FIG. 1 . Thearray camera 100 includes animager array 102 configured to communicate with aprocessor 108. The processor receives image data generated by the imager array. Theimager array 102 includes multiplefocal planes 104. Imager arrays including multiple focal planes are discussed in U.S. patent application Ser. No. 13/106,797 entitled “Architectures for System on Chip Array Cameras,” to Pain et al., the disclosure of which is incorporated by reference in its entirety. - Turning now to
FIG. 1B , an array camera module in accordance with an embodiment of the invention is conceptually illustrated. Thearray camera module 150 includes animager array 152 including an array offocal planes 154 along with a correspondingoptic array 156 including an array of lens stacks 158. Within the array of lens stacks, eachlens stack 158 creates an optical channel that forms an image of the scene on an array of light sensitive pixels within a correspondingfocal plane 154. Each pairing of alens stack 158 andfocal plane 154 forms asingle camera 104 within the camera module. Each pixel within afocal plane 154 of acamera 104 generates image data that can be sent from thecamera 104 to theprocessor 108. In many embodiments, the lens stack within each optical channel is configured so that pixels of eachfocal plane 158 sample the same object space or region within the scene. In several embodiments, the lens stacks are configured so that the pixels that sample the same object space do so with sub-pixel offsets to provide sampling diversity that can be utilized to recover increased resolution through the use of super-resolution processes. The term sampling diversity refers to the fact that the images from different viewpoints sample the same object in the scene but with slight sub-pixel offsets. By processing the images with sub-pixel precision, additional information encoded due to the sub-pixel offsets can be recovered when compared to simply sampling the object space with a single image. - In the illustrated embodiment, the focal planes are configured in a 5×5 array. In other embodiments, any of a variety of array configurations can be utilized including linear arrays. Each
focal plane 154 of the imager array is capable of capturing an image of the scene. The sensor elements utilized in the focal planes can be individual light sensing elements such as, but not limited to, traditional CIS (CMOS Image Sensor) pixels, CCD (charge-coupled device) pixels, high dynamic range sensor elements, multispectral sensor elements and/or any other structure configured to generate an electrical signal indicative of light incident on the structure. In many embodiments, the sensor elements of each focal plane have similar physical properties and receive light via the same optical channel and color filter (where present). In other embodiments, the sensor elements have different characteristics and, in many instances, the characteristics of the sensor elements are related to the color filter applied to each sensor element. - In a variety of embodiments, color filters in individual cameras can be used to pattern the camera module with π filter groups as further discussed in U.S. Provisional patent application Ser. No. 13/875,248 entitled “Camera Modules Patterned with pi FilterGroups” filed May 1, 2013, the disclosure of which is incorporated by reference herein in its entirety. Any of a variety of color filter configurations can be utilized including the configuration in
FIG. 1C including eight Green cameras, four Blue cameras, and four Red cameras, where the cameras are distributed around the center of the camera. The cameras can be used to capture data with respect to different colors, or a specific portion of the spectrum. In contrast to applying color filters to the pixels of the camera, color filters in many embodiments of the invention are included in the lens stack. For example, a Green color camera can include a lens stack with a Green light filter that allows Green light to pass through the optical channel. In many embodiments, the pixels in each focal plane are the same and the light information captured by the pixels is differentiated by the color filters in the corresponding lens stack for each filter plane. Although a specific construction of a camera module with an optic array including color filters in the lens stacks is described above, camera modules including it filter groups can be implemented in a variety of ways including (but not limited to) by applying color filters to the pixels of the focal planes of the camera module similar to the manner in which color filters are applied to the pixels of a conventional color camera. In several embodiments, at least one of the cameras in the camera module can include uniform color filters applied to the pixels in its focal plane. In many embodiments, a Bayer filter pattern is applied to the pixels of one of the cameras in a camera module. In a number of embodiments, camera modules are constructed in which color filters are utilized in both the lens stacks and on the pixels of the imager array. - In several embodiments, the
processor 108 is configured to take the image data captured by the sensor and synthesize high resolution images. In a number of embodiments, the captured light field image data includes color information and depth information. In many embodiments of the invention, the image is synthesized from a reference viewpoint, typically that of a referencefocal plane 104 within theimager array 102. In many embodiments, an array of images is created using the image data captured by the focal planes in the imager array and can be considered to be a number of images of the scene captured from different viewpoints. In several embodiments, theprocessor 108 is configured to take the image data captured by the sensor and create a light field containing the captured image data. In several embodiments, the array of images is utilized by theprocessor 108 to synthesize high resolution images. - The
processor 108 is able to synthesize an image from a virtual viewpoint. In a number of embodiments, a virtual viewpoint is any viewpoint which is not the reference viewpoint. In several embodiments, the virtual viewpoint corresponds to a viewpoint of one of thefocal planes 104 in theimager array 102. In many embodiments, the processor is able to synthesize an image from a virtual viewpoint, which does not correspond to anyfocal plane 104 in theimager array 102. - Although a specific array camera architecture is illustrated in
FIGS. 1A, 1B , and 1C, alternative architectures can also be utilized in accordance with embodiments of the invention. Systems and methods for synthesizing images from virtual viewpoints in accordance with embodiments of the invention are discussed below. - Users may utilize a variety of devices to interact with images synthesized using light field image data, including (but not limited to) devices with sensors capable of capturing light fields and synthesizing images. A diagram of a device capable of synthesizing and manipulating images synthesized from light field image data in accordance with an embodiment of the invention is illustrated in
FIG. 2 . Thedevice 200 contains aprocessor 210 capable of being configured via software (e.g. an image synthesis application) to load and manipulate images synthesized from light field image data. In several embodiments, the image synthesis application configures the processor to synthesize images from virtual viewpoints based on the light field image data utilizing processes similar to those described below. In many embodiments of the invention, theprocessor 210 is connected to asensory array 220 capable of capturing light fields. An array camera includes theprocessor 210 and theimager array 220. - In many embodiments, the processor is connected to a
display 212 capable of displaying 2D images. In several embodiments, thedisplay 212 is capable of displaying 3D images. In a variety of embodiments, the display displays 3D images utilizing a stereo pair of images. The stereo pairs of images can include images synthesized from captured light field image data from different virtual viewpoints. In a number of embodiments, theprocessor 210 is connected to auser interface 214, such as a touchscreen interface, a pointing device, and/or a keyboard. In several embodiments, theuser interface 214 can be a camera or array camera capable of tracking user movements. As is described in more detail below, theuser interface 214 can utilize input received from the input devices to select one or more virtual viewpoints within the light field image data and synthesize images from the virtual viewpoints. - In many embodiments, the
processor 210 is connected to astorage device 216. Thestorage device 216 is capable of storing light fields and light field image files and delivering those images to theprocessor 210 for manipulation. In a number of embodiments, theprocessor 210 is connected to anetwork interface 218 capable of communication via a network. The network communication involves receiving and transmitting light fields and light field image files, where the light fields and light field image files can be stored in storage device 216 (if present) or loaded directly into theprocessor 210. - Although specific devices capable of synthesizing and manipulating images synthesized from light field image data is described above with respect to
FIG. 2 , any device capable of synthesizing images from different viewpoints using light field image data and/or generating images from different viewpoints using image data contained within a light field image file can be used in accordance with embodiments of the invention. Methods for synthesizing images from a virtual viewpoint in a light field in accordance with embodiments of the invention are discussed below. - As described above, an image can be synthesized from a reference viewpoint or from a virtual viewpoint in a captured light field. Images synthesized from a light field from a reference viewpoint and from virtual viewpoints in accordance with embodiments of the invention are illustrated in
FIGS. 3A, 3B, and 3C . Theimage 310 captures a scene containing aperson 317, abox 318, and atree 319. Theimage 310 has been captured from the reference viewpoint of a light field camera (not illustrated).Virtual viewpoints V1 312 andV2 314 are shown, where the dashed lines projection fromvirtual viewpoints V1 312 andV2 314 conceptually illustrate the sight lines from the virtual viewpoints. - An image synthesized from
virtual viewpoint V1 312 in accordance with an embodiment of the invention is illustrated inFIG. 3B . Theimage 320 illustrates thesame person 317,box 318, andtree 319 as inlight field image 310. However, from the perspective ofvirtual viewpoint V1 312, theperson 317 appears to be located in front ofbox 318 and thetree 319 appears closer tobox 318 than from the reference viewpoint of the light field camera. - An image synthesized from
virtual viewpoint V2 314 in accordance with embodiments of the invention is illustrated inFIG. 3C . Theimage 330 illustrates thesame person 317,box 318, andtree 319 as in theimage 310 and theimage 320. However, from the perspective ofvirtual viewpoint V2 314, theperson 317 appears to be located closer to thebox 318 and further to the right of thebox 318 than in the reference viewpoint orvirtual viewpoint V1 312 and thetree 319 appears closer tobox 318 than from the reference viewpoint of the light field camera, yet further behind thebox 318 than in the light field image synthesized usingvirtual viewpoint V1 312. - Although a specific illustration of images synthesized from a variety of viewpoints in a captured light field is described above and illustrated in
FIGS. 3A, 3B , and 3C, the above is only an example and any variety of light fields and virtual viewpoints can be utilized in accordance with embodiments of the invention. Methods for synthesizing images from virtual viewpoints in light fields in accordance with embodiments of the invention are described below. - Synthesizing an Image from a Virtual Viewpoint in a Light Field
- An image can be synthesized from a variety of viewpoints within a captured light field. A process for capturing a light field and using the light field to synthesize an image from a virtual viewpoint in accordance with an embodiment of the invention is illustrated in
FIG. 4 . Theprocess 400 includes selecting (410) a virtual viewpoint in a light field. In a number of embodiments, the displacement is calculated (412) for each image in the light field to the virtual viewpoint. In many embodiments, the depth map is computed (414) for the virtual viewpoint and an image is synthesized (416) from the virtual viewpoint. - In several embodiments, a light field includes an array of images. In many embodiments of the invention, the virtual viewpoint selected (410) corresponds to a camera (i.e. the viewpoint of an imager within an imager array) which captured one of the images in the array of images. In a number of embodiments, the virtual viewpoint selected (410) does not correspond to the viewpoints of any of the imagers within the imager array that captured the light field. In many embodiments, synthesizing (416) an image utilizes color and depth information from the array of images. In several embodiments, synthesizing (416) an image results in a color image being created. A variety of processes can be utilized to synthesize (416) an image from a light field, including those disclosed in U.S. patent application Ser. No. 12/967,807 incorporated by reference above.
- A specific process for synthesizing an image from a virtual viewpoint in a light field is described above; however, a variety of processes for synthesizing an image from light field image data can be utilized in accordance with embodiments of the invention. Processes involving rendering images from different viewpoints using images and metadata contained within light field image files in accordance with embodiments of the invention are discussed further below.
- As described above, images may be synthesized from light field image data. However, many devices lack the capability to process light fields directly. These devices can utilize light field image files in order to render images from virtual viewpoints without needing the capability to directly process a light field and/or to synthesize images from virtual viewpoints using super resolution processes. A process for rendering an image from a virtual viewpoint within a light field using the pixels, depth information, and information concerning occluded pixels contained within a light field image file is illustrated in
FIG. 4B . Theprocess 450 includes selecting (460) a virtual viewpoint within the light field. Light rays are projected (462). In a number of embodiments, a depth map is calculated (464). The depth map can be utilized in a variety of ways, including solid object detection in the synthesized image and/or determination of three-dimensional information for synthesizing a three-dimensional image. The image is generated (466). - In a variety of embodiments, virtual viewpoints are selected (460) utilizing processes similar to those described below. In many embodiments, the light rays are projected (462) from the virtual viewpoint selected (460) to the pixels and occluded pixels contained within the light field image file. For example, from the virtual viewpoint, some occluded pixels can be visible from the virtual viewpoint, but not from the reference viewpoint. Likewise, some of the pixels visible from the reference viewpoint can be occluded pixels relative to the virtual viewpoint. In several embodiments, the visible and occluded pixels are determined by computing the geometry of the pixels using depth information. In many embodiments, a rendered image may not contain a pixel corresponding to a light ray projected from a virtual viewpoint (i.e. missing pixels). In a number of embodiments, a variety of resampling techniques, such as bilinear resampling, can be utilized in accordance with embodiments of the invention to create or fill missing pixels. Additionally, missing pixels can be estimated using probabilistic methods including, but not limited to, expectation-maximization, Markov networks, or texture synthesis methods such as inpainting. Cues such as color, depth, and saliency can be used to infer the missing pixels as appropriate to the requirements of specific applications in accordance with embodiments of the invention. Images can be generated (466) utilizing a variety of processes, including synthesizing images utilizing a super-resolution process (such as the processes described above) and rendering an image utilizing projected pixel information and depth information (such as the processes described below) in accordance with many embodiments of the invention.
- By way of example, referring to
FIGS. 3A, 3B, and 3C ,images virtual viewpoints V1 312 andV2 314 using the pixels, depth information, and information concerning occluded pixels contained in a light field image file. Although a specific process for rendering an image based upon a light field image file is described above, a number of processes can be utilized for rendering images from light field image data utilizing virtual viewpoints in accordance with embodiments of the invention. Much of the discussion that follows references synthesizing images from virtual viewpoints. As can be readily appreciated, similar processes can be utilized in rendering an image from a light field image file in accordance with embodiments of the invention. Methods for projecting light rays and creating images from virtual viewpoints using captured light fields in accordance with embodiments of the invention are described below. - In accordance with many embodiments of the invention, when rendering an image from a different perspective to an image synthesized from light field image data, the displacement of pixels and/or objects between the different viewpoints is calculated. Calculating the displacement of an object in a captured scene involves projecting light rays from a virtual viewpoint to the object. A process for projecting light rays in accordance with an embodiment of the invention is illustrated in
FIG. 5 . Theprocess 500 includes determining (510) the location of the camera. Displacements are determined (512) for each pixel in an object in the captured scene. The depth for each pixel in the object is determined (514). The pixels are mapped (516) to the appropriate location. - In many embodiments, the location of the camera is determined (510) based on a virtual viewpoint and a pinhole camera model. In several embodiments, a pinhole camera model enables determining (512) displacements of one or more pixels utilizing the formula:
-
- where, for camera C1j, pixel (x1, y1) maps to pixel (x2, y2), {tx,ty,tz} are the displacements from camera Cij to the virtual viewpoint, and z is the distance from the camera C1j to the object.
- In a variety of embodiments, the location of the camera is determined (510) utilizing a rotation matrix utilizing the formula:
-
- where R is a 3×3 rotation matrix, T is a 3×1 translation vector, p is a three-dimensional point in space with respect to one camera,
-
- are the coordinates of p in the virtual camera, and the subscripts indicate the ith coordinate of vector v. In many embodiments, the coordinates are normalized for focal length and pixel size of the imager array capturing the light field image data.
- In a variety of embodiments, the depth for each pixel is determined (514) based on the location determined (510) of the camera. In many embodiments, pixels are mapped (516) based on the determined (510) camera location and the determined (514) pixel depth information. Processes that can be utilized to determine (514) pixel depths and map (516) pixels in accordance with embodiments of the invention are described below.
- Although a specific process for projecting light rays is described above, a number of processes in accordance with embodiments of the invention can be utilized in accordance with embodiments of the invention, including those embodiments which calculate the displacement of an image without projecting light rays. Systems and methods for calculating depth and rendering images from virtual viewpoints using images synthesized from light field image data and depth maps in accordance with embodiments of the invention are described below.
- In accordance with several embodiments of the invention, synthesizing an image from a virtual viewpoint using a light field having an array of images involves computing the depth for each image in the array of images from the virtual viewpoint. A process for computing the depth for each image in an array of images from a virtual viewpoint in accordance with an embodiment of the invention is illustrated in
FIG. 6 . Theprocess 600 involves shifting (610) images into a stack of images for a particular depth level. The variance of each pixel in the image stack is computed (612). Steps (610) and (612) are repeated until each depth level in the light field has been considered (614). The depth level is determined (616). - In a number of embodiments, only the green images are shifted (610) into the stack of images; in other embodiments, any variety of subsets of images, including all images in the stack of images, may be shifted (610). In many embodiments, determining (616) the depth level involves minimizing the variance for a particular pixel across the image stack. In several embodiments, images are shifted to different positions using bilinear resampling, although a variety of resampling techniques may be utilized in accordance with embodiments of the invention.
- Although a specific process for computing the depth for an array of images is described above, a number of processes can be utilized in accordance with embodiments of the invention. In particular, processes that utilize depth information with respect to a reference viewpoint in an array of images can be utilized as appropriate to the requirements of specific applications in accordance with embodiments of the invention. Systems and methods for projecting pixels based on a virtual viewpoint and synthesizing images using virtual viewpoints in accordance with embodiments of the invention are described below.
- In accordance with embodiments of the invention, synthesizing a light field from a virtual viewpoint from a light field having an array of images involves projecting each pixel in the array of images based on the virtual viewpoint. A process for projecting pixels based on a virtual viewpoint in accordance with an embodiment of the invention is illustrated in
FIG. 7 . Theprocess 700 includes determining (710) the depth for each pixel using the virtual viewpoint. Pixels are projected (712) based on the virtual viewpoint. In many embodiments, if multiple pixels project to the same location, the pixels are combined (714). - In a number of embodiments, processes similar to those described above with regard to
FIG. 6 are utilized to determine (710) the depth for each pixel. In several embodiments, the process described above with regard toFIG. 5 is utilized to project (712) pixels based on the virtual viewpoint. In many embodiments, the combination (714) of pixels utilizes the weighted average of the pixels. In a number of embodiments, the weighted average of the pixels is the inverse of the distance from the camera from which the pixel originated to the virtual viewpoint. In many embodiments of the invention, the projected (712) pixels are utilized to render an image from the virtual viewpoint. In a number of embodiments, the projected (712) pixels are utilized to determine occlusion for one or more pixels based upon the projection (712) of the pixels from the virtual viewpoint. - In many embodiments, the projected (712) pixels are utilized to synthesize stereoscopic images (or video) from a captured light field image. The synthesis of stereoscopic images from virtual viewpoints includes synthesizing two or more images from different virtual viewpoints. In a variety of embodiments, two images are synthesized from two virtual viewpoints differing in their x-coordinates. The baseline (e.g. the distance between optical centers of the two synthesized images) can be selected to adjust the depth impression, comfort, proper display size, or optimal viewing distance of the stereoscopic image created using the synthesized images. These images can be rendered with parallel or converging optical axes as appropriate to the requirements of specific applications in accordance with embodiments of the invention.
- These stereoscopic images can be used for many applications, such as gaming, augmented reality using captured images, and vivid depth presentation of images on 3D displays. Stereo images can also be consumed on 2D displays, as the rapid alternation of left and right stereo views also gives a vivid depth sensation. Additionally, images synthesized from multiple virtual viewpoints are utilized to improve the depth map associated with the captured light field image data by providing additional information regarding the depth of particular pixels within the captured scene and/or by providing information regarding occluded pixels.
- Although specific processes for projecting pixels in two and three dimensions based on a virtual viewpoint is described above, a variety of processes can be utilized in accordance with embodiments of the invention. Systems and methods for manipulating virtual viewpoints and synthesizing images in accordance with embodiments of the invention are described below.
- When viewing a light field, it is often desirable to provide users with the ability to change the viewpoint of the light field. In other words, users have the ability to change the viewpoint within a light field of an image being displayed. A process for interacting with the viewpoint of an image in accordance with an embodiment of the invention is illustrated in
FIG. 8 . Theprocess 800 includes receiving (810) input. In many embodiments, the input is received (810) using a touchscreen interface. In several embodiments, the point selected on the touchscreen interface is used as the virtual viewpoint for the image. An image is synthesized (814) from a light field or rendered using a light field image file based upon the selected virtual viewpoint. - In many embodiments, a number of images are interpolated (812) between the original viewpoint and the selected virtual viewpoint. In several embodiments, interpolating (812) the image involves smoothly transitioning from the original viewpoint to the virtual viewpoint. In several embodiments, smoothly transitioning from the original viewpoint to the virtual viewpoint involves generating a number of images corresponding to a set of viewpoints interpolated (812) between the original viewpoint and the virtual viewpoint.
- In accordance with a number of embodiments of the invention, interpolating (812) between viewpoints allows for a high degree of user interaction. In many embodiments of the invention, devices capable of displaying and manipulating light fields and images rendered using light field image files include head and/or gaze tracking capabilities. In these embodiments, the
process 800 allows for users to view a variety of virtual viewpoints within a light field corresponding to where the user is looking. Furthermore, users may step into or out of a light field by generating images using virtual viewpoints in front of or behind the reference viewpoint; the detection of which can be performed using the head and/or gaze tracking capabilities of the device. In many embodiments, the head and/or gaze tracking may be performed on a plurality of images simultaneously in order to provide a 3D experience. - In accordance with embodiments of the invention, devices capable of manipulating light fields and images rendered using light field image files may include the capability of detecting the location of the device in space. These spatial location capabilities can be utilized using a variety of sensors as appropriate to the requirements of specific applications in accordance with embodiments of the invention, including, but not limited to, accelerometers, gyroscopes, and altimeters. Devices with spatial location determination abilities allow users to tilt and pan around the light field by generating images using virtual viewpoints determined using the spatial location ability of the device.
- In a number of embodiments, devices capable of manipulating light fields and/or images rendered using light field image files contain several of the capabilities described above; these devices can combine the above described techniques in a variety of ways to allow users to determine virtual viewpoints for synthesizing light field images and/or images rendered using light field image files. Specific processes for interacting with a light field and/or an image rendered using a light field image file are described above; however, a variety of processes not described above can be utilized in accordance with embodiments of the invention.
- Although the present invention has been described in certain specific aspects, many additional modifications and variations would be apparent to those skilled in the art. It is therefore to be understood that the present invention can be practiced otherwise than specifically described. Thus, embodiments of the present invention should be considered in all respects as illustrative and not restrictive. Accordingly, the scope of the invention should be determined not by the embodiments illustrated, but by the appended claims and their equivalents.
Claims (21)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/876,024 US10390005B2 (en) | 2012-09-28 | 2015-10-06 | Generating images from light fields utilizing virtual viewpoints |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201261707691P | 2012-09-28 | 2012-09-28 | |
US14/042,275 US20140092281A1 (en) | 2012-09-28 | 2013-09-30 | Generating Images from Light Fields Utilizing Virtual Viewpoints |
US14/876,024 US10390005B2 (en) | 2012-09-28 | 2015-10-06 | Generating images from light fields utilizing virtual viewpoints |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/042,275 Continuation US20140092281A1 (en) | 2012-09-28 | 2013-09-30 | Generating Images from Light Fields Utilizing Virtual Viewpoints |
Publications (2)
Publication Number | Publication Date |
---|---|
US20160255333A1 true US20160255333A1 (en) | 2016-09-01 |
US10390005B2 US10390005B2 (en) | 2019-08-20 |
Family
ID=50384817
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/042,275 Abandoned US20140092281A1 (en) | 2012-09-28 | 2013-09-30 | Generating Images from Light Fields Utilizing Virtual Viewpoints |
US14/876,024 Active US10390005B2 (en) | 2012-09-28 | 2015-10-06 | Generating images from light fields utilizing virtual viewpoints |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/042,275 Abandoned US20140092281A1 (en) | 2012-09-28 | 2013-09-30 | Generating Images from Light Fields Utilizing Virtual Viewpoints |
Country Status (4)
Country | Link |
---|---|
US (2) | US20140092281A1 (en) |
EP (2) | EP4307659A1 (en) |
CN (1) | CN104685860A (en) |
WO (1) | WO2014052974A2 (en) |
Cited By (61)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9706132B2 (en) | 2012-05-01 | 2017-07-11 | Fotonation Cayman Limited | Camera modules patterned with pi filter groups |
US9712759B2 (en) | 2008-05-20 | 2017-07-18 | Fotonation Cayman Limited | Systems and methods for generating depth maps using a camera arrays incorporating monochrome and color cameras |
US9743051B2 (en) | 2013-02-24 | 2017-08-22 | Fotonation Cayman Limited | Thin form factor computational array cameras and modular array cameras |
US9749568B2 (en) | 2012-11-13 | 2017-08-29 | Fotonation Cayman Limited | Systems and methods for array camera focal plane control |
US9749547B2 (en) | 2008-05-20 | 2017-08-29 | Fotonation Cayman Limited | Capturing and processing of images using camera array incorperating Bayer cameras having different fields of view |
US9754422B2 (en) | 2012-02-21 | 2017-09-05 | Fotonation Cayman Limited | Systems and method for performing depth based image editing |
US9774789B2 (en) | 2013-03-08 | 2017-09-26 | Fotonation Cayman Limited | Systems and methods for high dynamic range imaging using array cameras |
US9794476B2 (en) | 2011-09-19 | 2017-10-17 | Fotonation Cayman Limited | Systems and methods for controlling aliasing in images captured by an array camera for use in super resolution processing using pixel apertures |
US9800856B2 (en) | 2013-03-13 | 2017-10-24 | Fotonation Cayman Limited | Systems and methods for synthesizing images from image data captured by an array camera using restricted depth of field depth maps in which depth estimation precision varies |
US9807382B2 (en) | 2012-06-28 | 2017-10-31 | Fotonation Cayman Limited | Systems and methods for detecting defective camera arrays and optic arrays |
US9811753B2 (en) | 2011-09-28 | 2017-11-07 | Fotonation Cayman Limited | Systems and methods for encoding light field image files |
US9813616B2 (en) | 2012-08-23 | 2017-11-07 | Fotonation Cayman Limited | Feature based high resolution motion estimation from low resolution images captured using an array source |
US9858673B2 (en) | 2012-08-21 | 2018-01-02 | Fotonation Cayman Limited | Systems and methods for estimating depth and visibility from a reference viewpoint for pixels in a set of images captured from different viewpoints |
US9888194B2 (en) | 2013-03-13 | 2018-02-06 | Fotonation Cayman Limited | Array camera architecture implementing quantum film image sensors |
US9898856B2 (en) | 2013-09-27 | 2018-02-20 | Fotonation Cayman Limited | Systems and methods for depth-assisted perspective distortion correction |
US9924092B2 (en) | 2013-11-07 | 2018-03-20 | Fotonation Cayman Limited | Array cameras incorporating independently aligned lens stacks |
US9942474B2 (en) | 2015-04-17 | 2018-04-10 | Fotonation Cayman Limited | Systems and methods for performing high speed video capture and depth estimation using array cameras |
US9955070B2 (en) | 2013-03-15 | 2018-04-24 | Fotonation Cayman Limited | Systems and methods for synthesizing high resolution images using image deconvolution based on motion and depth information |
US9986224B2 (en) | 2013-03-10 | 2018-05-29 | Fotonation Cayman Limited | System and methods for calibration of an array camera |
WO2018094804A1 (en) * | 2016-11-23 | 2018-05-31 | 深圳市元征科技股份有限公司 | Image processing method and device |
US10009538B2 (en) | 2013-02-21 | 2018-06-26 | Fotonation Cayman Limited | Systems and methods for generating compressed light field representation data using captured light fields, array geometry, and parallax information |
US10091405B2 (en) | 2013-03-14 | 2018-10-02 | Fotonation Cayman Limited | Systems and methods for reducing motion blur in images or video in ultra low light with array cameras |
US10089740B2 (en) | 2014-03-07 | 2018-10-02 | Fotonation Limited | System and methods for depth regularization and semiautomatic interactive matting using RGB-D images |
US10119808B2 (en) | 2013-11-18 | 2018-11-06 | Fotonation Limited | Systems and methods for estimating depth from projected texture using camera arrays |
US10122993B2 (en) | 2013-03-15 | 2018-11-06 | Fotonation Limited | Autofocus system for a conventional camera that uses depth information from an array camera |
US10127682B2 (en) | 2013-03-13 | 2018-11-13 | Fotonation Limited | System and methods for calibration of an array camera |
US10182216B2 (en) | 2013-03-15 | 2019-01-15 | Fotonation Limited | Extended color processing on pelican array cameras |
US10218889B2 (en) | 2011-05-11 | 2019-02-26 | Fotonation Limited | Systems and methods for transmitting and receiving array camera image data |
US10250871B2 (en) | 2014-09-29 | 2019-04-02 | Fotonation Limited | Systems and methods for dynamic calibration of array cameras |
US10261219B2 (en) | 2012-06-30 | 2019-04-16 | Fotonation Limited | Systems and methods for manufacturing camera modules using active alignment of lens stack arrays and sensors |
US10306120B2 (en) | 2009-11-20 | 2019-05-28 | Fotonation Limited | Capturing and processing of images captured by camera arrays incorporating cameras with telephoto and conventional lenses to generate depth maps |
US10366472B2 (en) | 2010-12-14 | 2019-07-30 | Fotonation Limited | Systems and methods for synthesizing high resolution images using images captured by an array of independently controllable imagers |
US10388069B2 (en) | 2015-09-09 | 2019-08-20 | Futurewei Technologies, Inc. | Methods and systems for light field augmented reality/virtual reality on mobile devices |
US10412314B2 (en) | 2013-03-14 | 2019-09-10 | Fotonation Limited | Systems and methods for photometric normalization in array cameras |
US10455168B2 (en) | 2010-05-12 | 2019-10-22 | Fotonation Limited | Imager array interfaces |
US10455218B2 (en) | 2013-03-15 | 2019-10-22 | Fotonation Limited | Systems and methods for estimating depth using stereo array cameras |
US10482618B2 (en) | 2017-08-21 | 2019-11-19 | Fotonation Limited | Systems and methods for hybrid depth regularization |
CN110637273A (en) * | 2017-05-10 | 2019-12-31 | 微软技术许可有限责任公司 | Presenting applications within a virtual environment |
WO2020040521A1 (en) * | 2018-08-21 | 2020-02-27 | 삼성전자 주식회사 | Method for synthesizing intermediate view of light field, system for synthesizing intermediate view of light field, and method for compressing light field |
KR20200021891A (en) * | 2018-08-21 | 2020-03-02 | 삼성전자주식회사 | Method for the synthesis of intermediate views of a light field, system for the synthesis of intermediate views of a light field, and method for the compression of a light field |
US10708492B2 (en) | 2013-11-26 | 2020-07-07 | Fotonation Limited | Array camera configurations incorporating constituent array cameras and constituent cameras |
CN111669603A (en) * | 2019-03-07 | 2020-09-15 | 阿里巴巴集团控股有限公司 | Multi-angle free visual angle data processing method and device, medium, terminal and equipment |
US11143865B1 (en) * | 2017-12-05 | 2021-10-12 | Apple Inc. | Lens array for shifting perspective of an imaging system |
US11182872B2 (en) * | 2018-11-02 | 2021-11-23 | Electronics And Telecommunications Research Institute | Plenoptic data storage system and operating method thereof |
US11270110B2 (en) | 2019-09-17 | 2022-03-08 | Boston Polarimetrics, Inc. | Systems and methods for surface modeling using polarization cues |
US11290658B1 (en) | 2021-04-15 | 2022-03-29 | Boston Polarimetrics, Inc. | Systems and methods for camera exposure control |
US11302012B2 (en) | 2019-11-30 | 2022-04-12 | Boston Polarimetrics, Inc. | Systems and methods for transparent object segmentation using polarization cues |
US11521347B2 (en) | 2019-03-07 | 2022-12-06 | Alibaba Group Holding Limited | Method, apparatus, medium, and device for generating multi-angle free-respective image data |
US11525906B2 (en) | 2019-10-07 | 2022-12-13 | Intrinsic Innovation Llc | Systems and methods for augmentation of sensor systems and imaging systems with polarization |
US11580667B2 (en) | 2020-01-29 | 2023-02-14 | Intrinsic Innovation Llc | Systems and methods for characterizing object pose detection and measurement systems |
US11689813B2 (en) | 2021-07-01 | 2023-06-27 | Intrinsic Innovation Llc | Systems and methods for high dynamic range imaging using crossed polarizers |
US11792538B2 (en) | 2008-05-20 | 2023-10-17 | Adeia Imaging Llc | Capturing and processing of images including occlusions focused on an image sensor by a lens stack array |
US11797863B2 (en) | 2020-01-30 | 2023-10-24 | Intrinsic Innovation Llc | Systems and methods for synthesizing data for training statistical models on different imaging modalities including polarized images |
US11953700B2 (en) | 2020-05-27 | 2024-04-09 | Intrinsic Innovation Llc | Multi-aperture polarization optical systems using beam splitters |
US11954886B2 (en) | 2021-04-15 | 2024-04-09 | Intrinsic Innovation Llc | Systems and methods for six-degree of freedom pose estimation of deformable objects |
US12020455B2 (en) | 2021-03-10 | 2024-06-25 | Intrinsic Innovation Llc | Systems and methods for high dynamic range image reconstruction |
US12067746B2 (en) | 2021-05-07 | 2024-08-20 | Intrinsic Innovation Llc | Systems and methods for using computer vision to pick up small objects |
US12069227B2 (en) | 2021-03-10 | 2024-08-20 | Intrinsic Innovation Llc | Multi-modal and multi-spectral stereo camera arrays |
US12172310B2 (en) | 2021-06-29 | 2024-12-24 | Intrinsic Innovation Llc | Systems and methods for picking objects using 3-D geometry and segmentation |
US12175741B2 (en) | 2021-06-22 | 2024-12-24 | Intrinsic Innovation Llc | Systems and methods for a vision guided end effector |
US12293535B2 (en) | 2021-08-03 | 2025-05-06 | Intrinsic Innovation Llc | Systems and methods for training pose estimators in computer vision |
Families Citing this family (59)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130265459A1 (en) | 2011-06-28 | 2013-10-10 | Pelican Imaging Corporation | Optical arrangements for use with an array camera |
US9113043B1 (en) | 2011-10-24 | 2015-08-18 | Disney Enterprises, Inc. | Multi-perspective stereoscopy from light fields |
US9165401B1 (en) * | 2011-10-24 | 2015-10-20 | Disney Enterprises, Inc. | Multi-perspective stereoscopy from light fields |
US9214013B2 (en) | 2012-09-14 | 2015-12-15 | Pelican Imaging Corporation | Systems and methods for correcting user identified artifacts in light field images |
EP4307659A1 (en) | 2012-09-28 | 2024-01-17 | Adeia Imaging LLC | Generating images from light fields utilizing virtual viewpoints |
US9521416B1 (en) | 2013-03-11 | 2016-12-13 | Kip Peli P1 Lp | Systems and methods for image data compression |
US9106784B2 (en) | 2013-03-13 | 2015-08-11 | Pelican Imaging Corporation | Systems and methods for controlling aliasing in images captured by an array camera for use in super-resolution processing |
WO2014150856A1 (en) | 2013-03-15 | 2014-09-25 | Pelican Imaging Corporation | Array camera implementing quantum dot color filters |
US9633442B2 (en) | 2013-03-15 | 2017-04-25 | Fotonation Cayman Limited | Array cameras including an array camera module augmented with a separate camera |
US10242400B1 (en) | 2013-10-25 | 2019-03-26 | Appliance Computing III, Inc. | User interface for image-based rendering of virtual tours |
US9686471B2 (en) * | 2013-11-01 | 2017-06-20 | Light Labs Inc. | Methods and apparatus relating to image stabilization |
US9390505B2 (en) * | 2013-12-12 | 2016-07-12 | Qualcomm Incorporated | Method and apparatus for generating plenoptic depth maps |
US10244223B2 (en) | 2014-01-10 | 2019-03-26 | Ostendo Technologies, Inc. | Methods for full parallax compressed light field 3D imaging systems |
US9247117B2 (en) | 2014-04-07 | 2016-01-26 | Pelican Imaging Corporation | Systems and methods for correcting for warpage of a sensor array in an array camera module by introducing warpage into a focal plane of a lens stack array |
US9521319B2 (en) | 2014-06-18 | 2016-12-13 | Pelican Imaging Corporation | Array cameras and array camera modules including spectral filters disposed outside of a constituent image sensor |
US9507995B2 (en) | 2014-08-29 | 2016-11-29 | X Development Llc | Combination of stereo and structured-light processing |
EP3070942B1 (en) | 2015-03-17 | 2023-11-22 | InterDigital CE Patent Holdings | Method and apparatus for displaying light field video data |
KR20170140187A (en) | 2015-04-23 | 2017-12-20 | 오스텐도 테크놀로지스 인코포레이티드 | Method for fully parallax compression optical field synthesis using depth information |
WO2016172384A1 (en) | 2015-04-23 | 2016-10-27 | Ostendo Technologies, Inc. | Methods and apparatus for full parallax light field display systems |
EP3107007B1 (en) | 2015-06-17 | 2020-05-20 | InterDigital CE Patent Holdings | Method and apparatus for data retrieval in a lightfield database |
US9911225B2 (en) | 2015-09-29 | 2018-03-06 | Siemens Healthcare Gmbh | Live capturing of light map image sequences for image-based lighting of medical data |
US9549174B1 (en) | 2015-10-14 | 2017-01-17 | Zspace, Inc. | Head tracked stereoscopic display system that uses light field type data |
US10448030B2 (en) * | 2015-11-16 | 2019-10-15 | Ostendo Technologies, Inc. | Content adaptive light field compression |
KR102446442B1 (en) | 2015-11-24 | 2022-09-23 | 삼성전자주식회사 | Digital photographing apparatus and method of operation thereof |
CN105704476B (en) * | 2016-01-14 | 2018-03-20 | 东南大学 | A kind of virtual visual point image frequency domain fast acquiring method based on edge reparation |
US10003783B2 (en) * | 2016-02-26 | 2018-06-19 | Infineon Technologies Ag | Apparatus for generating a three-dimensional color image and a method for producing a three-dimensional color image |
EP3220351A1 (en) * | 2016-03-14 | 2017-09-20 | Thomson Licensing | Method and device for processing lightfield data |
JP6911045B2 (en) | 2016-04-12 | 2021-07-28 | クイッディエント・エルエルシー | Everyday scene restoration engine |
US10453431B2 (en) | 2016-04-28 | 2019-10-22 | Ostendo Technologies, Inc. | Integrated near-far light field display systems |
US11612307B2 (en) | 2016-11-24 | 2023-03-28 | University Of Washington | Light field capture and rendering for head-mounted displays |
CN107424187B (en) * | 2017-04-17 | 2023-10-24 | 奥比中光科技集团股份有限公司 | Depth calculation processor, data processing method and 3D image device |
EP3416371A1 (en) * | 2017-06-12 | 2018-12-19 | Thomson Licensing | Method for displaying, on a 2d display device, a content derived from light field data |
EP3416381A1 (en) | 2017-06-12 | 2018-12-19 | Thomson Licensing | Method and apparatus for providing information to a user observing a multi view content |
EP3698324B1 (en) * | 2017-10-20 | 2022-09-21 | Toyota Motor Europe | Method and system for processing an image and determining viewpoints of objects |
CN107580207A (en) * | 2017-10-31 | 2018-01-12 | 武汉华星光电技术有限公司 | The generation method and generating means of light field 3D display cell picture |
EP3779892A4 (en) * | 2018-04-12 | 2021-05-05 | Toppan Printing Co., Ltd. | Light-field image generation system, image display system, shape information acquisition server, image generation server, display device, light-field image generation method and image display method |
CN112470189B (en) * | 2018-04-17 | 2024-03-29 | 上海科技大学 | Occlusion cancellation for light field systems |
WO2019213450A1 (en) | 2018-05-02 | 2019-11-07 | Quidient, Llc | A codec for processing scenes of almost unlimited detail |
KR102545980B1 (en) | 2018-07-19 | 2023-06-21 | 액티브 서지컬, 인크. | Systems and methods for multi-modal sensing of depth in vision systems for automated surgical robots |
US10623660B1 (en) | 2018-09-27 | 2020-04-14 | Eloupes, Inc. | Camera array for a mediated-reality system |
CN113993474A (en) | 2018-12-28 | 2022-01-28 | 艾科缇弗外科公司 | Systems and methods for optimizing accessibility, workspace, and dexterity in minimally invasive surgery |
US10565773B1 (en) | 2019-01-15 | 2020-02-18 | Nokia Technologies Oy | Efficient light field video streaming |
KR102461919B1 (en) | 2019-04-01 | 2022-11-01 | 구글 엘엘씨 | Technology to capture and edit dynamic depth images |
JP2022526626A (en) | 2019-04-08 | 2022-05-25 | アクティブ サージカル, インコーポレイテッド | Systems and methods for medical imaging |
US12292564B2 (en) | 2019-04-08 | 2025-05-06 | Activ Surgical, Inc. | Systems and methods for medical imaging |
WO2020214821A1 (en) | 2019-04-19 | 2020-10-22 | Activ Surgical, Inc. | Systems and methods for trocar kinematics |
CN110246146B (en) * | 2019-04-29 | 2021-07-30 | 北京邮电大学 | Method and device for generating full parallax light field content based on multiple depth image rendering |
CN112087614A (en) * | 2019-06-12 | 2020-12-15 | 上海麦界信息技术有限公司 | Method, device and computer readable medium for generating two-dimensional light field image |
EP4017340A4 (en) | 2019-08-21 | 2023-12-13 | Activ Surgical, Inc. | Systems and methods for medical imaging |
EP3816942A1 (en) * | 2019-10-29 | 2021-05-05 | Koninklijke Philips N.V. | An image processing method for setting transparency values and color values of pixels in a virtual image |
CN111427166B (en) * | 2020-03-31 | 2022-07-05 | 京东方科技集团股份有限公司 | Light field display method and system, storage medium and display panel |
US10949986B1 (en) | 2020-05-12 | 2021-03-16 | Proprio, Inc. | Methods and systems for imaging a scene, such as a medical scene, and tracking objects within the scene |
WO2021249955A1 (en) * | 2020-06-09 | 2021-12-16 | Interdigital Vc Holdings France, Sas | Metadata to describe color filter on top on camera sensors |
US11651475B2 (en) * | 2020-07-24 | 2023-05-16 | Samsung Electronics Co., Ltd. | Image restoration method and device |
CN113724309B (en) * | 2021-08-27 | 2024-06-14 | 杭州海康威视数字技术股份有限公司 | Image generation method, device, equipment and storage medium |
US12261988B2 (en) | 2021-11-08 | 2025-03-25 | Proprio, Inc. | Methods for generating stereoscopic views in multicamera systems, and associated devices and systems |
US20220198768A1 (en) * | 2022-03-09 | 2022-06-23 | Intel Corporation | Methods and apparatus to control appearance of views in free viewpoint media |
JP7626104B2 (en) * | 2022-06-20 | 2025-02-04 | トヨタ自動車株式会社 | System and terminal device |
CN116528065B (en) * | 2023-06-30 | 2023-09-26 | 深圳臻像科技有限公司 | An efficient light field acquisition and generation method for virtual scene content |
Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6097394A (en) * | 1997-04-28 | 2000-08-01 | Board Of Trustees, Leland Stanford, Jr. University | Method and system for light field rendering |
US20020006113A1 (en) * | 2000-07-06 | 2002-01-17 | Young Choi | Switching circuit for board duplication apparatus |
US20020061131A1 (en) * | 2000-10-18 | 2002-05-23 | Sawhney Harpreet Singh | Method and apparatus for synthesizing new video and/or still imagery from a collection of real video and/or still imagery |
US20040019637A1 (en) * | 2002-07-26 | 2004-01-29 | International Business Machines Corporaion | Interactive one to many communication in a cooperating community of users |
US20040196379A1 (en) * | 2003-04-04 | 2004-10-07 | Stmicroelectronics, Inc. | Compound camera and methods for implementing auto-focus, depth-of-field and high-resolution functions |
US20050219264A1 (en) * | 2004-04-01 | 2005-10-06 | Microsoft Corporation | Pop-up light field |
US20100141802A1 (en) * | 2008-12-08 | 2010-06-10 | Timothy Knight | Light Field Data Acquisition Devices, and Methods of Using and Manufacturing Same |
US20110255592A1 (en) * | 2008-10-27 | 2011-10-20 | Lg Electronics Inc. | Virtual view image synthesis method and apparatus |
US20110267348A1 (en) * | 2010-04-29 | 2011-11-03 | Dennis Lin | Systems and methods for generating a virtual camera viewpoint for an image |
US20120057040A1 (en) * | 2010-05-11 | 2012-03-08 | Byung Kwan Park | Apparatus and method for processing light field data using a mask with an attenuation pattern |
US20120062756A1 (en) * | 2004-12-17 | 2012-03-15 | Dong Tian | Method and System for Processing Multiview Videos for View Synthesis Using Skip and Direct Modes |
US8842201B2 (en) * | 2011-01-06 | 2014-09-23 | Sony Corporation | Image capturing device including lens array and processing |
Family Cites Families (833)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4124798A (en) | 1965-12-09 | 1978-11-07 | Thompson Kenneth B | Optical viewing apparatus |
US4198646A (en) | 1978-10-13 | 1980-04-15 | Hughes Aircraft Company | Monolithic imager for near-IR |
US4323925A (en) | 1980-07-07 | 1982-04-06 | Avco Everett Research Laboratory, Inc. | Method and apparatus for arraying image sensor modules |
JPS5769476A (en) | 1980-10-16 | 1982-04-28 | Fuji Xerox Co Ltd | Reader control system |
JPS5925483A (en) | 1982-08-04 | 1984-02-09 | Hitachi Denshi Ltd | solid-state imaging device |
US4652909A (en) | 1982-09-14 | 1987-03-24 | New York Institute Of Technology | Television camera and recording system for high definition television having imagers of different frame rate |
US4460449A (en) | 1983-01-03 | 1984-07-17 | Amerace Corporation | Apparatus for making a tool |
JPS6140709A (en) | 1984-08-02 | 1986-02-27 | みのる産業株式会社 | Rotary mesh seeder |
EP0289885A1 (en) | 1987-05-08 | 1988-11-09 | Siemens Aktiengesellschaft | Aperture system for production of several partical probes with changeable cross-section |
JPS6437177A (en) | 1987-08-03 | 1989-02-07 | Canon Kk | Image pickup device |
DE58902538D1 (en) | 1988-05-19 | 1992-12-03 | Siemens Ag | METHOD FOR OBSERVING A SCENE AND DEVICE FOR IMPLEMENTING THE METHOD. |
JPH02285772A (en) | 1989-04-26 | 1990-11-26 | Toshiba Corp | Picture reader |
US5070414A (en) | 1988-09-20 | 1991-12-03 | Kabushiki Kaisha Toshiba | Method and apparatus for reading image information formed on material |
US5157499A (en) | 1990-06-29 | 1992-10-20 | Kabushiki Kaisha N A C | High-speed video camera using solid-state image sensor |
US5144448A (en) | 1990-07-31 | 1992-09-01 | Vidar Systems Corporation | Scanning apparatus using multiple CCD arrays and related method |
US5325449A (en) | 1992-05-15 | 1994-06-28 | David Sarnoff Research Center, Inc. | Method for fusing images and apparatus therefor |
JP3032382B2 (en) | 1992-07-13 | 2000-04-17 | シャープ株式会社 | Digital signal sampling frequency converter |
JPH06129851A (en) | 1992-10-13 | 1994-05-13 | Sumitomo Electric Ind Ltd | Calibration method for stereo camera |
WO1994020875A2 (en) | 1993-03-03 | 1994-09-15 | Street Graham S B | Method and apparatus for image alignment |
US5659424A (en) | 1993-05-25 | 1997-08-19 | Hitachi, Ltd. | Projecting lens and image display device |
JPH0715457A (en) | 1993-06-18 | 1995-01-17 | Hitachi Ltd | Digital communication switching system |
JPH0756112Y2 (en) | 1993-10-11 | 1995-12-25 | 中日本炉工業株式会社 | Vacuum degassing low pressure casting equipment |
EP0677821A3 (en) | 1994-04-14 | 1996-03-06 | Hewlett Packard Co | Magnify a digital image using feedback. |
AU1742895A (en) | 1994-06-09 | 1996-01-04 | Kollmorgen Instrument Corporation | Stereoscopic electro-optical system for automated inspection and/or alignment of imaging devices on a production assembly line |
US20020195548A1 (en) | 2001-06-06 | 2002-12-26 | Dowski Edward Raymond | Wavefront coding interference contrast imaging systems |
US5629524A (en) | 1995-02-21 | 1997-05-13 | Advanced Scientific Concepts, Inc. | High speed crystallography detector |
US5933190A (en) | 1995-04-18 | 1999-08-03 | Imec Vzw | Pixel structure, image sensor using such pixel structure and corresponding peripheral circuitry |
US5963664A (en) | 1995-06-22 | 1999-10-05 | Sarnoff Corporation | Method and system for image combination using a parallax-based technique |
US6005607A (en) | 1995-06-29 | 1999-12-21 | Matsushita Electric Industrial Co., Ltd. | Stereoscopic computer graphics image generating apparatus and stereoscopic TV apparatus |
GB2302978A (en) | 1995-07-04 | 1997-02-05 | Sharp Kk | LIquid crystal device |
WO1997018633A1 (en) | 1995-11-07 | 1997-05-22 | California Institute Of Technology | Capacitively coupled successive approximation ultra low power analog-to-digital converter |
JP3502713B2 (en) | 1995-12-21 | 2004-03-02 | 本田技研工業株式会社 | Vehicle distance measuring device |
JPH09181913A (en) | 1995-12-26 | 1997-07-11 | Olympus Optical Co Ltd | Camera system |
US6124974A (en) | 1996-01-26 | 2000-09-26 | Proxemics | Lenslet array systems and methods |
US5973844A (en) | 1996-01-26 | 1999-10-26 | Proxemics | Lenslet array systems and methods |
US6493465B2 (en) | 1996-02-21 | 2002-12-10 | Canon Kabushiki Kaisha | Matching point extracting method and apparatus therefor |
US5832312A (en) | 1996-02-22 | 1998-11-03 | Eastman Kodak Company | Watertight body for accommodating a photographic camera |
MY118360A (en) | 1996-04-30 | 2004-10-30 | Nippon Telegraph & Telephone | Scheme for detecting shot boundaries in compressed video data using inter-frame/inter field prediction coding and intra-frame/intra-field coding |
US6002743A (en) | 1996-07-17 | 1999-12-14 | Telymonde; Timothy D. | Method and apparatus for image acquisition from a plurality of cameras |
GB9616262D0 (en) | 1996-08-02 | 1996-09-11 | Philips Electronics Nv | Post-processing generation of focus/defocus effects for computer graphics images |
US6141048A (en) | 1996-08-19 | 2000-10-31 | Eastman Kodak Company | Compact image capture device |
US6137535A (en) | 1996-11-04 | 2000-10-24 | Eastman Kodak Company | Compact digital camera with segmented fields of view |
US5808350A (en) | 1997-01-03 | 1998-09-15 | Raytheon Company | Integrated IR, visible and NIR sensor and methods of fabricating same |
JPH10232626A (en) | 1997-02-20 | 1998-09-02 | Canon Inc | Stereoscopic image display device |
JPH10253351A (en) | 1997-03-14 | 1998-09-25 | Kyocera Corp | Distance measuring device |
US5801919A (en) | 1997-04-04 | 1998-09-01 | Gateway 2000, Inc. | Adjustably mounted camera assembly for portable computers |
US6515701B2 (en) | 1997-07-24 | 2003-02-04 | Polaroid Corporation | Focal plane exposure control system for CMOS area image sensors |
US6563537B1 (en) | 1997-07-31 | 2003-05-13 | Fuji Photo Film Co., Ltd. | Image signal interpolation |
JP3430935B2 (en) | 1997-10-20 | 2003-07-28 | 富士ゼロックス株式会社 | Image reading device and lens |
NO305728B1 (en) | 1997-11-14 | 1999-07-12 | Reidar E Tangen | Optoelectronic camera and method of image formatting in the same |
JP4243779B2 (en) | 1997-11-14 | 2009-03-25 | 株式会社ニコン | Diffusion plate manufacturing method, diffusion plate, microlens array manufacturing method, and microlens array |
US6069365A (en) | 1997-11-25 | 2000-05-30 | Alan Y. Chow | Optical processor based imaging system |
JPH11242189A (en) | 1997-12-25 | 1999-09-07 | Olympus Optical Co Ltd | Method and device for forming image |
US6721008B2 (en) | 1998-01-22 | 2004-04-13 | Eastman Kodak Company | Integrated CMOS active pixel digital camera |
JPH11223708A (en) | 1998-02-09 | 1999-08-17 | Nikon Corp | Indentator and production of micro-optical element array |
US6054703A (en) | 1998-03-20 | 2000-04-25 | Syscan, Inc. | Sensing module for accelerating signal readout from image sensors |
US6160909A (en) | 1998-04-01 | 2000-12-12 | Canon Kabushiki Kaisha | Depth control for stereoscopic images |
JP3745117B2 (en) * | 1998-05-08 | 2006-02-15 | キヤノン株式会社 | Image processing apparatus and image processing method |
JP3931936B2 (en) | 1998-05-11 | 2007-06-20 | セイコーエプソン株式会社 | Microlens array substrate, method for manufacturing the same, and display device |
JP3284190B2 (en) | 1998-05-14 | 2002-05-20 | 富士重工業株式会社 | Image correction device for stereo camera |
US6205241B1 (en) | 1998-06-01 | 2001-03-20 | Canon Kabushiki Kaisha | Compression of stereoscopic images |
US6137100A (en) | 1998-06-08 | 2000-10-24 | Photobit Corporation | CMOS image sensor with different pixel sizes for different colors |
US6069351A (en) | 1998-07-16 | 2000-05-30 | Intel Corporation | Focal plane processor for scaling information from image sensors |
US6903770B1 (en) | 1998-07-27 | 2005-06-07 | Sanyo Electric Co., Ltd. | Digital camera which produces a single image based on two exposures |
US6340994B1 (en) | 1998-08-12 | 2002-01-22 | Pixonics, Llc | System and method for using temporal gamma and reverse super-resolution to process images for use in digital display systems |
US6269175B1 (en) | 1998-08-28 | 2001-07-31 | Sarnoff Corporation | Method and apparatus for enhancing regions of aligned images using flow estimation |
US6879735B1 (en) | 1998-09-14 | 2005-04-12 | University Of Utah Reasearch Foundation | Method of digital image enhancement and sharpening |
US6310650B1 (en) | 1998-09-23 | 2001-10-30 | Honeywell International Inc. | Method and apparatus for calibrating a tiled display |
GB2343320B (en) | 1998-10-31 | 2003-03-26 | Ibm | Camera system for three dimentional images and video |
JP3596314B2 (en) | 1998-11-02 | 2004-12-02 | 日産自動車株式会社 | Object edge position measuring device and moving object traffic judging device |
US6611289B1 (en) | 1999-01-15 | 2003-08-26 | Yanbin Yu | Digital cameras using multiple sensors with multiple lenses |
JP3875423B2 (en) | 1999-01-19 | 2007-01-31 | 日本放送協会 | Solid-state imaging device and video signal output device therefor |
US6603513B1 (en) | 1999-02-16 | 2003-08-05 | Micron Technology, Inc. | Using a single control line to provide select and reset signals to image sensors in two rows of a digital imaging device |
JP3634677B2 (en) * | 1999-02-19 | 2005-03-30 | キヤノン株式会社 | Image interpolation method, image processing method, image display method, image processing apparatus, image display apparatus, and computer program storage medium |
US6563540B2 (en) | 1999-02-26 | 2003-05-13 | Intel Corporation | Light sensor with increased dynamic range |
US20020063807A1 (en) | 1999-04-19 | 2002-05-30 | Neal Margulis | Method for Performing Image Transforms in a Digital Display System |
US6819358B1 (en) | 1999-04-26 | 2004-11-16 | Microsoft Corporation | Error calibration for digital image sensors and apparatus using the same |
US6864916B1 (en) | 1999-06-04 | 2005-03-08 | The Trustees Of Columbia University In The City Of New York | Apparatus and method for high dynamic range imaging using spatially varying exposures |
JP2001008235A (en) | 1999-06-25 | 2001-01-12 | Minolta Co Ltd | Image input method for reconfiguring three-dimensional data and multiple-lens data input device |
JP2001042042A (en) | 1999-07-27 | 2001-02-16 | Canon Inc | Imaging device |
US7015954B1 (en) | 1999-08-09 | 2006-03-21 | Fuji Xerox Co., Ltd. | Automatic video system using multiple cameras |
US6647142B1 (en) | 1999-08-19 | 2003-11-11 | Mitsubishi Electric Research Laboratories, Inc. | Badge identification system |
US6771833B1 (en) | 1999-08-20 | 2004-08-03 | Eastman Kodak Company | Method and system for enhancing digital images |
US6628330B1 (en) | 1999-09-01 | 2003-09-30 | Neomagic Corp. | Color interpolator and horizontal/vertical edge enhancer using two line buffer and alternating even/odd filters for digital camera |
US6358862B1 (en) | 1999-09-02 | 2002-03-19 | Micron Technology, Inc | Passivation integrity improvements |
JP3280001B2 (en) | 1999-09-16 | 2002-04-30 | 富士重工業株式会社 | Stereo image misalignment adjustment device |
US6639596B1 (en) | 1999-09-20 | 2003-10-28 | Microsoft Corporation | Stereo reconstruction from multiperspective panoramas |
US6628845B1 (en) | 1999-10-20 | 2003-09-30 | Nec Laboratories America, Inc. | Method for subpixel registration of images |
US6774941B1 (en) | 1999-10-26 | 2004-08-10 | National Semiconductor Corporation | CCD output processing stage that amplifies signals from colored pixels based on the conversion efficiency of the colored pixels |
US6671399B1 (en) | 1999-10-27 | 2003-12-30 | Canon Kabushiki Kaisha | Fast epipolar line adjustment of stereo pairs |
US6674892B1 (en) | 1999-11-01 | 2004-01-06 | Canon Kabushiki Kaisha | Correcting an epipolar axis for skew and offset |
JP2001195050A (en) | 1999-11-05 | 2001-07-19 | Mitsubishi Electric Corp | Graphic accelerator |
WO2001039512A1 (en) | 1999-11-26 | 2001-05-31 | Sanyo Electric Co., Ltd. | Device and method for converting two-dimensional video to three-dimensional video |
JP3950926B2 (en) | 1999-11-30 | 2007-08-01 | エーユー オプトロニクス コーポレイション | Image display method, host device, image display device, and display interface |
JP3728160B2 (en) | 1999-12-06 | 2005-12-21 | キヤノン株式会社 | Depth image measuring apparatus and method, and mixed reality presentation system |
US7068851B1 (en) | 1999-12-10 | 2006-06-27 | Ricoh Co., Ltd. | Multiscale sharpening and smoothing with wavelets |
FI107680B (en) | 1999-12-22 | 2001-09-14 | Nokia Oyj | Procedure for transmitting video images, data transmission systems, transmitting video terminal and receiving video terminal |
US6476805B1 (en) | 1999-12-23 | 2002-11-05 | Microsoft Corporation | Techniques for spatial displacement estimation and multi-resolution operations on light fields |
US6502097B1 (en) | 1999-12-23 | 2002-12-31 | Microsoft Corporation | Data structure for efficient access to variable-size data objects |
JP2001194114A (en) | 2000-01-14 | 2001-07-19 | Sony Corp | Image processing apparatus and method and program providing medium |
US6523046B2 (en) | 2000-02-25 | 2003-02-18 | Microsoft Corporation | Infrastructure and method for supporting generic multimedia metadata |
JP2001264033A (en) | 2000-03-17 | 2001-09-26 | Sony Corp | Three-dimensional shape-measuring apparatus and its method, three-dimensional modeling device and its method, and program providing medium |
US6571466B1 (en) | 2000-03-27 | 2003-06-03 | Amkor Technology, Inc. | Flip chip image sensor package fabrication method |
JP2001277260A (en) | 2000-03-30 | 2001-10-09 | Seiko Epson Corp | Microlens array, method for manufacturing the same, master for manufacturing the same, and display device |
WO2001075949A1 (en) | 2000-04-04 | 2001-10-11 | Advantest Corporation | Multibeam exposure apparatus comprising multiaxis electron lens and method for manufacturing semiconductor device |
WO2001082593A1 (en) | 2000-04-24 | 2001-11-01 | The Government Of The United States Of America, As Represented By The Secretary Of The Navy | Apparatus and method for color image fusion |
JP2001337263A (en) | 2000-05-25 | 2001-12-07 | Olympus Optical Co Ltd | Range-finding device |
JP4501239B2 (en) | 2000-07-13 | 2010-07-14 | ソニー株式会社 | Camera calibration apparatus and method, and storage medium |
WO2002009424A2 (en) | 2000-07-21 | 2002-01-31 | The Trustees Of Columbia University In The City Of New York | Method and apparatus for image mosaicing |
AU2001277110A1 (en) | 2000-07-21 | 2002-02-05 | The Trustees Of Columbia University In The City Of New York | Method and apparatus for reducing distortion in images |
US7154546B1 (en) | 2000-08-07 | 2006-12-26 | Micron Technology, Inc. | Pixel optimization for color |
US7200261B2 (en) | 2000-08-25 | 2007-04-03 | Fujifilm Corporation | Parallax image capturing apparatus and parallax image processing apparatus |
US6734905B2 (en) | 2000-10-20 | 2004-05-11 | Micron Technology, Inc. | Dynamic range extension for CMOS image sensors |
US6774889B1 (en) | 2000-10-24 | 2004-08-10 | Microsoft Corporation | System and method for transforming an ordinary computer monitor screen into a touch screen |
US7262799B2 (en) | 2000-10-25 | 2007-08-28 | Canon Kabushiki Kaisha | Image sensing apparatus and its control method, control program, and storage medium |
US6476971B1 (en) | 2000-10-31 | 2002-11-05 | Eastman Kodak Company | Method of manufacturing a microlens array mold and a microlens array |
JP3918499B2 (en) | 2000-11-01 | 2007-05-23 | セイコーエプソン株式会社 | Gap measuring method, gap measuring device, shape measuring method, shape measuring device, and liquid crystal device manufacturing method |
US6788338B1 (en) | 2000-11-20 | 2004-09-07 | Petko Dimitrov Dinev | High resolution video camera apparatus having two image sensors and signal processing |
US7490774B2 (en) | 2003-11-13 | 2009-02-17 | Metrologic Instruments, Inc. | Hand-supportable imaging based bar code symbol reader employing automatic light exposure measurement and illumination control subsystem integrated therein |
JP2002171537A (en) | 2000-11-30 | 2002-06-14 | Canon Inc | Compound image pickup system, image pickup device and electronic device |
WO2002045003A1 (en) | 2000-12-01 | 2002-06-06 | Imax Corporation | Techniques and systems for developing high-resolution imagery |
EP1356431A2 (en) | 2000-12-05 | 2003-10-29 | YEDA RESEARCH AND DEVELOPMENT Co. LTD. | Apparatus and method for alignment of spatial or temporal non-overlapping image sequences |
JP2002252338A (en) | 2000-12-18 | 2002-09-06 | Canon Inc | Imaging device and imaging system |
JP2002195910A (en) | 2000-12-26 | 2002-07-10 | Omron Corp | System for testing optical part |
JP2002209226A (en) | 2000-12-28 | 2002-07-26 | Canon Inc | Image pickup device |
US7805680B2 (en) | 2001-01-03 | 2010-09-28 | Nokia Corporation | Statistical metering and filtering of content via pixel-based metadata |
JP3957460B2 (en) | 2001-01-15 | 2007-08-15 | 沖電気工業株式会社 | Transmission header compression apparatus, moving picture encoding apparatus, and moving picture transmission system |
JP2002250607A (en) | 2001-02-27 | 2002-09-06 | Optex Co Ltd | Object detection sensor |
US6635941B2 (en) | 2001-03-21 | 2003-10-21 | Canon Kabushiki Kaisha | Structure of semiconductor device with improved reliability |
JP2002324743A (en) | 2001-04-24 | 2002-11-08 | Canon Inc | Exposing method and equipment thereof |
US6443579B1 (en) | 2001-05-02 | 2002-09-03 | Kenneth Myers | Field-of-view controlling arrangements |
US20020167537A1 (en) | 2001-05-11 | 2002-11-14 | Miroslav Trajkovic | Motion-based tracking with pan-tilt-zoom camera |
US7235785B2 (en) | 2001-05-11 | 2007-06-26 | Irvine Sensors Corp. | Imaging device with multiple fields of view incorporating memory-based temperature compensation of an uncooled focal plane array |
US7738013B2 (en) | 2001-05-29 | 2010-06-15 | Samsung Electronics Co., Ltd. | Systems and methods for power conservation in a CMOS imager |
US7420602B2 (en) | 2001-05-29 | 2008-09-02 | Samsung Semiconductor Israel R&D Center (Sirc) | Cmos imager for cellular applications and methods of using such |
US6482669B1 (en) | 2001-05-30 | 2002-11-19 | Taiwan Semiconductor Manufacturing Company | Colors only process to reduce package yield loss |
US6525302B2 (en) | 2001-06-06 | 2003-02-25 | The Regents Of The University Of Colorado | Wavefront coding phase contrast imaging systems |
US20030025227A1 (en) | 2001-08-02 | 2003-02-06 | Zograph, Llc | Reproduction of relief patterns |
US8675119B2 (en) | 2001-08-09 | 2014-03-18 | Trustees Of Columbia University In The City Of New York | Adaptive imaging using digital light processing |
DE60141901D1 (en) | 2001-08-31 | 2010-06-02 | St Microelectronics Srl | Noise filter for Bavarian pattern image data |
JP3978706B2 (en) | 2001-09-20 | 2007-09-19 | セイコーエプソン株式会社 | Manufacturing method of fine structure |
JP2003139910A (en) | 2001-10-30 | 2003-05-14 | Sony Corp | Optical element, method and device for manufacturing the same, and liquid crystal display device and image projection type display device using the same |
DE10153237A1 (en) | 2001-10-31 | 2003-05-15 | Lfk Gmbh | Method and device for the automated determination of the modulation transfer function (MTF) of focal plane array (FPA) cameras |
JP3705766B2 (en) | 2001-11-28 | 2005-10-12 | 独立行政法人科学技術振興機構 | Image input device |
EP1468314A4 (en) | 2001-12-18 | 2006-12-13 | Univ Rochester | ILLUSTRATION USING AN ASPHARIAN MULTI-FUNGI CASE FOR MAINTAINING AN ADVANCED SHARPNESS |
US7212228B2 (en) | 2002-01-16 | 2007-05-01 | Advanced Telecommunications Research Institute International | Automatic camera calibration method |
US7302118B2 (en) | 2002-02-07 | 2007-11-27 | Microsoft Corporation | Transformation of images |
US20030179418A1 (en) | 2002-03-19 | 2003-09-25 | Eastman Kodak Company | Producing a defective pixel map from defective cluster pixels in an area array image sensor |
US8369607B2 (en) | 2002-03-27 | 2013-02-05 | Sanyo Electric Co., Ltd. | Method and apparatus for processing three-dimensional images |
JP2003298920A (en) | 2002-03-29 | 2003-10-17 | Fuji Photo Film Co Ltd | Digital camera |
US20030188659A1 (en) | 2002-04-05 | 2003-10-09 | Canadian Bank Note Company Limited | Method and apparatus for reproducing a color image based on monochrome images derived therefrom |
AU2003226047A1 (en) | 2002-04-10 | 2003-10-27 | Pan-X Imaging, Inc. | A digital imaging system |
US6856314B2 (en) | 2002-04-18 | 2005-02-15 | Stmicroelectronics, Inc. | Method and system for 3D reconstruction of multiple views with altering search path and occlusion modeling |
US6917702B2 (en) | 2002-04-24 | 2005-07-12 | Mitsubishi Electric Research Labs, Inc. | Calibration of multiple cameras for a turntable-based 3D scanner |
JP3567327B2 (en) | 2002-05-08 | 2004-09-22 | 富士写真光機株式会社 | Imaging lens |
US6783900B2 (en) | 2002-05-13 | 2004-08-31 | Micron Technology, Inc. | Color filter imaging array and method of formation |
JP2004048644A (en) | 2002-05-21 | 2004-02-12 | Sony Corp | Information processor, information processing system and interlocutor display method |
JP2003347192A (en) | 2002-05-24 | 2003-12-05 | Toshiba Corp | Energy beam exposure method and exposure device |
US7129981B2 (en) | 2002-06-27 | 2006-10-31 | International Business Machines Corporation | Rendering system and method for images having differing foveal area and peripheral view area resolutions |
JP2004088713A (en) | 2002-06-27 | 2004-03-18 | Olympus Corp | Imaging lens unit and imaging device |
JP4147059B2 (en) | 2002-07-03 | 2008-09-10 | 株式会社トプコン | Calibration data measuring device, measuring method and measuring program, computer-readable recording medium, and image data processing device |
JP2004037924A (en) | 2002-07-04 | 2004-02-05 | Minolta Co Ltd | Imaging apparatus |
US8111289B2 (en) | 2002-07-15 | 2012-02-07 | Magna B.S.P. Ltd. | Method and apparatus for implementing multipurpose monitoring system |
US20040012689A1 (en) | 2002-07-16 | 2004-01-22 | Fairchild Imaging | Charge coupled devices in tiled arrays |
JP2004078296A (en) | 2002-08-09 | 2004-03-11 | Victor Co Of Japan Ltd | Picture generation device |
AU2003274951A1 (en) * | 2002-08-30 | 2004-03-19 | Orasee Corp. | Multi-dimensional image system for digital image input and output |
US7447380B2 (en) | 2002-09-12 | 2008-11-04 | Inoe Technologies, Llc | Efficient method for creating a viewpoint from plurality of images |
US20040050104A1 (en) | 2002-09-18 | 2004-03-18 | Eastman Kodak Company | Forming information transfer lens array |
US20040207836A1 (en) | 2002-09-27 | 2004-10-21 | Rajeshwar Chhibber | High dynamic range optical inspection system and method |
US7084904B2 (en) | 2002-09-30 | 2006-08-01 | Microsoft Corporation | Foveated wide-angle imaging system and method for capturing and viewing wide-angle images in real time |
US7477781B1 (en) | 2002-10-10 | 2009-01-13 | Dalsa Corporation | Method and apparatus for adaptive pixel correction of multi-color matrix |
JP4171786B2 (en) | 2002-10-25 | 2008-10-29 | コニカミノルタホールディングス株式会社 | Image input device |
US7742088B2 (en) | 2002-11-19 | 2010-06-22 | Fujifilm Corporation | Image sensor and digital camera |
KR101237945B1 (en) | 2002-11-21 | 2013-02-28 | 비젼 Ⅲ 이미징 인코퍼레이티드 | Critical Alignment Of Parallax Images For Autostereoscopic Display |
US20040105021A1 (en) | 2002-12-02 | 2004-06-03 | Bolymedia Holdings Co., Ltd. | Color filter patterns for image sensors |
US20040114807A1 (en) | 2002-12-13 | 2004-06-17 | Dan Lelescu | Statistical representation and coding of light field data |
US6878918B2 (en) | 2003-01-09 | 2005-04-12 | Dialdg Semiconductor Gmbh | APS pixel with reset noise suppression and programmable binning capability |
US7340099B2 (en) | 2003-01-17 | 2008-03-04 | University Of New Brunswick | System and method for image fusion |
DE10301941B4 (en) | 2003-01-20 | 2005-11-17 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Camera and method for optical recording of a screen |
US7379592B2 (en) | 2003-01-21 | 2008-05-27 | United States Of America As Represented By The Secretary Of The Navy | System and method for significant dust detection and enhancement of dust images over land and ocean |
US7515747B2 (en) | 2003-01-31 | 2009-04-07 | The Circle For The Promotion Of Science And Engineering | Method for creating high resolution color image, system for creating high resolution color image and program creating high resolution color image |
US7005637B2 (en) | 2003-01-31 | 2006-02-28 | Intevac, Inc. | Backside thinning of image array devices |
US7308157B2 (en) | 2003-02-03 | 2007-12-11 | Photon Dynamics, Inc. | Method and apparatus for optical inspection of a display |
US7595817B1 (en) | 2003-02-12 | 2009-09-29 | The Research Foundation Of State University Of New York | Linear system based, qualitative independent motion detection from compressed MPEG surveillance video |
US20040165090A1 (en) | 2003-02-13 | 2004-08-26 | Alex Ning | Auto-focus (AF) lens and process |
JP2004266369A (en) | 2003-02-21 | 2004-09-24 | Sony Corp | Solid-state image pickup unit and its driving method |
US7106914B2 (en) | 2003-02-27 | 2006-09-12 | Microsoft Corporation | Bayesian image super resolution |
US7148861B2 (en) | 2003-03-01 | 2006-12-12 | The Boeing Company | Systems and methods for providing enhanced vision imaging with decreased latency |
US8218052B2 (en) | 2003-03-07 | 2012-07-10 | Iconix Video, Inc. | High frame rate high definition imaging system and method |
US7218320B2 (en) | 2003-03-13 | 2007-05-15 | Sony Corporation | System and method for capturing facial and body motion |
US6801719B1 (en) | 2003-03-14 | 2004-10-05 | Eastman Kodak Company | Camera using beam splitter with micro-lens image amplification |
US7206449B2 (en) | 2003-03-19 | 2007-04-17 | Mitsubishi Electric Research Laboratories, Inc. | Detecting silhouette edges in images |
US7373005B2 (en) | 2003-04-10 | 2008-05-13 | Micron Technology, Inc. | Compression system for integrated sensor devices |
US7097311B2 (en) | 2003-04-19 | 2006-08-29 | University Of Kentucky Research Foundation | Super-resolution overlay in multi-projector displays |
US6958862B1 (en) | 2003-04-21 | 2005-10-25 | Foveon, Inc. | Use of a lenslet array with a vertically stacked pixel array |
US7428330B2 (en) | 2003-05-02 | 2008-09-23 | Microsoft Corporation | Cyclopean virtual imaging via generalized probabilistic smoothing |
SE525665C2 (en) | 2003-05-08 | 2005-03-29 | Forskarpatent I Syd Ab | Matrix of pixels and electronic imaging device comprising said matrix of pixels |
WO2004102958A1 (en) | 2003-05-13 | 2004-11-25 | Xceed Imaging Ltd. | Optical method and system for enhancing image resolution |
JP2004348674A (en) | 2003-05-26 | 2004-12-09 | Noritsu Koki Co Ltd | Region detection method and its device |
US20040239782A1 (en) | 2003-05-30 | 2004-12-02 | William Equitz | System and method for efficient improvement of image quality in cameras |
US20040240052A1 (en) | 2003-06-02 | 2004-12-02 | Pentax Corporation | Multiple-focal imaging device, and a mobile device having the multiple-focal-length imaging device |
JP2004363478A (en) | 2003-06-06 | 2004-12-24 | Sanyo Electric Co Ltd | Manufacturing method of semiconductor device |
KR100539234B1 (en) | 2003-06-11 | 2005-12-27 | 삼성전자주식회사 | A CMOS type image sensor module having transparent polymeric encapsulation material |
US6818934B1 (en) | 2003-06-24 | 2004-11-16 | Omnivision International Holding Ltd | Image sensor having micro-lens array separated with trench structures and method of making |
US7362918B2 (en) | 2003-06-24 | 2008-04-22 | Microsoft Corporation | System and method for de-noising multiple copies of a signal |
US7388609B2 (en) | 2003-07-07 | 2008-06-17 | Zoran Corporation | Dynamic identification and correction of defective pixels |
US7090135B2 (en) | 2003-07-07 | 2006-08-15 | Symbol Technologies, Inc. | Imaging arrangement and barcode imager for imaging an optical code or target at a plurality of focal planes |
US20050007461A1 (en) | 2003-07-11 | 2005-01-13 | Novatek Microelectronic Co. | Correction system and method of analog front end |
JP3731589B2 (en) | 2003-07-18 | 2006-01-05 | ソニー株式会社 | Imaging device and synchronization signal generator |
US7233737B2 (en) | 2003-08-12 | 2007-06-19 | Micron Technology, Inc. | Fixed-focus camera module and associated method of assembly |
US7643703B2 (en) | 2003-09-03 | 2010-01-05 | Battelle Energy Alliance, Llc | Image change detection systems, methods, and articles of manufacture |
EP1671258A4 (en) | 2003-09-04 | 2008-03-19 | Sarnoff Corp | Method and apparatus for performing iris recognition from an image |
US7161606B2 (en) | 2003-09-08 | 2007-01-09 | Honda Motor Co., Ltd. | Systems and methods for directly generating a view using a layered approach |
JP4015090B2 (en) | 2003-09-08 | 2007-11-28 | 株式会社東芝 | Stereoscopic display device and image display method |
JP4020850B2 (en) | 2003-10-06 | 2007-12-12 | 株式会社東芝 | Magnetic recording medium manufacturing method, manufacturing apparatus, imprint stamper and manufacturing method thereof |
US7079251B2 (en) | 2003-10-16 | 2006-07-18 | 4D Technology Corporation | Calibration and error correction in multi-channel imaging |
JP4071793B2 (en) | 2003-10-22 | 2008-04-02 | 松下電器産業株式会社 | IMAGING DEVICE, MANUFACTURING METHOD THEREOF, AND PORTABLE DEVICE |
US7840067B2 (en) | 2003-10-24 | 2010-11-23 | Arcsoft, Inc. | Color matching and color correction for images forming a panoramic image |
JP4118916B2 (en) | 2003-11-11 | 2008-07-16 | オリンパス株式会社 | Multispectral imaging device |
JP4235539B2 (en) | 2003-12-01 | 2009-03-11 | 独立行政法人科学技術振興機構 | Image composition apparatus and image composition method |
US7453510B2 (en) | 2003-12-11 | 2008-11-18 | Nokia Corporation | Imaging device |
US20050128509A1 (en) | 2003-12-11 | 2005-06-16 | Timo Tokkonen | Image creating method and imaging device |
US7328288B2 (en) | 2003-12-11 | 2008-02-05 | Canon Kabushiki Kaisha | Relay apparatus for relaying communication from CPU to peripheral device |
JP3859158B2 (en) | 2003-12-16 | 2006-12-20 | セイコーエプソン株式会社 | Microlens concave substrate, microlens substrate, transmissive screen, and rear projector |
US7123298B2 (en) | 2003-12-18 | 2006-10-17 | Avago Technologies Sensor Ip Pte. Ltd. | Color image sensor with imaging elements imaging on respective regions of sensor elements |
US7511749B2 (en) | 2003-12-18 | 2009-03-31 | Aptina Imaging Corporation | Color image sensor having imaging element array forming images on respective regions of sensor elements |
US7376250B2 (en) | 2004-01-05 | 2008-05-20 | Honda Motor Co., Ltd. | Apparatus, method and program for moving object detection |
US7496293B2 (en) | 2004-01-14 | 2009-02-24 | Elbit Systems Ltd. | Versatile camera for various visibility conditions |
US7773143B2 (en) | 2004-04-08 | 2010-08-10 | Tessera North America, Inc. | Thin color camera having sub-pixel resolution |
US8134637B2 (en) | 2004-01-28 | 2012-03-13 | Microsoft Corporation | Method and system to increase X-Y resolution in a depth (Z) camera using red, blue, green (RGB) sensing |
US7453688B2 (en) | 2004-01-29 | 2008-11-18 | Inventec Corporation | Multimedia device for portable computers |
US20050185711A1 (en) | 2004-02-20 | 2005-08-25 | Hanspeter Pfister | 3D television system and method |
SE527889C2 (en) | 2004-03-17 | 2006-07-04 | Thomas Jeff Adamo | Apparatus for imaging an object |
JP2006047944A (en) | 2004-03-24 | 2006-02-16 | Fuji Photo Film Co Ltd | Photographing lens |
WO2005096218A1 (en) | 2004-03-31 | 2005-10-13 | Canon Kabushiki Kaisha | Imaging system performance measurement |
JP4665422B2 (en) | 2004-04-02 | 2011-04-06 | ソニー株式会社 | Imaging device |
US8634014B2 (en) | 2004-04-05 | 2014-01-21 | Hewlett-Packard Development Company, L.P. | Imaging device analysis systems and imaging device analysis methods |
US7091531B2 (en) | 2004-04-07 | 2006-08-15 | Micron Technology, Inc. | High dynamic range pixel amplifier |
US8049806B2 (en) | 2004-09-27 | 2011-11-01 | Digitaloptics Corporation East | Thin camera and associated methods |
US7620265B1 (en) | 2004-04-12 | 2009-11-17 | Equinox Corporation | Color invariant image fusion of visible and thermal infrared video |
US7292735B2 (en) | 2004-04-16 | 2007-11-06 | Microsoft Corporation | Virtual image artifact detection |
US7773404B2 (en) | 2005-01-07 | 2010-08-10 | Invisage Technologies, Inc. | Quantum dot optical devices with enhanced gain and sensitivity and methods of making same |
US8218625B2 (en) | 2004-04-23 | 2012-07-10 | Dolby Laboratories Licensing Corporation | Encoding, decoding and representing high dynamic range images |
US20060034531A1 (en) | 2004-05-10 | 2006-02-16 | Seiko Epson Corporation | Block noise level evaluation method for compressed images and control method of imaging device utilizing the evaluation method |
CN1953708B (en) | 2004-05-14 | 2010-06-16 | 皇家飞利浦电子股份有限公司 | System and method for diagnosing breast cancer |
JP4610411B2 (en) | 2004-05-17 | 2011-01-12 | ミツビシ・エレクトリック・リサーチ・ラボラトリーズ・インコーポレイテッド | Method for generating a stylized image of a scene containing objects |
US7355793B2 (en) | 2004-05-19 | 2008-04-08 | The Regents Of The University Of California | Optical system applicable to improving the dynamic range of Shack-Hartmann sensors |
US20050265633A1 (en) | 2004-05-25 | 2005-12-01 | Sarnoff Corporation | Low latency pyramid processor for image processing systems |
JP2005354124A (en) | 2004-06-08 | 2005-12-22 | Seiko Epson Corp | Generation of high pixel density image from multiple low pixel density images |
US20060013318A1 (en) | 2004-06-22 | 2006-01-19 | Jennifer Webb | Video error detection, recovery, and concealment |
US7330593B2 (en) | 2004-06-25 | 2008-02-12 | Stmicroelectronics, Inc. | Segment based image matching method and system |
JP4408755B2 (en) | 2004-06-28 | 2010-02-03 | Necエレクトロニクス株式会社 | Deinterleaving device, mobile communication terminal, and deinterleaving method |
JP4479373B2 (en) | 2004-06-28 | 2010-06-09 | ソニー株式会社 | Image sensor |
US7447382B2 (en) | 2004-06-30 | 2008-11-04 | Intel Corporation | Computing a higher resolution image from multiple lower resolution images using model-based, robust Bayesian estimation |
JP2006033493A (en) | 2004-07-16 | 2006-02-02 | Matsushita Electric Ind Co Ltd | Imaging device |
US7189954B2 (en) | 2004-07-19 | 2007-03-13 | Micron Technology, Inc. | Microelectronic imagers with optical devices and methods of manufacturing such microelectronic imagers |
JP2006033570A (en) | 2004-07-20 | 2006-02-02 | Olympus Corp | Image generating device |
US8027531B2 (en) | 2004-07-21 | 2011-09-27 | The Board Of Trustees Of The Leland Stanford Junior University | Apparatus and method for capturing a scene using staggered triggering of dense camera arrays |
GB0416496D0 (en) | 2004-07-23 | 2004-08-25 | Council Of The Central Lab Of | Imaging device |
US20060023197A1 (en) | 2004-07-27 | 2006-02-02 | Joel Andrew H | Method and system for automated production of autostereoscopic and animated prints and transparencies from digital and non-digital media |
US7068432B2 (en) | 2004-07-27 | 2006-06-27 | Micron Technology, Inc. | Controlling lens shape in a microlens array |
DE102004036469A1 (en) | 2004-07-28 | 2006-02-16 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Camera module, array based thereon and method for its production |
US20060028476A1 (en) | 2004-08-03 | 2006-02-09 | Irwin Sobel | Method and system for providing extensive coverage of an object using virtual cameras |
JP2006050263A (en) | 2004-08-04 | 2006-02-16 | Olympus Corp | Image generation method and device |
KR20070064319A (en) | 2004-08-06 | 2007-06-20 | 유니버시티 오브 워싱톤 | Variable Stare Viewing Scanning Optical Display |
US7430339B2 (en) | 2004-08-09 | 2008-09-30 | Microsoft Corporation | Border matting by dynamic programming |
US7609302B2 (en) | 2004-08-11 | 2009-10-27 | Micron Technology, Inc. | Correction of non-uniform sensitivity in an image array |
US7061693B2 (en) | 2004-08-16 | 2006-06-13 | Xceed Imaging Ltd. | Optical method and system for extended depth of focus |
US7645635B2 (en) | 2004-08-16 | 2010-01-12 | Micron Technology, Inc. | Frame structure and semiconductor attach process for use therewith for fabrication of image sensor packages and the like, and resulting packages |
EP1797523A4 (en) | 2004-08-23 | 2009-07-22 | Sarnoff Corp | METHOD AND APPARATUS FOR PRODUCING A MERGED IMAGE |
US7795577B2 (en) | 2004-08-25 | 2010-09-14 | Richard Ian Olsen | Lens frame and optical focus assembly for imager module |
US20060054782A1 (en) | 2004-08-25 | 2006-03-16 | Olsen Richard I | Apparatus for multiple camera devices and method of operating same |
US7916180B2 (en) | 2004-08-25 | 2011-03-29 | Protarius Filo Ag, L.L.C. | Simultaneous multiple field of view digital cameras |
US7564019B2 (en) | 2005-08-25 | 2009-07-21 | Richard Ian Olsen | Large dynamic range cameras |
US8124929B2 (en) | 2004-08-25 | 2012-02-28 | Protarius Filo Ag, L.L.C. | Imager module optical focus and assembly method |
CN100489599C (en) | 2004-08-26 | 2009-05-20 | 财团法人秋田企业活性化中心 | Liquid crystal lens |
JP4057597B2 (en) | 2004-08-26 | 2008-03-05 | 独立行政法人科学技術振興機構 | Optical element |
US20060046204A1 (en) | 2004-08-31 | 2006-03-02 | Sharp Laboratories Of America, Inc. | Directly patternable microlens |
JP2006080852A (en) | 2004-09-09 | 2006-03-23 | Olympus Corp | Apparatus and method of processing image, electronic camera, scanner and image processing program |
US20060055811A1 (en) | 2004-09-14 | 2006-03-16 | Frtiz Bernard S | Imaging system having modules with adaptive optical elements |
US7145124B2 (en) | 2004-09-15 | 2006-12-05 | Raytheon Company | Multispectral imaging chip using photonic crystals |
JP4202991B2 (en) | 2004-09-29 | 2008-12-24 | 株式会社東芝 | Stereo image data recording method and display reproduction method |
JP3977368B2 (en) | 2004-09-30 | 2007-09-19 | クラリオン株式会社 | Parking assistance system |
DE102004049676A1 (en) | 2004-10-12 | 2006-04-20 | Infineon Technologies Ag | Method for computer-aided motion estimation in a plurality of temporally successive digital images, arrangement for computer-aided motion estimation, computer program element and computer-readable storage medium |
JP2006119368A (en) | 2004-10-21 | 2006-05-11 | Konica Minolta Opto Inc | Wide-angle optical system, imaging lens device, monitor camera and digital equipment |
JP4534715B2 (en) | 2004-10-22 | 2010-09-01 | 株式会社ニコン | Imaging apparatus and image processing program |
DE102004052994C5 (en) | 2004-11-03 | 2010-08-26 | Vistec Electron Beam Gmbh | Multi-beam modulator for a particle beam and use of the multi-beam modulator for maskless substrate structuring |
KR100603601B1 (en) | 2004-11-08 | 2006-07-24 | 한국전자통신연구원 | Multi-view content generation device and method |
US7598996B2 (en) | 2004-11-16 | 2009-10-06 | Aptina Imaging Corporation | System and method for focusing a digital camera |
CN101111748B (en) | 2004-12-03 | 2014-12-17 | 弗卢克公司 | Visible light and ir combined image camera with a laser pointer |
US7483065B2 (en) | 2004-12-15 | 2009-01-27 | Aptina Imaging Corporation | Multi-lens imaging systems and methods using optical filters having mosaic patterns |
EP1851527A2 (en) | 2005-01-07 | 2007-11-07 | GestureTek, Inc. | Creating 3d images of objects by illuminating with infrared patterns |
US7073908B1 (en) | 2005-01-11 | 2006-07-11 | Anthony Italo Provitola | Enhancement of depth perception |
US7767949B2 (en) | 2005-01-18 | 2010-08-03 | Rearden, Llc | Apparatus and method for capturing still images and video using coded aperture techniques |
US7671321B2 (en) | 2005-01-18 | 2010-03-02 | Rearden, Llc | Apparatus and method for capturing still images and video using coded lens imaging techniques |
US7602997B2 (en) | 2005-01-19 | 2009-10-13 | The United States Of America As Represented By The Secretary Of The Army | Method of super-resolving images |
US7408627B2 (en) | 2005-02-08 | 2008-08-05 | Canesta, Inc. | Methods and system to quantify depth data accuracy in three-dimensional sensors using single frame capture |
US7965314B1 (en) | 2005-02-09 | 2011-06-21 | Flir Systems, Inc. | Foveal camera systems and methods |
US7561191B2 (en) | 2005-02-18 | 2009-07-14 | Eastman Kodak Company | Camera phone using multiple lenses and image sensors to provide an extended zoom range |
CN101189487B (en) | 2005-03-11 | 2010-08-11 | 形创有限公司 | Auto-referenced system and apparatus for three-dimensional scanning |
JP2006258930A (en) | 2005-03-15 | 2006-09-28 | Nikon Corp | Method for manufacturing microlens and method for manufacturing die for microlens |
WO2006102181A1 (en) | 2005-03-21 | 2006-09-28 | Massachusetts Institute Of Technology (Mit) | Real-time, continuous-wave terahertz imaging using a microbolometer focal-plane array |
WO2006100903A1 (en) | 2005-03-23 | 2006-09-28 | Matsushita Electric Industrial Co., Ltd. | On-vehicle imaging device |
WO2006101064A1 (en) | 2005-03-24 | 2006-09-28 | Matsushita Electric Industrial Co., Ltd. | Imaging device and lens array used therein |
US7297917B2 (en) | 2005-03-24 | 2007-11-20 | Micron Technology, Inc. | Readout technique for increasing or maintaining dynamic range in image sensors |
US7683950B2 (en) | 2005-04-26 | 2010-03-23 | Eastman Kodak Company | Method and apparatus for correcting a channel dependent color aberration in a digital image |
US7956871B2 (en) | 2005-04-28 | 2011-06-07 | Samsung Electronics Co., Ltd. | Color disparity correction in image sensors methods and circuits |
US7656428B2 (en) | 2005-05-05 | 2010-02-02 | Avago Technologies General Ip (Singapore) Pte. Ltd. | Imaging device employing optical motion sensor as gyroscope |
WO2006123581A1 (en) | 2005-05-18 | 2006-11-23 | Hitachi Medical Corporation | Radiograph and image processing program |
US8411182B2 (en) | 2005-06-02 | 2013-04-02 | Xerox Corporation | System for controlling integration times of photosensors in an imaging device |
US7968888B2 (en) | 2005-06-08 | 2011-06-28 | Panasonic Corporation | Solid-state image sensor and manufacturing method thereof |
JP2006345233A (en) | 2005-06-09 | 2006-12-21 | Fujifilm Holdings Corp | Imaging device and digital camera |
KR100813961B1 (en) | 2005-06-14 | 2008-03-14 | 삼성전자주식회사 | Video receiver |
US7364306B2 (en) | 2005-06-20 | 2008-04-29 | Digital Display Innovations, Llc | Field sequential light source modulation for a digital display system |
JP4826152B2 (en) | 2005-06-23 | 2011-11-30 | 株式会社ニコン | Image composition method and imaging apparatus |
US20070102622A1 (en) | 2005-07-01 | 2007-05-10 | Olsen Richard I | Apparatus for multiple camera devices and method of operating same |
JP4577126B2 (en) | 2005-07-08 | 2010-11-10 | オムロン株式会社 | Projection pattern generation apparatus and generation method for stereo correspondence |
WO2007014293A1 (en) | 2005-07-25 | 2007-02-01 | The Regents Of The University Of California | Digital imaging system and method to produce mosaic images |
CA2553473A1 (en) | 2005-07-26 | 2007-01-26 | Wa James Tam | Generating a depth map from a tw0-dimensional source image for stereoscopic and multiview imaging |
US7969488B2 (en) | 2005-08-03 | 2011-06-28 | Micron Technologies, Inc. | Correction of cluster defects in imagers |
US7929801B2 (en) | 2005-08-15 | 2011-04-19 | Sony Corporation | Depth information for auto focus using two pictures and two-dimensional Gaussian scale space theory |
US20070041391A1 (en) | 2005-08-18 | 2007-02-22 | Micron Technology, Inc. | Method and apparatus for controlling imager output data rate |
US20070040922A1 (en) | 2005-08-22 | 2007-02-22 | Micron Technology, Inc. | HDR/AB on multi-way shared pixels |
US7964835B2 (en) | 2005-08-25 | 2011-06-21 | Protarius Filo Ag, L.L.C. | Digital cameras with direct luminance and chrominance detection |
US20070258006A1 (en) | 2005-08-25 | 2007-11-08 | Olsen Richard I | Solid state camera optics frame and assembly |
US20070083114A1 (en) | 2005-08-26 | 2007-04-12 | The University Of Connecticut | Systems and methods for image resolution enhancement |
JP4804856B2 (en) | 2005-09-29 | 2011-11-02 | 富士フイルム株式会社 | Single focus lens |
US8009209B2 (en) | 2005-09-30 | 2011-08-30 | Simon Fraser University | Methods and apparatus for detecting defects in imaging arrays by image analysis |
WO2007044725A2 (en) | 2005-10-07 | 2007-04-19 | The Board Of Trustees Of The Leland Stanford Junior University | Microscopy arrangements and approaches |
US8300085B2 (en) | 2005-10-14 | 2012-10-30 | Microsoft Corporation | Occlusion handling in stereo imaging |
JP4773179B2 (en) | 2005-10-14 | 2011-09-14 | 富士フイルム株式会社 | Imaging device |
US7806604B2 (en) | 2005-10-20 | 2010-10-05 | Honeywell International Inc. | Face detection and tracking in a wide field of view |
KR100730406B1 (en) * | 2005-11-16 | 2007-06-19 | 광운대학교 산학협력단 | Stereoscopic image display device using intermediate element image |
JP4389865B2 (en) | 2005-11-17 | 2009-12-24 | ソニー株式会社 | SIGNAL PROCESSING DEVICE FOR SOLID-STATE IMAGING ELEMENT, SIGNAL PROCESSING METHOD, AND IMAGING DEVICE |
US7599547B2 (en) | 2005-11-30 | 2009-10-06 | Microsoft Corporation | Symmetric stereo model for handling occlusion |
CN101356831B (en) | 2005-11-30 | 2010-09-01 | 意大利电信股份公司 | Method for determining disperse optical parallax field of stereo vision |
JP4516516B2 (en) | 2005-12-07 | 2010-08-04 | 本田技研工業株式会社 | Person detection device, person detection method, and person detection program |
TWI296480B (en) | 2005-12-19 | 2008-05-01 | Quanta Comp Inc | Image camera of an electronic device |
JP4501855B2 (en) | 2005-12-22 | 2010-07-14 | ソニー株式会社 | Image signal processing apparatus, imaging apparatus, image signal processing method, and computer program |
JP2007180730A (en) | 2005-12-27 | 2007-07-12 | Eastman Kodak Co | Digital camera and data management method |
JP2009522591A (en) | 2005-12-30 | 2009-06-11 | ノキア コーポレイション | Method and apparatus for controlling autofocus of a video camera by tracking a region of interest |
US7855786B2 (en) | 2006-01-09 | 2010-12-21 | Bae Systems Spectral Solutions Llc | Single camera multi-spectral imager |
US7675080B2 (en) | 2006-01-10 | 2010-03-09 | Aptina Imaging Corp. | Uniform color filter arrays in a moat |
US8194169B2 (en) | 2006-01-20 | 2012-06-05 | Panasonic Corporation | Compound eye camera module and method of producing the same |
DE102006004802B4 (en) | 2006-01-23 | 2008-09-25 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Image acquisition system and method for producing at least one image capture system |
JP4834412B2 (en) | 2006-02-03 | 2011-12-14 | 富士フイルム株式会社 | Solid-state imaging device and electronic endoscope using the same |
US20070201859A1 (en) | 2006-02-24 | 2007-08-30 | Logitech Europe S.A. | Method and system for use of 3D sensors in an image capture device |
US7391572B2 (en) | 2006-03-01 | 2008-06-24 | International Business Machines Corporation | Hybrid optical/electronic structures fabricated by a common molding process |
US7924483B2 (en) | 2006-03-06 | 2011-04-12 | Smith Scott T | Fused multi-array color image sensor |
US7616254B2 (en) | 2006-03-16 | 2009-11-10 | Sony Corporation | Simple method for calculating camera defocus from an image scene |
US8360574B2 (en) | 2006-03-20 | 2013-01-29 | High Performance Optics, Inc. | High performance selective light wavelength filtering providing improved contrast sensitivity |
US7606484B1 (en) | 2006-03-23 | 2009-10-20 | Flir Systems, Inc. | Infrared and near-infrared camera hyperframing |
JP4615468B2 (en) | 2006-03-23 | 2011-01-19 | 富士フイルム株式会社 | Imaging device |
US7342212B2 (en) | 2006-03-31 | 2008-03-11 | Micron Technology, Inc. | Analog vertical sub-sampling in an active pixel sensor (APS) image sensor |
US7916934B2 (en) | 2006-04-04 | 2011-03-29 | Mitsubishi Electric Research Laboratories, Inc. | Method and system for acquiring, encoding, decoding and displaying 3D light fields |
US8044994B2 (en) | 2006-04-04 | 2011-10-25 | Mitsubishi Electric Research Laboratories, Inc. | Method and system for decoding and displaying 3D light fields |
US20070242141A1 (en) | 2006-04-14 | 2007-10-18 | Sony Corporation And Sony Electronics Inc. | Adjustable neutral density filter system for dynamic range compression from scene to imaging sensor |
CN101064780B (en) | 2006-04-30 | 2012-07-04 | 台湾新力国际股份有限公司 | Method and apparatus for improving image stitching accuracy using lens distortion correction |
US20070263114A1 (en) | 2006-05-01 | 2007-11-15 | Microalign Technologies, Inc. | Ultra-thin digital imaging device of high resolution for mobile electronic devices and method of imaging |
US7580620B2 (en) | 2006-05-08 | 2009-08-25 | Mitsubishi Electric Research Laboratories, Inc. | Method for deblurring images using optimized temporal coding patterns |
US9736346B2 (en) | 2006-05-09 | 2017-08-15 | Stereo Display, Inc | Imaging system improving image resolution of the system with low resolution image sensor |
US7889264B2 (en) | 2006-05-12 | 2011-02-15 | Ricoh Co., Ltd. | End-to-end design of superresolution electro-optic imaging systems |
US7916362B2 (en) | 2006-05-22 | 2011-03-29 | Eastman Kodak Company | Image sensor with improved light sensitivity |
US8139142B2 (en) | 2006-06-01 | 2012-03-20 | Microsoft Corporation | Video manipulation of red, green, blue, distance (RGB-Z) data including segmentation, up-sampling, and background substitution techniques |
IES20070229A2 (en) | 2006-06-05 | 2007-10-03 | Fotonation Vision Ltd | Image acquisition method and apparatus |
JP4631811B2 (en) | 2006-06-12 | 2011-02-16 | 株式会社日立製作所 | Imaging device |
JP5106870B2 (en) | 2006-06-14 | 2012-12-26 | 株式会社東芝 | Solid-state image sensor |
FR2902530A1 (en) | 2006-06-19 | 2007-12-21 | St Microelectronics Rousset | Polymer lens fabricating method for e.g. complementary MOS imager, involves realizing opaque zones on convex lens by degrading molecular structure of polymer material, where zones form diaphragm and diffraction network that forms filter |
TWI362550B (en) | 2007-06-21 | 2012-04-21 | Ether Precision Inc | The method for manufacturing the image captures unit |
US7925117B2 (en) | 2006-06-27 | 2011-04-12 | Honeywell International Inc. | Fusion of sensor data and synthetic data to form an integrated image |
KR100793369B1 (en) | 2006-07-06 | 2008-01-11 | 삼성전자주식회사 | Image sensor with improved resolution and image detection method using the same |
US20080024683A1 (en) | 2006-07-31 | 2008-01-31 | Niranjan Damera-Venkata | Overlapped multi-projector system with dithering |
JP2008039852A (en) | 2006-08-01 | 2008-02-21 | Agc Techno Glass Co Ltd | Glass optical element and its manufacturing method |
US20080030592A1 (en) | 2006-08-01 | 2008-02-07 | Eastman Kodak Company | Producing digital image with different resolution portions |
US8406562B2 (en) | 2006-08-11 | 2013-03-26 | Geo Semiconductor Inc. | System and method for automated calibration and correction of display geometry and color |
EP1892688B1 (en) | 2006-08-24 | 2010-09-01 | Valeo Vision | Method for determining the passing of a vehicle in a bottleneck |
US8687087B2 (en) | 2006-08-29 | 2014-04-01 | Csr Technology Inc. | Digital camera with selectively increased dynamic range by control of parameters during image acquisition |
US8306063B2 (en) | 2006-08-29 | 2012-11-06 | EXFO Services Assurance, Inc. | Real-time transport protocol stream detection system and method |
KR100746360B1 (en) | 2006-08-31 | 2007-08-06 | 삼성전기주식회사 | Stamper manufacturing method |
NO326372B1 (en) | 2006-09-21 | 2008-11-17 | Polight As | Polymer Lens |
WO2008039802A2 (en) | 2006-09-25 | 2008-04-03 | Ophthonix, Incorporated | Method for correction of chromatic aberration and achromatic lens |
JP4403162B2 (en) | 2006-09-29 | 2010-01-20 | 株式会社東芝 | Stereoscopic image display device and method for producing stereoscopic image |
US20080080028A1 (en) | 2006-10-02 | 2008-04-03 | Micron Technology, Inc. | Imaging method, apparatus and system having extended depth of field |
US8031258B2 (en) | 2006-10-04 | 2011-10-04 | Omnivision Technologies, Inc. | Providing multiple video signals from single sensor |
JP5244806B2 (en) | 2006-10-11 | 2013-07-24 | ポライト エイエス | Small adjustable lens design |
EP2074445B1 (en) | 2006-10-11 | 2017-04-12 | poLight AS | Method for manufacturing adjustable lens |
US8073196B2 (en) | 2006-10-16 | 2011-12-06 | University Of Southern California | Detection and tracking of moving objects from a moving platform in presence of strong parallax |
US7702229B2 (en) | 2006-10-18 | 2010-04-20 | Eastman Kodak Company | Lens array assisted focus detection |
JP4349456B2 (en) | 2006-10-23 | 2009-10-21 | ソニー株式会社 | Solid-state image sensor |
JP4942221B2 (en) | 2006-10-25 | 2012-05-30 | 国立大学法人東京工業大学 | High resolution virtual focal plane image generation method |
US7888159B2 (en) | 2006-10-26 | 2011-02-15 | Omnivision Technologies, Inc. | Image sensor having curved micro-mirrors over the sensing photodiode and method for fabricating |
JP4452951B2 (en) | 2006-11-02 | 2010-04-21 | 富士フイルム株式会社 | Distance image generation method and apparatus |
KR20080043106A (en) | 2006-11-13 | 2008-05-16 | 삼성전자주식회사 | Optical lens and manufacturing method thereof |
US8059162B2 (en) | 2006-11-15 | 2011-11-15 | Sony Corporation | Imaging apparatus and method, and method for designing imaging apparatus |
US20080118241A1 (en) | 2006-11-16 | 2008-05-22 | Tekolste Robert | Control of stray light in camera systems employing an optics stack and associated methods |
CN201043890Y (en) | 2006-11-17 | 2008-04-02 | 中国科学院上海光学精密机械研究所 | Single Aperture Multiple Imaging Optical Imaging Ranging Device |
SG176440A1 (en) | 2006-11-21 | 2011-12-29 | Mantisvision Ltd | 3d geometric modeling and 3d video content creation |
DE102006055641B4 (en) * | 2006-11-22 | 2013-01-31 | Visumotion Gmbh | Arrangement and method for recording and reproducing images of a scene and / or an object |
KR20080047002A (en) | 2006-11-24 | 2008-05-28 | 엘지이노텍 주식회사 | Lens assembly of camera module and its manufacturing method |
JP4406937B2 (en) | 2006-12-01 | 2010-02-03 | 富士フイルム株式会社 | Imaging device |
US20100265385A1 (en) | 2009-04-18 | 2010-10-21 | Knight Timothy J | Light Field Camera Image, File and Configuration Data, and Methods of Using, Storing and Communicating Same |
US8559705B2 (en) | 2006-12-01 | 2013-10-15 | Lytro, Inc. | Interactive refocusing of electronic images |
JP5040493B2 (en) | 2006-12-04 | 2012-10-03 | ソニー株式会社 | Imaging apparatus and imaging method |
US8242426B2 (en) | 2006-12-12 | 2012-08-14 | Dolby Laboratories Licensing Corporation | Electronic camera having multiple sensors for capturing high dynamic range images and related methods |
US7646549B2 (en) | 2006-12-18 | 2010-01-12 | Xceed Imaging Ltd | Imaging system and method for providing extended depth of focus, range extraction and super resolved imaging |
US8213500B2 (en) | 2006-12-21 | 2012-07-03 | Sharp Laboratories Of America, Inc. | Methods and systems for processing film grain noise |
TWI324015B (en) | 2006-12-22 | 2010-04-21 | Ind Tech Res Inst | Autofocus searching method |
US8103111B2 (en) | 2006-12-26 | 2012-01-24 | Olympus Imaging Corp. | Coding method, electronic camera, recording medium storing coded program, and decoding method |
US20080158259A1 (en) | 2006-12-28 | 2008-07-03 | Texas Instruments Incorporated | Image warping and lateral color correction |
US20080158698A1 (en) | 2006-12-29 | 2008-07-03 | Chao-Chi Chang | Lens barrel array and lens array and the method of making the same |
US7973823B2 (en) | 2006-12-29 | 2011-07-05 | Nokia Corporation | Method and system for image pre-processing |
US20080165257A1 (en) | 2007-01-05 | 2008-07-10 | Micron Technology, Inc. | Configurable pixel array system and method |
JP4993578B2 (en) | 2007-01-15 | 2012-08-08 | オリンパスイメージング株式会社 | Image file playback device, image file processing and editing device |
US8655052B2 (en) | 2007-01-26 | 2014-02-18 | Intellectual Discovery Co., Ltd. | Methodology for 3D scene reconstruction from 2D image sequences |
JP5024992B2 (en) | 2007-02-02 | 2012-09-12 | 株式会社ジャパンディスプレイセントラル | Display device |
US7792423B2 (en) | 2007-02-06 | 2010-09-07 | Mitsubishi Electric Research Laboratories, Inc. | 4D light field cameras |
US7956988B1 (en) | 2007-02-06 | 2011-06-07 | Alpha Technology, LLC | Light detection and ranging systems and related methods |
JP4969474B2 (en) | 2007-02-09 | 2012-07-04 | オリンパスイメージング株式会社 | Decoding method, decoding device, and decoding program |
JP4386083B2 (en) | 2007-02-27 | 2009-12-16 | トヨタ自動車株式会社 | Parking assistance device |
JP4153013B1 (en) | 2007-03-06 | 2008-09-17 | シャープ株式会社 | Imaging lens, imaging unit, and portable information terminal including the same |
US7755679B2 (en) | 2007-03-07 | 2010-07-13 | Altasens, Inc. | Apparatus and method for reducing edge effect in an image sensor |
US7859588B2 (en) | 2007-03-09 | 2010-12-28 | Eastman Kodak Company | Method and apparatus for operating a dual lens camera to augment an image |
US7729602B2 (en) | 2007-03-09 | 2010-06-01 | Eastman Kodak Company | Camera using multiple lenses and image sensors operable in a default imaging mode |
US7683962B2 (en) | 2007-03-09 | 2010-03-23 | Eastman Kodak Company | Camera using multiple lenses and image sensors in a rangefinder configuration to provide a range map |
US7676146B2 (en) | 2007-03-09 | 2010-03-09 | Eastman Kodak Company | Camera using multiple lenses and image sensors to provide improved focusing capability |
JP2008242658A (en) | 2007-03-26 | 2008-10-09 | Funai Electric Co Ltd | 3D object imaging device |
JP4915859B2 (en) | 2007-03-26 | 2012-04-11 | 船井電機株式会社 | Object distance deriving device |
US7738017B2 (en) | 2007-03-27 | 2010-06-15 | Aptina Imaging Corporation | Method and apparatus for automatic linear shift parallax correction for multi-array image systems |
US8055466B2 (en) | 2007-03-30 | 2011-11-08 | Mitutoyo Corporation | Global calibration for stereo vision probe |
US8165418B2 (en) | 2007-03-30 | 2012-04-24 | Brother Kogyo Kabushiki Kaisha | Image processor |
TWI433052B (en) | 2007-04-02 | 2014-04-01 | Primesense Ltd | Depth mapping using projected patterns |
US8213711B2 (en) | 2007-04-03 | 2012-07-03 | Her Majesty The Queen In Right Of Canada As Represented By The Minister Of Industry, Through The Communications Research Centre Canada | Method and graphical user interface for modifying depth maps |
US8098941B2 (en) | 2007-04-03 | 2012-01-17 | Aptina Imaging Corporation | Method and apparatus for parallelization of image compression encoders |
JP2008258885A (en) | 2007-04-04 | 2008-10-23 | Texas Instr Japan Ltd | Imaging apparatus and driving method of imaging apparatus |
CN101281282A (en) | 2007-04-04 | 2008-10-08 | 鸿富锦精密工业(深圳)有限公司 | Lens module |
CN103839955B (en) | 2007-04-18 | 2016-05-25 | 因维萨热技术公司 | For material, the system and method for electrooptical device |
WO2009023044A2 (en) | 2007-04-24 | 2009-02-19 | 21 Ct, Inc. | Method and system for fast dense stereoscopic ranging |
KR100869219B1 (en) | 2007-05-03 | 2008-11-18 | 동부일렉트로닉스 주식회사 | Image sensor and its manufacturing method |
US8462220B2 (en) | 2007-05-09 | 2013-06-11 | Aptina Imaging Corporation | Method and apparatus for improving low-light performance for small pixel image sensors |
US7812869B2 (en) | 2007-05-11 | 2010-10-12 | Aptina Imaging Corporation | Configurable pixel array system and method |
JP4341695B2 (en) | 2007-05-17 | 2009-10-07 | ソニー株式会社 | Image input processing device, imaging signal processing circuit, and imaging signal noise reduction method |
JP4337911B2 (en) | 2007-05-24 | 2009-09-30 | ソニー株式会社 | Imaging device, imaging circuit, and imaging method |
US20080298674A1 (en) | 2007-05-29 | 2008-12-04 | Image Masters Inc. | Stereoscopic Panoramic imaging system |
US7733575B2 (en) | 2007-05-31 | 2010-06-08 | Artificial Muscle, Inc. | Optical systems employing compliant electroactive materials |
US8290358B1 (en) | 2007-06-25 | 2012-10-16 | Adobe Systems Incorporated | Methods and apparatus for light-field imaging |
JP5406182B2 (en) | 2007-06-26 | 2014-02-05 | コーニンクレッカ フィリップス エヌ ヴェ | Method and system for encoding a three-dimensional video signal, included three-dimensional video signal, and method and signal for a decoder for a three-dimensional video signal |
US8125619B2 (en) | 2007-07-25 | 2012-02-28 | Eminent Electronic Technology Corp. | Integrated ambient light sensor and distance sensor |
JP5006727B2 (en) | 2007-07-26 | 2012-08-22 | 株式会社リコー | Image processing apparatus and digital camera |
WO2009020977A1 (en) | 2007-08-06 | 2009-02-12 | Adobe Systems Incorporated | Method and apparatus for radiance capture by multiplexing in the frequency domain |
EP2034338A1 (en) | 2007-08-11 | 2009-03-11 | ETH Zurich | Liquid Lens System |
EP2026563A1 (en) | 2007-08-14 | 2009-02-18 | Deutsche Thomson OHG | System and method for detecting defective pixels |
US7782364B2 (en) | 2007-08-21 | 2010-08-24 | Aptina Imaging Corporation | Multi-array sensor with integrated sub-array for parallax detection and photometer functionality |
US20090066693A1 (en) | 2007-09-06 | 2009-03-12 | Roc Carson | Encoding A Depth Map Into An Image Using Analysis Of Two Consecutive Captured Frames |
US7973834B2 (en) | 2007-09-24 | 2011-07-05 | Jianwen Yang | Electro-optical foveated imaging and tracking system |
US20090079862A1 (en) | 2007-09-25 | 2009-03-26 | Micron Technology, Inc. | Method and apparatus providing imaging auto-focus utilizing absolute blur value |
US20090086074A1 (en) | 2007-09-27 | 2009-04-02 | Omnivision Technologies, Inc. | Dual mode camera solution apparatus, system, and method |
US7940311B2 (en) | 2007-10-03 | 2011-05-10 | Nokia Corporation | Multi-exposure pattern for enhancing dynamic range of images |
JP5172267B2 (en) | 2007-10-09 | 2013-03-27 | 富士フイルム株式会社 | Imaging device |
US8049289B2 (en) | 2007-10-11 | 2011-11-01 | Dongbu Hitek Co., Ltd. | Image sensor and method for manufacturing the same |
US8938009B2 (en) | 2007-10-12 | 2015-01-20 | Qualcomm Incorporated | Layered encoded bitstream structure |
US7956924B2 (en) | 2007-10-18 | 2011-06-07 | Adobe Systems Incorporated | Fast computational camera based on two arrays of lenses |
US7787112B2 (en) | 2007-10-22 | 2010-08-31 | Visiongate, Inc. | Depth of field extension for optical tomography |
US7920193B2 (en) | 2007-10-23 | 2011-04-05 | Aptina Imaging Corporation | Methods, systems and apparatuses using barrier self-calibration for high dynamic range imagers |
US7777804B2 (en) | 2007-10-26 | 2010-08-17 | Omnivision Technologies, Inc. | High dynamic range sensor with reduced line memory for color interpolation |
WO2009061814A2 (en) | 2007-11-05 | 2009-05-14 | University Of Florida Research Foundation, Inc. | Lossless data compression and real-time decompression |
US20090128644A1 (en) | 2007-11-15 | 2009-05-21 | Camp Jr William O | System and method for generating a photograph |
US7852461B2 (en) | 2007-11-15 | 2010-12-14 | Microsoft International Holdings B.V. | Dual mode depth imaging |
US8351685B2 (en) | 2007-11-16 | 2013-01-08 | Gwangju Institute Of Science And Technology | Device and method for estimating depth map, and method for generating intermediate image and method for encoding multi-view video using the same |
US8126279B2 (en) | 2007-11-19 | 2012-02-28 | The University Of Arizona | Lifting-based view compensated compression and remote visualization of volume rendered images |
KR20090055803A (en) | 2007-11-29 | 2009-06-03 | 광주과학기술원 | Method and apparatus for generating multiview depth map and method for generating variance in multiview image |
JP5010445B2 (en) | 2007-11-29 | 2012-08-29 | パナソニック株式会社 | Manufacturing method of mold for microlens array |
US8384803B2 (en) | 2007-12-13 | 2013-02-26 | Keigo Iizuka | Camera system and method for amalgamating images to create an omni-focused image |
TWI353778B (en) | 2007-12-21 | 2011-12-01 | Ind Tech Res Inst | Moving object detection apparatus and method |
US20110031381A1 (en) | 2007-12-28 | 2011-02-10 | Hiok-Nam Tay | Light guide array for an image sensor |
TWI362628B (en) | 2007-12-28 | 2012-04-21 | Ind Tech Res Inst | Methof for producing an image with depth by using 2d image |
JP4413261B2 (en) | 2008-01-10 | 2010-02-10 | シャープ株式会社 | Imaging apparatus and optical axis control method |
JP5198295B2 (en) | 2008-01-15 | 2013-05-15 | 富士フイルム株式会社 | Image sensor position adjustment method, camera module manufacturing method and apparatus, and camera module |
US8189065B2 (en) | 2008-01-23 | 2012-05-29 | Adobe Systems Incorporated | Methods and apparatus for full-resolution light-field capture and rendering |
US7962033B2 (en) | 2008-01-23 | 2011-06-14 | Adobe Systems Incorporated | Methods and apparatus for full-resolution light-field capture and rendering |
JP4956452B2 (en) | 2008-01-25 | 2012-06-20 | 富士重工業株式会社 | Vehicle environment recognition device |
GB0802290D0 (en) | 2008-02-08 | 2008-03-12 | Univ Kent Canterbury | Camera adapter based optical imaging apparatus |
US8319301B2 (en) | 2008-02-11 | 2012-11-27 | Omnivision Technologies, Inc. | Self-aligned filter for an image sensor |
JP2009206922A (en) | 2008-02-28 | 2009-09-10 | Funai Electric Co Ltd | Compound-eye imaging apparatus |
CN101520532A (en) | 2008-02-29 | 2009-09-02 | 鸿富锦精密工业(深圳)有限公司 | Composite lens |
KR101564383B1 (en) | 2008-03-03 | 2015-10-30 | 에이브이오 유에스에이 홀딩 2 코포레이션 | Object matching for tracking, indexing, and search |
JPWO2009119229A1 (en) | 2008-03-26 | 2011-07-21 | コニカミノルタホールディングス株式会社 | 3D imaging apparatus and calibration method for 3D imaging apparatus |
US8497905B2 (en) | 2008-04-11 | 2013-07-30 | nearmap australia pty ltd. | Systems and methods of capturing large area images in detail including cascaded cameras and/or calibration features |
US8259208B2 (en) | 2008-04-15 | 2012-09-04 | Sony Corporation | Method and apparatus for performing touch-based adjustments within imaging devices |
US7843554B2 (en) | 2008-04-25 | 2010-11-30 | Rockwell Collins, Inc. | High dynamic range sensor system and method |
US8280194B2 (en) | 2008-04-29 | 2012-10-02 | Sony Corporation | Reduced hardware implementation for a two-picture depth map algorithm |
US8155456B2 (en) | 2008-04-29 | 2012-04-10 | Adobe Systems Incorporated | Method and apparatus for block-based compression of light-field images |
US8724921B2 (en) | 2008-05-05 | 2014-05-13 | Aptina Imaging Corporation | Method of capturing high dynamic range images with objects in the scene |
US8633996B2 (en) | 2008-05-09 | 2014-01-21 | Rambus Inc. | Image sensor having nonlinear response |
JP2009273035A (en) | 2008-05-09 | 2009-11-19 | Toshiba Corp | Image compression apparatus, image decompression apparatus, and image processor |
US8208543B2 (en) | 2008-05-19 | 2012-06-26 | Microsoft Corporation | Quantization and differential coding of alpha image data |
US8866920B2 (en) | 2008-05-20 | 2014-10-21 | Pelican Imaging Corporation | Capturing and processing of images using monolithic camera array with heterogeneous imagers |
DK3876510T3 (en) | 2008-05-20 | 2024-11-11 | Adeia Imaging Llc | CAPTURE AND PROCESSING OF IMAGES USING MONOLITHIC CAMERA ARRAY WITH HETEROGENEOUS IMAGES |
US8442355B2 (en) | 2008-05-23 | 2013-05-14 | Samsung Electronics Co., Ltd. | System and method for generating a multi-dimensional image |
US8125559B2 (en) | 2008-05-25 | 2012-02-28 | Avistar Communications Corporation | Image formation for large photosensor array surfaces |
US8131097B2 (en) | 2008-05-28 | 2012-03-06 | Aptina Imaging Corporation | Method and apparatus for extended depth-of-field image restoration |
US8244058B1 (en) | 2008-05-30 | 2012-08-14 | Adobe Systems Incorporated | Method and apparatus for managing artifacts in frequency domain processing of light-field images |
JP2009300268A (en) | 2008-06-13 | 2009-12-24 | Nippon Hoso Kyokai <Nhk> | Three-dimensional information detection device |
US7710667B2 (en) | 2008-06-25 | 2010-05-04 | Aptina Imaging Corp. | Imaging module with symmetrical lens system and method of manufacture |
WO2009157273A1 (en) | 2008-06-25 | 2009-12-30 | コニカミノルタオプト株式会社 | Imaging optical system, and imaging lens manufacturing method |
KR101000531B1 (en) | 2008-06-26 | 2010-12-14 | 에스디씨마이크로 주식회사 | CC management system using wireless LAN that increases data transmission range |
US7916396B2 (en) | 2008-06-27 | 2011-03-29 | Micron Technology, Inc. | Lens master devices, lens structures, imaging devices, and methods and apparatuses of making the same |
US8326069B2 (en) | 2008-06-30 | 2012-12-04 | Intel Corporation | Computing higher resolution images from multiple lower resolution images |
US7773317B2 (en) | 2008-07-01 | 2010-08-10 | Aptina Imaging Corp. | Lens system with symmetrical optics |
US7920339B2 (en) | 2008-07-02 | 2011-04-05 | Aptina Imaging Corporation | Method and apparatus providing singlet wafer lens system with field flattener |
US8456517B2 (en) | 2008-07-09 | 2013-06-04 | Primesense Ltd. | Integrated processor for 3D mapping |
KR101445185B1 (en) | 2008-07-10 | 2014-09-30 | 삼성전자주식회사 | Flexible image capturing apparatus having a plurality of image capturing units and method of manufacturing the same |
CN102112845B (en) | 2008-08-06 | 2013-09-11 | 形创有限公司 | System for adaptive three-dimensional scanning of surface characteristics |
CN102124742B (en) | 2008-08-20 | 2013-09-11 | 汤姆逊许可公司 | Refined depth map |
CN101656259A (en) | 2008-08-20 | 2010-02-24 | 鸿富锦精密工业(深圳)有限公司 | Image sensor packaging structure, packaging method and camera module |
US7924312B2 (en) | 2008-08-22 | 2011-04-12 | Fluke Corporation | Infrared and visible-light image registration |
CN102138102A (en) | 2008-09-01 | 2011-07-27 | 兰斯维克托公司 | Wafer-level fabrication of liquid crystal optoelectronic devices |
JP5105482B2 (en) | 2008-09-01 | 2012-12-26 | 船井電機株式会社 | Optical condition design method and compound eye imaging apparatus |
US8098297B2 (en) | 2008-09-03 | 2012-01-17 | Sony Corporation | Pre- and post-shutter signal image capture and sort for digital camera |
KR20100028344A (en) | 2008-09-04 | 2010-03-12 | 삼성전자주식회사 | Method and apparatus for editing image of portable terminal |
JP5238429B2 (en) | 2008-09-25 | 2013-07-17 | 株式会社東芝 | Stereoscopic image capturing apparatus and stereoscopic image capturing system |
US8553093B2 (en) | 2008-09-30 | 2013-10-08 | Sony Corporation | Method and apparatus for super-resolution imaging using digital imaging devices |
KR101468267B1 (en) | 2008-10-02 | 2014-12-15 | 프라운호퍼-게젤샤프트 추르 푀르데룽 데어 안제반텐 포르슝 에 파우 | Intermediate view synthesis and multi-view data signal extraction |
US9619917B2 (en) | 2008-10-03 | 2017-04-11 | Apple Inc. | Depth of field for a camera in a media-editing application |
US9064476B2 (en) | 2008-10-04 | 2015-06-23 | Microsoft Technology Licensing, Llc | Image super-resolution using gradient profile prior |
US8310525B2 (en) | 2008-10-07 | 2012-11-13 | Seiko Epson Corporation | One-touch projector alignment for 3D stereo display |
JP5547739B2 (en) | 2008-10-15 | 2014-07-16 | イノベイティブ テクノロジー ディストリビューターズ エルエルシー | Digital processing method and system for optical flow determination |
JP2010096723A (en) | 2008-10-20 | 2010-04-30 | Funai Electric Co Ltd | Device for deriving distance of object |
US8436909B2 (en) | 2008-10-21 | 2013-05-07 | Stmicroelectronics S.R.L. | Compound camera sensor and related method of processing digital images |
US8063975B2 (en) | 2008-10-29 | 2011-11-22 | Jabil Circuit, Inc. | Positioning wafer lenses on electronic imagers |
KR101502597B1 (en) | 2008-11-13 | 2015-03-13 | 삼성전자주식회사 | Wide depth of field 3d display apparatus and method |
WO2010057081A1 (en) | 2008-11-14 | 2010-05-20 | The Scripps Research Institute | Image analysis platform for identifying artifacts in samples and laboratory consumables |
AU2008246243B2 (en) | 2008-11-19 | 2011-12-22 | Canon Kabushiki Kaisha | DVC as generic file format for plenoptic camera |
WO2010065344A1 (en) | 2008-11-25 | 2010-06-10 | Refocus Imaging, Inc. | System of and method for video refocusing |
JP4852591B2 (en) | 2008-11-27 | 2012-01-11 | 富士フイルム株式会社 | Stereoscopic image processing apparatus, method, recording medium, and stereoscopic imaging apparatus |
US8013904B2 (en) | 2008-12-09 | 2011-09-06 | Seiko Epson Corporation | View projection matrix based high performance low latency display pipeline |
JP5311016B2 (en) | 2008-12-10 | 2013-10-09 | コニカミノルタ株式会社 | Stereo camera unit and stereo matching method |
KR101200490B1 (en) | 2008-12-10 | 2012-11-12 | 한국전자통신연구원 | Image registration device and method |
US8149323B2 (en) | 2008-12-18 | 2012-04-03 | Qualcomm Incorporated | System and method to autofocus assisted by autoexposure control |
JP4631966B2 (en) | 2008-12-22 | 2011-02-16 | ソニー株式会社 | Image processing apparatus, image processing method, and program |
CN101770060B (en) | 2008-12-27 | 2014-03-26 | 鸿富锦精密工业(深圳)有限公司 | Camera module and assembly method thereof |
US8405742B2 (en) | 2008-12-30 | 2013-03-26 | Massachusetts Institute Of Technology | Processing images having different focus |
US20100177411A1 (en) | 2009-01-09 | 2010-07-15 | Shashikant Hegde | Wafer level lens replication on micro-electrical-mechanical systems |
WO2010081010A2 (en) | 2009-01-09 | 2010-07-15 | New York University | Methods, computer-accessible medium and systems for facilitating dark flash photography |
CN102272796B (en) | 2009-01-09 | 2014-03-12 | 柯尼卡美能达控股株式会社 | Motion vector generation apparatus and motion vector generation method |
US8189089B1 (en) | 2009-01-20 | 2012-05-29 | Adobe Systems Incorporated | Methods and apparatus for reducing plenoptic camera artifacts |
US8315476B1 (en) | 2009-01-20 | 2012-11-20 | Adobe Systems Incorporated | Super-resolution with the focused plenoptic camera |
US8300108B2 (en) | 2009-02-02 | 2012-10-30 | L-3 Communications Cincinnati Electronics Corporation | Multi-channel imaging devices comprising unit cells |
US20100194860A1 (en) | 2009-02-03 | 2010-08-05 | Bit Cauldron Corporation | Method of stereoscopic 3d image capture using a mobile device, cradle or dongle |
US8761491B2 (en) | 2009-02-06 | 2014-06-24 | Himax Technologies Limited | Stereo-matching processor using belief propagation |
US8290301B2 (en) | 2009-02-06 | 2012-10-16 | Raytheon Company | Optimized imaging system for collection of high resolution imagery |
KR101776955B1 (en) | 2009-02-10 | 2017-09-08 | 소니 주식회사 | Solid-state imaging device, method of manufacturing the same, and electronic apparatus |
WO2010095440A1 (en) | 2009-02-20 | 2010-08-26 | パナソニック株式会社 | Recording medium, reproduction device, and integrated circuit |
US8520970B2 (en) | 2010-04-23 | 2013-08-27 | Flir Systems Ab | Infrared resolution and contrast enhancement with fusion |
US8207759B2 (en) | 2009-03-12 | 2012-06-26 | Fairchild Semiconductor Corporation | MIPI analog switch for automatic selection of multiple inputs based on clock voltages |
US8542287B2 (en) | 2009-03-19 | 2013-09-24 | Digitaloptics Corporation | Dual sensor camera |
US8450821B2 (en) | 2009-03-26 | 2013-05-28 | Micron Technology, Inc. | Method and apparatus providing combined spacer and optical lens element |
US8106949B2 (en) | 2009-03-26 | 2012-01-31 | Seiko Epson Corporation | Small memory footprint light transport matrix capture |
US7901095B2 (en) | 2009-03-27 | 2011-03-08 | Seiko Epson Corporation | Resolution scalable view projection |
JP4529010B1 (en) | 2009-03-30 | 2010-08-25 | シャープ株式会社 | Imaging device |
JP5222205B2 (en) | 2009-04-03 | 2013-06-26 | Kddi株式会社 | Image processing apparatus, method, and program |
WO2010116367A1 (en) | 2009-04-07 | 2010-10-14 | Nextvision Stabilized Systems Ltd | Continuous electronic zoom for an imaging system with multiple imaging devices having different fixed fov |
US20100259610A1 (en) | 2009-04-08 | 2010-10-14 | Celsia, Llc | Two-Dimensional Display Synced with Real World Object Movement |
US8294099B2 (en) | 2009-04-10 | 2012-10-23 | Bae Systems Information And Electronic Systems Integration Inc. | On-wafer butted microbolometer imaging array |
JP5463718B2 (en) | 2009-04-16 | 2014-04-09 | ソニー株式会社 | Imaging device |
US8717417B2 (en) | 2009-04-16 | 2014-05-06 | Primesense Ltd. | Three-dimensional mapping and imaging |
US20120249550A1 (en) | 2009-04-18 | 2012-10-04 | Lytro, Inc. | Selective Transmission of Image Data Based on Device Attributes |
US8908058B2 (en) | 2009-04-18 | 2014-12-09 | Lytro, Inc. | Storage and transmission of pictures including multiple frames |
EP2244484B1 (en) * | 2009-04-22 | 2012-03-28 | Raytrix GmbH | Digital imaging method for synthesizing an image using data recorded with a plenoptic camera |
CN101527046B (en) | 2009-04-28 | 2012-09-05 | 青岛海信数字多媒体技术国家重点实验室有限公司 | Motion detection method, device and system |
US8271544B2 (en) | 2009-05-01 | 2012-09-18 | Creative Technology Ltd | Data file having more than one mode of operation |
DE102009003110A1 (en) | 2009-05-14 | 2010-11-18 | Robert Bosch Gmbh | Image processing method for determining depth information from at least two input images recorded by means of a stereo camera system |
US8203633B2 (en) | 2009-05-27 | 2012-06-19 | Omnivision Technologies, Inc. | Four-channel color filter array pattern |
US10091439B2 (en) | 2009-06-03 | 2018-10-02 | Flir Systems, Inc. | Imager with array of multiple infrared imaging modules |
KR20100130423A (en) | 2009-06-03 | 2010-12-13 | 삼성전자주식회사 | Wafer-level lens module and imaging module having same |
US8766808B2 (en) | 2010-03-09 | 2014-07-01 | Flir Systems, Inc. | Imager with multiple sensor arrays |
CN101931742B (en) | 2009-06-18 | 2013-04-24 | 鸿富锦精密工业(深圳)有限公司 | Image sensing module and image capture module |
US20100321640A1 (en) | 2009-06-22 | 2010-12-23 | Industrial Technology Research Institute | Projection display chip |
JP5254893B2 (en) | 2009-06-26 | 2013-08-07 | キヤノン株式会社 | Image conversion method and apparatus, and pattern identification method and apparatus |
WO2011008443A2 (en) | 2009-06-29 | 2011-01-20 | Lensvector Inc. | Wafer level camera module with active optical element |
JP2011030184A (en) | 2009-07-01 | 2011-02-10 | Sony Corp | Image processing apparatus, and image processing method |
US8212197B2 (en) | 2009-07-02 | 2012-07-03 | Xerox Corporation | Image sensor with integration time compensation |
JP2011017764A (en) | 2009-07-07 | 2011-01-27 | Konica Minolta Opto Inc | Imaging lens, imaging apparatus and portable terminal |
US8345144B1 (en) | 2009-07-15 | 2013-01-01 | Adobe Systems Incorporated | Methods and apparatus for rich image capture with focused plenoptic cameras |
US20110019243A1 (en) | 2009-07-21 | 2011-01-27 | Constant Jr Henry J | Stereoscopic form reader |
CN101964866B (en) | 2009-07-24 | 2013-03-20 | 鸿富锦精密工业(深圳)有限公司 | Computation and image pickup type digital camera |
US8436893B2 (en) | 2009-07-31 | 2013-05-07 | 3Dmedia Corporation | Methods, systems, and computer-readable storage media for selecting image capture positions to generate three-dimensional (3D) images |
US20110032341A1 (en) | 2009-08-04 | 2011-02-10 | Ignatov Artem Konstantinovich | Method and system to transform stereo content |
US8577183B2 (en) | 2009-08-05 | 2013-11-05 | Raytheon Company | Resolution on demand |
CN102483511B (en) | 2009-08-11 | 2014-11-12 | 乙太精密有限公司 | Method and device for aligning a lens with an optical system |
JP2011044801A (en) | 2009-08-19 | 2011-03-03 | Toshiba Corp | Image processor |
US8154632B2 (en) | 2009-08-24 | 2012-04-10 | Lifesize Communications, Inc. | Detection of defective pixels in an image sensor |
KR101680300B1 (en) | 2009-08-31 | 2016-11-28 | 삼성전자주식회사 | Liquid lens and method for manufacturing the same |
US9274699B2 (en) | 2009-09-03 | 2016-03-01 | Obscura Digital | User interface for a large scale multi-user, multi-touch system |
US8411146B2 (en) | 2009-09-04 | 2013-04-02 | Lockheed Martin Corporation | Single camera color and infrared polarimetric imaging |
EP2476248A1 (en) | 2009-09-07 | 2012-07-18 | Nokia Corp. | An apparatus |
FR2950153B1 (en) | 2009-09-15 | 2011-12-23 | Commissariat Energie Atomique | OPTICAL DEVICE WITH DEFORMABLE MEMBRANE WITH PIEZOELECTRIC ACTUATION |
US20140076336A1 (en) | 2009-09-17 | 2014-03-20 | Ascentia Health, Inc. | Ear insert for relief of tmj discomfort and headaches |
CN102549507B (en) | 2009-10-02 | 2014-08-20 | 皇家飞利浦电子股份有限公司 | Selecting viewpoints for generating additional views in 3D video |
US8199165B2 (en) | 2009-10-14 | 2012-06-12 | Hewlett-Packard Development Company, L.P. | Methods and systems for object segmentation in digital images |
DE102009049387B4 (en) | 2009-10-14 | 2016-05-25 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Apparatus, image processing apparatus and method for optical imaging |
EP2499829B1 (en) | 2009-10-14 | 2019-04-17 | Dolby International AB | Methods and devices for depth map processing |
US8502909B2 (en) | 2009-10-19 | 2013-08-06 | Pixar | Super light-field lens |
US20110207074A1 (en) | 2009-10-26 | 2011-08-25 | Olaf Andrew Hall-Holt | Dental imaging system and method |
EP2494402B1 (en) | 2009-10-30 | 2018-04-18 | Hewlett-Packard Development Company, L.P. | Stereo display systems |
WO2011053711A1 (en) | 2009-10-30 | 2011-05-05 | Invisage Technologies, Inc. | Systems and methods for color binning |
WO2011055655A1 (en) | 2009-11-05 | 2011-05-12 | コニカミノルタオプト株式会社 | Image pickup device, optical unit, wafer lens laminated body, and method for manufacturing wafer lens laminated body |
WO2011058876A1 (en) | 2009-11-13 | 2011-05-19 | 富士フイルム株式会社 | Distance measuring device, distance measuring method, distance measuring program, distance measuring system, and image capturing device |
TR200908688A2 (en) | 2009-11-17 | 2011-06-21 | Vestel Elektron�K San. Ve T�C. A.�. | Noise reduction with depth compensation in multi-image video. |
JP5399215B2 (en) | 2009-11-18 | 2014-01-29 | シャープ株式会社 | Multi-lens camera device and electronic information device |
US8514491B2 (en) * | 2009-11-20 | 2013-08-20 | Pelican Imaging Corporation | Capturing and processing of images using monolithic camera array with heterogeneous imagers |
WO2011066275A2 (en) | 2009-11-25 | 2011-06-03 | Massachusetts Institute Of Technology | Actively addressable aperture light field camera |
KR101608970B1 (en) | 2009-11-27 | 2016-04-05 | 삼성전자주식회사 | Apparatus and method for processing image using light field data |
US8730338B2 (en) | 2009-12-01 | 2014-05-20 | Nokia Corporation | Set of camera modules hinged on a body and functionally connected to a single actuator |
US8400555B1 (en) | 2009-12-01 | 2013-03-19 | Adobe Systems Incorporated | Focused plenoptic camera employing microlenses with different focal lengths |
JP5446797B2 (en) | 2009-12-04 | 2014-03-19 | 株式会社リコー | Imaging device |
US8446492B2 (en) | 2009-12-10 | 2013-05-21 | Honda Motor Co., Ltd. | Image capturing device, method of searching for occlusion region, and program |
JP5387377B2 (en) | 2009-12-14 | 2014-01-15 | ソニー株式会社 | Image processing apparatus, image processing method, and program |
US9030530B2 (en) | 2009-12-15 | 2015-05-12 | Thomson Licensing | Stereo-image quality and disparity/depth indications |
KR101281961B1 (en) | 2009-12-21 | 2013-07-03 | 한국전자통신연구원 | Method and apparatus for editing depth video |
US20110153248A1 (en) | 2009-12-23 | 2011-06-23 | Yeming Gu | Ophthalmic quality metric system |
EP2518995B1 (en) | 2009-12-24 | 2018-08-22 | Sharp Kabushiki Kaisha | Multocular image pickup apparatus and multocular image pickup method |
JP4983905B2 (en) | 2009-12-25 | 2012-07-25 | カシオ計算機株式会社 | Imaging apparatus, 3D modeling data generation method, and program |
KR101643607B1 (en) | 2009-12-30 | 2016-08-10 | 삼성전자주식회사 | Method and apparatus for generating of image data |
CN102117576A (en) | 2009-12-31 | 2011-07-06 | 鸿富锦精密工业(深圳)有限公司 | Digital photo frame |
CN102131044B (en) | 2010-01-20 | 2014-03-26 | 鸿富锦精密工业(深圳)有限公司 | Camera module |
US8649008B2 (en) | 2010-02-04 | 2014-02-11 | University Of Southern California | Combined spectral and polarimetry imaging and diagnostics |
US8593512B2 (en) | 2010-02-05 | 2013-11-26 | Creative Technology Ltd | Device and method for scanning an object on a working surface |
US8326142B2 (en) | 2010-02-12 | 2012-12-04 | Sri International | Optical image systems |
JP5387856B2 (en) | 2010-02-16 | 2014-01-15 | ソニー株式会社 | Image processing apparatus, image processing method, image processing program, and imaging apparatus |
US8648918B2 (en) | 2010-02-18 | 2014-02-11 | Sony Corporation | Method and system for obtaining a point spread function using motion information |
US20130033579A1 (en) | 2010-02-19 | 2013-02-07 | Dual Aperture Inc. | Processing multi-aperture image data |
KR20110097690A (en) | 2010-02-23 | 2011-08-31 | 삼성전자주식회사 | Method and apparatus for providing multiview still image service, Method and apparatus for receiving multiview still image service |
KR101802238B1 (en) | 2010-02-23 | 2017-11-29 | 삼성전자주식회사 | Apparatus and method for generating a three-dimension image data in portable terminal |
KR20130000401A (en) | 2010-02-28 | 2013-01-02 | 오스터하우트 그룹 인코포레이티드 | Local advertising content on an interactive head-mounted eyepiece |
JP5776173B2 (en) | 2010-03-01 | 2015-09-09 | 株式会社リコー | Imaging device and distance measuring device |
US8860833B2 (en) | 2010-03-03 | 2014-10-14 | Adobe Systems Incorporated | Blended rendering of focused plenoptic camera data |
US20110222757A1 (en) | 2010-03-10 | 2011-09-15 | Gbo 3D Technology Pte. Ltd. | Systems and methods for 2D image and spatial data capture for 3D stereo imaging |
US20110221950A1 (en) | 2010-03-12 | 2011-09-15 | Doeke Jolt Oostra | Camera device, wafer scale package |
WO2011116203A1 (en) | 2010-03-17 | 2011-09-22 | Pelican Imaging Corporation | Fabrication process for mastering imaging lens arrays |
EP2391142B1 (en) | 2010-03-19 | 2013-12-25 | FUJIFILM Corporation | Imaging device, method and program, and recording medium using same |
US8736733B2 (en) | 2010-03-19 | 2014-05-27 | Invisage Technologies, Inc. | Dark current reduction in image sensors via dynamic electrical biasing |
WO2011114683A1 (en) | 2010-03-19 | 2011-09-22 | パナソニック株式会社 | Stereovision-image position matching apparatus, stereovision-image position matching method, and program therefor |
US8896668B2 (en) | 2010-04-05 | 2014-11-25 | Qualcomm Incorporated | Combining data from multiple image sensors |
US20110242342A1 (en) | 2010-04-05 | 2011-10-06 | Qualcomm Incorporated | Combining data from multiple image sensors |
US8600186B2 (en) | 2010-04-26 | 2013-12-03 | City University Of Hong Kong | Well focused catadioptric image acquisition |
US20110267264A1 (en) | 2010-04-29 | 2011-11-03 | Mccarthy John | Display system with multiple optical sensors |
US20130250150A1 (en) | 2010-05-03 | 2013-09-26 | Michael R. Malone | Devices and methods for high-resolution image and video capture |
US9256974B1 (en) | 2010-05-04 | 2016-02-09 | Stephen P Hines | 3-D motion-parallax portable display software application |
US8885890B2 (en) | 2010-05-07 | 2014-11-11 | Microsoft Corporation | Depth map confidence filtering |
KR20110124473A (en) | 2010-05-11 | 2011-11-17 | 삼성전자주식회사 | 3D image generating device and method for multi-view image |
US20130147979A1 (en) | 2010-05-12 | 2013-06-13 | Pelican Imaging Corporation | Systems and methods for extending dynamic range of imager arrays by controlling pixel analog gain |
JP5545016B2 (en) | 2010-05-12 | 2014-07-09 | ソニー株式会社 | Imaging device |
US8928793B2 (en) | 2010-05-12 | 2015-01-06 | Pelican Imaging Corporation | Imager array interfaces |
WO2011142774A1 (en) | 2010-05-14 | 2011-11-17 | Omnivision Technologies, Inc. | Alternative color image array and associated methods |
US8576293B2 (en) | 2010-05-18 | 2013-11-05 | Aptina Imaging Corporation | Multi-channel imager |
SG176327A1 (en) | 2010-05-20 | 2011-12-29 | Sony Corp | A system and method of image processing |
US8602887B2 (en) | 2010-06-03 | 2013-12-10 | Microsoft Corporation | Synthesis of information from multiple audiovisual sources |
US20120062697A1 (en) | 2010-06-09 | 2012-03-15 | Chemimage Corporation | Hyperspectral imaging sensor for tracking moving targets |
DE102010024666A1 (en) | 2010-06-18 | 2011-12-22 | Hella Kgaa Hueck & Co. | Method for optical self-diagnosis of a camera system and apparatus for carrying out such a method |
US20110310980A1 (en) | 2010-06-22 | 2011-12-22 | Qualcomm Mems Technologies, Inc. | Apparatus and methods for processing frames of video data across a display interface using a block-based encoding scheme and a tag id |
KR20120000485A (en) | 2010-06-25 | 2012-01-02 | 삼성전자주식회사 | Depth Image Coding Apparatus and Method Using Prediction Mode |
CN101883291B (en) | 2010-06-29 | 2012-12-19 | 上海大学 | Method for drawing viewpoints by reinforcing interested region |
US8493432B2 (en) | 2010-06-29 | 2013-07-23 | Mitsubishi Electric Research Laboratories, Inc. | Digital refocusing for wide-angle images using axial-cone cameras |
EP2403234A1 (en) * | 2010-06-29 | 2012-01-04 | Koninklijke Philips Electronics N.V. | Method and system for constructing a compound image from data obtained by an array of image capturing devices |
CN102959970B (en) | 2010-06-30 | 2015-04-15 | 富士胶片株式会社 | Device, method, and program for determining obstacle within imaging range when capturing images displayed in three-dimensional view |
US9406132B2 (en) | 2010-07-16 | 2016-08-02 | Qualcomm Incorporated | Vision-based quality metric for three dimensional video |
GB2482022A (en) | 2010-07-16 | 2012-01-18 | St Microelectronics Res & Dev | Method for measuring resolution and aberration of lens and sensor |
US8386964B2 (en) | 2010-07-21 | 2013-02-26 | Microsoft Corporation | Interactive image matting |
US20120019700A1 (en) | 2010-07-26 | 2012-01-26 | American Technologies Network Corporation | Optical system with automatic mixing of daylight and thermal vision digital video signals |
US20120026342A1 (en) | 2010-07-27 | 2012-02-02 | Xiaoguang Yu | Electronic system communicating with image sensor |
US20120026451A1 (en) | 2010-07-29 | 2012-02-02 | Lensvector Inc. | Tunable liquid crystal lens with single sided contacts |
CN102375199B (en) | 2010-08-11 | 2015-06-03 | 鸿富锦精密工业(深圳)有限公司 | Camera module |
US8428342B2 (en) | 2010-08-12 | 2013-04-23 | At&T Intellectual Property I, L.P. | Apparatus and method for providing three dimensional media content |
US8836793B1 (en) | 2010-08-13 | 2014-09-16 | Opto-Knowledge Systems, Inc. | True color night vision (TCNV) fusion |
US8493482B2 (en) | 2010-08-18 | 2013-07-23 | Apple Inc. | Dual image sensor image processing system and method |
US8749694B2 (en) | 2010-08-27 | 2014-06-10 | Adobe Systems Incorporated | Methods and apparatus for rendering focused plenoptic camera data using super-resolved demosaicing |
US8665341B2 (en) | 2010-08-27 | 2014-03-04 | Adobe Systems Incorporated | Methods and apparatus for rendering output images with simulated artistic effects from focused plenoptic camera data |
US8724000B2 (en) | 2010-08-27 | 2014-05-13 | Adobe Systems Incorporated | Methods and apparatus for super-resolution in integral photography |
GB2483434A (en) | 2010-08-31 | 2012-03-14 | Sony Corp | Detecting stereoscopic disparity by comparison with subset of pixel change points |
US20120056982A1 (en) | 2010-09-08 | 2012-03-08 | Microsoft Corporation | Depth camera based on structured light and stereo vision |
US9013550B2 (en) | 2010-09-09 | 2015-04-21 | Qualcomm Incorporated | Online reference generation and tracking for multi-user augmented reality |
JP2013538534A (en) | 2010-09-14 | 2013-10-10 | トムソン ライセンシング | Compression method and apparatus for occlusion data |
WO2012037075A1 (en) | 2010-09-14 | 2012-03-22 | Thomson Licensing | Method of presenting three-dimensional content with disparity adjustments |
US9013634B2 (en) | 2010-09-14 | 2015-04-21 | Adobe Systems Incorporated | Methods and apparatus for video completion |
US8780251B2 (en) | 2010-09-20 | 2014-07-15 | Canon Kabushiki Kaisha | Image capture with focus adjustment |
WO2012039043A1 (en) | 2010-09-22 | 2012-03-29 | 富士通株式会社 | Stereo image generating unit, method of generating stereo image, and stereo image generating computer program |
US20120086803A1 (en) | 2010-10-11 | 2012-04-12 | Malzbender Thomas G | Method and system for distance estimation using projected symbol sequences |
US20140192238A1 (en) | 2010-10-24 | 2014-07-10 | Linx Computational Imaging Ltd. | System and Method for Imaging and Image Processing |
JP5657343B2 (en) | 2010-10-28 | 2015-01-21 | 株式会社ザクティ | Electronics |
WO2012056437A1 (en) | 2010-10-29 | 2012-05-03 | École Polytechnique Fédérale De Lausanne (Epfl) | Omnidirectional sensor array system |
US9137503B2 (en) | 2010-11-03 | 2015-09-15 | Sony Corporation | Lens and color filter arrangement, super-resolution camera system and method |
US9065991B2 (en) | 2010-11-04 | 2015-06-23 | Lensvector Inc. | Methods of adjustment free manufacture of focus free camera modules |
US20120113232A1 (en) | 2010-11-10 | 2012-05-10 | Sony Pictures Technologies Inc. | Multiple camera system and method for selectable interaxial separation |
MY150361A (en) | 2010-12-03 | 2013-12-31 | Mimos Berhad | Method of image segmentation using intensity and depth information |
WO2012078126A1 (en) | 2010-12-08 | 2012-06-14 | Thomson Licensing | System and method for trinocular depth acquisition with triangular sensor |
US8878950B2 (en) | 2010-12-14 | 2014-11-04 | Pelican Imaging Corporation | Systems and methods for synthesizing high resolution images using super-resolution processes |
JP5963422B2 (en) | 2010-12-17 | 2016-08-03 | キヤノン株式会社 | Imaging apparatus, display apparatus, computer program, and stereoscopic image display system |
US9177381B2 (en) | 2010-12-22 | 2015-11-03 | Nani Holdings IP, LLC | Depth estimate determination, systems and methods |
US8682107B2 (en) | 2010-12-22 | 2014-03-25 | Electronics And Telecommunications Research Institute | Apparatus and method for creating 3D content for oriental painting |
US8565709B2 (en) | 2010-12-30 | 2013-10-22 | Apple Inc. | Digital signal filter |
WO2012100230A1 (en) | 2011-01-20 | 2012-07-26 | Fivefocal, Llc | Passively athermalized infrared imaging system and methods of manufacturing same |
US8717467B2 (en) | 2011-01-25 | 2014-05-06 | Aptina Imaging Corporation | Imaging systems with array cameras for depth sensing |
US8581995B2 (en) | 2011-01-25 | 2013-11-12 | Aptina Imaging Corporation | Method and apparatus for parallax correction in fused array imaging systems |
JP5594477B2 (en) | 2011-01-26 | 2014-09-24 | Nltテクノロジー株式会社 | Image display device, image display method, and program |
CN103415860B (en) | 2011-01-27 | 2019-07-12 | 苹果公司 | The method for determining the method for the corresponding relationship between the first and second images and determining video camera posture |
US20120200726A1 (en) | 2011-02-09 | 2012-08-09 | Research In Motion Limited | Method of Controlling the Depth of Field for a Small Sensor Camera Using an Extension for EDOF |
CA2767023C (en) | 2011-02-09 | 2014-09-09 | Research In Motion Limited | Increased low light sensitivity for image sensors by combining quantum dot sensitivity to visible and infrared light |
US20140176592A1 (en) | 2011-02-15 | 2014-06-26 | Lytro, Inc. | Configuring two-dimensional image processing based on light-field parameters |
WO2012117583A1 (en) | 2011-02-28 | 2012-09-07 | 富士フイルム株式会社 | Color imaging device |
US8406548B2 (en) | 2011-02-28 | 2013-03-26 | Sony Corporation | Method and apparatus for performing a blur rendering process on an image |
US8537245B2 (en) | 2011-03-04 | 2013-09-17 | Hand Held Products, Inc. | Imaging and decoding device with quantum dot imager |
CA2769358C (en) | 2011-03-08 | 2016-06-07 | Research In Motion Limited | Quantum dot image sensor with dummy pixels used for intensity calculations |
US9565449B2 (en) | 2011-03-10 | 2017-02-07 | Qualcomm Incorporated | Coding multiview video plus depth content |
US20120249853A1 (en) | 2011-03-28 | 2012-10-04 | Marc Krolczyk | Digital camera for reviewing related images |
US8824821B2 (en) | 2011-03-28 | 2014-09-02 | Sony Corporation | Method and apparatus for performing user inspired visual effects rendering on an image |
US8422770B2 (en) | 2011-03-30 | 2013-04-16 | Mckesson Financial Holdings | Method, apparatus and computer program product for displaying normalized medical images |
US9030528B2 (en) | 2011-04-04 | 2015-05-12 | Apple Inc. | Multi-zone imaging sensor and lens array |
FR2974449A1 (en) | 2011-04-22 | 2012-10-26 | Commissariat Energie Atomique | IMAGEUR INTEGRATED CIRCUIT AND STEREOSCOPIC IMAGE CAPTURE DEVICE |
US20120274626A1 (en) | 2011-04-29 | 2012-11-01 | Himax Media Solutions, Inc. | Stereoscopic Image Generating Apparatus and Method |
WO2012149971A1 (en) | 2011-05-04 | 2012-11-08 | Sony Ericsson Mobile Communications Ab | Method, graphical user interface, and computer program product for processing of a light field image |
KR101973822B1 (en) | 2011-05-11 | 2019-04-29 | 포토네이션 케이맨 리미티드 | Systems and methods for transmitting and receiving array camera image data |
US8843346B2 (en) | 2011-05-13 | 2014-09-23 | Amazon Technologies, Inc. | Using spatial information with device interaction |
US8629901B2 (en) | 2011-05-19 | 2014-01-14 | National Taiwan University | System and method of revising depth of a 3D image pair |
US20120293489A1 (en) | 2011-05-20 | 2012-11-22 | Himax Technologies Limited | Nonlinear depth remapping system and method thereof |
JP5797016B2 (en) | 2011-05-30 | 2015-10-21 | キヤノン株式会社 | Image processing apparatus, image processing method, and program |
US9635274B2 (en) | 2011-06-15 | 2017-04-25 | Microsoft Technology Licensing, Llc | High resolution multispectral image capture |
JP2013005259A (en) | 2011-06-17 | 2013-01-07 | Sony Corp | Image processing apparatus, image processing method, and program |
US20130265459A1 (en) | 2011-06-28 | 2013-10-10 | Pelican Imaging Corporation | Optical arrangements for use with an array camera |
KR20140045458A (en) | 2011-06-28 | 2014-04-16 | 펠리칸 이매징 코포레이션 | Optical arrangements for use with an array camera |
US8773513B2 (en) | 2011-07-01 | 2014-07-08 | Seiko Epson Corporation | Context and epsilon stereo constrained correspondence matching |
US9300946B2 (en) * | 2011-07-08 | 2016-03-29 | Personify, Inc. | System and method for generating a depth map and fusing images from a camera array |
JP2013024886A (en) | 2011-07-14 | 2013-02-04 | Sanyo Electric Co Ltd | Imaging apparatus |
JP5780865B2 (en) | 2011-07-14 | 2015-09-16 | キヤノン株式会社 | Image processing apparatus, imaging system, and image processing system |
US9363535B2 (en) | 2011-07-22 | 2016-06-07 | Qualcomm Incorporated | Coding motion depth maps with depth range variation |
US9264689B2 (en) | 2011-08-04 | 2016-02-16 | Semiconductor Components Industries, Llc | Systems and methods for color compensation in multi-view video |
US8432435B2 (en) * | 2011-08-10 | 2013-04-30 | Seiko Epson Corporation | Ray image modeling for fast catadioptric light field rendering |
US8866951B2 (en) | 2011-08-24 | 2014-10-21 | Aptina Imaging Corporation | Super-resolution imaging systems |
US9009952B2 (en) | 2011-08-29 | 2015-04-21 | Asm Technology Singapore Pte. Ltd. | Apparatus for assembling a lens module and an image sensor to form a camera module, and a method of assembling the same |
US8704895B2 (en) | 2011-08-29 | 2014-04-22 | Qualcomm Incorporated | Fast calibration of displays using spectral-based colorimetrically calibrated multicolor camera |
US20130070060A1 (en) | 2011-09-19 | 2013-03-21 | Pelican Imaging Corporation | Systems and methods for determining depth from multiple views of a scene that include aliasing using hypothesized fusion |
US9100639B2 (en) | 2011-09-20 | 2015-08-04 | Panasonic Intellectual Property Management Co., Ltd. | Light field imaging device and image processing device |
EP2760209B1 (en) | 2011-09-21 | 2017-07-12 | FUJIFILM Corporation | Image processing device, method, program and recording medium, stereoscopic image capture device, portable electronic apparatus, printer, and stereoscopic image player device |
CN104081414B (en) | 2011-09-28 | 2017-08-01 | Fotonation开曼有限公司 | System and method for coding and decoding light field image file |
US8908083B2 (en) | 2011-09-28 | 2014-12-09 | Apple Inc. | Dynamic autofocus operations |
JP5831105B2 (en) | 2011-09-30 | 2015-12-09 | ソニー株式会社 | Imaging apparatus and imaging method |
WO2013055960A1 (en) | 2011-10-11 | 2013-04-18 | Pelican Imaging Corporation | Lens stack arrays including adaptive optical elements |
EP2592823A3 (en) | 2011-10-12 | 2013-06-19 | Canon Kabushiki Kaisha | Image-capturing device |
US20130107072A1 (en) | 2011-10-31 | 2013-05-02 | Ankit Kumar | Multi-resolution ip camera |
JP5149435B1 (en) | 2011-11-04 | 2013-02-20 | 株式会社東芝 | Video processing apparatus and video processing method |
US9692991B2 (en) | 2011-11-04 | 2017-06-27 | Qualcomm Incorporated | Multispectral imaging system |
EP2590138B1 (en) | 2011-11-07 | 2019-09-11 | Flir Systems AB | Gas visualization arrangements, devices, and methods |
EP2780755A4 (en) | 2011-11-15 | 2015-09-02 | Technion Res & Dev Foundation | Method and system for transmitting light |
US20130121559A1 (en) | 2011-11-16 | 2013-05-16 | Sharp Laboratories Of America, Inc. | Mobile device with three dimensional augmented reality |
JP6019568B2 (en) | 2011-11-28 | 2016-11-02 | ソニー株式会社 | Image processing apparatus and method, recording medium, and program |
US9661310B2 (en) | 2011-11-28 | 2017-05-23 | ArcSoft Hanzhou Co., Ltd. | Image depth recovering method and stereo image fetching device thereof |
EP2600316A1 (en) | 2011-11-29 | 2013-06-05 | Inria Institut National de Recherche en Informatique et en Automatique | Method, system and software program for shooting and editing a film comprising at least one image of a 3D computer-generated animation |
KR101862404B1 (en) | 2011-12-09 | 2018-05-29 | 엘지이노텍 주식회사 | Apparatus and method for eliminating noise of stereo image |
US9117295B2 (en) | 2011-12-20 | 2015-08-25 | Adobe Systems Incorporated | Refinement of depth maps by fusion of multiple estimates |
US8941722B2 (en) | 2012-01-03 | 2015-01-27 | Sony Corporation | Automatic intelligent focus control of video |
WO2013119706A1 (en) | 2012-02-06 | 2013-08-15 | Pelican Imaging Corporation | Systems and methods for extending dynamic range of imager arrays by controlling pixel analog gain |
US9172889B2 (en) | 2012-02-09 | 2015-10-27 | Semiconductor Components Industries, Llc | Imaging systems and methods for generating auto-exposed high-dynamic-range images |
US9412206B2 (en) | 2012-02-21 | 2016-08-09 | Pelican Imaging Corporation | Systems and methods for the manipulation of captured light field image data |
JP5860304B2 (en) | 2012-02-23 | 2016-02-16 | キヤノン株式会社 | Imaging apparatus, control method therefor, program, and storage medium |
JP5924978B2 (en) | 2012-02-28 | 2016-05-25 | キヤノン株式会社 | Image processing apparatus and image processing method |
JP6112824B2 (en) | 2012-02-28 | 2017-04-12 | キヤノン株式会社 | Image processing method and apparatus, and program. |
EP2637139A1 (en) | 2012-03-05 | 2013-09-11 | Thomson Licensing | Method and apparatus for bi-layer segmentation |
US9338439B2 (en) | 2012-04-02 | 2016-05-10 | Intel Corporation | Systems, methods, and computer program products for runtime adjustment of image warping parameters in a multi-camera system |
US9156168B2 (en) | 2012-04-13 | 2015-10-13 | Automation Engineering, Inc. | Active alignment using continuous motion sweeps and temporal interpolation |
US8994845B2 (en) | 2012-04-27 | 2015-03-31 | Blackberry Limited | System and method of adjusting a camera based on image data |
CN104335246B (en) | 2012-05-01 | 2018-09-04 | Fotonation开曼有限公司 | The camera model of pattern is formed with pi optical filters group |
US9210392B2 (en) | 2012-05-01 | 2015-12-08 | Pelican Imaging Coporation | Camera modules patterned with pi filter groups |
CN104272729A (en) | 2012-05-02 | 2015-01-07 | 皇家飞利浦有限公司 | Quality metric for processing 3d video |
EP2820838B1 (en) | 2012-05-09 | 2020-01-08 | Lytro, Inc. | Optimization of optical systems for improved light field capture and manipulation |
US9179126B2 (en) | 2012-06-01 | 2015-11-03 | Ostendo Technologies, Inc. | Spatio-temporal light field cameras |
WO2013182873A1 (en) | 2012-06-08 | 2013-12-12 | Nokia Corporation | A multi-frame image calibrator |
EP2677734A3 (en) | 2012-06-18 | 2016-01-13 | Sony Mobile Communications AB | Array camera imaging system and method |
US9100635B2 (en) | 2012-06-28 | 2015-08-04 | Pelican Imaging Corporation | Systems and methods for detecting defective camera arrays and optic arrays |
JP5929553B2 (en) | 2012-06-28 | 2016-06-08 | ソニー株式会社 | Image processing apparatus, imaging apparatus, image processing method, and program |
US20140002674A1 (en) | 2012-06-30 | 2014-01-02 | Pelican Imaging Corporation | Systems and Methods for Manufacturing Camera Modules Using Active Alignment of Lens Stack Arrays and Sensors |
US8896594B2 (en) | 2012-06-30 | 2014-11-25 | Microsoft Corporation | Depth sensing with depth-adaptive illumination |
US9147251B2 (en) | 2012-08-03 | 2015-09-29 | Flyby Media, Inc. | Systems and methods for efficient 3D tracking of weakly textured planar surfaces for augmented reality applications |
US8988566B2 (en) | 2012-08-09 | 2015-03-24 | Omnivision Technologies, Inc. | Lens array for partitioned image sensor having color filters |
CN104662589B (en) | 2012-08-21 | 2017-08-04 | 派力肯影像公司 | Systems and methods for parallax detection and correction in images captured using an array camera |
US20140055632A1 (en) | 2012-08-23 | 2014-02-27 | Pelican Imaging Corporation | Feature based high resolution motion estimation from low resolution images captured using an array source |
CN104584545B (en) * | 2012-08-31 | 2017-05-31 | 索尼公司 | Image processing apparatus, image processing method and information processor |
US9214013B2 (en) | 2012-09-14 | 2015-12-15 | Pelican Imaging Corporation | Systems and methods for correcting user identified artifacts in light field images |
US9373088B2 (en) | 2012-09-17 | 2016-06-21 | The Board Of Trustees Of The Leland Stanford Junior University | Brain machine interface utilizing a discrete action state decoder in parallel with a continuous decoder for a neural prosthetic device |
US9143673B2 (en) | 2012-09-19 | 2015-09-22 | Google Inc. | Imaging device with a plurality of pixel arrays |
EP4307659A1 (en) | 2012-09-28 | 2024-01-17 | Adeia Imaging LLC | Generating images from light fields utilizing virtual viewpoints |
TW201415879A (en) | 2012-10-12 | 2014-04-16 | Wintek Corp | Image capture device |
US9609190B2 (en) | 2012-10-31 | 2017-03-28 | Invisage Technologies, Inc. | Devices, methods, and systems for expanded-field-of-view image and video capture |
US9143711B2 (en) | 2012-11-13 | 2015-09-22 | Pelican Imaging Corporation | Systems and methods for array camera focal plane control |
US9538152B2 (en) | 2012-11-28 | 2017-01-03 | Corephotonics Ltd. | High resolution thin multi-aperture imaging systems |
US9001226B1 (en) | 2012-12-04 | 2015-04-07 | Lytro, Inc. | Capturing and relighting images using multiple devices |
US9088369B2 (en) | 2012-12-28 | 2015-07-21 | Synergy Microwave Corporation | Self injection locked phase locked looped optoelectronic oscillator |
US20140183334A1 (en) | 2013-01-03 | 2014-07-03 | Visera Technologies Company Limited | Image sensor for light field device and manufacturing method thereof |
US9671595B2 (en) | 2013-01-05 | 2017-06-06 | Light Labs Inc. | Methods and apparatus for using multiple optical chains in paralell |
KR20140094395A (en) | 2013-01-22 | 2014-07-30 | 삼성전자주식회사 | photographing device for taking a picture by a plurality of microlenses and method thereof |
US9769365B1 (en) | 2013-02-15 | 2017-09-19 | Red.Com, Inc. | Dense field imaging |
US9462164B2 (en) | 2013-02-21 | 2016-10-04 | Pelican Imaging Corporation | Systems and methods for generating compressed light field representation data using captured light fields, array geometry, and parallax information |
US9374512B2 (en) | 2013-02-24 | 2016-06-21 | Pelican Imaging Corporation | Thin form factor computational array cameras and modular array cameras |
US20150002734A1 (en) | 2013-07-01 | 2015-01-01 | Motorola Mobility Llc | Electronic Device with Modulated Light Flash Operation for Rolling Shutter Image Sensor |
US9917998B2 (en) | 2013-03-08 | 2018-03-13 | Fotonation Cayman Limited | Systems and methods for measuring scene information while capturing images using array cameras |
US8866912B2 (en) | 2013-03-10 | 2014-10-21 | Pelican Imaging Corporation | System and methods for calibration of an array camera using a single captured image |
US9521416B1 (en) | 2013-03-11 | 2016-12-13 | Kip Peli P1 Lp | Systems and methods for image data compression |
US9519972B2 (en) | 2013-03-13 | 2016-12-13 | Kip Peli P1 Lp | Systems and methods for synthesizing images from image data captured by an array camera using restricted depth of field depth maps in which depth estimation precision varies |
US9106784B2 (en) | 2013-03-13 | 2015-08-11 | Pelican Imaging Corporation | Systems and methods for controlling aliasing in images captured by an array camera for use in super-resolution processing |
WO2014160142A1 (en) | 2013-03-13 | 2014-10-02 | Pelican Imaging Corporation | Systems and methods for using alignment to increase sampling diversity of cameras in an array camera module |
US9124831B2 (en) | 2013-03-13 | 2015-09-01 | Pelican Imaging Corporation | System and methods for calibration of an array camera |
US9888194B2 (en) | 2013-03-13 | 2018-02-06 | Fotonation Cayman Limited | Array camera architecture implementing quantum film image sensors |
WO2014153098A1 (en) | 2013-03-14 | 2014-09-25 | Pelican Imaging Corporation | Photmetric normalization in array cameras |
US9578259B2 (en) | 2013-03-14 | 2017-02-21 | Fotonation Cayman Limited | Systems and methods for reducing motion blur in images or video in ultra low light with array cameras |
US10122993B2 (en) | 2013-03-15 | 2018-11-06 | Fotonation Limited | Autofocus system for a conventional camera that uses depth information from an array camera |
WO2014150856A1 (en) | 2013-03-15 | 2014-09-25 | Pelican Imaging Corporation | Array camera implementing quantum dot color filters |
US20140267286A1 (en) | 2013-03-15 | 2014-09-18 | Pelican Imaging Corporation | Systems and Methods for Providing an Array Projector |
US9445003B1 (en) | 2013-03-15 | 2016-09-13 | Pelican Imaging Corporation | Systems and methods for synthesizing high resolution images using image deconvolution based on motion and depth information |
US9633442B2 (en) | 2013-03-15 | 2017-04-25 | Fotonation Cayman Limited | Array cameras including an array camera module augmented with a separate camera |
JP2016524125A (en) | 2013-03-15 | 2016-08-12 | ペリカン イメージング コーポレイション | System and method for stereoscopic imaging using a camera array |
WO2014144157A1 (en) | 2013-03-15 | 2014-09-18 | Pelican Imaging Corporation | Optical arrangements for use with an array camera |
US9497429B2 (en) | 2013-03-15 | 2016-11-15 | Pelican Imaging Corporation | Extended color processing on pelican array cameras |
WO2015048694A2 (en) | 2013-09-27 | 2015-04-02 | Pelican Imaging Corporation | Systems and methods for depth-assisted perspective distortion correction |
US20150098079A1 (en) | 2013-10-09 | 2015-04-09 | Hilti Aktiengesellschaft | System and method for camera based position and orientation measurement |
US20150104101A1 (en) | 2013-10-14 | 2015-04-16 | Apple Inc. | Method and ui for z depth image segmentation |
US9426343B2 (en) | 2013-11-07 | 2016-08-23 | Pelican Imaging Corporation | Array cameras incorporating independently aligned lens stacks |
US10119808B2 (en) | 2013-11-18 | 2018-11-06 | Fotonation Limited | Systems and methods for estimating depth from projected texture using camera arrays |
EP3075140B1 (en) | 2013-11-26 | 2018-06-13 | FotoNation Cayman Limited | Array camera configurations incorporating multiple constituent array cameras |
KR102317607B1 (en) | 2013-11-27 | 2021-10-25 | 아이에프피 에너지스 누벨 | Process for the production of carbon black from at least one fcc slurry cut, comprising a specific hydrotreatment |
US9979878B2 (en) | 2014-02-21 | 2018-05-22 | Light Labs Inc. | Intuitive camera user interface methods and apparatus |
JP6211435B2 (en) | 2014-02-26 | 2017-10-11 | 株式会社アドバンテスト | Manufacturing method of semiconductor device |
WO2015134996A1 (en) | 2014-03-07 | 2015-09-11 | Pelican Imaging Corporation | System and methods for depth regularization and semiautomatic interactive matting using rgb-d images |
US9521319B2 (en) | 2014-06-18 | 2016-12-13 | Pelican Imaging Corporation | Array cameras and array camera modules including spectral filters disposed outside of a constituent image sensor |
US9992483B2 (en) | 2014-09-03 | 2018-06-05 | Intel Corporation | Imaging architecture for depth camera mode with mode switching |
US10250871B2 (en) | 2014-09-29 | 2019-04-02 | Fotonation Limited | Systems and methods for dynamic calibration of array cameras |
-
2013
- 2013-09-30 EP EP23177057.9A patent/EP4307659A1/en active Pending
- 2013-09-30 WO PCT/US2013/062720 patent/WO2014052974A2/en active Application Filing
- 2013-09-30 US US14/042,275 patent/US20140092281A1/en not_active Abandoned
- 2013-09-30 CN CN201380050549.5A patent/CN104685860A/en active Pending
- 2013-09-30 EP EP13841613.6A patent/EP2901671A4/en not_active Ceased
-
2015
- 2015-10-06 US US14/876,024 patent/US10390005B2/en active Active
Patent Citations (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6097394A (en) * | 1997-04-28 | 2000-08-01 | Board Of Trustees, Leland Stanford, Jr. University | Method and system for light field rendering |
US20020006113A1 (en) * | 2000-07-06 | 2002-01-17 | Young Choi | Switching circuit for board duplication apparatus |
US7085409B2 (en) * | 2000-10-18 | 2006-08-01 | Sarnoff Corporation | Method and apparatus for synthesizing new video and/or still imagery from a collection of real video and/or still imagery |
US20020061131A1 (en) * | 2000-10-18 | 2002-05-23 | Sawhney Harpreet Singh | Method and apparatus for synthesizing new video and/or still imagery from a collection of real video and/or still imagery |
US20040019637A1 (en) * | 2002-07-26 | 2004-01-29 | International Business Machines Corporaion | Interactive one to many communication in a cooperating community of users |
US20040196379A1 (en) * | 2003-04-04 | 2004-10-07 | Stmicroelectronics, Inc. | Compound camera and methods for implementing auto-focus, depth-of-field and high-resolution functions |
US7425984B2 (en) * | 2003-04-04 | 2008-09-16 | Stmicroelectronics, Inc. | Compound camera and methods for implementing auto-focus, depth-of-field and high-resolution functions |
US20050219264A1 (en) * | 2004-04-01 | 2005-10-06 | Microsoft Corporation | Pop-up light field |
US20120062756A1 (en) * | 2004-12-17 | 2012-03-15 | Dong Tian | Method and System for Processing Multiview Videos for View Synthesis Using Skip and Direct Modes |
US20110255592A1 (en) * | 2008-10-27 | 2011-10-20 | Lg Electronics Inc. | Virtual view image synthesis method and apparatus |
US20100141802A1 (en) * | 2008-12-08 | 2010-06-10 | Timothy Knight | Light Field Data Acquisition Devices, and Methods of Using and Manufacturing Same |
US20110267348A1 (en) * | 2010-04-29 | 2011-11-03 | Dennis Lin | Systems and methods for generating a virtual camera viewpoint for an image |
US20120057040A1 (en) * | 2010-05-11 | 2012-03-08 | Byung Kwan Park | Apparatus and method for processing light field data using a mask with an attenuation pattern |
US8842201B2 (en) * | 2011-01-06 | 2014-09-23 | Sony Corporation | Image capturing device including lens array and processing |
Cited By (112)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11792538B2 (en) | 2008-05-20 | 2023-10-17 | Adeia Imaging Llc | Capturing and processing of images including occlusions focused on an image sensor by a lens stack array |
US9712759B2 (en) | 2008-05-20 | 2017-07-18 | Fotonation Cayman Limited | Systems and methods for generating depth maps using a camera arrays incorporating monochrome and color cameras |
US11412158B2 (en) | 2008-05-20 | 2022-08-09 | Fotonation Limited | Capturing and processing of images including occlusions focused on an image sensor by a lens stack array |
US10027901B2 (en) | 2008-05-20 | 2018-07-17 | Fotonation Cayman Limited | Systems and methods for generating depth maps using a camera arrays incorporating monochrome and color cameras |
US9749547B2 (en) | 2008-05-20 | 2017-08-29 | Fotonation Cayman Limited | Capturing and processing of images using camera array incorperating Bayer cameras having different fields of view |
US12022207B2 (en) | 2008-05-20 | 2024-06-25 | Adeia Imaging Llc | Capturing and processing of images including occlusions focused on an image sensor by a lens stack array |
US12041360B2 (en) | 2008-05-20 | 2024-07-16 | Adeia Imaging Llc | Capturing and processing of images including occlusions focused on an image sensor by a lens stack array |
US10142560B2 (en) | 2008-05-20 | 2018-11-27 | Fotonation Limited | Capturing and processing of images including occlusions focused on an image sensor by a lens stack array |
US10306120B2 (en) | 2009-11-20 | 2019-05-28 | Fotonation Limited | Capturing and processing of images captured by camera arrays incorporating cameras with telephoto and conventional lenses to generate depth maps |
US10455168B2 (en) | 2010-05-12 | 2019-10-22 | Fotonation Limited | Imager array interfaces |
US11423513B2 (en) | 2010-12-14 | 2022-08-23 | Fotonation Limited | Systems and methods for synthesizing high resolution images using images captured by an array of independently controllable imagers |
US12243190B2 (en) | 2010-12-14 | 2025-03-04 | Adeia Imaging Llc | Systems and methods for synthesizing high resolution images using images captured by an array of independently controllable imagers |
US11875475B2 (en) | 2010-12-14 | 2024-01-16 | Adeia Imaging Llc | Systems and methods for synthesizing high resolution images using images captured by an array of independently controllable imagers |
US10366472B2 (en) | 2010-12-14 | 2019-07-30 | Fotonation Limited | Systems and methods for synthesizing high resolution images using images captured by an array of independently controllable imagers |
US10742861B2 (en) | 2011-05-11 | 2020-08-11 | Fotonation Limited | Systems and methods for transmitting and receiving array camera image data |
US10218889B2 (en) | 2011-05-11 | 2019-02-26 | Fotonation Limited | Systems and methods for transmitting and receiving array camera image data |
US9794476B2 (en) | 2011-09-19 | 2017-10-17 | Fotonation Cayman Limited | Systems and methods for controlling aliasing in images captured by an array camera for use in super resolution processing using pixel apertures |
US10375302B2 (en) | 2011-09-19 | 2019-08-06 | Fotonation Limited | Systems and methods for controlling aliasing in images captured by an array camera for use in super resolution processing using pixel apertures |
US9811753B2 (en) | 2011-09-28 | 2017-11-07 | Fotonation Cayman Limited | Systems and methods for encoding light field image files |
US10430682B2 (en) | 2011-09-28 | 2019-10-01 | Fotonation Limited | Systems and methods for decoding image files containing depth maps stored as metadata |
US11729365B2 (en) | 2011-09-28 | 2023-08-15 | Adela Imaging LLC | Systems and methods for encoding image files containing depth maps stored as metadata |
US10019816B2 (en) | 2011-09-28 | 2018-07-10 | Fotonation Cayman Limited | Systems and methods for decoding image files containing depth maps stored as metadata |
US20180197035A1 (en) | 2011-09-28 | 2018-07-12 | Fotonation Cayman Limited | Systems and Methods for Encoding Image Files Containing Depth Maps Stored as Metadata |
US10275676B2 (en) | 2011-09-28 | 2019-04-30 | Fotonation Limited | Systems and methods for encoding image files containing depth maps stored as metadata |
US10984276B2 (en) | 2011-09-28 | 2021-04-20 | Fotonation Limited | Systems and methods for encoding image files containing depth maps stored as metadata |
US12052409B2 (en) | 2011-09-28 | 2024-07-30 | Adela Imaging LLC | Systems and methods for encoding image files containing depth maps stored as metadata |
US10311649B2 (en) | 2012-02-21 | 2019-06-04 | Fotonation Limited | Systems and method for performing depth based image editing |
US9754422B2 (en) | 2012-02-21 | 2017-09-05 | Fotonation Cayman Limited | Systems and method for performing depth based image editing |
US9706132B2 (en) | 2012-05-01 | 2017-07-11 | Fotonation Cayman Limited | Camera modules patterned with pi filter groups |
US9807382B2 (en) | 2012-06-28 | 2017-10-31 | Fotonation Cayman Limited | Systems and methods for detecting defective camera arrays and optic arrays |
US10334241B2 (en) | 2012-06-28 | 2019-06-25 | Fotonation Limited | Systems and methods for detecting defective camera arrays and optic arrays |
US11022725B2 (en) | 2012-06-30 | 2021-06-01 | Fotonation Limited | Systems and methods for manufacturing camera modules using active alignment of lens stack arrays and sensors |
US10261219B2 (en) | 2012-06-30 | 2019-04-16 | Fotonation Limited | Systems and methods for manufacturing camera modules using active alignment of lens stack arrays and sensors |
US12002233B2 (en) | 2012-08-21 | 2024-06-04 | Adeia Imaging Llc | Systems and methods for estimating depth and visibility from a reference viewpoint for pixels in a set of images captured from different viewpoints |
US9858673B2 (en) | 2012-08-21 | 2018-01-02 | Fotonation Cayman Limited | Systems and methods for estimating depth and visibility from a reference viewpoint for pixels in a set of images captured from different viewpoints |
US10380752B2 (en) | 2012-08-21 | 2019-08-13 | Fotonation Limited | Systems and methods for estimating depth and visibility from a reference viewpoint for pixels in a set of images captured from different viewpoints |
US9813616B2 (en) | 2012-08-23 | 2017-11-07 | Fotonation Cayman Limited | Feature based high resolution motion estimation from low resolution images captured using an array source |
US10462362B2 (en) | 2012-08-23 | 2019-10-29 | Fotonation Limited | Feature based high resolution motion estimation from low resolution images captured using an array source |
US9749568B2 (en) | 2012-11-13 | 2017-08-29 | Fotonation Cayman Limited | Systems and methods for array camera focal plane control |
US10009538B2 (en) | 2013-02-21 | 2018-06-26 | Fotonation Cayman Limited | Systems and methods for generating compressed light field representation data using captured light fields, array geometry, and parallax information |
US9743051B2 (en) | 2013-02-24 | 2017-08-22 | Fotonation Cayman Limited | Thin form factor computational array cameras and modular array cameras |
US9774831B2 (en) | 2013-02-24 | 2017-09-26 | Fotonation Cayman Limited | Thin form factor computational array cameras and modular array cameras |
US9774789B2 (en) | 2013-03-08 | 2017-09-26 | Fotonation Cayman Limited | Systems and methods for high dynamic range imaging using array cameras |
US9917998B2 (en) | 2013-03-08 | 2018-03-13 | Fotonation Cayman Limited | Systems and methods for measuring scene information while capturing images using array cameras |
US9986224B2 (en) | 2013-03-10 | 2018-05-29 | Fotonation Cayman Limited | System and methods for calibration of an array camera |
US11570423B2 (en) | 2013-03-10 | 2023-01-31 | Adeia Imaging Llc | System and methods for calibration of an array camera |
US11272161B2 (en) | 2013-03-10 | 2022-03-08 | Fotonation Limited | System and methods for calibration of an array camera |
US10225543B2 (en) | 2013-03-10 | 2019-03-05 | Fotonation Limited | System and methods for calibration of an array camera |
US11985293B2 (en) | 2013-03-10 | 2024-05-14 | Adeia Imaging Llc | System and methods for calibration of an array camera |
US10958892B2 (en) | 2013-03-10 | 2021-03-23 | Fotonation Limited | System and methods for calibration of an array camera |
US10127682B2 (en) | 2013-03-13 | 2018-11-13 | Fotonation Limited | System and methods for calibration of an array camera |
US9800856B2 (en) | 2013-03-13 | 2017-10-24 | Fotonation Cayman Limited | Systems and methods for synthesizing images from image data captured by an array camera using restricted depth of field depth maps in which depth estimation precision varies |
US9888194B2 (en) | 2013-03-13 | 2018-02-06 | Fotonation Cayman Limited | Array camera architecture implementing quantum film image sensors |
US10412314B2 (en) | 2013-03-14 | 2019-09-10 | Fotonation Limited | Systems and methods for photometric normalization in array cameras |
US10547772B2 (en) | 2013-03-14 | 2020-01-28 | Fotonation Limited | Systems and methods for reducing motion blur in images or video in ultra low light with array cameras |
US10091405B2 (en) | 2013-03-14 | 2018-10-02 | Fotonation Cayman Limited | Systems and methods for reducing motion blur in images or video in ultra low light with array cameras |
US10182216B2 (en) | 2013-03-15 | 2019-01-15 | Fotonation Limited | Extended color processing on pelican array cameras |
US10542208B2 (en) | 2013-03-15 | 2020-01-21 | Fotonation Limited | Systems and methods for synthesizing high resolution images using image deconvolution based on motion and depth information |
US10638099B2 (en) | 2013-03-15 | 2020-04-28 | Fotonation Limited | Extended color processing on pelican array cameras |
US10674138B2 (en) | 2013-03-15 | 2020-06-02 | Fotonation Limited | Autofocus system for a conventional camera that uses depth information from an array camera |
US10122993B2 (en) | 2013-03-15 | 2018-11-06 | Fotonation Limited | Autofocus system for a conventional camera that uses depth information from an array camera |
US10455218B2 (en) | 2013-03-15 | 2019-10-22 | Fotonation Limited | Systems and methods for estimating depth using stereo array cameras |
US9955070B2 (en) | 2013-03-15 | 2018-04-24 | Fotonation Cayman Limited | Systems and methods for synthesizing high resolution images using image deconvolution based on motion and depth information |
US9898856B2 (en) | 2013-09-27 | 2018-02-20 | Fotonation Cayman Limited | Systems and methods for depth-assisted perspective distortion correction |
US10540806B2 (en) | 2013-09-27 | 2020-01-21 | Fotonation Limited | Systems and methods for depth-assisted perspective distortion correction |
US9924092B2 (en) | 2013-11-07 | 2018-03-20 | Fotonation Cayman Limited | Array cameras incorporating independently aligned lens stacks |
US11486698B2 (en) | 2013-11-18 | 2022-11-01 | Fotonation Limited | Systems and methods for estimating depth from projected texture using camera arrays |
US10119808B2 (en) | 2013-11-18 | 2018-11-06 | Fotonation Limited | Systems and methods for estimating depth from projected texture using camera arrays |
US10767981B2 (en) | 2013-11-18 | 2020-09-08 | Fotonation Limited | Systems and methods for estimating depth from projected texture using camera arrays |
US10708492B2 (en) | 2013-11-26 | 2020-07-07 | Fotonation Limited | Array camera configurations incorporating constituent array cameras and constituent cameras |
US10089740B2 (en) | 2014-03-07 | 2018-10-02 | Fotonation Limited | System and methods for depth regularization and semiautomatic interactive matting using RGB-D images |
US10574905B2 (en) | 2014-03-07 | 2020-02-25 | Fotonation Limited | System and methods for depth regularization and semiautomatic interactive matting using RGB-D images |
US10250871B2 (en) | 2014-09-29 | 2019-04-02 | Fotonation Limited | Systems and methods for dynamic calibration of array cameras |
US11546576B2 (en) | 2014-09-29 | 2023-01-03 | Adeia Imaging Llc | Systems and methods for dynamic calibration of array cameras |
US9942474B2 (en) | 2015-04-17 | 2018-04-10 | Fotonation Cayman Limited | Systems and methods for performing high speed video capture and depth estimation using array cameras |
US10388069B2 (en) | 2015-09-09 | 2019-08-20 | Futurewei Technologies, Inc. | Methods and systems for light field augmented reality/virtual reality on mobile devices |
WO2018094804A1 (en) * | 2016-11-23 | 2018-05-31 | 深圳市元征科技股份有限公司 | Image processing method and device |
CN110637273A (en) * | 2017-05-10 | 2019-12-31 | 微软技术许可有限责任公司 | Presenting applications within a virtual environment |
US10482618B2 (en) | 2017-08-21 | 2019-11-19 | Fotonation Limited | Systems and methods for hybrid depth regularization |
US11983893B2 (en) | 2017-08-21 | 2024-05-14 | Adeia Imaging Llc | Systems and methods for hybrid depth regularization |
US11562498B2 (en) | 2017-08-21 | 2023-01-24 | Adela Imaging LLC | Systems and methods for hybrid depth regularization |
US10818026B2 (en) | 2017-08-21 | 2020-10-27 | Fotonation Limited | Systems and methods for hybrid depth regularization |
US20220146820A1 (en) * | 2017-12-05 | 2022-05-12 | Apple Inc. | Lens Array for Shifting Perspective of an Imaging System |
US11921286B2 (en) * | 2017-12-05 | 2024-03-05 | Apple Inc. | Lens array for shifting perspective of an imaging system |
US11143865B1 (en) * | 2017-12-05 | 2021-10-12 | Apple Inc. | Lens array for shifting perspective of an imaging system |
WO2020040521A1 (en) * | 2018-08-21 | 2020-02-27 | 삼성전자 주식회사 | Method for synthesizing intermediate view of light field, system for synthesizing intermediate view of light field, and method for compressing light field |
US11533464B2 (en) | 2018-08-21 | 2022-12-20 | Samsung Electronics Co., Ltd. | Method for synthesizing intermediate view of light field, system for synthesizing intermediate view of light field, and method for compressing light field |
KR20200021891A (en) * | 2018-08-21 | 2020-03-02 | 삼성전자주식회사 | Method for the synthesis of intermediate views of a light field, system for the synthesis of intermediate views of a light field, and method for the compression of a light field |
KR102658359B1 (en) * | 2018-08-21 | 2024-04-17 | 삼성전자주식회사 | Method for the synthesis of intermediate views of a light field, system for the synthesis of intermediate views of a light field, and method for the compression of a light field |
US11182872B2 (en) * | 2018-11-02 | 2021-11-23 | Electronics And Telecommunications Research Institute | Plenoptic data storage system and operating method thereof |
CN111669603A (en) * | 2019-03-07 | 2020-09-15 | 阿里巴巴集团控股有限公司 | Multi-angle free visual angle data processing method and device, medium, terminal and equipment |
US11521347B2 (en) | 2019-03-07 | 2022-12-06 | Alibaba Group Holding Limited | Method, apparatus, medium, and device for generating multi-angle free-respective image data |
US11699273B2 (en) | 2019-09-17 | 2023-07-11 | Intrinsic Innovation Llc | Systems and methods for surface modeling using polarization cues |
US11270110B2 (en) | 2019-09-17 | 2022-03-08 | Boston Polarimetrics, Inc. | Systems and methods for surface modeling using polarization cues |
US11525906B2 (en) | 2019-10-07 | 2022-12-13 | Intrinsic Innovation Llc | Systems and methods for augmentation of sensor systems and imaging systems with polarization |
US12099148B2 (en) | 2019-10-07 | 2024-09-24 | Intrinsic Innovation Llc | Systems and methods for surface normals sensing with polarization |
US11982775B2 (en) | 2019-10-07 | 2024-05-14 | Intrinsic Innovation Llc | Systems and methods for augmentation of sensor systems and imaging systems with polarization |
US11302012B2 (en) | 2019-11-30 | 2022-04-12 | Boston Polarimetrics, Inc. | Systems and methods for transparent object segmentation using polarization cues |
US11842495B2 (en) | 2019-11-30 | 2023-12-12 | Intrinsic Innovation Llc | Systems and methods for transparent object segmentation using polarization cues |
US11580667B2 (en) | 2020-01-29 | 2023-02-14 | Intrinsic Innovation Llc | Systems and methods for characterizing object pose detection and measurement systems |
US11797863B2 (en) | 2020-01-30 | 2023-10-24 | Intrinsic Innovation Llc | Systems and methods for synthesizing data for training statistical models on different imaging modalities including polarized images |
US11953700B2 (en) | 2020-05-27 | 2024-04-09 | Intrinsic Innovation Llc | Multi-aperture polarization optical systems using beam splitters |
US12020455B2 (en) | 2021-03-10 | 2024-06-25 | Intrinsic Innovation Llc | Systems and methods for high dynamic range image reconstruction |
US12069227B2 (en) | 2021-03-10 | 2024-08-20 | Intrinsic Innovation Llc | Multi-modal and multi-spectral stereo camera arrays |
US11290658B1 (en) | 2021-04-15 | 2022-03-29 | Boston Polarimetrics, Inc. | Systems and methods for camera exposure control |
US11954886B2 (en) | 2021-04-15 | 2024-04-09 | Intrinsic Innovation Llc | Systems and methods for six-degree of freedom pose estimation of deformable objects |
US11683594B2 (en) | 2021-04-15 | 2023-06-20 | Intrinsic Innovation Llc | Systems and methods for camera exposure control |
US12067746B2 (en) | 2021-05-07 | 2024-08-20 | Intrinsic Innovation Llc | Systems and methods for using computer vision to pick up small objects |
US12175741B2 (en) | 2021-06-22 | 2024-12-24 | Intrinsic Innovation Llc | Systems and methods for a vision guided end effector |
US12172310B2 (en) | 2021-06-29 | 2024-12-24 | Intrinsic Innovation Llc | Systems and methods for picking objects using 3-D geometry and segmentation |
US11689813B2 (en) | 2021-07-01 | 2023-06-27 | Intrinsic Innovation Llc | Systems and methods for high dynamic range imaging using crossed polarizers |
US12293535B2 (en) | 2021-08-03 | 2025-05-06 | Intrinsic Innovation Llc | Systems and methods for training pose estimators in computer vision |
Also Published As
Publication number | Publication date |
---|---|
EP2901671A4 (en) | 2016-08-24 |
WO2014052974A2 (en) | 2014-04-03 |
EP4307659A1 (en) | 2024-01-17 |
WO2014052974A3 (en) | 2014-06-12 |
US10390005B2 (en) | 2019-08-20 |
US20140092281A1 (en) | 2014-04-03 |
CN104685860A (en) | 2015-06-03 |
EP2901671A2 (en) | 2015-08-05 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10390005B2 (en) | Generating images from light fields utilizing virtual viewpoints | |
JP6767543B2 (en) | Image imaging and processing using a monolithic camera array with different types of imaging devices | |
JP6021541B2 (en) | Image processing apparatus and method | |
US9031343B2 (en) | Systems and methods for encoding light field image files having a depth map | |
US20190089947A1 (en) | Autofocus System for a Conventional Camera That Uses Depth Information from an Array Camera | |
JP5140210B2 (en) | Imaging apparatus and image processing method | |
US9341935B2 (en) | Image capturing device | |
US20130229544A1 (en) | Image processing device | |
US20160269632A1 (en) | Image processing system and image processing method | |
WO2013108339A1 (en) | Stereo imaging device | |
EP2580914A2 (en) | Video camera providing videos with perceived depth | |
WO2012029299A1 (en) | Image capture device, playback device, and image-processing method | |
US10349040B2 (en) | Storing data retrieved from different sensors for generating a 3-D image | |
WO2011156131A1 (en) | Forming video with perceived depth | |
US20150009294A1 (en) | Image processing device and method, and imaging device | |
JP6128748B2 (en) | Image processing apparatus and method | |
WO2017118662A1 (en) | Spherical virtual reality camera | |
JP2010181826A (en) | Three-dimensional image forming apparatus | |
JP6021489B2 (en) | Imaging apparatus, image processing apparatus and method thereof | |
JP2013150071A (en) | Encoder, encoding method, program and storage medium | |
JP2000222116A (en) | Position recognition method for display image, position recognition device therefor and virtual image stereoscopic synthesis device | |
Sabater et al. | Light-field demultiplexing and disparity estimation | |
JP6257260B2 (en) | Imaging apparatus and control method thereof | |
WO2012014695A1 (en) | Three-dimensional imaging device and imaging method for same | |
JP2013175821A (en) | Image processing device, image processing method, and program |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: DBD CREDIT FUNDING LLC, NEW YORK Free format text: SECURITY INTEREST;ASSIGNOR:PELICAN IMAGING CORPORATION;REEL/FRAME:037565/0417 Effective date: 20151221 Owner name: KIP PELI P1 LP, NEW YORK Free format text: SECURITY INTEREST;ASSIGNOR:PELICAN IMAGING CORPORATION;REEL/FRAME:037565/0439 Effective date: 20151221 Owner name: KIP PELI P1 LP, NEW YORK Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:PELICAN IMAGING CORPORATION;REEL/FRAME:037565/0385 Effective date: 20151221 |
|
AS | Assignment |
Owner name: DBD CREDIT FUNDING LLC, NEW YORK Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE ASSIGNOR AND ASSIGNEE PREVIOUSLY RECORDED AT REEL: 037565 FRAME: 0439. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY INTEREST;ASSIGNOR:KIP PELI P1 LP;REEL/FRAME:037591/0377 Effective date: 20151221 |
|
AS | Assignment |
Owner name: DRAWBRIDGE OPPORTUNITIES FUND LP, NEW YORK Free format text: SECURITY INTEREST;ASSIGNOR:DBD CREDIT FUNDING LLC;REEL/FRAME:038982/0151 Effective date: 20160608 Owner name: DRAWBRIDGE OPPORTUNITIES FUND LP, NEW YORK Free format text: SECURITY INTEREST;ASSIGNOR:DBD CREDIT FUNDING LLC;REEL/FRAME:039117/0345 Effective date: 20160608 |
|
AS | Assignment |
Owner name: PELICAN IMAGING CORPORATION, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:NISENZON, SEMYON;JAIN, ANKIT K.;SIGNING DATES FROM 20131121 TO 20131122;REEL/FRAME:039899/0724 |
|
AS | Assignment |
Owner name: DRAWBRIDGE SPECIAL OPPORTUNITIES FUND LP, NEW YORK Free format text: CHANGE OF NAME;ASSIGNOR:DBD CREDIT FUNDING LLC;REEL/FRAME:040494/0930 Effective date: 20161019 Owner name: DRAWBRIDGE SPECIAL OPPORTUNITIES FUND LP, NEW YORK Free format text: CHANGE OF NAME;ASSIGNOR:DBD CREDIT FUNDING LLC;REEL/FRAME:040423/0725 Effective date: 20161019 |
|
AS | Assignment |
Owner name: PELICAN IMAGING CORPORATION, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:KIP PELI P1 LP;REEL/FRAME:040674/0677 Effective date: 20161031 Owner name: FOTONATION CAYMAN LIMITED, UNITED STATES Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:PELICAN IMAGING CORPORATION;REEL/FRAME:040675/0025 Effective date: 20161031 |
|
AS | Assignment |
Owner name: FOTONATION LIMITED, IRELAND Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:FOTONATION CAYMAN LIMITED;REEL/FRAME:046539/0815 Effective date: 20180516 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |
|
AS | Assignment |
Owner name: BANK OF AMERICA, N.A., AS COLLATERAL AGENT, NORTH CAROLINA Free format text: SECURITY INTEREST;ASSIGNORS:ADEIA GUIDES INC.;ADEIA IMAGING LLC;ADEIA MEDIA HOLDINGS LLC;AND OTHERS;REEL/FRAME:063529/0272 Effective date: 20230501 |