US20160219217A1 - Camera Field Of View Effects Based On Device Orientation And Scene Content - Google Patents
Camera Field Of View Effects Based On Device Orientation And Scene Content Download PDFInfo
- Publication number
- US20160219217A1 US20160219217A1 US14/602,832 US201514602832A US2016219217A1 US 20160219217 A1 US20160219217 A1 US 20160219217A1 US 201514602832 A US201514602832 A US 201514602832A US 2016219217 A1 US2016219217 A1 US 2016219217A1
- Authority
- US
- United States
- Prior art keywords
- image
- view
- orientation
- field
- display
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 230000000694 effects Effects 0.000 title description 3
- 238000012545 processing Methods 0.000 claims description 11
- 238000000034 method Methods 0.000 abstract description 19
- 230000003247 decreasing effect Effects 0.000 abstract 1
- 238000004891 communication Methods 0.000 description 4
- 238000001514 detection method Methods 0.000 description 4
- 238000004590 computer program Methods 0.000 description 3
- 238000011161 development Methods 0.000 description 3
- 238000010586 diagram Methods 0.000 description 3
- 210000001061 forehead Anatomy 0.000 description 3
- 230000006870 function Effects 0.000 description 2
- 210000003128 head Anatomy 0.000 description 2
- 238000012805 post-processing Methods 0.000 description 2
- 230000035755 proliferation Effects 0.000 description 2
- 238000013481 data capture Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 210000000887 face Anatomy 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 230000006855 networking Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- H04N5/23238—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/63—Control of cameras or camera modules by using electronic viewfinders
- H04N23/631—Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters
- H04N23/632—Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters for displaying or modifying preview images prior to image capturing, e.g. variety of image resolutions or capturing parameters
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/222—Studio circuitry; Studio devices; Studio equipment
- H04N5/262—Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
- H04N5/2628—Alteration of picture size, shape, position or orientation, e.g. zooming, rotation, rolling, perspective, translation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/80—Camera processing pipelines; Components thereof
-
- H04N5/23293—
Definitions
- This disclosure relates generally to the field of image capture, and more particularly, to acquiring images with a wide “field of view” (FOV) image sensor.
- FOV field of view
- Field of view refers to the angular extent of a given scene that is imaged by a camera.
- FOV is typically measured in terms of a number of degrees, and may be expressed as a vertical FOV, horizontal FOV, and/or diagonal FOV.
- the diagonal FOV of the image sensor is often referred to herein, as it is a more relevant measure of the camera's optics since it attempts to cover the corners of the image, where “roll off,” i.e., vignetting, problems associated with pixels at the corners of the image sensor may become more pronounced.
- a typical 35 mm camera with a lens having a focal length of 50 mm will have a horizontal FOV of 39.6°, a vertical FOV of 27.0°, and a diagonal FOV of 46.8°.
- lenses with focal lengths shorter than the sensor diagonal will distort perspective, such that objects that are closer to the camera appear to be larger than they would with a normal lens, and distant objects appear to be smaller and further away. Because of this distortion, wide angle lenses are not typically used for portraits. Likewise a normal lens (e.g., with a focal length approximately equal to the sensor diagonal) is not typically used for landscape photography because of the limited field of view.
- Those cameras that are optimized for portraits and video conference streaming are not optimal for landscapes and group photos because of their limited field of view.
- cameras on devices that face in the opposite direction of the device's preview display screen e.g., “rear-facing” cameras
- the field of view of a given camera also may influence how the user composes the shot and the quality of the ultimate captured image.
- image sensors use a fixed FOV, whether the camera device is being held in “portrait” orientation or in “landscape” orientation.
- FOV field-of-V
- FOV field-of-V
- the FOV may not be sufficiently wide to capture all the members of the group, even when the camera device is being held by the user at a fully-extended arm's length (estimated at around 50 cm).
- the camera may be so close to the user's face that the resulting captured image has unwanted perspective distortions, such as exaggerated nose, cheek, and/or forehead sizes, as well as the gaze of the user appearing “cross-eyed.”
- a standard consumer electronic device camera with a wide angle lens e.g., a “rear-facing” camera
- taking a portrait image requires moving within roughly 30 cm of the subject's face, which could be uncomfortable for the subject and result in a distorted image.
- a standard camera with a normal lens i.e., non-wide angle lens
- an image capture device comprising an image sensor designed to read out a 16:9 aspect ratio (or some other wide-angle aspect ratio), for example, 3840 ⁇ 2160 pixels.
- a lens with a 79° diagonal FOV would cover the full image sensor diagonal and could be used to produce wide angle landscape images.
- the full, wide diagonal FOV may be displayed to the user on a preview screen of the camera device, and the landscape image may be captured in a wide, 16:9 aspect ratio.
- a 4:3 aspect ratio image (or some other non-wide-angle aspect ratio), e.g., using an effective 68° diagonal FOV, may be displayed and captured.
- the user can hold the camera device in portrait orientation to take self-portraits or make personal video calls.
- the user wants to include other people in the photo, include scenery in the photo, or make a video call with multiple people in view, he or she may hold the phone in landscape orientation, allowing the camera will configure itself appropriately and output the wider diagonal FOV, higher resolution and 16:9 aspect ratio image(s).
- the camera may be configured to be optimized for a portrait or self-portrait image, e.g., using the reduced FOV techniques described above.
- the FOV may be reduced to a 38° or less horizontal FOV (simulating a 50 mm equivalent focal length) so that the user will move the camera device further away, in order to properly frame his or her face in the shot on the camera device's preview display screen.
- FOV field-to-view
- lens distortion will be reduced, producing a more flattering and natural-looking image.
- the aspect ratio of the captured portrait orientation image may be cropped to a 4:3 aspect ratio, with the crop window shifted over the image sensor so that the user will be inclined to position the camera so that his or her eyes are closer to the lens plane, thus producing a more natural-looking gaze in the portrait or self-portrait image.
- the image data capture by the image sensor may be optically zoomed, cropped, scaled, and/or shifted before being displayed to the user on the camera device's preview display screen, so that the user will naturally be inclined to hold the camera device in a position that will produce an “optimal” portrait or self-portrait image, i.e., avoid as much as possible of the unwanted distortion and unnatural gazes that are typically produced by smaller FOV cameras.
- FIG. 1A illustrates an exemplary self-portrait image taken in landscape orientation using a camera device having a wide diagonal FOV lens, in accordance with one embodiment.
- FIG. 1B illustrates an exemplary self-portrait image taken in portrait orientation using a camera device having a wide diagonal FOV lens, in accordance with one embodiment.
- FIG. 2A illustrates an exemplary self-portrait image taken in portrait orientation at bent-arm's length, in accordance with one embodiment.
- FIG. 2B illustrates an exemplary self-portrait image taken in portrait orientation at fully-extended arm's length, in accordance with one embodiment.
- FIG. 2C illustrates an exemplary 4:3 aspect ratio self-portrait image crop taken from a portrait orientation image captured by a 16:9 aspect ratio wide diagonal FOV image sensor, in accordance with one embodiment.
- FIG. 3A illustrates an exemplary scaled 4:3 aspect ratio self-portrait image crop taken from a portrait orientation image captured by a 16:9 aspect ratio wide diagonal FOV image sensor and displayed on a camera device's preview display screen, in accordance with one embodiment.
- FIG. 3B illustrates an exemplary shifted 4:3 aspect ratio self-portrait image crop taken from a portrait orientation image captured by a 16:9 aspect ratio wide diagonal FOV image sensor and displayed on a camera device's preview display screen, in accordance with one embodiment.
- FIG. 4 is a flowchart illustrating a method of capturing enhanced self-portrait images using a wide diagonal FOV camera, in accordance with one embodiment.
- FIG. 5 illustrates a simplified functional block diagram of an illustrative electronic image capture and display device, according to one embodiment.
- Systems, methods and program storage devices are disclosed, which provide instructions to cause one or more cameras and/or processing units to capture enhanced self-portrait images using a wide diagonal FOV camera.
- the techniques disclosed herein are applicable to any number of electronic devices with cameras and displays: such as digital cameras, digital video cameras, mobile phones, personal data assistants (PDAs), portable music players, monitors, as well as desktop, laptop, and tablet computer displays.
- PDAs personal data assistants
- portable music players portable music players
- monitors as well as desktop, laptop, and tablet computer displays.
- image capture device 100 may, e.g., comprise a mobile device having a preview display screen 102 , a physical and/or “virtual” shutter button of some sort 106 , and a front-facing camera 104 (i.e., a camera lens facing in the same direction as the device's preview display screen 102 ). As shown in FIG.
- the image capture device 100 captures images having a horizontal field of view whose extent is labeled “HFOV 1 .”
- a wide FOV lens will be considered one wherein the horizontal FOV of the lens is greater than 63°.
- the horizontal FOV of the lens may be as great as 71°.
- dashed lines 110 is an example of the type of image that would be captured by a camera that did not have as wide of a HFOV, i.e., an “HFOV 2 ” that was less than “HFOV 1 .”
- the image bounded by dashed lines 110 would have a horizontal FOV of closer to 57° in the example depicted.
- the smaller horizontal FOV image would have cut off a portion of the head of one of the subjects in image 125 , and thus produced an undesirable image.
- one or more transparent or semi-transparent user interface (UI) elements 108 may be applied to the preview display screen 102 of device 100 in order allow the user to have access to the virtual shutter 106 (and/or other UI-based photographic control elements), while still being able to view the full wide FOV preview image streamed to the preview display screen.
- UI user interface
- FIG. 1B an exemplary self-portrait image taken in portrait orientation 175 using a camera device 150 having a wide diagonal FOV lens is illustrated, in accordance with one embodiment.
- the image capture device 150 captures images having a horizontal field of view whose extent is labeled “HFOV 2 .”
- the horizontal field of view “HFOV 2 ” corresponds to the image represented by dashed lines 110 in FIG. 1A that would have been captured if a smaller horizontal FOV was used.
- a smaller horizontal (and thus diagonal) FOV is often more appropriate for “selfie” images taken with the image capture device in portrait orientation, as there is often only a single subject in the capture image.
- one embodiment disclosed herein comprises displaying a different cropping of the image captured by the image sensor to the preview display screen of the image capture device based on whether the device is being held in a portrait or a landscape orientation. In this way, the user is able to keep the device at a comfortable viewing distance when snapping the photo, no matter what orientation the device is being held in.
- the cropping of the image captured by the image sensor in portrait orientation may be configured to crop in on the “interesting” part of image, e.g., where the face is likely to be, in portrait mode for self-portraits and video conferencing, etc.
- face detection routines structured light, IR cameras, or other sensors could be used to try and determine what object or objects the user is intending to be the subject of interest in the composed portrait image.
- one or more opaque or semi-opaque user interface (UI) elements may be applied to the preview display screen 102 of device 100 in order to limit the FOV of the preview image streamed to the preview display screen to the desired portrait orientation aspect ratio.
- Image 225 represents a portrait orientation “selfie” image taken by user 200 with his arm partially bent, e.g., at a distance (D 1 ) of about 30 cm from his face. It has been empirically found that this is a comfortable reading distance and a distance that users will instinctively tend to hold the device when taking portrait orientation “selfie” images. However, this “default” distance that users tend to hold the image capture device at away from their face may lead to perspective distortions in a resultant image 225 captured by a wide FOV lens.
- holding the camera too close to the user's face may cause distortion in the user's nose ( 215 ), cheeks/chin ( 220 ), and/or forehead ( 205 ), as well as causing the gaze of the user to appear somewhat “cross-eyed” ( 210 ). These distortions often result in an unnatural and unflattering appearance for the user in the “selfie” image.
- Image 275 represents a portrait orientation “selfie” image taken by user 200 with his arm fully-extended, e.g., at a distance (D 2 ) of about 50 cm from his face. It has been empirically found that this distance is more commonly used when users are attempting to take a “landscape” orientation “group” image composed of more than one subject.
- the improved image capture device may intelligently crop the larger FOV image captured by the front-facing camera in order to frame the face of the subject in the center of the image that will be streamed to the image capture device's preview display screen. In doing so, the user's gaze will appear to be “closer” to the camera and user's nose, cheeks/chin, and forehead will appear less distorted, resulting in a more natural and flattering appearance for the user in the “selfie” image.
- FIG. 2C an exemplary 4:3 aspect ratio self-portrait image crop 275 taken from a portrait orientation image captured by a 16:9 aspect ratio wide diagonal FOV image sensor 285 is illustrated, in accordance with one embodiment.
- FIG. 2C illustrates the particular cropped part 275 of the exemplary wide FOV captured image 285 that will be displayed on the preview display screen of the device in FIG. 2B .
- the cropping of the image captured by the image sensor in portrait orientation may be configured to crop in on the “interesting” part of image, e.g., where the face is likely to be.
- the selected cropped region of the full, wide FOV image may be based on some factor(s) other than face detection, such as object recognition, focus, exposure, and/or coloration.
- the selected cropped region of the full, wide FOV image may be selected manually by the user or system in post-processing, i.e., after the image has been captured.
- a face detection algorithm could be used as a post-process to create optimized portrait images from a source image.
- the full, e.g., 16:9 aspect ratio source image could then be saved to memory, along with the metadata defining the determined cropped view.
- an exemplary scaled 4:3 aspect ratio self-portrait image crop 330 taken from a portrait orientation image captured by a 16:9 aspect ratio wide diagonal FOV image sensor 300 and displayed on a camera device 320 's preview display screen 102 is illustrated, in accordance with one embodiment.
- an intelligent image capture device may optically zoom, crop, scale, and/or shift a captured image before it is displayed to the user on the device's preview display screen based, at least in part on scene content, so that the user will naturally be inclined to hold the camera device in a position that will produce a more “optimal” self-portrait image, i.e., avoid the unwanted distortion and unnatural gazes referred to above.
- the image capture device has intelligently determined the relevant portion of the image 310 having a particular location within the image and a particular shape (i.e., a box sized to frame the location of the user's face) and then scaled up a cropped portion 315 of the relevant portion 310 and displayed it on the device's preview screen, resulting in the “zoomed in” look to the 4:3 aspect ratio self-portrait image crop 330 displayed to the user on the preview display screen 102 of camera device 320 .
- a particular shape i.e., a box sized to frame the location of the user's face
- the determination of the size and shape of the relevant portion of the image 310 may be based, at least in part, on the understanding that a user taking a portrait or self-portrait image will naturally move the device in a particular direction/orientation, so that the subject's entire face will appear on the preview display screen 102 before the image is captured.
- the “zoomed in” look to the user's face on the preview display screen 102 will cause the user to end up holding the device at distance more akin to the D 2 distance in FIG. 2B , resulting in a more optimal “selfie” image with less distortion in the user's face.
- this “zoomed in” effect could be active at all times that the device is being held in portrait orientation while the front-facing camera is active so that a user never takes a “selfie” at the “too close” range, such as the D 1 distance in FIG. 2A .
- FIG. 3B an exemplary shifted 4:3 aspect ratio self-portrait image crop 370 taken from a portrait orientation image captured by a 16:9 aspect ratio wide diagonal FOV image sensor 340 and displayed on a camera device 360 's preview display screen 102 is illustrated, in accordance with one embodiment.
- FIG. 3B an exemplary shifted 4:3 aspect ratio self-portrait image crop 370 taken from a portrait orientation image captured by a 16:9 aspect ratio wide diagonal FOV image sensor 340 and displayed on a camera device 360 's preview display screen 102 is illustrated, in accordance with one embodiment.
- FIG. 3B an exemplary shifted 4:3 aspect ratio self-portrait image crop 370 taken from a portrait orientation image captured by a 16:9 aspect ratio wide diagonal FOV image sensor 340 and displayed on a camera device 360 's preview display screen 102 is illustrated, in accordance with one embodiment.
- FIG. 3B an exemplary shifted 4:3 aspect ratio self-portrait image crop 370 taken from a portrait orientation image captured by
- the user is holding the device too low with respect to his face to capture an optimal self-portrait (which are generally taken with the camera perpendicular to—or above—the user's eye level), and so the image capture device has intelligently determined the relevant portion of the image 350 having a particular location within the image and a particular shape (i.e., a box sized to frame the location of the user's face) and then cropped out a shifted portion of the image 355 (i.e., a portion shifted down from the current location of the user's face) and displayed it on the device's preview screen, resulting in the “chopped off” look to the 4:3 aspect ratio self-portrait image crop 370 displayed to the user on the preview display screen 102 of camera device 360 (i.e., the top portion of the subject's head is “chopped off” from the image shown on preview display screen 102 ).
- a shifted portion of the image 355 i.e., a portion shifted down from the current location of the user's face
- the amount of scaling and/or shifting applied to the relevant portion of the image that is displayed on the image capture device's preview display screen will be determined, based at least in part, on the amount of re-positioning and/or re-orienting of the image capture device that needs to be performed by the user in order to result in a more optimal self-portrait image.
- the intelligent image capture device could alternately (or additionally) provide other UI cues, e.g., arrows, haptic feedback, visual or audible instructions, etc. to coax the user to move the device into a more optimal selfie-taking position.
- UI cues e.g., arrows, haptic feedback, visual or audible instructions, etc.
- such UI cues could gradually fade from the UI as the image capture device is moved into the more optimal position.
- the intelligent image capture device could alternately (or additionally) utilize one or more positional sensors within the device (e.g., gyrometer, accelerometer, barometer, altimeter, etc.) to coax the user into holding the device at a more optimal selfie-taking position (e.g., a more optimal tilt, height, angle, etc., with respect to the user).
- a more optimal selfie-taking position e.g., a more optimal tilt, height, angle, etc., with respect to the user.
- the device could automatically be capturing images and then save and present all images to the user for later selection or a preferred image(s), and/or the device could save and present only those images that are determined to be the most optimally-composed.
- the process may start with the camera device beginning to operate in an “image capture” mode using a wide diagonal FOV lens (Step 405 ). On some camera devices, this may comprise displaying a “live feed” of the images being captured by the camera device's image sensor on a preview display screen on the device.
- the camera device may determine whether the device is in landscape orientation (or, alternatively, in portrait orientation) (Step 410 ). This determination may be made by using any one or more of a number of positional sensors disposed within the image capture device, e.g., an accelerometer, gyrometer, or the like.
- the process may proceed to Step 430 .
- the determination of the image capture device's orientation may be based on whether the device has been rotated more than a threshold angle off its vertical axis, such that it is then positioned in a second orientation that is substantially orthogonal to its first orientation. For example, the image capture device may be deemed to be in a “portrait” orientation until it is rotated more than 60° (in any direction) off its vertical axis, at which point it would be deemed to be in a “landscape” orientation.
- the process may configure the camera to be optimized for a “group” photo.
- this may simply entail using the full diagonal FOV of the image sensor. In other embodiments, this may entail adding one or more transparent UI elements on the sides of the preview display screen, as described above in reference to FIG. 1A . In still other embodiments, this may entail performing face detection, face recognition, auto focusing, auto exposure, etc.
- Step 410 the process may proceed to Step 415 .
- the camera device may reduce its FOV, e.g., by any of the methods discussed above, such as optically zooming, cropping and/or scaling the image obtained by the image sensor to have a smaller effective FOV.
- reducing the device's FOV may also optionally comprise cropping the camera device to an aspect ratio of 4:3, or another aspect ratio desirable for portraits (Step 420 ). [Dashed line boxes in FIG.
- Step 4 indicate that a particular Step may be optionally performed in some embodiments.
- the image capture device may apply one or more FOV effects, e.g., optically zooming, cropping, scaling, or shifting, the image captured by the image sensor in order to coax the user to hold the image capture device in a more optimal portrait-taking position (Step 425 ).
- FOV effects e.g., optically zooming, cropping, scaling, or shifting
- Step 435 the process may proceed to Step 435 , wherein the image is captured by the image capture device according to how the camera has been configured.
- any desired post-processing may be performed on the captured image, e.g., cropping, shifting, scaling, color balancing, etc. (Step 440 ).
- the captured image is returned to the calling application and/or stored in memory, either on-board the image capture device or communicatively coupled to the image capture device (Step 445 ).
- FIG. 5 is a simplified functional block diagram of an illustrative electronic device for image capture and display, according to one embodiment.
- Electronic device 500 may include processor 505 , display 510 , user interface 515 , graphics hardware 520 , device sensors 525 (e.g., proximity sensor/ambient light sensor, accelerometer and/or gyroscope), microphone 530 , audio codec(s) 535 , speaker(s) 540 , communications circuitry 545 , digital image capture unit 550 , video codec(s) 555 , memory 560 , storage 565 , and communications bus 570 .
- Electronic device 500 may be, for example, a personal digital assistant (PDA), personal music player, a mobile telephone, or a notebook, laptop or tablet computer system.
- PDA personal digital assistant
- Processor 505 may execute instructions necessary to carry out or control the operation of many functions performed by device 500 (e.g., such as the generation and/or processing of time-lapse video in accordance with operation 100 ).
- Processor 505 may, for instance, drive display 510 and receive user input from user interface 515 .
- User interface 515 can take a variety of forms, such as a button, keypad, dial, a click wheel, keyboard, display screen and/or a touch screen.
- Processor 505 may be a system-on-chip such as those found in mobile devices and include a dedicated graphics processing unit (GPU).
- GPU graphics processing unit
- Processor 505 may be based on reduced instruction-set computer (RISC) or complex instruction-set computer (CISC) architectures or any other suitable architecture and may include one or more processing cores.
- Graphics hardware 520 may be special purpose computational hardware for processing graphics and/or assisting processor 505 process graphics information.
- graphics hardware 520 may include a programmable graphics processing unit (GPU).
- Sensor and camera circuitry 550 may capture still and video images that may be processed to generate images in accordance with this disclosure.
- the image sensor may comprise a lens with a wide diagonal FOV, such as 75° or greater (corresponding to a 35 mm equivalent 28 mm focal length lens).
- Output from camera circuitry 550 may be processed, at least in part, by video codec(s) 555 and/or processor 505 and/or graphics hardware 520 , and/or a dedicated image processing unit incorporated within circuitry 550 . Images so captured may be stored in memory 560 and/or storage 565 .
- Memory 560 may include one or more different types of media used by processor 505 , graphics hardware 520 , and image capture circuitry 550 to perform device functions.
- memory 560 may include memory cache, read-only memory (ROM), and/or random access memory (RAM).
- Storage 565 may store media (e.g., audio, image and video files), computer program instructions or software, preference information, device profile information, and any other suitable data.
- Storage 565 may include one more non-transitory storage mediums including, for example, magnetic disks (fixed, floppy, and removable) and tape, optical media such as CD-ROMs and digital video disks (DVDs), and semiconductor memory devices such as Electrically Programmable Read-Only Memory (EPROM), and Electrically Erasable Programmable Read-Only Memory (EEPROM).
- Memory 560 and storage 565 may be used to retain computer program instructions or code organized into one or more modules and written in any desired computer programming language. When executed by, for example, processor 505 such computer program code may implement one or more of the methods described herein.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Human Computer Interaction (AREA)
- Studio Devices (AREA)
Abstract
Description
- This disclosure relates generally to the field of image capture, and more particularly, to acquiring images with a wide “field of view” (FOV) image sensor.
- “Field of view,” as used herein, refers to the angular extent of a given scene that is imaged by a camera. FOV is typically measured in terms of a number of degrees, and may be expressed as a vertical FOV, horizontal FOV, and/or diagonal FOV. The diagonal FOV of the image sensor is often referred to herein, as it is a more relevant measure of the camera's optics since it attempts to cover the corners of the image, where “roll off,” i.e., vignetting, problems associated with pixels at the corners of the image sensor may become more pronounced. For reference, a typical 35 mm camera with a lens having a focal length of 50 mm will have a horizontal FOV of 39.6°, a vertical FOV of 27.0°, and a diagonal FOV of 46.8°.
- For a given camera-to-subject distance, lenses with focal lengths shorter than the sensor diagonal (commonly known as wide angle lenses) will distort perspective, such that objects that are closer to the camera appear to be larger than they would with a normal lens, and distant objects appear to be smaller and further away. Because of this distortion, wide angle lenses are not typically used for portraits. Likewise a normal lens (e.g., with a focal length approximately equal to the sensor diagonal) is not typically used for landscape photography because of the limited field of view.
- The advent of portable integrated computing devices has caused a wide proliferation of cameras and video devices. These integrated computing devices commonly take the form of smartphones or tablets and typically include general purpose computers, cameras, sophisticated user interfaces including touch sensitive screens, and wireless communications abilities through Wi-Fi, LTE, HSDPA and other cell-based or wireless technologies. The wide proliferation of these integrated devices provides opportunities to use the devices' capabilities to perform tasks that would otherwise require dedicated hardware and software. For example, as noted above, integrated devices such as smartphones and tablets typically have one or two embedded cameras. These cameras generally amount to lens/camera hardware modules that may be controlled through the general purpose computer using firmware and/or software (e.g., “Apps”) and a user interface including the touch-screen fixed buttons and touchless control such as voice control.
- The integration of cameras into communication devices such as smartphones and tablets has enabled people to share images and videos in ways never before possible. It is now very popular to acquire and immediately share photos with other people either by sending the photos via text message, SMS, or email, or by uploading the photos to an Internet-based service, such as a social networking site or a photo sharing site.
- Along with the rise in popularity of portable integrated computing devices with integrated cameras has come a rise in the popularity of consumer photography. In particular, users often take so-called “selfies,” which are portrait or self-portrait photographs of the user of the device (and/or others in the user's proximity), typically taken with a front-facing camera on the device, i.e., a camera that faces in the same direction as the camera device's preview display screen. Most prior art cameras are optimized for either wide angle general photography or for self-portraits and video streaming use cases. Those cameras that are optimized for wide angles are optimized for group and landscape compositions, but are not optimal for portraits, due, e.g., to the distortion that occurs when subjects are at short distances from the camera. Those cameras that are optimized for portraits and video conference streaming (e.g., “front-facing” cameras) are not optimal for landscapes and group photos because of their limited field of view. Moreover, cameras on devices that face in the opposite direction of the device's preview display screen (e.g., “rear-facing” cameras) are typically designed with a short focal length and wide field of view, which is not optimal for portraits, due, e.g., to distortion caused at short distances from the camera. The field of view of a given camera also may influence how the user composes the shot and the quality of the ultimate captured image.
- Traditionally, image sensors use a fixed FOV, whether the camera device is being held in “portrait” orientation or in “landscape” orientation. For camera devices with a relatively limited FOV, e.g., a horizontal FOV of 57° or less, this may result in sub-optimal “selfie” photo-taking conditions. For example, when attempting to compose a “group” selfie while holding the camera device in a “landscape” orientation, the FOV may not be sufficiently wide to capture all the members of the group, even when the camera device is being held by the user at a fully-extended arm's length (estimated at around 50 cm). Conversely, when attempting to compose a traditional, “self-only” selfie while holding the camera device in a “portrait” orientation, typically at a bent-arm's length (estimated at around 30 cm), the camera may be so close to the user's face that the resulting captured image has unwanted perspective distortions, such as exaggerated nose, cheek, and/or forehead sizes, as well as the gaze of the user appearing “cross-eyed.” For a standard consumer electronic device camera with a wide angle lens (e.g., a “rear-facing” camera), taking a portrait image requires moving within roughly 30 cm of the subject's face, which could be uncomfortable for the subject and result in a distorted image. A standard camera with a normal lens (i.e., non-wide angle lens) may not have a wide enough field of view to capture typical scenes.
- Thus, in some embodiments described herein, an image capture device is disclosed, comprising an image sensor designed to read out a 16:9 aspect ratio (or some other wide-angle aspect ratio), for example, 3840×2160 pixels. In such an exemplary device, a lens with a 79° diagonal FOV would cover the full image sensor diagonal and could be used to produce wide angle landscape images. When the camera device is held in landscape orientation and/or a landscape photo is otherwise selected by the user or the system, the full, wide diagonal FOV may be displayed to the user on a preview screen of the camera device, and the landscape image may be captured in a wide, 16:9 aspect ratio.
- When the camera device is held in portrait orientation and/or a portrait photo is selected by the user or the system, a 4:3 aspect ratio image (or some other non-wide-angle aspect ratio), e.g., using an effective 68° diagonal FOV, may be displayed and captured. In this way, the user can hold the camera device in portrait orientation to take self-portraits or make personal video calls. If the user wants to include other people in the photo, include scenery in the photo, or make a video call with multiple people in view, he or she may hold the phone in landscape orientation, allowing the camera will configure itself appropriately and output the wider diagonal FOV, higher resolution and 16:9 aspect ratio image(s).
- In other embodiments described herein, if the camera device is detected to be in the portrait orientation, the camera may be configured to be optimized for a portrait or self-portrait image, e.g., using the reduced FOV techniques described above. In one exemplary embodiment, the FOV may be reduced to a 38° or less horizontal FOV (simulating a 50 mm equivalent focal length) so that the user will move the camera device further away, in order to properly frame his or her face in the shot on the camera device's preview display screen. As a result, lens distortion will be reduced, producing a more flattering and natural-looking image. In still other embodiments, the aspect ratio of the captured portrait orientation image may be cropped to a 4:3 aspect ratio, with the crop window shifted over the image sensor so that the user will be inclined to position the camera so that his or her eyes are closer to the lens plane, thus producing a more natural-looking gaze in the portrait or self-portrait image. In still other embodiments, the image data capture by the image sensor may be optically zoomed, cropped, scaled, and/or shifted before being displayed to the user on the camera device's preview display screen, so that the user will naturally be inclined to hold the camera device in a position that will produce an “optimal” portrait or self-portrait image, i.e., avoid as much as possible of the unwanted distortion and unnatural gazes that are typically produced by smaller FOV cameras.
- Further embodiments include methods and non-transitory program storage devices, readable by a programmable control device and comprising instructions stored thereon to cause one or more processing units to implement the functionality described herein.
-
FIG. 1A illustrates an exemplary self-portrait image taken in landscape orientation using a camera device having a wide diagonal FOV lens, in accordance with one embodiment. -
FIG. 1B illustrates an exemplary self-portrait image taken in portrait orientation using a camera device having a wide diagonal FOV lens, in accordance with one embodiment. -
FIG. 2A illustrates an exemplary self-portrait image taken in portrait orientation at bent-arm's length, in accordance with one embodiment. -
FIG. 2B illustrates an exemplary self-portrait image taken in portrait orientation at fully-extended arm's length, in accordance with one embodiment. -
FIG. 2C illustrates an exemplary 4:3 aspect ratio self-portrait image crop taken from a portrait orientation image captured by a 16:9 aspect ratio wide diagonal FOV image sensor, in accordance with one embodiment. -
FIG. 3A illustrates an exemplary scaled 4:3 aspect ratio self-portrait image crop taken from a portrait orientation image captured by a 16:9 aspect ratio wide diagonal FOV image sensor and displayed on a camera device's preview display screen, in accordance with one embodiment. -
FIG. 3B illustrates an exemplary shifted 4:3 aspect ratio self-portrait image crop taken from a portrait orientation image captured by a 16:9 aspect ratio wide diagonal FOV image sensor and displayed on a camera device's preview display screen, in accordance with one embodiment. -
FIG. 4 is a flowchart illustrating a method of capturing enhanced self-portrait images using a wide diagonal FOV camera, in accordance with one embodiment. -
FIG. 5 illustrates a simplified functional block diagram of an illustrative electronic image capture and display device, according to one embodiment. - Systems, methods and program storage devices are disclosed, which provide instructions to cause one or more cameras and/or processing units to capture enhanced self-portrait images using a wide diagonal FOV camera. The techniques disclosed herein are applicable to any number of electronic devices with cameras and displays: such as digital cameras, digital video cameras, mobile phones, personal data assistants (PDAs), portable music players, monitors, as well as desktop, laptop, and tablet computer displays.
- In the following description, for purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding of the inventive concept. As part of this description, some of this disclosure's drawings represent structures and devices in block diagram form in order to avoid obscuring the invention. In the interest of clarity, not all features of an actual implementation are described in this specification. Moreover, the language used in this disclosure has been principally selected for readability and instructional purposes, and may not have been selected to delineate or circumscribe the inventive subject matter, resort to the claims being necessary to determine such inventive subject matter. Reference in this disclosure to “one embodiment” or to “an embodiment” means that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one implementation of the invention, and multiple references to “one embodiment” or “an embodiment” should not be understood as necessarily all referring to the same embodiment.
- It will be appreciated that, in the development of any actual implementation (as in any development project), numerous decisions must be made to achieve the developers' specific goals (e.g., compliance with system- and business-related constraints), and that these goals may vary from one implementation to another. It will also be appreciated that such development efforts might be complex and time-consuming, but would nevertheless be a routine undertaking for those of ordinary skill in the design of an implementation of image processing systems having the benefit of this disclosure.
- Referring now to
FIG. 1A , an exemplary self-portrait image taken inlandscape orientation 125 using acamera device 100 having a wide diagonal FOV lens is illustrated, in accordance with one embodiment. In some embodiments,image capture device 100 may, e.g., comprise a mobile device having apreview display screen 102, a physical and/or “virtual” shutter button of somesort 106, and a front-facing camera 104 (i.e., a camera lens facing in the same direction as the device's preview display screen 102). As shown inFIG. 1A , theimage capture device 100 captures images having a horizontal field of view whose extent is labeled “HFOV1.” In some embodiments, a wide FOV lens will be considered one wherein the horizontal FOV of the lens is greater than 63°. In some embodiments, in fact, the horizontal FOV of the lens may be as great as 71°. Also shown inFIG. 1A via dashedlines 110 is an example of the type of image that would be captured by a camera that did not have as wide of a HFOV, i.e., an “HFOV2” that was less than “HFOV1.” For example, the image bounded by dashedlines 110 would have a horizontal FOV of closer to 57° in the example depicted. As may now be appreciated, the smaller horizontal FOV image would have cut off a portion of the head of one of the subjects inimage 125, and thus produced an undesirable image. - In some embodiments, one or more transparent or semi-transparent user interface (UI)
elements 108 may be applied to thepreview display screen 102 ofdevice 100 in order allow the user to have access to the virtual shutter 106 (and/or other UI-based photographic control elements), while still being able to view the full wide FOV preview image streamed to the preview display screen. - Referring now to
FIG. 1B , an exemplary self-portrait image taken inportrait orientation 175 using acamera device 150 having a wide diagonal FOV lens is illustrated, in accordance with one embodiment. As shown inFIG. 1B , theimage capture device 150 captures images having a horizontal field of view whose extent is labeled “HFOV2.” The horizontal field of view “HFOV2” corresponds to the image represented by dashedlines 110 inFIG. 1A that would have been captured if a smaller horizontal FOV was used. Unlike in the landscape orientation example shown inFIG. 1A , a smaller horizontal (and thus diagonal) FOV is often more appropriate for “selfie” images taken with the image capture device in portrait orientation, as there is often only a single subject in the capture image. - One of the difficulties associated with using a wide FOV lens is that the user has to bring the image capture device closer to his or her face in order to take a picture with his or her face taking up a reasonable amount of the image. However, the closer the user places the camera to his or her face, the more perspective distortion is introduced into the capture image, due, e.g., to the greater disparity between the tip of the user's nose and the rest of the user's face. This is one reason why wide FOV lens have typically been disfavored in the prior art for front-facing cameras. In order to avoid the user having to bring the camera very close to his or her face to take a “selfie” in portrait orientation, one embodiment disclosed herein comprises displaying a different cropping of the image captured by the image sensor to the preview display screen of the image capture device based on whether the device is being held in a portrait or a landscape orientation. In this way, the user is able to keep the device at a comfortable viewing distance when snapping the photo, no matter what orientation the device is being held in.
- In some embodiments, the cropping of the image captured by the image sensor in portrait orientation may be configured to crop in on the “interesting” part of image, e.g., where the face is likely to be, in portrait mode for self-portraits and video conferencing, etc. In addition to software-implemented face detection routines, structured light, IR cameras, or other sensors could be used to try and determine what object or objects the user is intending to be the subject of interest in the composed portrait image.
- In some embodiments, one or more opaque or semi-opaque user interface (UI) elements (e.g., 155/160) may be applied to the
preview display screen 102 ofdevice 100 in order to limit the FOV of the preview image streamed to the preview display screen to the desired portrait orientation aspect ratio. - Referring now to
FIG. 2A , an exemplary self-portrait image 225 taken in portrait orientation at bent-arm's length is illustrated, in accordance with one embodiment. Image 225 represents a portrait orientation “selfie” image taken byuser 200 with his arm partially bent, e.g., at a distance (D1) of about 30 cm from his face. It has been empirically found that this is a comfortable reading distance and a distance that users will instinctively tend to hold the device when taking portrait orientation “selfie” images. However, this “default” distance that users tend to hold the image capture device at away from their face may lead to perspective distortions in a resultant image 225 captured by a wide FOV lens. In particular, holding the camera too close to the user's face may cause distortion in the user's nose (215), cheeks/chin (220), and/or forehead (205), as well as causing the gaze of the user to appear somewhat “cross-eyed” (210). These distortions often result in an unnatural and unflattering appearance for the user in the “selfie” image. - Referring now to
FIG. 2B , an exemplary self-portrait image 275 taken in portrait orientation at fully-extended arm's length is illustrated, in accordance with one embodiment.Image 275 represents a portrait orientation “selfie” image taken byuser 200 with his arm fully-extended, e.g., at a distance (D2) of about 50 cm from his face. It has been empirically found that this distance is more commonly used when users are attempting to take a “landscape” orientation “group” image composed of more than one subject. By holding a wide FOV camera lens at this longer distance, D2, the improved image capture device may intelligently crop the larger FOV image captured by the front-facing camera in order to frame the face of the subject in the center of the image that will be streamed to the image capture device's preview display screen. In doing so, the user's gaze will appear to be “closer” to the camera and user's nose, cheeks/chin, and forehead will appear less distorted, resulting in a more natural and flattering appearance for the user in the “selfie” image. - Referring now to
FIG. 2C , an exemplary 4:3 aspect ratio self-portrait image crop 275 taken from a portrait orientation image captured by a 16:9 aspect ratio wide diagonalFOV image sensor 285 is illustrated, in accordance with one embodiment.FIG. 2C illustrates the particular croppedpart 275 of the exemplary wide FOV capturedimage 285 that will be displayed on the preview display screen of the device inFIG. 2B . As mentioned above, according to some embodiments, the cropping of the image captured by the image sensor in portrait orientation may be configured to crop in on the “interesting” part of image, e.g., where the face is likely to be. In some embodiments, the selected cropped region of the full, wide FOV image may be based on some factor(s) other than face detection, such as object recognition, focus, exposure, and/or coloration. In still other embodiments, the selected cropped region of the full, wide FOV image may be selected manually by the user or system in post-processing, i.e., after the image has been captured. For example, a face detection algorithm could be used as a post-process to create optimized portrait images from a source image. The full, e.g., 16:9 aspect ratio source image, could then be saved to memory, along with the metadata defining the determined cropped view. - Referring now to
FIG. 3A , an exemplary scaled 4:3 aspect ratio self-portrait image crop 330 taken from a portrait orientation image captured by a 16:9 aspect ratio wide diagonalFOV image sensor 300 and displayed on acamera device 320'spreview display screen 102 is illustrated, in accordance with one embodiment. As explained above, according to some embodiments, an intelligent image capture device may optically zoom, crop, scale, and/or shift a captured image before it is displayed to the user on the device's preview display screen based, at least in part on scene content, so that the user will naturally be inclined to hold the camera device in a position that will produce a more “optimal” self-portrait image, i.e., avoid the unwanted distortion and unnatural gazes referred to above. - In the example of
FIG. 3A , the user is holding the device too close to his face to capture an optimal self-portrait, and so the image capture device has intelligently determined the relevant portion of theimage 310 having a particular location within the image and a particular shape (i.e., a box sized to frame the location of the user's face) and then scaled up a croppedportion 315 of therelevant portion 310 and displayed it on the device's preview screen, resulting in the “zoomed in” look to the 4:3 aspect ratio self-portrait image crop 330 displayed to the user on thepreview display screen 102 ofcamera device 320. In some embodiments, the determination of the size and shape of the relevant portion of theimage 310 may be based, at least in part, on the understanding that a user taking a portrait or self-portrait image will naturally move the device in a particular direction/orientation, so that the subject's entire face will appear on thepreview display screen 102 before the image is captured. In the example ofFIG. 3A , the “zoomed in” look to the user's face on thepreview display screen 102 will cause the user to end up holding the device at distance more akin to the D2 distance inFIG. 2B , resulting in a more optimal “selfie” image with less distortion in the user's face. According to some embodiments, this “zoomed in” effect could be active at all times that the device is being held in portrait orientation while the front-facing camera is active so that a user never takes a “selfie” at the “too close” range, such as the D1 distance inFIG. 2A . - Referring now to
FIG. 3B , an exemplary shifted 4:3 aspect ratio self-portrait image crop 370 taken from a portrait orientation image captured by a 16:9 aspect ratio wide diagonalFOV image sensor 340 and displayed on acamera device 360'spreview display screen 102 is illustrated, in accordance with one embodiment. In the example ofFIG. 3B , the user is holding the device too low with respect to his face to capture an optimal self-portrait (which are generally taken with the camera perpendicular to—or above—the user's eye level), and so the image capture device has intelligently determined the relevant portion of theimage 350 having a particular location within the image and a particular shape (i.e., a box sized to frame the location of the user's face) and then cropped out a shifted portion of the image 355 (i.e., a portion shifted down from the current location of the user's face) and displayed it on the device's preview screen, resulting in the “chopped off” look to the 4:3 aspect ratio self-portrait image crop 370 displayed to the user on thepreview display screen 102 of camera device 360 (i.e., the top portion of the subject's head is “chopped off” from the image shown on preview display screen 102). This will naturally lead the user to move the device farther “upwards” with respect to his face so that his entire face will appear on thepreview display screen 102. In doing so, the user will end up holding the device at a more optimal “selfie” image-taking angle, resulting in less distortion in the user's face and a more flattering overall appearance in the resultant captured image. It is to be understood that the shifting could take place in any direction or combination of directions in order to better position the user for a more optimal portrait image, and that the aforementioned scaling techniques could likewise be used in combination with the described shifting techniques in order to better position the user for a more optimal portrait or self-portrait image. Further, the amount of scaling and/or shifting applied to the relevant portion of the image that is displayed on the image capture device's preview display screen will be determined, based at least in part, on the amount of re-positioning and/or re-orienting of the image capture device that needs to be performed by the user in order to result in a more optimal self-portrait image. - In some embodiments, the intelligent image capture device could alternately (or additionally) provide other UI cues, e.g., arrows, haptic feedback, visual or audible instructions, etc. to coax the user to move the device into a more optimal selfie-taking position. In other embodiment, such UI cues could gradually fade from the UI as the image capture device is moved into the more optimal position. In yet other embodiments, the intelligent image capture device could alternately (or additionally) utilize one or more positional sensors within the device (e.g., gyrometer, accelerometer, barometer, altimeter, etc.) to coax the user into holding the device at a more optimal selfie-taking position (e.g., a more optimal tilt, height, angle, etc., with respect to the user). In still other embodiments, as the user is moving the image capture device to the determined more optimal height, angle, FOV, etc., the device could automatically be capturing images and then save and present all images to the user for later selection or a preferred image(s), and/or the device could save and present only those images that are determined to be the most optimally-composed.
- Referring now to
FIG. 4 , aflowchart 400 illustrating a method of capturing enhanced self-portrait images using an image capture device having a wide diagonal FOV camera is illustrated, in accordance with one embodiment. First, the process may start with the camera device beginning to operate in an “image capture” mode using a wide diagonal FOV lens (Step 405). On some camera devices, this may comprise displaying a “live feed” of the images being captured by the camera device's image sensor on a preview display screen on the device. During this image capture mode, the camera device may determine whether the device is in landscape orientation (or, alternatively, in portrait orientation) (Step 410). This determination may be made by using any one or more of a number of positional sensors disposed within the image capture device, e.g., an accelerometer, gyrometer, or the like. - If, at
Step 410 it is determined that the image capture device is in landscape orientation, the process may proceed to Step 430. In some implementations, the determination of the image capture device's orientation may be based on whether the device has been rotated more than a threshold angle off its vertical axis, such that it is then positioned in a second orientation that is substantially orthogonal to its first orientation. For example, the image capture device may be deemed to be in a “portrait” orientation until it is rotated more than 60° (in any direction) off its vertical axis, at which point it would be deemed to be in a “landscape” orientation. AtStep 430, the process may configure the camera to be optimized for a “group” photo. According to some embodiments, this may simply entail using the full diagonal FOV of the image sensor. In other embodiments, this may entail adding one or more transparent UI elements on the sides of the preview display screen, as described above in reference toFIG. 1A . In still other embodiments, this may entail performing face detection, face recognition, auto focusing, auto exposure, etc. - If, instead, at
Step 410 it is determined that the image capture device is not in landscape orientation (i.e., is in portrait orientation), the process may proceed to Step 415. AtStep 415, the camera device may reduce its FOV, e.g., by any of the methods discussed above, such as optically zooming, cropping and/or scaling the image obtained by the image sensor to have a smaller effective FOV. In some embodiments, reducing the device's FOV may also optionally comprise cropping the camera device to an aspect ratio of 4:3, or another aspect ratio desirable for portraits (Step 420). [Dashed line boxes inFIG. 4 indicate that a particular Step may be optionally performed in some embodiments.] Next the image capture device may apply one or more FOV effects, e.g., optically zooming, cropping, scaling, or shifting, the image captured by the image sensor in order to coax the user to hold the image capture device in a more optimal portrait-taking position (Step 425). These techniques are described in greater detail above with reference toFIGS. 3A and 3B . - Once the camera of the image capture device has been configured for the a landscape or portrait-orientation image capture, the process may proceed to Step 435, wherein the image is captured by the image capture device according to how the camera has been configured. Next, any desired post-processing may be performed on the captured image, e.g., cropping, shifting, scaling, color balancing, etc. (Step 440). Finally, the captured image is returned to the calling application and/or stored in memory, either on-board the image capture device or communicatively coupled to the image capture device (Step 445).
-
FIG. 5 is a simplified functional block diagram of an illustrative electronic device for image capture and display, according to one embodiment.Electronic device 500 may includeprocessor 505,display 510, user interface 515,graphics hardware 520, device sensors 525 (e.g., proximity sensor/ambient light sensor, accelerometer and/or gyroscope),microphone 530, audio codec(s) 535, speaker(s) 540,communications circuitry 545, digitalimage capture unit 550, video codec(s) 555,memory 560,storage 565, andcommunications bus 570.Electronic device 500 may be, for example, a personal digital assistant (PDA), personal music player, a mobile telephone, or a notebook, laptop or tablet computer system. -
Processor 505 may execute instructions necessary to carry out or control the operation of many functions performed by device 500 (e.g., such as the generation and/or processing of time-lapse video in accordance with operation 100).Processor 505 may, for instance,drive display 510 and receive user input from user interface 515. User interface 515 can take a variety of forms, such as a button, keypad, dial, a click wheel, keyboard, display screen and/or a touch screen.Processor 505 may be a system-on-chip such as those found in mobile devices and include a dedicated graphics processing unit (GPU).Processor 505 may be based on reduced instruction-set computer (RISC) or complex instruction-set computer (CISC) architectures or any other suitable architecture and may include one or more processing cores.Graphics hardware 520 may be special purpose computational hardware for processing graphics and/or assistingprocessor 505 process graphics information. In one embodiment,graphics hardware 520 may include a programmable graphics processing unit (GPU). - Sensor and
camera circuitry 550 may capture still and video images that may be processed to generate images in accordance with this disclosure. As mentioned above, the image sensor may comprise a lens with a wide diagonal FOV, such as 75° or greater (corresponding to a 35 mm equivalent 28 mm focal length lens). Output fromcamera circuitry 550 may be processed, at least in part, by video codec(s) 555 and/orprocessor 505 and/orgraphics hardware 520, and/or a dedicated image processing unit incorporated withincircuitry 550. Images so captured may be stored inmemory 560 and/orstorage 565.Memory 560 may include one or more different types of media used byprocessor 505,graphics hardware 520, andimage capture circuitry 550 to perform device functions. For example,memory 560 may include memory cache, read-only memory (ROM), and/or random access memory (RAM).Storage 565 may store media (e.g., audio, image and video files), computer program instructions or software, preference information, device profile information, and any other suitable data.Storage 565 may include one more non-transitory storage mediums including, for example, magnetic disks (fixed, floppy, and removable) and tape, optical media such as CD-ROMs and digital video disks (DVDs), and semiconductor memory devices such as Electrically Programmable Read-Only Memory (EPROM), and Electrically Erasable Programmable Read-Only Memory (EEPROM).Memory 560 andstorage 565 may be used to retain computer program instructions or code organized into one or more modules and written in any desired computer programming language. When executed by, for example,processor 505 such computer program code may implement one or more of the methods described herein. - It is to be understood that the above description is intended to be illustrative, and not restrictive. The material has been presented to enable any person skilled in the art to make and use the invention as claimed and is provided in the context of particular embodiments, variations of which will be readily apparent to those skilled in the art (e.g., some of the disclosed embodiments may be used in combination with each other). In addition, it will be understood that some of the operations identified herein may be performed in different orders. The scope of the invention therefore should be determined with reference to the appended claims, along with the full scope of equivalents to which such claims are entitled. In the appended claims, the terms “including” and “in which” are used as the plain-English equivalents of the respective terms “comprising” and “wherein.”
Claims (28)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/602,832 US9712751B2 (en) | 2015-01-22 | 2015-01-22 | Camera field of view effects based on device orientation and scene content |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/602,832 US9712751B2 (en) | 2015-01-22 | 2015-01-22 | Camera field of view effects based on device orientation and scene content |
Publications (2)
Publication Number | Publication Date |
---|---|
US20160219217A1 true US20160219217A1 (en) | 2016-07-28 |
US9712751B2 US9712751B2 (en) | 2017-07-18 |
Family
ID=56432979
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/602,832 Active 2035-09-10 US9712751B2 (en) | 2015-01-22 | 2015-01-22 | Camera field of view effects based on device orientation and scene content |
Country Status (1)
Country | Link |
---|---|
US (1) | US9712751B2 (en) |
Cited By (54)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170150097A1 (en) * | 2015-11-20 | 2017-05-25 | Microsoft Technology Licensing, Llc | Communication System |
CN106909224A (en) * | 2017-02-28 | 2017-06-30 | 维沃移动通信有限公司 | The method and mobile terminal of a kind of display image preview interface |
US9813623B2 (en) | 2015-10-30 | 2017-11-07 | Essential Products, Inc. | Wide field of view camera for integration with a mobile device |
US9819865B2 (en) | 2015-10-30 | 2017-11-14 | Essential Products, Inc. | Imaging device and method for generating an undistorted wide view image |
US9906721B2 (en) * | 2015-10-30 | 2018-02-27 | Essential Products, Inc. | Apparatus and method to record a 360 degree image |
US10091458B2 (en) | 2015-11-20 | 2018-10-02 | Microsoft Technology Licensing, Llc | Communication system |
US10114221B1 (en) * | 2017-07-03 | 2018-10-30 | Essential Products, Inc. | Toggling a 360-degree camera to selfie view |
US20180349685A1 (en) * | 2016-09-02 | 2018-12-06 | VeriHelp, Inc. | Identity verification via validated facial recognition and graph database |
US20180352150A1 (en) * | 2017-05-31 | 2018-12-06 | The Procter & Gamble Company | System And Method For Guiding A User To Take A Selfie |
WO2019041163A1 (en) * | 2017-08-30 | 2019-03-07 | 深圳传音通讯有限公司 | Method and system for realising both automatic watermarking and square photography |
WO2019045245A1 (en) * | 2017-08-30 | 2019-03-07 | Samsung Electronics Co., Ltd. | Synchronizing image captures in multiple sensor devices |
US20190122702A1 (en) * | 2016-03-31 | 2019-04-25 | Sony Corporation | Information processing device, information processing method, and computer program |
US20190191087A1 (en) * | 2016-08-19 | 2019-06-20 | Lg Electronics Inc. | Mobile terminal |
US20190199924A1 (en) * | 2016-09-14 | 2019-06-27 | Huawei Technologies Co., Ltd. | Photographing method and terminal |
CN110072058A (en) * | 2019-05-28 | 2019-07-30 | 珠海格力电器股份有限公司 | Image shooting device and method and terminal |
US10400929B2 (en) | 2017-09-27 | 2019-09-03 | Quick Fitting, Inc. | Fitting device, arrangement and method |
US10462345B2 (en) | 2017-08-11 | 2019-10-29 | Essential Products, Inc. | Deformable structure that compensates for displacement of a camera module of a camera accessory |
US10523870B2 (en) * | 2017-12-21 | 2019-12-31 | Elliptic Laboratories As | Contextual display |
US10645294B1 (en) | 2019-05-06 | 2020-05-05 | Apple Inc. | User interfaces for capturing and managing visual media |
US10725722B1 (en) * | 2017-09-08 | 2020-07-28 | Apple Inc. | Modular multiple display electronic devices |
US10818007B2 (en) | 2017-05-31 | 2020-10-27 | The Procter & Gamble Company | Systems and methods for determining apparent skin age |
CN111953919A (en) * | 2019-05-17 | 2020-11-17 | 成都鼎桥通信技术有限公司 | Video recording method and device of handheld terminal in video single call |
AU2020267151B2 (en) * | 2019-05-06 | 2020-12-03 | Apple Inc. | User interfaces for capturing and managing visual media |
US10893217B2 (en) * | 2017-12-28 | 2021-01-12 | Canon Kabushiki Kaisha | Electronic apparatus and method for clipping a range out of a wide field view image |
US10893216B2 (en) | 2017-12-28 | 2021-01-12 | Canon Kabushiki Kaisha | Electronic apparatus and method for controlling same |
JP2021051751A (en) * | 2019-05-06 | 2021-04-01 | アップル インコーポレイテッドApple Inc. | User interfaces for capturing and managing visual media |
US10969047B1 (en) | 2020-01-29 | 2021-04-06 | Quick Fitting Holding Company, Llc | Electrical conduit fitting and assembly |
US11035510B1 (en) | 2020-01-31 | 2021-06-15 | Quick Fitting Holding Company, Llc | Electrical conduit fitting and assembly |
US11054973B1 (en) | 2020-06-01 | 2021-07-06 | Apple Inc. | User interfaces for managing media |
US11055762B2 (en) | 2016-03-21 | 2021-07-06 | The Procter & Gamble Company | Systems and methods for providing customized product recommendations |
US11102414B2 (en) | 2015-04-23 | 2021-08-24 | Apple Inc. | Digital viewfinder user interface for multiple cameras |
US11112964B2 (en) | 2018-02-09 | 2021-09-07 | Apple Inc. | Media capture lock affordance for graphical user interface |
US11128792B2 (en) | 2018-09-28 | 2021-09-21 | Apple Inc. | Capturing and displaying images with multiple focal planes |
US11144099B1 (en) * | 2018-12-28 | 2021-10-12 | Facebook, Inc. | Systems and methods for providing content |
US11165949B2 (en) | 2016-06-12 | 2021-11-02 | Apple Inc. | User interface for capturing photos with different camera magnifications |
US11178335B2 (en) | 2018-05-07 | 2021-11-16 | Apple Inc. | Creative camera |
US11202012B2 (en) * | 2012-04-25 | 2021-12-14 | Sony Corporation | Imaging apparatus and display control method for self-portrait photography |
US11204692B2 (en) | 2017-06-04 | 2021-12-21 | Apple Inc. | User interface camera effects |
US11212449B1 (en) | 2020-09-25 | 2021-12-28 | Apple Inc. | User interfaces for media capture and management |
US20220038621A1 (en) * | 2019-05-17 | 2022-02-03 | Samsung Electronics Co., Ltd. | Device for automatically capturing photo or video about specific moment, and operation method thereof |
WO2022068505A1 (en) * | 2020-09-30 | 2022-04-07 | 华为技术有限公司 | Photographing method and electronic device |
US11321857B2 (en) | 2018-09-28 | 2022-05-03 | Apple Inc. | Displaying and editing images with depth information |
US11350026B1 (en) | 2021-04-30 | 2022-05-31 | Apple Inc. | User interfaces for altering visual media |
US20220321754A1 (en) * | 2021-03-31 | 2022-10-06 | Samsung Electronics Co., Ltd. | Method for measuring illuminance using camera and supporting the same |
US11468625B2 (en) | 2018-09-11 | 2022-10-11 | Apple Inc. | User interfaces for simulated depth effects |
CN115185431A (en) * | 2020-06-01 | 2022-10-14 | 苹果公司 | User interface for managing media |
CN115225806A (en) * | 2021-04-19 | 2022-10-21 | 苹果公司 | Cinematic image framing for wide field of view (FOV) cameras |
EP4117274A1 (en) * | 2017-03-17 | 2023-01-11 | Samsung Electronics Co., Ltd. | Method for providing different indicator for image based on shooting mode and electronic device thereof |
US11706521B2 (en) | 2019-05-06 | 2023-07-18 | Apple Inc. | User interfaces for capturing and managing visual media |
US11722764B2 (en) | 2018-05-07 | 2023-08-08 | Apple Inc. | Creative camera |
US11770601B2 (en) | 2019-05-06 | 2023-09-26 | Apple Inc. | User interfaces for capturing and managing visual media |
US11778339B2 (en) | 2021-04-30 | 2023-10-03 | Apple Inc. | User interfaces for altering visual media |
US20230410324A1 (en) * | 2022-06-21 | 2023-12-21 | Lenovo (Singapore) Pte. Ltd. | Auto-cropping of images based on device motion |
US12112024B2 (en) | 2021-06-01 | 2024-10-08 | Apple Inc. | User interfaces for managing media styles |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180013975A1 (en) * | 2016-07-11 | 2018-01-11 | Kombikam LLC | Multiple device recording process and system |
WO2021175920A1 (en) | 2020-03-06 | 2021-09-10 | Telefonaktiebolaget Lm Ericsson (Publ) | Methods providing video conferencing with adjusted/modified video and related video conferencing nodes |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060064716A1 (en) * | 2000-07-24 | 2006-03-23 | Vivcom, Inc. | Techniques for navigating multiple video streams |
US20060072847A1 (en) * | 2004-10-01 | 2006-04-06 | Microsoft Corporation | System for automatic image cropping based on image saliency |
US7034848B2 (en) * | 2001-01-05 | 2006-04-25 | Hewlett-Packard Development Company, L.P. | System and method for automatically cropping graphical images |
US20070236762A1 (en) * | 2006-03-30 | 2007-10-11 | Fujifilm Corporation | Automatic trimming method, apparatus and program |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8249388B2 (en) | 2008-11-24 | 2012-08-21 | Microsoft Corporation | Identifying portions of an image for cropping |
-
2015
- 2015-01-22 US US14/602,832 patent/US9712751B2/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060064716A1 (en) * | 2000-07-24 | 2006-03-23 | Vivcom, Inc. | Techniques for navigating multiple video streams |
US7034848B2 (en) * | 2001-01-05 | 2006-04-25 | Hewlett-Packard Development Company, L.P. | System and method for automatically cropping graphical images |
US20060072847A1 (en) * | 2004-10-01 | 2006-04-06 | Microsoft Corporation | System for automatic image cropping based on image saliency |
US20070236762A1 (en) * | 2006-03-30 | 2007-10-11 | Fujifilm Corporation | Automatic trimming method, apparatus and program |
Cited By (103)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11202012B2 (en) * | 2012-04-25 | 2021-12-14 | Sony Corporation | Imaging apparatus and display control method for self-portrait photography |
US11490017B2 (en) | 2015-04-23 | 2022-11-01 | Apple Inc. | Digital viewfinder user interface for multiple cameras |
US12149831B2 (en) | 2015-04-23 | 2024-11-19 | Apple Inc. | Digital viewfinder user interface for multiple cameras |
US11711614B2 (en) | 2015-04-23 | 2023-07-25 | Apple Inc. | Digital viewfinder user interface for multiple cameras |
US11102414B2 (en) | 2015-04-23 | 2021-08-24 | Apple Inc. | Digital viewfinder user interface for multiple cameras |
US9819865B2 (en) | 2015-10-30 | 2017-11-14 | Essential Products, Inc. | Imaging device and method for generating an undistorted wide view image |
US9906721B2 (en) * | 2015-10-30 | 2018-02-27 | Essential Products, Inc. | Apparatus and method to record a 360 degree image |
US9813623B2 (en) | 2015-10-30 | 2017-11-07 | Essential Products, Inc. | Wide field of view camera for integration with a mobile device |
US10218904B2 (en) | 2015-10-30 | 2019-02-26 | Essential Products, Inc. | Wide field of view camera for integration with a mobile device |
US10091458B2 (en) | 2015-11-20 | 2018-10-02 | Microsoft Technology Licensing, Llc | Communication system |
US20170150097A1 (en) * | 2015-11-20 | 2017-05-25 | Microsoft Technology Licensing, Llc | Communication System |
US11055762B2 (en) | 2016-03-21 | 2021-07-06 | The Procter & Gamble Company | Systems and methods for providing customized product recommendations |
US10679677B2 (en) * | 2016-03-31 | 2020-06-09 | Sony Corporation | Information processing device and information processing method |
US20190122702A1 (en) * | 2016-03-31 | 2019-04-25 | Sony Corporation | Information processing device, information processing method, and computer program |
US11962889B2 (en) | 2016-06-12 | 2024-04-16 | Apple Inc. | User interface for camera effects |
US12132981B2 (en) | 2016-06-12 | 2024-10-29 | Apple Inc. | User interface for camera effects |
US11641517B2 (en) | 2016-06-12 | 2023-05-02 | Apple Inc. | User interface for camera effects |
US11165949B2 (en) | 2016-06-12 | 2021-11-02 | Apple Inc. | User interface for capturing photos with different camera magnifications |
US11245837B2 (en) | 2016-06-12 | 2022-02-08 | Apple Inc. | User interface for camera effects |
US20190191087A1 (en) * | 2016-08-19 | 2019-06-20 | Lg Electronics Inc. | Mobile terminal |
US10863087B2 (en) * | 2016-08-19 | 2020-12-08 | Lg Electronics Inc. | Mobile terminal selecting normal mode or wide angle camera mode based on movement of mobile terminal |
US20180349685A1 (en) * | 2016-09-02 | 2018-12-06 | VeriHelp, Inc. | Identity verification via validated facial recognition and graph database |
US20190199924A1 (en) * | 2016-09-14 | 2019-06-27 | Huawei Technologies Co., Ltd. | Photographing method and terminal |
US10834318B2 (en) * | 2016-09-14 | 2020-11-10 | Huawei Technologies Co., Ltd. | Automatic photographing method and terminal based on use posture |
CN106909224A (en) * | 2017-02-28 | 2017-06-30 | 维沃移动通信有限公司 | The method and mobile terminal of a kind of display image preview interface |
EP3579080A4 (en) * | 2017-02-28 | 2020-02-05 | Vivo Mobile Communication Co., Ltd. | METHOD FOR DISPLAYING AN IMAGE PRE-INTERFACE AND MOBILE TERMINAL |
US11997382B2 (en) | 2017-03-17 | 2024-05-28 | Samsung Electronics Co., Ltd. | Method for providing different indicator for image based on shooting mode and electronic device thereof |
EP4117274A1 (en) * | 2017-03-17 | 2023-01-11 | Samsung Electronics Co., Ltd. | Method for providing different indicator for image based on shooting mode and electronic device thereof |
US11570359B2 (en) | 2017-03-17 | 2023-01-31 | Samsung Electronics Co., Ltd. | Method for providing different indicator for image based on shooting mode and electronic device thereof |
US10818007B2 (en) | 2017-05-31 | 2020-10-27 | The Procter & Gamble Company | Systems and methods for determining apparent skin age |
US10574883B2 (en) * | 2017-05-31 | 2020-02-25 | The Procter & Gamble Company | System and method for guiding a user to take a selfie |
US20180352150A1 (en) * | 2017-05-31 | 2018-12-06 | The Procter & Gamble Company | System And Method For Guiding A User To Take A Selfie |
US11687224B2 (en) | 2017-06-04 | 2023-06-27 | Apple Inc. | User interface camera effects |
US11204692B2 (en) | 2017-06-04 | 2021-12-21 | Apple Inc. | User interface camera effects |
US10114221B1 (en) * | 2017-07-03 | 2018-10-30 | Essential Products, Inc. | Toggling a 360-degree camera to selfie view |
US10462345B2 (en) | 2017-08-11 | 2019-10-29 | Essential Products, Inc. | Deformable structure that compensates for displacement of a camera module of a camera accessory |
US10863057B2 (en) | 2017-08-30 | 2020-12-08 | Samsung Electronics Co., Ltd. | Synchronizing image captures in multiple sensor devices |
WO2019045245A1 (en) * | 2017-08-30 | 2019-03-07 | Samsung Electronics Co., Ltd. | Synchronizing image captures in multiple sensor devices |
WO2019041163A1 (en) * | 2017-08-30 | 2019-03-07 | 深圳传音通讯有限公司 | Method and system for realising both automatic watermarking and square photography |
US10725722B1 (en) * | 2017-09-08 | 2020-07-28 | Apple Inc. | Modular multiple display electronic devices |
US10400929B2 (en) | 2017-09-27 | 2019-09-03 | Quick Fitting, Inc. | Fitting device, arrangement and method |
NO344671B1 (en) * | 2017-12-21 | 2020-03-02 | Elliptic Laboratories As | Contextual display |
US10523870B2 (en) * | 2017-12-21 | 2019-12-31 | Elliptic Laboratories As | Contextual display |
US10893217B2 (en) * | 2017-12-28 | 2021-01-12 | Canon Kabushiki Kaisha | Electronic apparatus and method for clipping a range out of a wide field view image |
US10893216B2 (en) | 2017-12-28 | 2021-01-12 | Canon Kabushiki Kaisha | Electronic apparatus and method for controlling same |
GB2571431B (en) * | 2017-12-28 | 2021-12-22 | Canon Kk | Electronic apparatus and method for controlling the same |
US11977731B2 (en) | 2018-02-09 | 2024-05-07 | Apple Inc. | Media capture lock affordance for graphical user interface |
US11112964B2 (en) | 2018-02-09 | 2021-09-07 | Apple Inc. | Media capture lock affordance for graphical user interface |
US11178335B2 (en) | 2018-05-07 | 2021-11-16 | Apple Inc. | Creative camera |
US12170834B2 (en) | 2018-05-07 | 2024-12-17 | Apple Inc. | Creative camera |
US11722764B2 (en) | 2018-05-07 | 2023-08-08 | Apple Inc. | Creative camera |
US12154218B2 (en) | 2018-09-11 | 2024-11-26 | Apple Inc. | User interfaces simulated depth effects |
US11468625B2 (en) | 2018-09-11 | 2022-10-11 | Apple Inc. | User interfaces for simulated depth effects |
US11321857B2 (en) | 2018-09-28 | 2022-05-03 | Apple Inc. | Displaying and editing images with depth information |
US11128792B2 (en) | 2018-09-28 | 2021-09-21 | Apple Inc. | Capturing and displaying images with multiple focal planes |
US11895391B2 (en) | 2018-09-28 | 2024-02-06 | Apple Inc. | Capturing and displaying images with multiple focal planes |
US11669985B2 (en) | 2018-09-28 | 2023-06-06 | Apple Inc. | Displaying and editing images with depth information |
US11144099B1 (en) * | 2018-12-28 | 2021-10-12 | Facebook, Inc. | Systems and methods for providing content |
US10652470B1 (en) | 2019-05-06 | 2020-05-12 | Apple Inc. | User interfaces for capturing and managing visual media |
US10791273B1 (en) | 2019-05-06 | 2020-09-29 | Apple Inc. | User interfaces for capturing and managing visual media |
JP2021051751A (en) * | 2019-05-06 | 2021-04-01 | アップル インコーポレイテッドApple Inc. | User interfaces for capturing and managing visual media |
AU2022221466B2 (en) * | 2019-05-06 | 2023-09-14 | Apple Inc. | User interfaces for capturing and managing visual media |
US12192617B2 (en) | 2019-05-06 | 2025-01-07 | Apple Inc. | User interfaces for capturing and managing visual media |
AU2021254567B2 (en) * | 2019-05-06 | 2021-12-02 | Apple Inc. | User interfaces for capturing and managing visual media |
JP7590612B2 (en) | 2019-05-06 | 2024-11-26 | アップル インコーポレイテッド | User interface for capturing and managing visual media |
US10735642B1 (en) | 2019-05-06 | 2020-08-04 | Apple Inc. | User interfaces for capturing and managing visual media |
US11706521B2 (en) | 2019-05-06 | 2023-07-18 | Apple Inc. | User interfaces for capturing and managing visual media |
US10645294B1 (en) | 2019-05-06 | 2020-05-05 | Apple Inc. | User interfaces for capturing and managing visual media |
US10674072B1 (en) * | 2019-05-06 | 2020-06-02 | Apple Inc. | User interfaces for capturing and managing visual media |
AU2020267151B8 (en) * | 2019-05-06 | 2020-12-24 | Apple Inc. | User interfaces for capturing and managing visual media |
US10681282B1 (en) * | 2019-05-06 | 2020-06-09 | Apple Inc. | User interfaces for capturing and managing visual media |
AU2020267151B2 (en) * | 2019-05-06 | 2020-12-03 | Apple Inc. | User interfaces for capturing and managing visual media |
US10735643B1 (en) | 2019-05-06 | 2020-08-04 | Apple Inc. | User interfaces for capturing and managing visual media |
US11223771B2 (en) * | 2019-05-06 | 2022-01-11 | Apple Inc. | User interfaces for capturing and managing visual media |
US11770601B2 (en) | 2019-05-06 | 2023-09-26 | Apple Inc. | User interfaces for capturing and managing visual media |
CN111953919A (en) * | 2019-05-17 | 2020-11-17 | 成都鼎桥通信技术有限公司 | Video recording method and device of handheld terminal in video single call |
US20220038621A1 (en) * | 2019-05-17 | 2022-02-03 | Samsung Electronics Co., Ltd. | Device for automatically capturing photo or video about specific moment, and operation method thereof |
US11977981B2 (en) * | 2019-05-17 | 2024-05-07 | Samsung Electronics Co., Ltd. | Device for automatically capturing photo or video about specific moment, and operation method thereof |
CN110072058A (en) * | 2019-05-28 | 2019-07-30 | 珠海格力电器股份有限公司 | Image shooting device and method and terminal |
US10969047B1 (en) | 2020-01-29 | 2021-04-06 | Quick Fitting Holding Company, Llc | Electrical conduit fitting and assembly |
US11035510B1 (en) | 2020-01-31 | 2021-06-15 | Quick Fitting Holding Company, Llc | Electrical conduit fitting and assembly |
CN115185431A (en) * | 2020-06-01 | 2022-10-14 | 苹果公司 | User interface for managing media |
US11054973B1 (en) | 2020-06-01 | 2021-07-06 | Apple Inc. | User interfaces for managing media |
US11617022B2 (en) | 2020-06-01 | 2023-03-28 | Apple Inc. | User interfaces for managing media |
US12081862B2 (en) | 2020-06-01 | 2024-09-03 | Apple Inc. | User interfaces for managing media |
US11330184B2 (en) | 2020-06-01 | 2022-05-10 | Apple Inc. | User interfaces for managing media |
US11212449B1 (en) | 2020-09-25 | 2021-12-28 | Apple Inc. | User interfaces for media capture and management |
US12155925B2 (en) | 2020-09-25 | 2024-11-26 | Apple Inc. | User interfaces for media capture and management |
EP4210322A4 (en) * | 2020-09-30 | 2024-01-24 | Huawei Technologies Co., Ltd. | PHOTOGRAPHY METHOD AND ELECTRONIC DEVICE |
WO2022068505A1 (en) * | 2020-09-30 | 2022-04-07 | 华为技术有限公司 | Photographing method and electronic device |
US20220321754A1 (en) * | 2021-03-31 | 2022-10-06 | Samsung Electronics Co., Ltd. | Method for measuring illuminance using camera and supporting the same |
US12022200B2 (en) * | 2021-03-31 | 2024-06-25 | Samsung Electronics Co., Ltd. | Method for measuring illuminance using camera and supporting the same |
CN115225806A (en) * | 2021-04-19 | 2022-10-21 | 苹果公司 | Cinematic image framing for wide field of view (FOV) cameras |
US12106534B2 (en) | 2021-04-19 | 2024-10-01 | Apple Inc. | Cinematic image framing for wide field of view (FOV) cameras |
US11539876B2 (en) | 2021-04-30 | 2022-12-27 | Apple Inc. | User interfaces for altering visual media |
US12101567B2 (en) | 2021-04-30 | 2024-09-24 | Apple Inc. | User interfaces for altering visual media |
US11778339B2 (en) | 2021-04-30 | 2023-10-03 | Apple Inc. | User interfaces for altering visual media |
US11416134B1 (en) | 2021-04-30 | 2022-08-16 | Apple Inc. | User interfaces for altering visual media |
US11418699B1 (en) | 2021-04-30 | 2022-08-16 | Apple Inc. | User interfaces for altering visual media |
US11350026B1 (en) | 2021-04-30 | 2022-05-31 | Apple Inc. | User interfaces for altering visual media |
US12112024B2 (en) | 2021-06-01 | 2024-10-08 | Apple Inc. | User interfaces for managing media styles |
US20230410324A1 (en) * | 2022-06-21 | 2023-12-21 | Lenovo (Singapore) Pte. Ltd. | Auto-cropping of images based on device motion |
US12190528B2 (en) * | 2022-06-21 | 2025-01-07 | Lenovo (Singapore) Pte. Ltd. | Auto-cropping of images based on device motion |
Also Published As
Publication number | Publication date |
---|---|
US9712751B2 (en) | 2017-07-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9712751B2 (en) | Camera field of view effects based on device orientation and scene content | |
CN103108120B (en) | Zoom control method and apparatus | |
JP5136669B2 (en) | Image processing apparatus, image processing method, and program | |
US9578260B2 (en) | Digital photographing apparatus and method of controlling the digital photographing apparatus | |
JP5567235B2 (en) | Image processing apparatus, photographing apparatus, program, and image processing method | |
US8786749B2 (en) | Digital photographing apparatus for displaying an icon corresponding to a subject feature and method of controlling the same | |
US9549126B2 (en) | Digital photographing apparatus and control method thereof | |
US20100208107A1 (en) | Imaging device and imaging device control method | |
US10158798B2 (en) | Imaging apparatus and method of controlling the same | |
US20100053363A1 (en) | Photographing method and apparatus | |
WO2017124899A1 (en) | Information processing method, apparatus and electronic device | |
US9111129B2 (en) | Subject detecting method and apparatus, and digital photographing apparatus | |
EP3945494A1 (en) | Video processing method, apparatus and storage medium | |
JP6096654B2 (en) | Image recording method, electronic device, and computer program | |
KR100926133B1 (en) | Method and apparatus for producing and taking digital contents | |
CN103888684A (en) | Image processing apparatus that combines a plurality of images | |
US9167150B2 (en) | Apparatus and method for processing image in mobile terminal having camera | |
US9001254B2 (en) | Digital photographing apparatus and method of controlling the same | |
WO2018196854A1 (en) | Photographing method, photographing apparatus and mobile terminal | |
CN115225806B (en) | Cinematic image framing for wide field of view (FOV) cameras | |
JP5370555B2 (en) | Imaging apparatus, imaging method, and program | |
CN109167921B (en) | Shooting method, shooting device, shooting terminal and storage medium | |
CN116803094A (en) | Image processing apparatus, control method, and program | |
US20250150654A1 (en) | Framework for Simultaneous Subject and Desk Capture During Videoconferencing | |
US8902345B2 (en) | Method and apparatus for synthesizing two or more images together with one another via divided display screen |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: APPLE INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:WILLIAMS, KENTA K.;MOTTA, RICARDO;REEL/FRAME:034790/0666 Effective date: 20141031 |
|
FEPP | Fee payment procedure |
Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |