US20070230944A1 - Plenoptic camera - Google Patents
Plenoptic camera Download PDFInfo
- Publication number
- US20070230944A1 US20070230944A1 US11/398,403 US39840306A US2007230944A1 US 20070230944 A1 US20070230944 A1 US 20070230944A1 US 39840306 A US39840306 A US 39840306A US 2007230944 A1 US2007230944 A1 US 2007230944A1
- Authority
- US
- United States
- Prior art keywords
- array
- object field
- camera
- lens
- image
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/0075—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00 with means for altering, e.g. increasing, the depth of field or depth of focus
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/95—Computational photography systems, e.g. light-field imaging systems
- H04N23/957—Light-field or plenoptic cameras or camera modules
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/95—Computational photography systems, e.g. light-field imaging systems
- H04N23/958—Computational photography systems, e.g. light-field imaging systems for extended depth of field imaging
- H04N23/959—Computational photography systems, e.g. light-field imaging systems for extended depth of field imaging by adjusting depth of field during image capture, e.g. maximising or setting range based on scene characteristics
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B3/00—Simple or compound lenses
- G02B3/0006—Arrays
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B3/00—Simple or compound lenses
- G02B3/0006—Arrays
- G02B3/0037—Arrays characterized by the distribution or form of lenses
- G02B3/0056—Arrays characterized by the distribution or form of lenses arranged along two different directions in a plane, e.g. honeycomb arrangement of lenses
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B5/00—Optical elements other than lenses
- G02B5/04—Prisms
- G02B5/045—Prism arrays
Definitions
- the present invention relates to cameras. More specifically, the present invention relates the design of a “plenoptic” camera, which captures information about the direction distribution of light rays entering the camera.
- a conventional camera does not capture information about the location of the aperture of the light rays entering the camera.
- a conventional digital camera captures a two-dimensional (2D) image representing a total amount of light which strikes each point on a photosensor within the camera.
- this 2D image contains no information about the directional distribution of the light that strikes the photosensor. This directional information at the pixels corresponds to locational information at the aperture.
- a “plenoptic” camera samples the four-dimensional (4D) optical phase space or light field and in doing so captures information about the directional distribution of the light rays.
- 4D four-dimensional
- a “plenoptic” camera samples the four-dimensional (4D) optical phase space or light field and in doing so captures information about the directional distribution of the light rays.
- Ng05 Ng, R., Levoy, M., Bredif, M., Duval, G., Horowitz, M. and Hanrahan, P., “Light Field Photography with a Hand-Held Plenoptic Camera,” Stanford University Computer Science Tech Report CSTR 2005-02, April 2005.
- the system described in [Ng05] uses a microlens array 106 comprised of about 100,000 lenslets which is placed a small distance (0.5 mm) from a CCD array 108 .
- Each lenslet splits a beam coming to it from the main lens 104 into (100) rays coming from different “pinhole” locations on the aperture of the main lens 108 .
- Each of these rays is recorded as a pixel, and the pixels under each lenslet collectively form a 100 -pixel image. If we call this 100 -pixel image a “macropixel,” then the plenoptic photograph captured by this camera will contain approximately 100,000 macropixels.
- a 16-megapixel sensor is used with an approximately 100,000 lenslet array to create a final output of approximately 300 ⁇ 300 macropixels, with one macropixel per lenslet.
- the macropixel created by each lenslet comprises approximately 150 pixels.
- only about 100 of these pixels are useful because of poor quality of edge pixels caused by a problem which is referred to as “vignetting.”
- These 100 pixels which comprise each macropixel make the captured data equivalent to 100 conventional images, one for each choice of the pixel inside a macropixel.
- the size of each picture produced by processing data from this camera is equal to the number of lenslets, and is hence 300 ⁇ 300.
- One embodiment of the present invention provides a plenoptic camera which captures information about the direction distribution of light rays entering the camera.
- this plenoptic camera includes a main lens which receives light from objects in an object field and directs the received light onto an image plane of the camera. It also includes a photodetector array located at the image plane of the camera, which captures the received light to produce an image.
- the plenoptic camera additionally includes an array of optical elements located between the object field and the main lens. Each optical element in this array receives light from the object field from a different angle than the other optical elements in the array, and consequently directs a different view of the object field into the main lens. In this way, the photodetector array receives a different view of the object field from each optical element in the array.
- a given optical element in the array of optical elements includes: a lens; a prism; or a lens and a prism.
- the lens is a negative lens with a negative focal length.
- the lens is an achromatic lens
- the prism is an achromatic prism
- the photodetector array is a Charge-Coupled Device (CCD) array.
- CCD Charge-Coupled Device
- One embodiment or the present invention additionally includes a processing mechanism configured to process the different views of the object field received by the photodetector array to produce a final image.
- the processing mechanism while producing the final image, is configured to use the different views of the object field to adjust one or more of the following: a plane-of-focus for the final image; a viewing angle for the final image; or a depth-of-field for the final image.
- the processing mechanism while processing the different views of the object field, is configured to perform view-morphing or interpolation operations between the different views to produce additional views of the object field which appear to be gathered from locations between the locations of the optical elements in the array.
- FIG. 1A illustrates a prior art plenoptic camera.
- FIG. 1B illustrates a plenoptic camera in accordance with an embodiment of the present invention.
- FIG. 2 illustrates a layout for the additional lenses and prisms in a plenoptic camera in accordance with an embodiment of the present invention.
- FIG. 3 illustrates an array of lenses and prisms in accordance with an embodiment of the present invention.
- FIG. 4 illustrates an array of lenses and an array of prisms in accordance with an embodiment of the present invention.
- FIG. 5 illustrates an exemplary scene in accordance with an embodiment of the present invention.
- FIG. 6 presents images of the exemplary scene taken through an array of lenses in accordance with an embodiment of the present invention.
- FIG. 7 presents images of the exemplary scene taken through an array of prisms in accordance with an embodiment of the present invention.
- FIG. 8 presents images of the exemplary scene taken through both the array of lenses and the array of prisms in accordance with an embodiment of the present invention.
- FIG. 9A illustrates an image of the exemplary scene which is generated so that both the foreground and background are in-focus in accordance with an embodiment of the present invention.
- FIG. 9B illustrates an image of the exemplary scene which is generated with the foreground in-focus in accordance with an embodiment of the present invention.
- FIG. 9C illustrates an image of the exemplary scene which is generated with the background in-focus in accordance with an embodiment of the present invention.
- FIG. 10 presents a flow chart illustrating how light is directed within a plenoptic camera in accordance with an embodiment of the present invention.
- a computer-readable storage medium which may be any device or medium that can store code and/or data for use by a computer system. This includes, but is not limited to, magnetic and optical storage devices such as disk drives, magnetic tape, CDs (compact discs), DVDs (digital versatile discs or digital video discs), or any device capable of storing data usable by a computer system.
- the size of each image for each macropixel is limited by the design of the camera.
- the present invention provides a new, easier-to-construct camera, which gives the designer more flexibility in the trade off between the number of views and the size of each view image.
- one embodiment of the present invention can capture a small number (10) of high-resolution images, whereas the system described in [Ng05] captures a large number (100) of low-resolution images.
- one embodiment of the present invention provides a design in which a small number of conventional lenses are placed in front of the main lens of the camera, instead of placing a much larger number of microlenses behind the main lens as is disclosed in [Ng05], thereby making plenoptic cameras much easier to build.
- the present invention makes it possible to reduce the number of effective images captured (from 100 in the prior art, down to 20, or even to 10) without loss of quality at the edge pixels. This is a major problem for the system disclosed in [Ng05], which limits image resolution to 300 ⁇ 300 macropixels. Using the same type of optical sensor, the present invention can achieve a significantly higher resolution for the final image (for example, five times the resolution). In other words, the present invention provides the flexibility to trade detail in the 3D information for detail in the 2D images. For simple scenes, composed of a few surfaces in 3D, a few images are sufficient to capture the complete 3D detail.
- the present invention places the array of lenses 114 in front of the main lens 116 as is illustrated in FIG. 1B . More specifically, one embodiment of the present invention achieves higher-resolution results by placing 19 lenses/prisms in front of the main lens instead of placing 90,000 lenses behind the main lens.
- an array of (about 10 to 100) lenses 114 and prisms 112 is placed 200 to 500 mm in front of the main lens 116 of a conventional camera. Note that the ratio of (width of array of lenses)/(distance to main lens) is ideally equal to the ⁇ -number of main lens 116 .
- Each lens is coupled with a corresponding achromatic prism, wherein the prism has a different angle for different lenses, depending on location of the lens.
- each prism is chosen to produce an angular deviation which is equal to the angle at which the main camera lens sees that prism. In this way, all prisms create images of the same object from the scene. Note that it is not essential to achieve precision with these prism angles and the arrangement of the prisms because small errors do not influence the final image quality. However, big errors should be avoided to ensure that pixels are not wasted due to random shifts of images which create gaps and/or overlaps.
- all of the lenses have the same negative focal length, for example ⁇ 100 mm. This focal length controls the field of view. Note that it is important for all lenses to have the same focal length if we want good focusing.
- the main lens 116 of the camera is focused on an array of virtual images which appear in front of the negative lenses. Note that each lens/prism in the array receives light from image field 110 from a different angle than the other lenses/prisms in the array, and consequently directs a different view of the image field 110 into the main lens 116 of the camera. In this way, CCD array 118 captures an array of pictures in which each picture provides a different view of the object field from a different lens/prism in the array.
- the array of pictures captured by CCD array 118 is processed by processing device 120 to produce a final image 122 .
- processing device 120 can be integrated into the camera or can be located outside of the camera.
- processing device 120 can achieve various “light-field” effects, such as refocusing an image, reducing noise, adjusting the viewing angle, and adjusting the depth-of-field for the final image.
- light-field effects such as refocusing an image, reducing noise, adjusting the viewing angle, and adjusting the depth-of-field for the final image.
- processing device 120 is additionally configured to perform view-morphing or interpolation operations between the different views to produce additional views of the object field which appear to be gathered from locations between the locations of the lenses/prisms in the array.
- view-morphing or interpolation operations between the different views to produce additional views of the object field which appear to be gathered from locations between the locations of the lenses/prisms in the array.
- the present invention can produce a large number of images (100) using a smaller number of lenses (20).
- the central lens has no prism because it is located on the main axis of the camera.
- the main camera lens has ⁇ -number ⁇ /2, which corresponds to 14 degrees.
- the prisms are chosen with deviation angles of 4 degrees, 7 degrees and 8 degrees as is illustrated in FIG. 2 .
- the lenses have diameter 25 mm and the total width of the array of lenses 114 is 125 mm.
- the array is positioned at distance of 250 mm from main lens 116 . (Note that this distance can be adjusted.) With a 16-megapixel CCD array, this embodiment is able to capture final images of about 600 ⁇ 600 pixels, which is 4 times better than the camera described in [Ng05] for the same camera resolution.
- a tube which looks like a telephoto lens can extend from the main lens to the array of lenses to prevent light from entering the system sideways and forming reflection spots on the prisms and lenses.
- One embodiment of the present invention can operate using prisms only and using lenses only, but it is preferable to use both lenses and prisms. To illustrate this, a number of pictures have been taken through the array of 7 negative lenses and the corresponding array of 6 prisms illustrated in FIG. 4 . These lenses and prisms are used to capture images of an exemplary scene which appears in FIG. 5 .
- FIG. 6 illustrates images of the exemplary scene which are taken through the array of lenses only. Note that these images are shifted with respect to each other and do not capture identical areas of the scene, although there is a small area of the scene near the cap of the tube which appears in all of the images.
- FIG. 7 presents images of the exemplary scene which are taken through the array of prisms only. Note that these prisms shift the images so the same part of the scene is captured in each image. However, the resulting field of view is quite narrow.
- FIG. 8 presents images of the exemplary scene which are taken through both the array of lenses and the array of prisms. Note that the prisms shift the images so that all the images are centered and the lenses expand the field of view. Also note that each two images form a stereo pair.
- the plane where the image is formed is further away from the camera. This makes the resulting system more compact, because it allows the array of lenses to be closer to the main lens.
- FIGS. 9A-9C illustrates how one embodiment of the system can virtually focus on different image planes after a picture has been taken in accordance with an embodiment of the present invention.
- the depth-of-field of the image is large, so both the bottle in the foreground and the tube in background are in-focus.
- the depth-of-field is reduced and the focal plane of the image is set to be nearer to the camera, so the bottle in the foreground is in-focus, while the tube in the background in out-of-focus.
- the focal plane of the image is set to be farther from the camera, so the tube in the background is in-focus, while the tube in the foreground is out-of-focus.
- FIG. 10 presents a flow chart illustrating how light is directed within a plenoptic camera in accordance with an embodiment of the present invention.
- light is received from objects in an object field at an array of optical elements located between the object field and the main lens of the camera (step 1002 ).
- Each optical element in this array receives light from the object field from a different angle, and consequently directs a different view of the object field into the main lens.
- step 1004 light is received from the array of optical elements at the main lens which directs the received light onto an image plane of the camera.
- step 1006 light is received from the main lens at a photodetector array located at the image place of the camera (step 1006 ), wherein the photodetector array receives a different view of the object field from each optical element in the array.
Abstract
Description
- 1. Field of the Invention
- The present invention relates to cameras. More specifically, the present invention relates the design of a “plenoptic” camera, which captures information about the direction distribution of light rays entering the camera.
- 2. Related Art
- Conventional cameras fail to capture a large amount of optical information. In particular, a conventional camera does not capture information about the location of the aperture of the light rays entering the camera. During operation, a conventional digital camera captures a two-dimensional (2D) image representing a total amount of light which strikes each point on a photosensor within the camera. However, this 2D image contains no information about the directional distribution of the light that strikes the photosensor. This directional information at the pixels corresponds to locational information at the aperture.
- In contrast, a “plenoptic” camera samples the four-dimensional (4D) optical phase space or light field and in doing so captures information about the directional distribution of the light rays. For example, see [Adelson92] Adelson, T., and Wang, J. Y. A. 1992, “Single lens stereo with a plenoptic camera,” IEEE Transactions on Pattern Analysis and Machine Intelligence 14, 2, February 1992, pp. 99-106. Also see [Ng05] Ng, R., Levoy, M., Bredif, M., Duval, G., Horowitz, M. and Hanrahan, P., “Light Field Photography with a Hand-Held Plenoptic Camera,” Stanford University Computer Science Tech Report CSTR 2005-02, April 2005. These papers describe plenoptic/light-field camera designs based on modifications to a conventional digital camera.
- Referring to
FIG. 1A , the system described in [Ng05] uses amicrolens array 106 comprised of about 100,000 lenslets which is placed a small distance (0.5 mm) from aCCD array 108. Each lenslet splits a beam coming to it from themain lens 104 into (100) rays coming from different “pinhole” locations on the aperture of themain lens 108. Each of these rays is recorded as a pixel, and the pixels under each lenslet collectively form a 100-pixel image. If we call this 100-pixel image a “macropixel,” then the plenoptic photograph captured by this camera will contain approximately 100,000 macropixels. By appropriately selecting a pixel from each macropixel, we can create conventional pictures taken with a virtual pinhole camera. Moreover, by mixing such images appropriately, we can refocus images originally taken out-of-focus, reduce noise, or achieve other “light-field” effects, as described in the papers above. - In the prototype described in [Ng05], a 16-megapixel sensor is used with an approximately 100,000 lenslet array to create a final output of approximately 300×300 macropixels, with one macropixel per lenslet. The macropixel created by each lenslet comprises approximately 150 pixels. However, only about 100 of these pixels are useful because of poor quality of edge pixels caused by a problem which is referred to as “vignetting.” These 100 pixels which comprise each macropixel make the captured data equivalent to 100 conventional images, one for each choice of the pixel inside a macropixel. The size of each picture produced by processing data from this camera is equal to the number of lenslets, and is hence 300×300.
- Unfortunately, an image with only 300×300 pixels has insufficient resolution for most practical uses. The number of pixels can be increased by increasing the number of lenslets and making them smaller. Unfortunately, the prior art cannot use the border pixels of each image. Note that a band of about 2 to 4 pixels along the border of the macropixel is lost depending upon whether the system is working with a Grayscale pattern or a Bayer pattern. When the image is small, these few border pixels comprise a large percentage of the image. For example, in a 10×10 color image, 4 pixels on each edge may be lost leaving only 2×2=4 central pixels. In this case, 96% of the information lost! Because of this problem, the system described in [Ng05] cannot reduce the size of each microlens and the image under it. Consequently, the number of microlenses, and hence the resolution of the image, is limited. (Currently, in a system that uses a 16-megapixel sensor, the number of microlenses is limited to less than 100,000.)
- Hence, what is needed is a method and an apparatus for increasing the resolution of a plenoptic camera without the above-described problems.
- One embodiment of the present invention provides a plenoptic camera which captures information about the direction distribution of light rays entering the camera. Like a conventional camera, this plenoptic camera includes a main lens which receives light from objects in an object field and directs the received light onto an image plane of the camera. It also includes a photodetector array located at the image plane of the camera, which captures the received light to produce an image. However, unlike a conventional camera, the plenoptic camera additionally includes an array of optical elements located between the object field and the main lens. Each optical element in this array receives light from the object field from a different angle than the other optical elements in the array, and consequently directs a different view of the object field into the main lens. In this way, the photodetector array receives a different view of the object field from each optical element in the array.
- In a variation on this embodiment, a given optical element in the array of optical elements includes: a lens; a prism; or a lens and a prism.
- In a further variation, the lens is a negative lens with a negative focal length.
- In a further variation, the lens is an achromatic lens, and the prism is an achromatic prism.
- In a variation on this embodiment, the photodetector array is a Charge-Coupled Device (CCD) array.
- One embodiment or the present invention additionally includes a processing mechanism configured to process the different views of the object field received by the photodetector array to produce a final image.
- In a further variation, while producing the final image, the processing mechanism is configured to use the different views of the object field to adjust one or more of the following: a plane-of-focus for the final image; a viewing angle for the final image; or a depth-of-field for the final image.
- In a further variation, while processing the different views of the object field, the processing mechanism is configured to perform view-morphing or interpolation operations between the different views to produce additional views of the object field which appear to be gathered from locations between the locations of the optical elements in the array.
-
FIG. 1A illustrates a prior art plenoptic camera. -
FIG. 1B illustrates a plenoptic camera in accordance with an embodiment of the present invention. -
FIG. 2 illustrates a layout for the additional lenses and prisms in a plenoptic camera in accordance with an embodiment of the present invention. -
FIG. 3 illustrates an array of lenses and prisms in accordance with an embodiment of the present invention. -
FIG. 4 illustrates an array of lenses and an array of prisms in accordance with an embodiment of the present invention. -
FIG. 5 illustrates an exemplary scene in accordance with an embodiment of the present invention. -
FIG. 6 presents images of the exemplary scene taken through an array of lenses in accordance with an embodiment of the present invention. -
FIG. 7 presents images of the exemplary scene taken through an array of prisms in accordance with an embodiment of the present invention. -
FIG. 8 presents images of the exemplary scene taken through both the array of lenses and the array of prisms in accordance with an embodiment of the present invention. -
FIG. 9A illustrates an image of the exemplary scene which is generated so that both the foreground and background are in-focus in accordance with an embodiment of the present invention. -
FIG. 9B illustrates an image of the exemplary scene which is generated with the foreground in-focus in accordance with an embodiment of the present invention. -
FIG. 9C illustrates an image of the exemplary scene which is generated with the background in-focus in accordance with an embodiment of the present invention. -
FIG. 10 presents a flow chart illustrating how light is directed within a plenoptic camera in accordance with an embodiment of the present invention. - The following description is presented to enable any person skilled in the art to make and use the invention, and is provided in the context of a particular application and its requirements. Various modifications to the disclosed embodiments will be readily apparent to those skilled in the art, and the general principles defined herein may be applied to other embodiments and applications without departing from the spirit and scope of the present invention. Thus, the present invention is not limited to the embodiments shown, but is to be accorded the widest scope consistent with the claims.
- The data structures and code described in this detailed description are typically stored on a computer-readable storage medium, which may be any device or medium that can store code and/or data for use by a computer system. This includes, but is not limited to, magnetic and optical storage devices such as disk drives, magnetic tape, CDs (compact discs), DVDs (digital versatile discs or digital video discs), or any device capable of storing data usable by a computer system.
- Overview
- In the system described in [Ng05], the size of each image for each macropixel is limited by the design of the camera. The present invention provides a new, easier-to-construct camera, which gives the designer more flexibility in the trade off between the number of views and the size of each view image. In particular, one embodiment of the present invention can capture a small number (10) of high-resolution images, whereas the system described in [Ng05] captures a large number (100) of low-resolution images.
- Hence, one embodiment of the present invention provides a design in which a small number of conventional lenses are placed in front of the main lens of the camera, instead of placing a much larger number of microlenses behind the main lens as is disclosed in [Ng05], thereby making plenoptic cameras much easier to build.
- Note that the present invention makes it possible to reduce the number of effective images captured (from 100 in the prior art, down to 20, or even to 10) without loss of quality at the edge pixels. This is a major problem for the system disclosed in [Ng05], which limits image resolution to 300×300 macropixels. Using the same type of optical sensor, the present invention can achieve a significantly higher resolution for the final image (for example, five times the resolution). In other words, the present invention provides the flexibility to trade detail in the 3D information for detail in the 2D images. For simple scenes, composed of a few surfaces in 3D, a few images are sufficient to capture the complete 3D detail. (For example, the human visual system uses only two images.) This speaks in favor of reducing the number of images, because 100 images is probably excessive, and 10 or 20 images is probably more appropriate. Embodiments of the present invention makes this possible, while it was not possible in the prior art.
- Instead of locating the array of lenses behind the main lens, as in the prior art system illustrated in
FIG. 1A , the present invention places the array oflenses 114 in front of themain lens 116 as is illustrated inFIG. 1B . More specifically, one embodiment of the present invention achieves higher-resolution results by placing 19 lenses/prisms in front of the main lens instead of placing 90,000 lenses behind the main lens. - In the embodiment of the present invention illustrated in
FIG. 1B , an array of (about 10 to 100)lenses 114 andprisms 112 is placed 200 to 500 mm in front of themain lens 116 of a conventional camera. Note that the ratio of (width of array of lenses)/(distance to main lens) is ideally equal to the ƒ-number ofmain lens 116. - Each lens is coupled with a corresponding achromatic prism, wherein the prism has a different angle for different lenses, depending on location of the lens. In particular, each prism is chosen to produce an angular deviation which is equal to the angle at which the main camera lens sees that prism. In this way, all prisms create images of the same object from the scene. Note that it is not essential to achieve precision with these prism angles and the arrangement of the prisms because small errors do not influence the final image quality. However, big errors should be avoided to ensure that pixels are not wasted due to random shifts of images which create gaps and/or overlaps.
- In one embodiment of the present invention, all of the lenses have the same negative focal length, for example −100 mm. This focal length controls the field of view. Note that it is important for all lenses to have the same focal length if we want good focusing.
- The
main lens 116 of the camera is focused on an array of virtual images which appear in front of the negative lenses. Note that each lens/prism in the array receives light fromimage field 110 from a different angle than the other lenses/prisms in the array, and consequently directs a different view of theimage field 110 into themain lens 116 of the camera. In this way,CCD array 118 captures an array of pictures in which each picture provides a different view of the object field from a different lens/prism in the array. - The array of pictures captured by
CCD array 118 is processed by processingdevice 120 to produce afinal image 122. (Note thatprocessing device 120 can be integrated into the camera or can be located outside of the camera.) By mixing these images appropriately,processing device 120 can achieve various “light-field” effects, such as refocusing an image, reducing noise, adjusting the viewing angle, and adjusting the depth-of-field for the final image. (For a description of the details of some of these light-field effects, please see [Adelson92] and [Ng05], also see U.S. Pat. No. 5,076,687, entitled “Optical Ranging Apparatus,” by inventor Edward H. Adelson.) - In one embodiment of the present invention,
processing device 120 is additionally configured to perform view-morphing or interpolation operations between the different views to produce additional views of the object field which appear to be gathered from locations between the locations of the lenses/prisms in the array. In this way, the present invention can produce a large number of images (100) using a smaller number of lenses (20). (This type of view-morphing operation is described in U.S. Pat. No. 6,351,269, entitled “Multiple Image Morphing,” by inventor Todor Georgiev.) - Note that generating these additional views greatly opens up the design space because the resulting system generates a large number of “high-resolution” images. This is an improvement over the system described in [Ng05], which captures a large number of “low-resolution” images.
- In an exemplary embodiment of the present invention, the array of
lenses 114 contains 19 lenses (with ƒ=−100 mm), and 18 prisms which are arranged in a hexagonal pattern as is illustrated inFIGS. 2 and 3 . Note that the central lens has no prism because it is located on the main axis of the camera. The main camera lens has ƒ-number ƒ/2, which corresponds to 14 degrees. To accommodate this ƒ-number, the prisms are chosen with deviation angles of 4 degrees, 7 degrees and 8 degrees as is illustrated inFIG. 2 . In the exemplary embodiment, the lenses have diameter 25 mm and the total width of the array oflenses 114 is 125 mm. Moreover, the array is positioned at distance of 250 mm frommain lens 116. (Note that this distance can be adjusted.) With a 16-megapixel CCD array, this embodiment is able to capture final images of about 600×600 pixels, which is 4 times better than the camera described in [Ng05] for the same camera resolution. - Note that a tube, which looks like a telephoto lens can extend from the main lens to the array of lenses to prevent light from entering the system sideways and forming reflection spots on the prisms and lenses.
- Lenses and Prisms
- One embodiment of the present invention can operate using prisms only and using lenses only, but it is preferable to use both lenses and prisms. To illustrate this, a number of pictures have been taken through the array of 7 negative lenses and the corresponding array of 6 prisms illustrated in
FIG. 4 . These lenses and prisms are used to capture images of an exemplary scene which appears inFIG. 5 . -
FIG. 6 illustrates images of the exemplary scene which are taken through the array of lenses only. Note that these images are shifted with respect to each other and do not capture identical areas of the scene, although there is a small area of the scene near the cap of the tube which appears in all of the images. -
FIG. 7 presents images of the exemplary scene which are taken through the array of prisms only. Note that these prisms shift the images so the same part of the scene is captured in each image. However, the resulting field of view is quite narrow. - Finally,
FIG. 8 presents images of the exemplary scene which are taken through both the array of lenses and the array of prisms. Note that the prisms shift the images so that all the images are centered and the lenses expand the field of view. Also note that each two images form a stereo pair. - By using a negative lens instead of a positive lens, the plane where the image is formed is further away from the camera. This makes the resulting system more compact, because it allows the array of lenses to be closer to the main lens.
- Generating a Resulting Image
- As mentioned above, the present invention can achieve various “light-field” effects, such as refocusing, reducing noise, adjusting the viewing angle, and adjusting the depth-of-field for the image. For example,
FIGS. 9A-9C illustrates how one embodiment of the system can virtually focus on different image planes after a picture has been taken in accordance with an embodiment of the present invention. InFIG. 9A , the depth-of-field of the image is large, so both the bottle in the foreground and the tube in background are in-focus. InFIG. 9B , the depth-of-field is reduced and the focal plane of the image is set to be nearer to the camera, so the bottle in the foreground is in-focus, while the tube in the background in out-of-focus. InFIG. 9C , the focal plane of the image is set to be farther from the camera, so the tube in the background is in-focus, while the tube in the foreground is out-of-focus. - Light Flow
-
FIG. 10 presents a flow chart illustrating how light is directed within a plenoptic camera in accordance with an embodiment of the present invention. First, light is received from objects in an object field at an array of optical elements located between the object field and the main lens of the camera (step 1002). Each optical element in this array receives light from the object field from a different angle, and consequently directs a different view of the object field into the main lens. - Next, light is received from the array of optical elements at the main lens which directs the received light onto an image plane of the camera (step 1004).
- Then, light is received from the main lens at a photodetector array located at the image place of the camera (step 1006), wherein the photodetector array receives a different view of the object field from each optical element in the array.
- Finally, the different views of the object field which are received by the photodetector array are processed to produce a final image (step 1008).
- The foregoing descriptions of embodiments of the present invention have been presented only for purposes of illustration and description: They are not intended to be exhaustive or to limit the present invention to the forms disclosed. Accordingly, many modifications and variations will be apparent to practitioners skilled in the art. Additionally, the above disclosure is not intended to limit the present invention. The scope of the present invention is defined by the appended claims.
Claims (20)
Priority Applications (7)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/398,403 US7620309B2 (en) | 2006-04-04 | 2006-04-04 | Plenoptic camera |
PCT/US2007/065855 WO2007115281A1 (en) | 2006-04-04 | 2007-04-03 | Improved plenoptic camera |
EP07760021.1A EP2008445B1 (en) | 2006-04-04 | 2007-04-03 | Improved plenoptic camera |
KR1020087027050A KR20090016453A (en) | 2006-04-04 | 2007-04-03 | Improved plenoptic camera |
CN2007800190343A CN101455071B (en) | 2006-04-04 | 2007-04-03 | Improved plenoptic camera |
JP2009504414A JP4981124B2 (en) | 2006-04-04 | 2007-04-03 | Improved plenoptic camera |
US12/574,183 US8238738B2 (en) | 2006-04-04 | 2009-10-06 | Plenoptic camera |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/398,403 US7620309B2 (en) | 2006-04-04 | 2006-04-04 | Plenoptic camera |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/574,183 Continuation US8238738B2 (en) | 2006-04-04 | 2009-10-06 | Plenoptic camera |
Publications (2)
Publication Number | Publication Date |
---|---|
US20070230944A1 true US20070230944A1 (en) | 2007-10-04 |
US7620309B2 US7620309B2 (en) | 2009-11-17 |
Family
ID=38222707
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/398,403 Expired - Fee Related US7620309B2 (en) | 2006-04-04 | 2006-04-04 | Plenoptic camera |
US12/574,183 Active 2027-03-27 US8238738B2 (en) | 2006-04-04 | 2009-10-06 | Plenoptic camera |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/574,183 Active 2027-03-27 US8238738B2 (en) | 2006-04-04 | 2009-10-06 | Plenoptic camera |
Country Status (6)
Country | Link |
---|---|
US (2) | US7620309B2 (en) |
EP (1) | EP2008445B1 (en) |
JP (1) | JP4981124B2 (en) |
KR (1) | KR20090016453A (en) |
CN (1) | CN101455071B (en) |
WO (1) | WO2007115281A1 (en) |
Cited By (125)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080124070A1 (en) * | 2006-11-28 | 2008-05-29 | Chia-Kai Liang | Camera using programmable aperture |
US20090027542A1 (en) * | 2007-07-13 | 2009-01-29 | Sony Corporation | Image pickup apparatus |
US20090140131A1 (en) * | 2005-06-23 | 2009-06-04 | Nikon Corporation | Image input apparatus, photodetection apparatus, and image synthesis method |
US20090160975A1 (en) * | 2007-12-19 | 2009-06-25 | Ncr Corporation | Methods and Apparatus for Improved Image Processing to Provide Retroactive Image Focusing and Improved Depth of Field in Retail Imaging Systems |
EP2133726A1 (en) | 2008-06-10 | 2009-12-16 | THOMSON Licensing | Multi-image capture system with improved depth image resolution |
US20100003024A1 (en) * | 2007-12-10 | 2010-01-07 | Amit Kumar Agrawal | Cameras with Varying Spatio-Angular-Temporal Resolutions |
US20100085468A1 (en) * | 2008-10-06 | 2010-04-08 | Park Byung-Kwan | Apparatus and method of capturing image |
US20100091133A1 (en) * | 2008-10-10 | 2010-04-15 | Lim Jae-Guyn | Image processing apparatus and method |
US20100128137A1 (en) * | 2008-11-21 | 2010-05-27 | Eastman Kodak Company | Extended depth of field for image sensor |
US20100128145A1 (en) * | 2008-11-25 | 2010-05-27 | Colvin Pitts | System of and Method for Video Refocusing |
US20100141802A1 (en) * | 2008-12-08 | 2010-06-10 | Timothy Knight | Light Field Data Acquisition Devices, and Methods of Using and Manufacturing Same |
US20100177979A1 (en) * | 2009-01-09 | 2010-07-15 | Canon Kabushiki Kaisha | Image processing apparatus and image processing method |
US20100194921A1 (en) * | 2009-02-05 | 2010-08-05 | Sony Corporation | Image pickup apparatus |
US20100310165A1 (en) * | 2009-06-09 | 2010-12-09 | Industrial Technology Research Institute | Image restoration method and apparatus |
US20110234841A1 (en) * | 2009-04-18 | 2011-09-29 | Lytro, Inc. | Storage and Transmission of Pictures Including Multiple Frames |
US20120140024A1 (en) * | 2010-12-03 | 2012-06-07 | Fly's Eye Imaging, LLC | Method of displaying an enhanced three-dimensional images |
US20120154651A1 (en) * | 2010-12-17 | 2012-06-21 | Canon Kabushiki Kaisha | Image sensing apparatus and method of controlling the image sensing apparatus |
US20120188391A1 (en) * | 2011-01-25 | 2012-07-26 | Scott Smith | Array camera having lenses with independent fields of view |
US20120327259A1 (en) * | 2011-06-24 | 2012-12-27 | Canon Kabushiki Kaisha | Image processing device, image processing method, image capturing device, and program |
US20130064453A1 (en) * | 2011-09-08 | 2013-03-14 | Casio Computer Co., Ltd. | Interpolation image generation apparatus, reconstructed image generation apparatus, method of generating interpolation image, and computer-readable recording medium storing program |
DE102011114325A1 (en) | 2011-09-24 | 2013-03-28 | Martin Häusler | Device for generating polarized optical images, for camera used in mobile phone, has lens system that transfers captured image field to camera sensor |
US20130169837A1 (en) * | 2011-12-28 | 2013-07-04 | Casio Computer Co., Ltd. | Device having image reconstructing function, method, and recording medium |
KR20140027815A (en) * | 2012-08-27 | 2014-03-07 | 삼성전자주식회사 | 3d image acquisition apparatus and method of obtaining color and depth images simultaneously |
US8749620B1 (en) | 2010-02-20 | 2014-06-10 | Lytro, Inc. | 3D light field cameras, images and files, and methods of using, operating, processing and viewing same |
US8768102B1 (en) | 2011-02-09 | 2014-07-01 | Lytro, Inc. | Downsampling light field images |
US8811769B1 (en) | 2012-02-28 | 2014-08-19 | Lytro, Inc. | Extended depth of field and variable center of perspective in light-field processing |
US8831377B2 (en) | 2012-02-28 | 2014-09-09 | Lytro, Inc. | Compensating for variation in microlens position during light-field image processing |
US8948545B2 (en) | 2012-02-28 | 2015-02-03 | Lytro, Inc. | Compensating for sensor saturation and microlens modulation during light-field image processing |
US8978984B2 (en) | 2013-02-28 | 2015-03-17 | Hand Held Products, Inc. | Indicia reading terminals and methods for decoding decodable indicia employing light field imaging |
US8988317B1 (en) | 2014-06-12 | 2015-03-24 | Lytro, Inc. | Depth determination for light field images |
WO2015041496A1 (en) * | 2013-09-23 | 2015-03-26 | 엘지이노텍 주식회사 | Camera module and manufacturing method for same |
US8995785B2 (en) | 2012-02-28 | 2015-03-31 | Lytro, Inc. | Light-field processing and analysis, camera control, and user interfaces and interaction on light-field capture devices |
US8997021B2 (en) | 2012-11-06 | 2015-03-31 | Lytro, Inc. | Parallax and/or three-dimensional effects for thumbnail image displays |
US9001226B1 (en) | 2012-12-04 | 2015-04-07 | Lytro, Inc. | Capturing and relighting images using multiple devices |
US9184199B2 (en) | 2011-08-01 | 2015-11-10 | Lytro, Inc. | Optical assembly including plenoptic microlens array |
US20150358529A1 (en) * | 2014-06-04 | 2015-12-10 | Canon Kabushiki Kaisha | Image processing device, its control method, and storage medium |
US20160057407A1 (en) * | 2013-02-13 | 2016-02-25 | Universität des Saarlandes | Plenoptic imaging device |
US9300932B2 (en) | 2012-05-09 | 2016-03-29 | Lytro, Inc. | Optimization of optical systems for improved light field capture and manipulation |
US9392153B2 (en) | 2013-12-24 | 2016-07-12 | Lytro, Inc. | Plenoptic camera resolution |
US20160231473A1 (en) * | 2015-02-09 | 2016-08-11 | Omnivision Technologies, Inc. | Wide-Angle Camera Using Achromatic Doublet Prism Array And Method of Manufacturing The Same |
US9420276B2 (en) | 2012-02-28 | 2016-08-16 | Lytro, Inc. | Calibration of light-field camera geometry via robust fitting |
US9456141B2 (en) | 2013-02-22 | 2016-09-27 | Lytro, Inc. | Light-field based autofocus |
DE102015006659A1 (en) | 2015-05-22 | 2016-11-24 | Martin Häusler | Plenoptic camera lens |
US9557741B1 (en) * | 2015-08-24 | 2017-01-31 | Ford Global Technologies, Llc | System and method for autonomous valet parking using plenoptic cameras |
CN106500629A (en) * | 2016-11-29 | 2017-03-15 | 深圳大学 | A kind of microscopic three-dimensional measurement apparatus and system |
US9607424B2 (en) | 2012-06-26 | 2017-03-28 | Lytro, Inc. | Depth-assigned content for depth-enhanced pictures |
US9635332B2 (en) | 2014-09-08 | 2017-04-25 | Lytro, Inc. | Saturated pixel recovery in light-field images |
US9667846B2 (en) * | 2012-11-27 | 2017-05-30 | Nokia Technologies Oy | Plenoptic camera apparatus, a method and a computer program |
CN107741644A (en) * | 2017-11-21 | 2018-02-27 | 杭州加速云信息技术有限公司 | A kind of imaging device for different visual angles imaging |
US10033986B2 (en) | 2015-05-26 | 2018-07-24 | Google Llc | Capturing light-field images with uneven and/or incomplete angular sampling |
US10057498B1 (en) * | 2013-03-15 | 2018-08-21 | Cognex Corporation | Light field vision system camera and methods for using the same |
US10092183B2 (en) | 2014-08-31 | 2018-10-09 | Dr. John Berestka | Systems and methods for analyzing the eye |
US10129524B2 (en) | 2012-06-26 | 2018-11-13 | Google Llc | Depth-assigned content for depth-enhanced virtual reality images |
US10205896B2 (en) | 2015-07-24 | 2019-02-12 | Google Llc | Automatic lens flare detection and correction for light-field images |
EP3444635A1 (en) * | 2017-08-14 | 2019-02-20 | Valeo Schalter und Sensoren GmbH | Receiver unit for a laser scanner device, laser scanner device, vehicle and method for capturing light |
US10234868B2 (en) | 2017-06-16 | 2019-03-19 | Ford Global Technologies, Llc | Mobile device initiation of vehicle remote-parking |
US10275898B1 (en) | 2015-04-15 | 2019-04-30 | Google Llc | Wedge-based light-field video capture |
US10275892B2 (en) | 2016-06-09 | 2019-04-30 | Google Llc | Multi-view scene segmentation and propagation |
US10281921B2 (en) | 2017-10-02 | 2019-05-07 | Ford Global Technologies, Llc | Autonomous parking of vehicles in perpendicular parking spots |
US10298834B2 (en) | 2006-12-01 | 2019-05-21 | Google Llc | Video refocusing |
US10334151B2 (en) | 2013-04-22 | 2019-06-25 | Google Llc | Phase detection autofocus using subaperture images |
US10341632B2 (en) | 2015-04-15 | 2019-07-02 | Google Llc. | Spatial random access enabled video system with a three-dimensional viewing volume |
US10336320B2 (en) | 2017-11-22 | 2019-07-02 | Ford Global Technologies, Llc | Monitoring of communication for vehicle remote park-assist |
US10348947B2 (en) * | 2016-09-07 | 2019-07-09 | Interdigital Ce Patent Holdings | Plenoptic imaging device equipped with an enhanced optical system |
US10354399B2 (en) | 2017-05-25 | 2019-07-16 | Google Llc | Multi-view back-projection to a light-field |
US10369988B2 (en) | 2017-01-13 | 2019-08-06 | Ford Global Technologies, Llc | Autonomous parking of vehicles inperpendicular parking spots |
US10384605B1 (en) | 2018-09-04 | 2019-08-20 | Ford Global Technologies, Llc | Methods and apparatus to facilitate pedestrian detection during remote-controlled maneuvers |
US10397545B2 (en) * | 2013-12-23 | 2019-08-27 | University Of Deleware | 3-D light field camera and photography method |
US10412373B2 (en) | 2015-04-15 | 2019-09-10 | Google Llc | Image capture for virtual reality displays |
US10419737B2 (en) | 2015-04-15 | 2019-09-17 | Google Llc | Data structures and delivery methods for expediting virtual reality playback |
US10440407B2 (en) | 2017-05-09 | 2019-10-08 | Google Llc | Adaptive control for immersive experience delivery |
US10444931B2 (en) | 2017-05-09 | 2019-10-15 | Google Llc | Vantage generation and interactive playback |
US10469873B2 (en) | 2015-04-15 | 2019-11-05 | Google Llc | Encoding and decoding virtual reality video |
US10474227B2 (en) | 2017-05-09 | 2019-11-12 | Google Llc | Generation of virtual reality with 6 degrees of freedom from limited viewer data |
US10493981B2 (en) | 2018-04-09 | 2019-12-03 | Ford Global Technologies, Llc | Input signal management for vehicle park-assist |
US10507868B2 (en) | 2018-02-22 | 2019-12-17 | Ford Global Technologies, Llc | Tire pressure monitoring for vehicle park-assist |
US10529233B1 (en) | 2018-09-24 | 2020-01-07 | Ford Global Technologies Llc | Vehicle and method for detecting a parking space via a drone |
US10540818B2 (en) | 2015-04-15 | 2020-01-21 | Google Llc | Stereo image generation and interactive playback |
US10545215B2 (en) | 2017-09-13 | 2020-01-28 | Google Llc | 4D camera tracking and optical stabilization |
US10546424B2 (en) | 2015-04-15 | 2020-01-28 | Google Llc | Layered content delivery for virtual and augmented reality experiences |
US10552947B2 (en) | 2012-06-26 | 2020-02-04 | Google Llc | Depth-based image blurring |
US10567464B2 (en) | 2015-04-15 | 2020-02-18 | Google Llc | Video compression with adaptive view-dependent lighting removal |
US10565734B2 (en) | 2015-04-15 | 2020-02-18 | Google Llc | Video capture, processing, calibration, computational fiber artifact removal, and light-field pipeline |
US10580304B2 (en) | 2017-10-02 | 2020-03-03 | Ford Global Technologies, Llc | Accelerometer-based external sound monitoring for voice controlled autonomous parking |
US10578676B2 (en) | 2017-11-28 | 2020-03-03 | Ford Global Technologies, Llc | Vehicle monitoring of mobile device state-of-charge |
US10585431B2 (en) | 2018-01-02 | 2020-03-10 | Ford Global Technologies, Llc | Mobile device tethering for a remote parking assist system of a vehicle |
US10585430B2 (en) | 2017-06-16 | 2020-03-10 | Ford Global Technologies, Llc | Remote park-assist authentication for vehicles |
US10583830B2 (en) | 2018-01-02 | 2020-03-10 | Ford Global Technologies, Llc | Mobile device tethering for a remote parking assist system of a vehicle |
US10594945B2 (en) | 2017-04-03 | 2020-03-17 | Google Llc | Generating dolly zoom effect using light field image data |
US10627811B2 (en) | 2017-11-07 | 2020-04-21 | Ford Global Technologies, Llc | Audio alerts for remote park-assist tethering |
US10628687B1 (en) | 2018-10-12 | 2020-04-21 | Ford Global Technologies, Llc | Parking spot identification for vehicle park-assist |
US10645281B1 (en) * | 2017-05-19 | 2020-05-05 | Shanghaitech University | Method and system for snapshot multi-spectral light field imaging |
US10679361B2 (en) | 2016-12-05 | 2020-06-09 | Google Llc | Multi-view rotoscope contour propagation |
US10683004B2 (en) | 2018-04-09 | 2020-06-16 | Ford Global Technologies, Llc | Input signal management for vehicle park-assist |
US10684773B2 (en) | 2018-01-03 | 2020-06-16 | Ford Global Technologies, Llc | Mobile device interface for trailer backup-assist |
US10683034B2 (en) | 2017-06-06 | 2020-06-16 | Ford Global Technologies, Llc | Vehicle remote parking systems and methods |
US10684627B2 (en) | 2018-02-06 | 2020-06-16 | Ford Global Technologies, Llc | Accelerometer-based external sound monitoring for position aware autonomous parking |
US10688918B2 (en) | 2018-01-02 | 2020-06-23 | Ford Global Technologies, Llc | Mobile device tethering for a remote parking assist system of a vehicle |
US10692192B2 (en) * | 2014-10-21 | 2020-06-23 | Connaught Electronics Ltd. | Method for providing image data from a camera system, camera system and motor vehicle |
US10717432B2 (en) | 2018-09-13 | 2020-07-21 | Ford Global Technologies, Llc | Park-assist based on vehicle door open positions |
US10732622B2 (en) | 2018-04-05 | 2020-08-04 | Ford Global Technologies, Llc | Advanced user interaction features for remote park assist |
US10737690B2 (en) | 2018-01-02 | 2020-08-11 | Ford Global Technologies, Llc | Mobile device tethering for a remote parking assist system of a vehicle |
US10747218B2 (en) | 2018-01-12 | 2020-08-18 | Ford Global Technologies, Llc | Mobile device tethering for remote parking assist |
US10759417B2 (en) | 2018-04-09 | 2020-09-01 | Ford Global Technologies, Llc | Input signal management for vehicle park-assist |
US10775781B2 (en) | 2017-06-16 | 2020-09-15 | Ford Global Technologies, Llc | Interface verification for vehicle remote park-assist |
US10793144B2 (en) | 2018-04-09 | 2020-10-06 | Ford Global Technologies, Llc | Vehicle remote park-assist communication counters |
US10814864B2 (en) | 2018-01-02 | 2020-10-27 | Ford Global Technologies, Llc | Mobile device tethering for a remote parking assist system of a vehicle |
US10821972B2 (en) | 2018-09-13 | 2020-11-03 | Ford Global Technologies, Llc | Vehicle remote parking assist systems and methods |
US10832429B2 (en) * | 2016-10-18 | 2020-11-10 | Photonic Sensors & Algorithms, S.L. | Device and method for obtaining distance information from views |
US10908603B2 (en) | 2018-10-08 | 2021-02-02 | Ford Global Technologies, Llc | Methods and apparatus to facilitate remote-controlled maneuvers |
US10917748B2 (en) | 2018-01-25 | 2021-02-09 | Ford Global Technologies, Llc | Mobile device tethering for vehicle systems based on variable time-of-flight and dead reckoning |
US10965862B2 (en) | 2018-01-18 | 2021-03-30 | Google Llc | Multi-camera navigation interface |
US10967851B2 (en) | 2018-09-24 | 2021-04-06 | Ford Global Technologies, Llc | Vehicle system and method for setting variable virtual boundary |
US10974717B2 (en) | 2018-01-02 | 2021-04-13 | Ford Global Technologies, I.LC | Mobile device tethering for a remote parking assist system of a vehicle |
US11097723B2 (en) | 2018-10-17 | 2021-08-24 | Ford Global Technologies, Llc | User interfaces for vehicle remote park assist |
US11137754B2 (en) | 2018-10-24 | 2021-10-05 | Ford Global Technologies, Llc | Intermittent delay mitigation for remote vehicle operation |
US11148661B2 (en) | 2018-01-02 | 2021-10-19 | Ford Global Technologies, Llc | Mobile device tethering for a remote parking assist system of a vehicle |
WO2021217031A1 (en) * | 2020-04-23 | 2021-10-28 | The Regents Of The University Of California | Ultrafast light field tomography |
US11169517B2 (en) | 2019-04-01 | 2021-11-09 | Ford Global Technologies, Llc | Initiation of vehicle remote park-assist with key fob |
US11188070B2 (en) | 2018-02-19 | 2021-11-30 | Ford Global Technologies, Llc | Mitigating key fob unavailability for remote parking assist systems |
US11195344B2 (en) | 2019-03-15 | 2021-12-07 | Ford Global Technologies, Llc | High phone BLE or CPU burden detection and notification |
US11275368B2 (en) | 2019-04-01 | 2022-03-15 | Ford Global Technologies, Llc | Key fobs for vehicle remote park-assist |
US11328446B2 (en) | 2015-04-15 | 2022-05-10 | Google Llc | Combining light-field data with active depth data for depth map generation |
US11456326B2 (en) * | 2018-03-15 | 2022-09-27 | Photonic Sensors & Algorithms, S.L. | Plenoptic camera for mobile devices |
US11789442B2 (en) | 2019-02-07 | 2023-10-17 | Ford Global Technologies, Llc | Anomalous input detection |
Families Citing this family (91)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP4545190B2 (en) * | 2005-03-24 | 2010-09-15 | パナソニック株式会社 | Imaging device |
US7620309B2 (en) * | 2006-04-04 | 2009-11-17 | Adobe Systems, Incorporated | Plenoptic camera |
US7969577B2 (en) * | 2006-09-14 | 2011-06-28 | Asml Netherlands B.V. | Inspection apparatus, an apparatus for projecting an image and a method of measuring a property of a substrate |
US20100265385A1 (en) * | 2009-04-18 | 2010-10-21 | Knight Timothy J | Light Field Camera Image, File and Configuration Data, and Methods of Using, Storing and Communicating Same |
US8559705B2 (en) * | 2006-12-01 | 2013-10-15 | Lytro, Inc. | Interactive refocusing of electronic images |
US7872796B2 (en) | 2007-01-25 | 2011-01-18 | Adobe Systems Incorporated | Light field microscope with lenslet array |
US8290358B1 (en) | 2007-06-25 | 2012-10-16 | Adobe Systems Incorporated | Methods and apparatus for light-field imaging |
US8019215B2 (en) * | 2007-08-06 | 2011-09-13 | Adobe Systems Incorporated | Method and apparatus for radiance capture by multiplexing in the frequency domain |
US7956924B2 (en) * | 2007-10-18 | 2011-06-07 | Adobe Systems Incorporated | Fast computational camera based on two arrays of lenses |
ES2372515B2 (en) * | 2008-01-15 | 2012-10-16 | Universidad De La Laguna | CHAMBER FOR THE REAL-TIME ACQUISITION OF THE VISUAL INFORMATION OF THREE-DIMENSIONAL SCENES. |
US7962033B2 (en) | 2008-01-23 | 2011-06-14 | Adobe Systems Incorporated | Methods and apparatus for full-resolution light-field capture and rendering |
US8189065B2 (en) | 2008-01-23 | 2012-05-29 | Adobe Systems Incorporated | Methods and apparatus for full-resolution light-field capture and rendering |
JP4941332B2 (en) | 2008-01-28 | 2012-05-30 | ソニー株式会社 | Imaging device |
US8155456B2 (en) * | 2008-04-29 | 2012-04-10 | Adobe Systems Incorporated | Method and apparatus for block-based compression of light-field images |
JP2009290268A (en) * | 2008-05-27 | 2009-12-10 | Sony Corp | Imaging apparatus |
US8244058B1 (en) | 2008-05-30 | 2012-08-14 | Adobe Systems Incorporated | Method and apparatus for managing artifacts in frequency domain processing of light-field images |
US7949252B1 (en) | 2008-12-11 | 2011-05-24 | Adobe Systems Incorporated | Plenoptic camera with large depth of field |
US8315476B1 (en) * | 2009-01-20 | 2012-11-20 | Adobe Systems Incorporated | Super-resolution with the focused plenoptic camera |
US8189089B1 (en) | 2009-01-20 | 2012-05-29 | Adobe Systems Incorporated | Methods and apparatus for reducing plenoptic camera artifacts |
US8345144B1 (en) * | 2009-07-15 | 2013-01-01 | Adobe Systems Incorporated | Methods and apparatus for rich image capture with focused plenoptic cameras |
US8228417B1 (en) * | 2009-07-15 | 2012-07-24 | Adobe Systems Incorporated | Focused plenoptic camera employing different apertures or filtering at different microlenses |
US20110115916A1 (en) * | 2009-11-16 | 2011-05-19 | Eiji Yafuso | System for mosaic image acquisition |
US8400555B1 (en) | 2009-12-01 | 2013-03-19 | Adobe Systems Incorporated | Focused plenoptic camera employing microlenses with different focal lengths |
CN102770873B (en) | 2010-01-22 | 2017-09-19 | 纽约市哥伦比亚大学理事会 | System, method and medium for recording image using optical diffuser |
US8665367B2 (en) * | 2010-01-29 | 2014-03-04 | Sharp Laboratories Of America, Inc. | Video resolution enhancement technique |
US8817015B2 (en) | 2010-03-03 | 2014-08-26 | Adobe Systems Incorporated | Methods, apparatus, and computer-readable storage media for depth-based rendering of focused plenoptic camera data |
US8358366B1 (en) | 2010-05-28 | 2013-01-22 | Adobe Systems Incorporate | Methods and apparatus for high-speed digital imaging |
US8493432B2 (en) * | 2010-06-29 | 2013-07-23 | Mitsubishi Electric Research Laboratories, Inc. | Digital refocusing for wide-angle images using axial-cone cameras |
US8803918B2 (en) | 2010-08-27 | 2014-08-12 | Adobe Systems Incorporated | Methods and apparatus for calibrating focused plenoptic camera data |
US8665341B2 (en) | 2010-08-27 | 2014-03-04 | Adobe Systems Incorporated | Methods and apparatus for rendering output images with simulated artistic effects from focused plenoptic camera data |
US8749694B2 (en) | 2010-08-27 | 2014-06-10 | Adobe Systems Incorporated | Methods and apparatus for rendering focused plenoptic camera data using super-resolved demosaicing |
US8724000B2 (en) | 2010-08-27 | 2014-05-13 | Adobe Systems Incorporated | Methods and apparatus for super-resolution in integral photography |
US20120056987A1 (en) * | 2010-09-03 | 2012-03-08 | Luke Fedoroff | 3d camera system and method |
US8675993B2 (en) | 2010-09-03 | 2014-03-18 | Adobe Systems Incorporated | Methods and apparatus for patch-based shape from shading |
US8780251B2 (en) * | 2010-09-20 | 2014-07-15 | Canon Kabushiki Kaisha | Image capture with focus adjustment |
DE102010047846B4 (en) * | 2010-09-30 | 2017-11-09 | Carl Zeiss Vision International Gmbh | Optical lens and method of manufacturing an optical lens |
US20140192238A1 (en) | 2010-10-24 | 2014-07-10 | Linx Computational Imaging Ltd. | System and Method for Imaging and Image Processing |
US8797432B2 (en) | 2011-01-20 | 2014-08-05 | Aptina Imaging Corporation | Imaging system with an array of image sensors |
US8803990B2 (en) | 2011-01-25 | 2014-08-12 | Aptina Imaging Corporation | Imaging system with multiple sensors for producing high-dynamic-range images |
GB2488519A (en) * | 2011-02-16 | 2012-09-05 | St Microelectronics Res & Dev | Multi-channel image sensor incorporating lenslet array and overlapping fields of view. |
US9030550B2 (en) | 2011-03-25 | 2015-05-12 | Adobe Systems Incorporated | Thin plenoptic cameras using solid immersion lenses |
JP5170276B2 (en) | 2011-03-31 | 2013-03-27 | カシオ計算機株式会社 | Image reconstruction device, image reconstruction method, and image reconstruction program |
US8716649B2 (en) * | 2011-05-05 | 2014-05-06 | Maxim Integrated Products, Inc. | Optical gesture sensor using a single illumination source |
US9229581B2 (en) | 2011-05-05 | 2016-01-05 | Maxim Integrated Products, Inc. | Method for detecting gestures using a multi-segment photodiode and one or fewer illumination sources |
US8531581B2 (en) * | 2011-05-23 | 2013-09-10 | Ricoh Co., Ltd. | Focusing and focus metrics for a plenoptic imaging system |
JP5818514B2 (en) * | 2011-05-27 | 2015-11-18 | キヤノン株式会社 | Image processing apparatus, image processing method, and program |
WO2012170111A1 (en) * | 2011-06-08 | 2012-12-13 | Lytro, Inc. | Storage and transmission of pictures including multiple frames |
TWI437266B (en) * | 2011-07-28 | 2014-05-11 | Univ Nat Taiwan Science Tech | Light harvesting lens module |
EP2751748B1 (en) | 2011-08-30 | 2019-05-08 | Digimarc Corporation | Methods and arrangements for identifying objects |
WO2013058735A1 (en) | 2011-10-18 | 2013-04-25 | Hewlett-Packard Development Company, L.P. | Depth mask assisted video stabilization |
US8928969B2 (en) | 2011-12-06 | 2015-01-06 | Ostendo Technologies, Inc. | Spatio-optical directional light modulator |
US8854724B2 (en) | 2012-03-27 | 2014-10-07 | Ostendo Technologies, Inc. | Spatio-temporal directional light modulator |
US8941750B2 (en) | 2011-12-27 | 2015-01-27 | Casio Computer Co., Ltd. | Image processing device for generating reconstruction image, image generating method, and storage medium |
EP2623964A1 (en) | 2012-02-06 | 2013-08-07 | Jürgen Kupper | X-ray device and x-ray method for studying a three-dimensional object |
US8953012B2 (en) | 2012-02-13 | 2015-02-10 | Raytheon Company | Multi-plenoptic system with image stacking and method for wide field-of-regard high-resolution imaging |
US9137441B2 (en) | 2012-02-16 | 2015-09-15 | Ricoh Co., Ltd. | Spatial reconstruction of plenoptic images |
JP5459337B2 (en) | 2012-03-21 | 2014-04-02 | カシオ計算機株式会社 | Imaging apparatus, image processing method, and program |
JP6019729B2 (en) * | 2012-05-11 | 2016-11-02 | ソニー株式会社 | Image processing apparatus, image processing method, and program |
US8628014B1 (en) * | 2012-05-15 | 2014-01-14 | John M. Hoffer, Jr. | Light field instruction symbol identifier and method of use |
US9179126B2 (en) | 2012-06-01 | 2015-11-03 | Ostendo Technologies, Inc. | Spatio-temporal light field cameras |
US8754829B2 (en) | 2012-08-04 | 2014-06-17 | Paul Lapstun | Scanning light field camera and display |
FR2994735B1 (en) | 2012-08-22 | 2014-08-01 | Onera (Off Nat Aerospatiale) | METHOD AND DEVICE FOR TELEMETRIC IMAGING |
JP2014220564A (en) * | 2013-05-01 | 2014-11-20 | キヤノン株式会社 | Image reproducing device, image reproducing method, program, and imaging device |
CN110061018B (en) | 2013-05-21 | 2023-11-28 | 弗托斯传感与算法公司 | Monolithic integration of plenoptic lenses on photosensor substrates |
WO2014204623A1 (en) | 2013-06-21 | 2014-12-24 | University Of South Florida | Full-color incoherent digital holography |
US9538075B2 (en) | 2013-12-30 | 2017-01-03 | Indiana University Research And Technology Corporation | Frequency domain processing techniques for plenoptic images |
WO2015137635A1 (en) * | 2014-03-13 | 2015-09-17 | Samsung Electronics Co., Ltd. | Image pickup apparatus and method for generating image having depth information |
KR102228456B1 (en) * | 2014-03-13 | 2021-03-16 | 삼성전자주식회사 | Image pickup apparatus and image pickup method of generating image having depth information |
US9305375B2 (en) | 2014-03-25 | 2016-04-05 | Lytro, Inc. | High-quality post-rendering depth blur |
WO2016007579A1 (en) | 2014-07-10 | 2016-01-14 | University Of South Florida | Systems and methods for performing self-interference incoherent digital holography |
TWI529661B (en) * | 2014-10-17 | 2016-04-11 | 國立臺灣大學 | Method of quickly building up depth map and image processing device |
US9444991B2 (en) | 2014-11-13 | 2016-09-13 | Lytro, Inc. | Robust layered light-field rendering |
EP3026884A1 (en) * | 2014-11-27 | 2016-06-01 | Thomson Licensing | Plenoptic camera comprising a light emitting device |
US10070055B2 (en) | 2015-03-25 | 2018-09-04 | Massachusetts Institute Of Technology | Devices and methods for optically multiplexed imaging |
CN106303166A (en) * | 2015-05-22 | 2017-01-04 | 电信科学技术研究院 | A kind of image capture device |
ITUA20161688A1 (en) | 2016-03-15 | 2017-09-15 | Univ Degli Studi Di Bari | Plenopic image acquisition device and procedure. |
US10783652B2 (en) | 2016-05-06 | 2020-09-22 | Arizona Board Of Regents On Behalf Of The University Of Arizona | Plenoptic imaging apparatus, method, and applications |
EP3264741A1 (en) * | 2016-06-30 | 2018-01-03 | Thomson Licensing | Plenoptic sub aperture view shuffling with improved resolution |
US10169849B2 (en) | 2016-11-11 | 2019-01-01 | International Business Machines Corporation | Contextual personalized focus for variable depth of field photographs on social networks |
GB201708861D0 (en) * | 2017-06-02 | 2017-07-19 | Purelifi Ltd | Concentrator height reduction |
WO2019018851A1 (en) * | 2017-07-21 | 2019-01-24 | California Institute Of Technology | Ultra-thin planar lens-less camera |
US11882371B2 (en) | 2017-08-11 | 2024-01-23 | California Institute Of Technology | Lensless 3-dimensional imaging using directional sensing elements |
US10218946B1 (en) * | 2017-08-29 | 2019-02-26 | Sony Interactive Entertainment LLC | High spectrum camera |
WO2019064047A1 (en) | 2017-09-26 | 2019-04-04 | Universita' Degli Studi Di Bari Aldo Moro | Device and process for the contemporary capture of standard and plenoptic images |
US20190137731A1 (en) * | 2017-11-03 | 2019-05-09 | Sony Corporation | Light field adapter for interchangeable lens cameras |
FR3090904B1 (en) | 2018-12-19 | 2021-02-19 | Office National Detudes Rech Aerospatiales | MULTI-CHANNEL MONOLITHIC OPTICAL COMPONENT |
CN113302541A (en) | 2019-03-22 | 2021-08-24 | 巴里大学 | Process and apparatus for capturing plenoptic images between arbitrary planes |
CN110928113B (en) * | 2019-12-03 | 2021-10-08 | 西北工业大学 | Light field acquisition device with variable spatial resolution |
EP3993385A1 (en) | 2020-10-29 | 2022-05-04 | Universitat de València | A multiperspective photography camera device |
US11947134B2 (en) | 2021-01-22 | 2024-04-02 | National Taiwan University | Device of generating 3D light-field image |
US11341224B2 (en) | 2021-01-29 | 2022-05-24 | Ut Battelle, Llc | Handheld multi-sensor biometric imaging device and processing pipeline |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4849782A (en) * | 1985-06-28 | 1989-07-18 | Canon Kabushiki Kaisha | Focus detecting device |
US5076687A (en) * | 1990-08-28 | 1991-12-31 | Massachusetts Institute Of Technology | Optical ranging apparatus |
US5400093A (en) * | 1992-12-28 | 1995-03-21 | U.S. Philips Corporation | Image projection system with autofocusing |
US5729011A (en) * | 1995-02-24 | 1998-03-17 | Olympus Optical Co., Ltd. | Spectroscopic apparatus and spectroscopic image recording apparatus |
US20010012149A1 (en) * | 1997-10-30 | 2001-08-09 | Shawn-Yu Lin | Optical elements comprising photonic crystals and applications thereof |
US20010050813A1 (en) * | 1996-05-09 | 2001-12-13 | Pierre Allio | Autostereoscopic imaging device and system comprising it |
US6339506B1 (en) * | 1998-11-06 | 2002-01-15 | Oni Systems Corp. | Microlens array with spatially varying optical property |
US6351269B1 (en) * | 1998-04-17 | 2002-02-26 | Adobe Systems Incorporated | Multiple image morphing |
Family Cites Families (57)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US725567A (en) | 1902-09-25 | 1903-04-14 | Frederic E Ives | Parallax stereogram and process of making same. |
US2039648A (en) * | 1933-05-06 | 1936-05-05 | Perser Corp | Camera for making parallax panoramagrams |
US3985419A (en) | 1970-10-05 | 1976-10-12 | Canon Kabushiki Kaisha | Method of making a synthetic focused image hologram |
JPS5040669B1 (en) * | 1970-11-13 | 1975-12-25 | ||
JPS5946376B2 (en) * | 1977-03-11 | 1984-11-12 | 富士写真フイルム株式会社 | stereoscopic photography device |
US4193093A (en) | 1978-08-03 | 1980-03-11 | The United States Of America As Represented By The Secretary Of The Navy | CCD camera interface circuit |
US4230942A (en) * | 1979-03-26 | 1980-10-28 | Honeywell Inc. | Crossed cylindrical lens |
US4580219A (en) * | 1983-05-02 | 1986-04-01 | General Electric Company | Method for reducing image artifacts due to projection measurement inconsistencies |
US4732453A (en) * | 1984-12-10 | 1988-03-22 | Integrated Images, Inc. | Integral photography apparatus and method of forming same |
US5361127A (en) | 1992-08-07 | 1994-11-01 | Hughes Aircraft Company | Multi-image single sensor depth recovery system |
KR950704670A (en) * | 1993-09-30 | 1995-11-20 | 가따다 데쯔야 | Confocal Optics |
JPH07270791A (en) * | 1994-03-25 | 1995-10-20 | Nikon Corp | Projector |
BR9612680A (en) * | 1996-06-03 | 1999-07-20 | Herman D Mims | Method and apparatus for three-dimensional photography |
US6137535A (en) * | 1996-11-04 | 2000-10-24 | Eastman Kodak Company | Compact digital camera with segmented fields of view |
GB9702006D0 (en) * | 1997-01-31 | 1997-03-19 | Univ Montfort | Lens arrangements |
US6097394A (en) * | 1997-04-28 | 2000-08-01 | Board Of Trustees, Leland Stanford, Jr. University | Method and system for light field rendering |
US7002749B2 (en) | 1997-07-08 | 2006-02-21 | Kremen Stanley H | Modular integral magnifier |
US7532372B2 (en) | 1997-07-08 | 2009-05-12 | Kremen Stanley H | Method for creating a holographic screen that reconstructs uniformly magnified three-dimensional images from projected integral photographs |
JP3869116B2 (en) * | 1998-04-27 | 2007-01-17 | 日本放送協会 | Stereoscopic image device |
US6268846B1 (en) * | 1998-06-22 | 2001-07-31 | Adobe Systems Incorporated | 3D graphics based on images and morphing |
JP3790371B2 (en) * | 1998-10-09 | 2006-06-28 | 日本放送協会 | Stereoscopic image device |
AU2421101A (en) | 1999-11-16 | 2001-05-30 | Agilent Technologies Inc. | Confocal imaging |
US6476805B1 (en) * | 1999-12-23 | 2002-11-05 | Microsoft Corporation | Techniques for spatial displacement estimation and multi-resolution operations on light fields |
US6738533B1 (en) * | 2000-03-29 | 2004-05-18 | Microsoft Corporation | Minimum sampling rate and minimum sampling curve for image-based rendering |
JP2001330769A (en) * | 2000-05-24 | 2001-11-30 | Canon Inc | Image pickup device and its control method |
US7061532B2 (en) | 2001-03-27 | 2006-06-13 | Hewlett-Packard Development Company, L.P. | Single sensor chip digital stereo camera |
US7260323B2 (en) * | 2002-06-12 | 2007-08-21 | Eastman Kodak Company | Imaging using silver halide films with micro-lens capture, scanning and digital reconstruction |
KR100721414B1 (en) | 2002-09-30 | 2007-05-23 | 도꾸리쯔교세이호징 가가꾸 기쥬쯔 신꼬 기꼬 | Confocal microscope, fluorescence measuring method and polarized light measuring method using confocal microscope |
US20040114807A1 (en) | 2002-12-13 | 2004-06-17 | Dan Lelescu | Statistical representation and coding of light field data |
JP2004200975A (en) * | 2002-12-18 | 2004-07-15 | Canon Inc | Imaging apparatus |
JP2004239932A (en) * | 2003-02-03 | 2004-08-26 | Noriji Ooishi | Stereoscopic image photographing device |
US7057806B2 (en) | 2003-05-09 | 2006-06-06 | 3M Innovative Properties Company | Scanning laser microscope with wavefront sensor |
US7702878B2 (en) * | 2004-03-19 | 2010-04-20 | Broadcom Corporation | Method and system for scalable video data width |
US7054067B2 (en) * | 2003-05-27 | 2006-05-30 | Nippon Hoso Kyokai | Three-dimensional image optical system |
DE10327733C5 (en) * | 2003-06-18 | 2012-04-19 | Limo Patentverwaltung Gmbh & Co. Kg | Device for shaping a light beam |
US7236304B2 (en) * | 2003-11-04 | 2007-06-26 | Matsushita Electric Industrial Co., Ltd. | Imaging Device |
JP4508619B2 (en) * | 2003-12-03 | 2010-07-21 | キヤノン株式会社 | Method for manufacturing solid-state imaging device |
WO2006039486A2 (en) * | 2004-10-01 | 2006-04-13 | The Board Of Trustees Of The Leland Stanford Junior University | Imaging arrangements and methods therefor |
US7653261B2 (en) * | 2004-11-12 | 2010-01-26 | Microsoft Corporation | Image tapestry |
US7636416B2 (en) * | 2005-02-25 | 2009-12-22 | Hitachi Medical Corporation | X-ray CT apparatus comprising a tube current control unit |
JP4545190B2 (en) * | 2005-03-24 | 2010-09-15 | パナソニック株式会社 | Imaging device |
JP4826152B2 (en) * | 2005-06-23 | 2011-11-30 | 株式会社ニコン | Image composition method and imaging apparatus |
WO2007031936A2 (en) * | 2005-09-13 | 2007-03-22 | Koninklijke Philips Electronics, N.V. | Automatic generation of optimal views for computed tomography thoracic diagnosis |
US8009209B2 (en) * | 2005-09-30 | 2011-08-30 | Simon Fraser University | Methods and apparatus for detecting defects in imaging arrays by image analysis |
US8248515B2 (en) * | 2006-02-07 | 2012-08-21 | The Board Of Trustees Of The Leland Stanford Junior University | Variable imaging arrangements and methods therefor |
US7620309B2 (en) * | 2006-04-04 | 2009-11-17 | Adobe Systems, Incorporated | Plenoptic camera |
US7676073B2 (en) * | 2006-08-29 | 2010-03-09 | Siemens Medical Solutions Usa, Inc. | System and method for reducing circular artifacts in tomographic imaging |
US8103111B2 (en) * | 2006-12-26 | 2012-01-24 | Olympus Imaging Corp. | Coding method, electronic camera, recording medium storing coded program, and decoding method |
JP4264569B2 (en) * | 2007-01-09 | 2009-05-20 | ソニー株式会社 | Imaging device |
US7860333B2 (en) * | 2007-01-09 | 2010-12-28 | University Of Utah Research Foundation | Systems and methods for deblurring data corrupted by shift variant blurring |
US7792423B2 (en) * | 2007-02-06 | 2010-09-07 | Mitsubishi Electric Research Laboratories, Inc. | 4D light field cameras |
CN100585453C (en) * | 2007-02-09 | 2010-01-27 | 奥林巴斯映像株式会社 | Decoding method and decoding apparatus |
US8019215B2 (en) * | 2007-08-06 | 2011-09-13 | Adobe Systems Incorporated | Method and apparatus for radiance capture by multiplexing in the frequency domain |
KR20090022596A (en) * | 2007-08-31 | 2009-03-04 | 삼성전기주식회사 | Diffraction type optical modulator and display apparatus including this |
US7962033B2 (en) * | 2008-01-23 | 2011-06-14 | Adobe Systems Incorporated | Methods and apparatus for full-resolution light-field capture and rendering |
US8189065B2 (en) * | 2008-01-23 | 2012-05-29 | Adobe Systems Incorporated | Methods and apparatus for full-resolution light-field capture and rendering |
KR101441586B1 (en) * | 2008-10-06 | 2014-09-23 | 삼성전자 주식회사 | Apparatus and method for capturing image |
-
2006
- 2006-04-04 US US11/398,403 patent/US7620309B2/en not_active Expired - Fee Related
-
2007
- 2007-04-03 CN CN2007800190343A patent/CN101455071B/en active Active
- 2007-04-03 EP EP07760021.1A patent/EP2008445B1/en not_active Expired - Fee Related
- 2007-04-03 JP JP2009504414A patent/JP4981124B2/en active Active
- 2007-04-03 WO PCT/US2007/065855 patent/WO2007115281A1/en active Application Filing
- 2007-04-03 KR KR1020087027050A patent/KR20090016453A/en active IP Right Grant
-
2009
- 2009-10-06 US US12/574,183 patent/US8238738B2/en active Active
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4849782A (en) * | 1985-06-28 | 1989-07-18 | Canon Kabushiki Kaisha | Focus detecting device |
US5076687A (en) * | 1990-08-28 | 1991-12-31 | Massachusetts Institute Of Technology | Optical ranging apparatus |
US5400093A (en) * | 1992-12-28 | 1995-03-21 | U.S. Philips Corporation | Image projection system with autofocusing |
US5729011A (en) * | 1995-02-24 | 1998-03-17 | Olympus Optical Co., Ltd. | Spectroscopic apparatus and spectroscopic image recording apparatus |
US20010050813A1 (en) * | 1996-05-09 | 2001-12-13 | Pierre Allio | Autostereoscopic imaging device and system comprising it |
US20010012149A1 (en) * | 1997-10-30 | 2001-08-09 | Shawn-Yu Lin | Optical elements comprising photonic crystals and applications thereof |
US6351269B1 (en) * | 1998-04-17 | 2002-02-26 | Adobe Systems Incorporated | Multiple image morphing |
US6339506B1 (en) * | 1998-11-06 | 2002-01-15 | Oni Systems Corp. | Microlens array with spatially varying optical property |
Cited By (177)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090140131A1 (en) * | 2005-06-23 | 2009-06-04 | Nikon Corporation | Image input apparatus, photodetection apparatus, and image synthesis method |
US7732744B2 (en) * | 2005-06-23 | 2010-06-08 | Nikon Corporation | Image input apparatus, photodetection apparatus, and image synthesis method |
US20080124070A1 (en) * | 2006-11-28 | 2008-05-29 | Chia-Kai Liang | Camera using programmable aperture |
US10298834B2 (en) | 2006-12-01 | 2019-05-21 | Google Llc | Video refocusing |
US20090027542A1 (en) * | 2007-07-13 | 2009-01-29 | Sony Corporation | Image pickup apparatus |
US8130310B2 (en) * | 2007-07-13 | 2012-03-06 | Sony Corporation | Image pickup apparatus |
US8229294B2 (en) * | 2007-12-10 | 2012-07-24 | Mitsubishi Electric Research Laboratories, Inc. | Cameras with varying spatio-angular-temporal resolutions |
US20100003024A1 (en) * | 2007-12-10 | 2010-01-07 | Amit Kumar Agrawal | Cameras with Varying Spatio-Angular-Temporal Resolutions |
US20090160975A1 (en) * | 2007-12-19 | 2009-06-25 | Ncr Corporation | Methods and Apparatus for Improved Image Processing to Provide Retroactive Image Focusing and Improved Depth of Field in Retail Imaging Systems |
WO2009150061A1 (en) * | 2008-06-10 | 2009-12-17 | Thomson Licensing | Multi-image capture system with improved depth image resolution |
EP2133726A1 (en) | 2008-06-10 | 2009-12-16 | THOMSON Licensing | Multi-image capture system with improved depth image resolution |
US8111320B2 (en) | 2008-06-10 | 2012-02-07 | Thomson Licensing | Multi-image capture system with improved depth image resolution |
US20110080491A1 (en) * | 2008-06-10 | 2011-04-07 | Valter Drazic | Multi-image capture system with improved depth image resolution |
US20100085468A1 (en) * | 2008-10-06 | 2010-04-08 | Park Byung-Kwan | Apparatus and method of capturing image |
US8947578B2 (en) | 2008-10-06 | 2015-02-03 | Samsung Electronics Co., Ltd. | Apparatus and method of capturing image |
US20100091133A1 (en) * | 2008-10-10 | 2010-04-15 | Lim Jae-Guyn | Image processing apparatus and method |
US8390728B2 (en) | 2008-10-10 | 2013-03-05 | Samsung Electronics Co., Ltd. | Image processing apparatus and method |
US20100128137A1 (en) * | 2008-11-21 | 2010-05-27 | Eastman Kodak Company | Extended depth of field for image sensor |
US8587681B2 (en) | 2008-11-21 | 2013-11-19 | Omnivision Technologies, Inc. | Extended depth of field for image sensor |
US8760566B2 (en) | 2008-11-25 | 2014-06-24 | Lytro, Inc. | Video refocusing |
US8570426B2 (en) | 2008-11-25 | 2013-10-29 | Lytro, Inc. | System of and method for video refocusing |
US20100128145A1 (en) * | 2008-11-25 | 2010-05-27 | Colvin Pitts | System of and Method for Video Refocusing |
US8446516B2 (en) | 2008-11-25 | 2013-05-21 | Lytro, Inc. | Generating and outputting video data from refocusable light field video data |
US20100129048A1 (en) * | 2008-11-25 | 2010-05-27 | Colvin Pitts | System and Method for Acquiring, Editing, Generating and Outputting Video Data |
US8614764B2 (en) | 2008-11-25 | 2013-12-24 | Lytro, Inc. | Acquiring, editing, generating and outputting video data |
US8279325B2 (en) | 2008-11-25 | 2012-10-02 | Lytro, Inc. | System and method for acquiring, editing, generating and outputting video data |
US20100141802A1 (en) * | 2008-12-08 | 2010-06-10 | Timothy Knight | Light Field Data Acquisition Devices, and Methods of Using and Manufacturing Same |
US8724014B2 (en) | 2008-12-08 | 2014-05-13 | Lytro, Inc. | Light field data acquisition |
US8289440B2 (en) * | 2008-12-08 | 2012-10-16 | Lytro, Inc. | Light field data acquisition devices, and methods of using and manufacturing same |
US9467607B2 (en) | 2008-12-08 | 2016-10-11 | Lytro, Inc. | Light field data acquisition |
US8976288B2 (en) * | 2008-12-08 | 2015-03-10 | Lytro, Inc. | Light field data acquisition |
US20140211077A1 (en) * | 2008-12-08 | 2014-07-31 | Lytro, Inc. | Light field data acquisition |
US8335393B2 (en) | 2009-01-09 | 2012-12-18 | Canon Kabushiki Kaisha | Image processing apparatus and image processing method |
US20100177979A1 (en) * | 2009-01-09 | 2010-07-15 | Canon Kabushiki Kaisha | Image processing apparatus and image processing method |
US8325241B2 (en) * | 2009-02-05 | 2012-12-04 | Sony Corporation | Image pickup apparatus that stores adjacent and contiguous pixel data before integration of same |
US20100194921A1 (en) * | 2009-02-05 | 2010-08-05 | Sony Corporation | Image pickup apparatus |
US8908058B2 (en) | 2009-04-18 | 2014-12-09 | Lytro, Inc. | Storage and transmission of pictures including multiple frames |
US20110234841A1 (en) * | 2009-04-18 | 2011-09-29 | Lytro, Inc. | Storage and Transmission of Pictures Including Multiple Frames |
US20100310165A1 (en) * | 2009-06-09 | 2010-12-09 | Industrial Technology Research Institute | Image restoration method and apparatus |
US8749620B1 (en) | 2010-02-20 | 2014-06-10 | Lytro, Inc. | 3D light field cameras, images and files, and methods of using, operating, processing and viewing same |
US20120140024A1 (en) * | 2010-12-03 | 2012-06-07 | Fly's Eye Imaging, LLC | Method of displaying an enhanced three-dimensional images |
US9124881B2 (en) * | 2010-12-03 | 2015-09-01 | Fly's Eye Imaging LLC | Method of displaying an enhanced three-dimensional images |
US8581998B2 (en) * | 2010-12-17 | 2013-11-12 | Canon Kabushiki Kaisha | Image sensing apparatus and method of controlling the image sensing apparatus |
US8711269B2 (en) | 2010-12-17 | 2014-04-29 | Canon Kabushiki Kaisha | Image sensing apparatus and method of controlling the image sensing apparatus |
US20120154651A1 (en) * | 2010-12-17 | 2012-06-21 | Canon Kabushiki Kaisha | Image sensing apparatus and method of controlling the image sensing apparatus |
US20120188391A1 (en) * | 2011-01-25 | 2012-07-26 | Scott Smith | Array camera having lenses with independent fields of view |
US8768102B1 (en) | 2011-02-09 | 2014-07-01 | Lytro, Inc. | Downsampling light field images |
US20120327259A1 (en) * | 2011-06-24 | 2012-12-27 | Canon Kabushiki Kaisha | Image processing device, image processing method, image capturing device, and program |
US8988546B2 (en) * | 2011-06-24 | 2015-03-24 | Canon Kabushiki Kaisha | Image processing device, image processing method, image capturing device, and program |
US9419049B2 (en) | 2011-08-01 | 2016-08-16 | Lytro, Inc. | Optical assembly including plenoptic microlens array |
US9305956B2 (en) | 2011-08-01 | 2016-04-05 | Lytro, Inc. | Optical assembly including plenoptic microlens array |
US9184199B2 (en) | 2011-08-01 | 2015-11-10 | Lytro, Inc. | Optical assembly including plenoptic microlens array |
US20130064453A1 (en) * | 2011-09-08 | 2013-03-14 | Casio Computer Co., Ltd. | Interpolation image generation apparatus, reconstructed image generation apparatus, method of generating interpolation image, and computer-readable recording medium storing program |
US8588516B2 (en) * | 2011-09-08 | 2013-11-19 | Casio Computer Co., Ltd. | Interpolation image generation apparatus, reconstructed image generation apparatus, method of generating interpolation image, and computer-readable recording medium storing program |
DE102011114325A1 (en) | 2011-09-24 | 2013-03-28 | Martin Häusler | Device for generating polarized optical images, for camera used in mobile phone, has lens system that transfers captured image field to camera sensor |
US20130169837A1 (en) * | 2011-12-28 | 2013-07-04 | Casio Computer Co., Ltd. | Device having image reconstructing function, method, and recording medium |
US8929685B2 (en) * | 2011-12-28 | 2015-01-06 | Casio Computer Co., Ltd. | Device having image reconstructing function, method, and recording medium |
US8811769B1 (en) | 2012-02-28 | 2014-08-19 | Lytro, Inc. | Extended depth of field and variable center of perspective in light-field processing |
US8831377B2 (en) | 2012-02-28 | 2014-09-09 | Lytro, Inc. | Compensating for variation in microlens position during light-field image processing |
US8995785B2 (en) | 2012-02-28 | 2015-03-31 | Lytro, Inc. | Light-field processing and analysis, camera control, and user interfaces and interaction on light-field capture devices |
US8948545B2 (en) | 2012-02-28 | 2015-02-03 | Lytro, Inc. | Compensating for sensor saturation and microlens modulation during light-field image processing |
US9420276B2 (en) | 2012-02-28 | 2016-08-16 | Lytro, Inc. | Calibration of light-field camera geometry via robust fitting |
US9386288B2 (en) | 2012-02-28 | 2016-07-05 | Lytro, Inc. | Compensating for sensor saturation and microlens modulation during light-field image processing |
US9172853B2 (en) | 2012-02-28 | 2015-10-27 | Lytro, Inc. | Microlens array architecture for avoiding ghosting in projected images |
US8971625B2 (en) | 2012-02-28 | 2015-03-03 | Lytro, Inc. | Generating dolly zoom effect using light field image data |
US9866810B2 (en) | 2012-05-09 | 2018-01-09 | Lytro, Inc. | Optimization of optical systems for improved light field capture and manipulation |
US9300932B2 (en) | 2012-05-09 | 2016-03-29 | Lytro, Inc. | Optimization of optical systems for improved light field capture and manipulation |
US10552947B2 (en) | 2012-06-26 | 2020-02-04 | Google Llc | Depth-based image blurring |
US10129524B2 (en) | 2012-06-26 | 2018-11-13 | Google Llc | Depth-assigned content for depth-enhanced virtual reality images |
US9607424B2 (en) | 2012-06-26 | 2017-03-28 | Lytro, Inc. | Depth-assigned content for depth-enhanced pictures |
KR101951318B1 (en) | 2012-08-27 | 2019-04-25 | 삼성전자주식회사 | 3D image acquisition apparatus and method of obtaining color and depth images simultaneously |
KR20140027815A (en) * | 2012-08-27 | 2014-03-07 | 삼성전자주식회사 | 3d image acquisition apparatus and method of obtaining color and depth images simultaneously |
US9451240B2 (en) | 2012-08-27 | 2016-09-20 | Samsung Electronics Co., Ltd. | 3-dimensional image acquisition apparatus and 3D image acquisition method for simultaneously obtaining color image and depth image |
US8997021B2 (en) | 2012-11-06 | 2015-03-31 | Lytro, Inc. | Parallax and/or three-dimensional effects for thumbnail image displays |
US9667846B2 (en) * | 2012-11-27 | 2017-05-30 | Nokia Technologies Oy | Plenoptic camera apparatus, a method and a computer program |
US9001226B1 (en) | 2012-12-04 | 2015-04-07 | Lytro, Inc. | Capturing and relighting images using multiple devices |
US20160057407A1 (en) * | 2013-02-13 | 2016-02-25 | Universität des Saarlandes | Plenoptic imaging device |
US9936188B2 (en) * | 2013-02-13 | 2018-04-03 | Universität des Saarlandes | Plenoptic imaging device |
US9456141B2 (en) | 2013-02-22 | 2016-09-27 | Lytro, Inc. | Light-field based autofocus |
US9235741B2 (en) | 2013-02-28 | 2016-01-12 | Hand Held Products, Inc. | Indicia reading terminals and methods employing light field imaging |
US8978984B2 (en) | 2013-02-28 | 2015-03-17 | Hand Held Products, Inc. | Indicia reading terminals and methods for decoding decodable indicia employing light field imaging |
US10057498B1 (en) * | 2013-03-15 | 2018-08-21 | Cognex Corporation | Light field vision system camera and methods for using the same |
US10334151B2 (en) | 2013-04-22 | 2019-06-25 | Google Llc | Phase detection autofocus using subaperture images |
US10151859B2 (en) | 2013-09-23 | 2018-12-11 | Lg Innotek Co., Ltd. | Camera module and manufacturing method for same |
WO2015041496A1 (en) * | 2013-09-23 | 2015-03-26 | 엘지이노텍 주식회사 | Camera module and manufacturing method for same |
US10397545B2 (en) * | 2013-12-23 | 2019-08-27 | University Of Deleware | 3-D light field camera and photography method |
US9628684B2 (en) | 2013-12-24 | 2017-04-18 | Lytro, Inc. | Light-field aberration correction |
US9392153B2 (en) | 2013-12-24 | 2016-07-12 | Lytro, Inc. | Plenoptic camera resolution |
US20150358529A1 (en) * | 2014-06-04 | 2015-12-10 | Canon Kabushiki Kaisha | Image processing device, its control method, and storage medium |
US9936121B2 (en) * | 2014-06-04 | 2018-04-03 | Canon Kabushiki Kaisha | Image processing device, control method of an image processing device, and storage medium that stores a program to execute a control method of an image processing device |
US8988317B1 (en) | 2014-06-12 | 2015-03-24 | Lytro, Inc. | Depth determination for light field images |
US11911109B2 (en) | 2014-08-31 | 2024-02-27 | Dr. John Berestka | Methods for analyzing the eye |
US10687703B2 (en) | 2014-08-31 | 2020-06-23 | John Berestka | Methods for analyzing the eye |
US10092183B2 (en) | 2014-08-31 | 2018-10-09 | Dr. John Berestka | Systems and methods for analyzing the eye |
US11452447B2 (en) | 2014-08-31 | 2022-09-27 | John Berestka | Methods for analyzing the eye |
US9635332B2 (en) | 2014-09-08 | 2017-04-25 | Lytro, Inc. | Saturated pixel recovery in light-field images |
US10692192B2 (en) * | 2014-10-21 | 2020-06-23 | Connaught Electronics Ltd. | Method for providing image data from a camera system, camera system and motor vehicle |
US9902120B2 (en) * | 2015-02-09 | 2018-02-27 | Omnivision Technologies, Inc. | Wide-angle camera using achromatic doublet prism array and method of manufacturing the same |
US20160231473A1 (en) * | 2015-02-09 | 2016-08-11 | Omnivision Technologies, Inc. | Wide-Angle Camera Using Achromatic Doublet Prism Array And Method of Manufacturing The Same |
US10567464B2 (en) | 2015-04-15 | 2020-02-18 | Google Llc | Video compression with adaptive view-dependent lighting removal |
US10540818B2 (en) | 2015-04-15 | 2020-01-21 | Google Llc | Stereo image generation and interactive playback |
US20190349573A1 (en) * | 2015-04-15 | 2019-11-14 | Google Llc | Image capture for virtual reality displays |
US11328446B2 (en) | 2015-04-15 | 2022-05-10 | Google Llc | Combining light-field data with active depth data for depth map generation |
US10565734B2 (en) | 2015-04-15 | 2020-02-18 | Google Llc | Video capture, processing, calibration, computational fiber artifact removal, and light-field pipeline |
US10951880B2 (en) * | 2015-04-15 | 2021-03-16 | Google Llc | Image capture for virtual reality displays |
US10419737B2 (en) | 2015-04-15 | 2019-09-17 | Google Llc | Data structures and delivery methods for expediting virtual reality playback |
US10341632B2 (en) | 2015-04-15 | 2019-07-02 | Google Llc. | Spatial random access enabled video system with a three-dimensional viewing volume |
US10275898B1 (en) | 2015-04-15 | 2019-04-30 | Google Llc | Wedge-based light-field video capture |
US10469873B2 (en) | 2015-04-15 | 2019-11-05 | Google Llc | Encoding and decoding virtual reality video |
US10546424B2 (en) | 2015-04-15 | 2020-01-28 | Google Llc | Layered content delivery for virtual and augmented reality experiences |
US10412373B2 (en) | 2015-04-15 | 2019-09-10 | Google Llc | Image capture for virtual reality displays |
DE102015006659A1 (en) | 2015-05-22 | 2016-11-24 | Martin Häusler | Plenoptic camera lens |
US10033986B2 (en) | 2015-05-26 | 2018-07-24 | Google Llc | Capturing light-field images with uneven and/or incomplete angular sampling |
US10205896B2 (en) | 2015-07-24 | 2019-02-12 | Google Llc | Automatic lens flare detection and correction for light-field images |
US9809218B2 (en) | 2015-08-24 | 2017-11-07 | Ford Global Technologies, Llc | System and method for autonomous valet parking using plenoptic cameras |
US10046760B2 (en) | 2015-08-24 | 2018-08-14 | Ford Global Technologies, Llc | System and method for autonomous valet parking using plenoptic cameras |
US9557741B1 (en) * | 2015-08-24 | 2017-01-31 | Ford Global Technologies, Llc | System and method for autonomous valet parking using plenoptic cameras |
US10275892B2 (en) | 2016-06-09 | 2019-04-30 | Google Llc | Multi-view scene segmentation and propagation |
US10348947B2 (en) * | 2016-09-07 | 2019-07-09 | Interdigital Ce Patent Holdings | Plenoptic imaging device equipped with an enhanced optical system |
US10832429B2 (en) * | 2016-10-18 | 2020-11-10 | Photonic Sensors & Algorithms, S.L. | Device and method for obtaining distance information from views |
CN106500629A (en) * | 2016-11-29 | 2017-03-15 | 深圳大学 | A kind of microscopic three-dimensional measurement apparatus and system |
CN106500629B (en) * | 2016-11-29 | 2022-09-27 | 深圳大学 | Microscopic three-dimensional measuring device and system |
US10679361B2 (en) | 2016-12-05 | 2020-06-09 | Google Llc | Multi-view rotoscope contour propagation |
US10369988B2 (en) | 2017-01-13 | 2019-08-06 | Ford Global Technologies, Llc | Autonomous parking of vehicles inperpendicular parking spots |
US10594945B2 (en) | 2017-04-03 | 2020-03-17 | Google Llc | Generating dolly zoom effect using light field image data |
US10440407B2 (en) | 2017-05-09 | 2019-10-08 | Google Llc | Adaptive control for immersive experience delivery |
US10474227B2 (en) | 2017-05-09 | 2019-11-12 | Google Llc | Generation of virtual reality with 6 degrees of freedom from limited viewer data |
US10444931B2 (en) | 2017-05-09 | 2019-10-15 | Google Llc | Vantage generation and interactive playback |
US10645281B1 (en) * | 2017-05-19 | 2020-05-05 | Shanghaitech University | Method and system for snapshot multi-spectral light field imaging |
US10354399B2 (en) | 2017-05-25 | 2019-07-16 | Google Llc | Multi-view back-projection to a light-field |
US10683034B2 (en) | 2017-06-06 | 2020-06-16 | Ford Global Technologies, Llc | Vehicle remote parking systems and methods |
US10585430B2 (en) | 2017-06-16 | 2020-03-10 | Ford Global Technologies, Llc | Remote park-assist authentication for vehicles |
US10234868B2 (en) | 2017-06-16 | 2019-03-19 | Ford Global Technologies, Llc | Mobile device initiation of vehicle remote-parking |
US10775781B2 (en) | 2017-06-16 | 2020-09-15 | Ford Global Technologies, Llc | Interface verification for vehicle remote park-assist |
EP3444635A1 (en) * | 2017-08-14 | 2019-02-20 | Valeo Schalter und Sensoren GmbH | Receiver unit for a laser scanner device, laser scanner device, vehicle and method for capturing light |
US10545215B2 (en) | 2017-09-13 | 2020-01-28 | Google Llc | 4D camera tracking and optical stabilization |
US10580304B2 (en) | 2017-10-02 | 2020-03-03 | Ford Global Technologies, Llc | Accelerometer-based external sound monitoring for voice controlled autonomous parking |
US10281921B2 (en) | 2017-10-02 | 2019-05-07 | Ford Global Technologies, Llc | Autonomous parking of vehicles in perpendicular parking spots |
US10627811B2 (en) | 2017-11-07 | 2020-04-21 | Ford Global Technologies, Llc | Audio alerts for remote park-assist tethering |
CN107741644A (en) * | 2017-11-21 | 2018-02-27 | 杭州加速云信息技术有限公司 | A kind of imaging device for different visual angles imaging |
US10336320B2 (en) | 2017-11-22 | 2019-07-02 | Ford Global Technologies, Llc | Monitoring of communication for vehicle remote park-assist |
US10578676B2 (en) | 2017-11-28 | 2020-03-03 | Ford Global Technologies, Llc | Vehicle monitoring of mobile device state-of-charge |
US10585431B2 (en) | 2018-01-02 | 2020-03-10 | Ford Global Technologies, Llc | Mobile device tethering for a remote parking assist system of a vehicle |
US10583830B2 (en) | 2018-01-02 | 2020-03-10 | Ford Global Technologies, Llc | Mobile device tethering for a remote parking assist system of a vehicle |
US10688918B2 (en) | 2018-01-02 | 2020-06-23 | Ford Global Technologies, Llc | Mobile device tethering for a remote parking assist system of a vehicle |
US10814864B2 (en) | 2018-01-02 | 2020-10-27 | Ford Global Technologies, Llc | Mobile device tethering for a remote parking assist system of a vehicle |
US11148661B2 (en) | 2018-01-02 | 2021-10-19 | Ford Global Technologies, Llc | Mobile device tethering for a remote parking assist system of a vehicle |
US10974717B2 (en) | 2018-01-02 | 2021-04-13 | Ford Global Technologies, I.LC | Mobile device tethering for a remote parking assist system of a vehicle |
US10737690B2 (en) | 2018-01-02 | 2020-08-11 | Ford Global Technologies, Llc | Mobile device tethering for a remote parking assist system of a vehicle |
US10684773B2 (en) | 2018-01-03 | 2020-06-16 | Ford Global Technologies, Llc | Mobile device interface for trailer backup-assist |
US10747218B2 (en) | 2018-01-12 | 2020-08-18 | Ford Global Technologies, Llc | Mobile device tethering for remote parking assist |
US10965862B2 (en) | 2018-01-18 | 2021-03-30 | Google Llc | Multi-camera navigation interface |
US10917748B2 (en) | 2018-01-25 | 2021-02-09 | Ford Global Technologies, Llc | Mobile device tethering for vehicle systems based on variable time-of-flight and dead reckoning |
US10684627B2 (en) | 2018-02-06 | 2020-06-16 | Ford Global Technologies, Llc | Accelerometer-based external sound monitoring for position aware autonomous parking |
US11188070B2 (en) | 2018-02-19 | 2021-11-30 | Ford Global Technologies, Llc | Mitigating key fob unavailability for remote parking assist systems |
US10507868B2 (en) | 2018-02-22 | 2019-12-17 | Ford Global Technologies, Llc | Tire pressure monitoring for vehicle park-assist |
GB2585782B (en) * | 2018-03-15 | 2022-11-09 | Photonic Sensors & Algorithms S L | Plenoptic camera for mobile devices |
US11456326B2 (en) * | 2018-03-15 | 2022-09-27 | Photonic Sensors & Algorithms, S.L. | Plenoptic camera for mobile devices |
US10732622B2 (en) | 2018-04-05 | 2020-08-04 | Ford Global Technologies, Llc | Advanced user interaction features for remote park assist |
US10759417B2 (en) | 2018-04-09 | 2020-09-01 | Ford Global Technologies, Llc | Input signal management for vehicle park-assist |
US10683004B2 (en) | 2018-04-09 | 2020-06-16 | Ford Global Technologies, Llc | Input signal management for vehicle park-assist |
US10793144B2 (en) | 2018-04-09 | 2020-10-06 | Ford Global Technologies, Llc | Vehicle remote park-assist communication counters |
US10493981B2 (en) | 2018-04-09 | 2019-12-03 | Ford Global Technologies, Llc | Input signal management for vehicle park-assist |
US10384605B1 (en) | 2018-09-04 | 2019-08-20 | Ford Global Technologies, Llc | Methods and apparatus to facilitate pedestrian detection during remote-controlled maneuvers |
US10821972B2 (en) | 2018-09-13 | 2020-11-03 | Ford Global Technologies, Llc | Vehicle remote parking assist systems and methods |
US10717432B2 (en) | 2018-09-13 | 2020-07-21 | Ford Global Technologies, Llc | Park-assist based on vehicle door open positions |
US10529233B1 (en) | 2018-09-24 | 2020-01-07 | Ford Global Technologies Llc | Vehicle and method for detecting a parking space via a drone |
US10967851B2 (en) | 2018-09-24 | 2021-04-06 | Ford Global Technologies, Llc | Vehicle system and method for setting variable virtual boundary |
US10908603B2 (en) | 2018-10-08 | 2021-02-02 | Ford Global Technologies, Llc | Methods and apparatus to facilitate remote-controlled maneuvers |
US10628687B1 (en) | 2018-10-12 | 2020-04-21 | Ford Global Technologies, Llc | Parking spot identification for vehicle park-assist |
US11097723B2 (en) | 2018-10-17 | 2021-08-24 | Ford Global Technologies, Llc | User interfaces for vehicle remote park assist |
US11137754B2 (en) | 2018-10-24 | 2021-10-05 | Ford Global Technologies, Llc | Intermittent delay mitigation for remote vehicle operation |
US11789442B2 (en) | 2019-02-07 | 2023-10-17 | Ford Global Technologies, Llc | Anomalous input detection |
US11195344B2 (en) | 2019-03-15 | 2021-12-07 | Ford Global Technologies, Llc | High phone BLE or CPU burden detection and notification |
US11169517B2 (en) | 2019-04-01 | 2021-11-09 | Ford Global Technologies, Llc | Initiation of vehicle remote park-assist with key fob |
US11275368B2 (en) | 2019-04-01 | 2022-03-15 | Ford Global Technologies, Llc | Key fobs for vehicle remote park-assist |
WO2021217031A1 (en) * | 2020-04-23 | 2021-10-28 | The Regents Of The University Of California | Ultrafast light field tomography |
Also Published As
Publication number | Publication date |
---|---|
EP2008445B1 (en) | 2017-10-18 |
JP4981124B2 (en) | 2012-07-18 |
CN101455071B (en) | 2013-04-03 |
US20100020187A1 (en) | 2010-01-28 |
WO2007115281A1 (en) | 2007-10-11 |
EP2008445A1 (en) | 2008-12-31 |
US7620309B2 (en) | 2009-11-17 |
JP2009532993A (en) | 2009-09-10 |
US8238738B2 (en) | 2012-08-07 |
CN101455071A (en) | 2009-06-10 |
KR20090016453A (en) | 2009-02-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US7620309B2 (en) | Plenoptic camera | |
US7956924B2 (en) | Fast computational camera based on two arrays of lenses | |
US8400555B1 (en) | Focused plenoptic camera employing microlenses with different focal lengths | |
US8947578B2 (en) | Apparatus and method of capturing image | |
US8380060B2 (en) | Methods and apparatus for full-resolution light-field capture and rendering | |
US10044919B2 (en) | Structures and methods for capturing images by a portable electronic device | |
US8265478B1 (en) | Plenoptic camera with large depth of field | |
US8379105B2 (en) | Methods and apparatus for full-resolution light-field capture and rendering | |
US8130310B2 (en) | Image pickup apparatus | |
Galstian | Smart mini-cameras | |
CN108432230B (en) | Imaging device and method for displaying an image of a scene | |
US8908054B1 (en) | Optics apparatus for hands-free focus | |
US10356349B2 (en) | Light field capture control methods and apparatuses, light field capture devices | |
Liang et al. | Light field acquisition using programmable aperture camera | |
US8953899B2 (en) | Method and system for rendering an image from a light-field camera | |
US20050254817A1 (en) | Autostereoscopic electronic camera | |
Oberdörster et al. | Digital focusing and refocusing with thin multi-aperture cameras | |
Sahin et al. | Light L16 computational camera | |
US20240111203A1 (en) | Imaging method and device for autofocusing | |
JP2013258449A (en) | Imaging apparatus and control method of the same |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: ADOBE SYSTEMS, INCORPORATED, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:GEORGIEV, TODOR G.;REEL/FRAME:017725/0887 Effective date: 20060404 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
CC | Certificate of correction | ||
FPAY | Fee payment |
Year of fee payment: 4 |
|
AS | Assignment |
Owner name: ADOBE SYSTEMS INCORPORATED, CALIFORNIA Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE ASSIGNEE NAME PREVIOUSLY RECORDED AT REEL: 017725 FRAME: 0887. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT;ASSIGNOR:GEORGIEV, TODOR G.;REEL/FRAME:038319/0024 Effective date: 20060404 |
|
FPAY | Fee payment |
Year of fee payment: 8 |
|
AS | Assignment |
Owner name: ADOBE INC., CALIFORNIA Free format text: CHANGE OF NAME;ASSIGNOR:ADOBE SYSTEMS INCORPORATED;REEL/FRAME:048867/0882 Effective date: 20181008 |
|
FEPP | Fee payment procedure |
Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
LAPS | Lapse for failure to pay maintenance fees |
Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20211117 |