The present disclosure relates to imaging systems and methods that include a multi-camera array. In particular, the disclosure relates to systems and methods that enable low-profile imaging systems of mobile devices while maintaining or improving image quality.
Many mobile devices, such as mobile phones and tablet computing devices, include cameras that may be operated by a user to capture still and/or video images. Because the mobile devices are typically designed to be relatively small, it can be important to design the cameras or imaging systems to be as thin as possible in order to maintain a low-profile mobile device. Folded optic image sensor arrays (“array cameras”) allow for the creation of low-profile image capture devices without shortening the focal length or decreasing the resolution of the image across the sensor array's field of view. By redirecting light toward each sensor in the array using a primary and secondary surface, and by positioning the lens assemblies used to focus the incoming light between the primary and secondary surfaces, the sensor array may be positioned on a flat substrate perpendicular to the lens assemblies. The longer focal length makes it possible to implement features such as optical zoom and to incorporate more complicated optics that require more space than commonly afforded by the traditional mobile camera, such as adding more optical elements.
Some array cameras employ a central mirror or prism with multiple facets to split incoming light comprising the target image into multiple portions for capture by the sensors in the array, wherein each facet directs a portion of the light from the target image toward a sensor in the array. Each portion of the split light may be passed through a lens assembly and reflected off of an additional surface positioned directly above or below a sensor, such that each sensor captures a portion of the image. The sensor fields of view can overlap to assist in stitching together the captured portions into a complete image.
The folded optic sensor arrays and image capture techniques described herein allow for the creation of low-profile image capture devices without shortening the focal length or decreasing the resolution of the image across the sensor array's field of view, and the captured images are free (or nearly so) of parallax artifacts. A challenge of array cameras is the quality degradation due to parallax between different views of same object as seen from different cameras of the array. Parallax prevents seamless stitching of the images captured by each camera into a final image completely free of artifacts. Camera views can partially overlap (for example by approximately 20%). Depending on the depth of the scene of in an image (e.g., distance from lens to object(s)) the image from one camera can be shifted relative to the image from another camera. The resulting parallax and tilt can cause “double image” ghosting in the image area corresponding to the overlapping fields of view when the images are stitched or fused together. Even if the array is structured such that there is no overlap in sensor fields of view, parallax results in discontinuous features in the image, such as lines and edges, when such features cross over the borders between sensor fields of view.
The above-described problems, among others, are addressed in various embodiments by the array cameras free of parallax and tilt artifacts as described herein. Some of the embodiments may employ a central mirror or prism, for example with multiple surfaces or facets, to split incoming light comprising the target image into multiple portions for capture by the sensors in the array. The mirror surfaces and sensors can be configured to avoid causing parallax and tilt artifacts in a captured image. For example, the planes of the central mirror surfaces of the array camera can be located at a midpoint along, and orthogonal to, a line between the corresponding camera location and the virtual camera location. Accordingly, the projected field-of-view (FOV) cones of all of the cameras in the array appear as if coming from the virtual camera location after folding of the incoming light by the mirrors.
Each portion of the split light may be passed through a lens assembly and reflected off of an optionally additional reflective surface positioned directly above or below a sensor, such that each sensor captures a portion of the image. In some circumstances, each sensor in the array may capture a portion of the image which overlaps slightly with the portions captured by neighboring sensors in the array, and these portions may be assembled into the target image, for example by linear blending or other image stitching techniques.
One innovation includes a system for capturing a target image of a scene, the system including an array of a plurality of cameras having a virtual center of projection, each of the plurality of cameras configured to capture one of a plurality of portions of a target image of a scene, and each of the plurality of cameras including an image sensor, a lens assembly including at least one lens, the lens assembly having a center of projection, the lens assembly positioned to focus light on the image sensor, the center of projection of the lens assembly located along a line passing through the virtual center of projection, and a mirror located with respect to the lens assembly to provide light to the lens assembly, the mirror further positioned on (or within) a mirror plane, (sometime referred herein as a “mirror plane”), the mirror plane positioned to intersect a point (for example, a midpoint) long the line passing through the virtual center of projection. The mirror plan may be positioned an angle orthogonal to the line. Various embodiments may include additional features.
The following are non-limiting examples of some features and embodiments of such systems. For example, the system may further include a central reflective element (for example, a pyramid-shaped reflective component) including a plurality of primary light re-directing surfaces configured to split light representing the target image of the scene into the plurality of portions, the mirror of each of the plurality of cameras forming one of the primary light folding surfaces. The central reflective element may include an apex formed by an intersection of each of the plurality of primary light re-directing surfaces. An optical axis of each of the plurality of cameras may pass through the apex. The apex and the virtual center of projection are located on a virtual optical axis passing through the apex and the virtual center of projection such that the virtual optical axis forms an optical axis of the array of the plurality of cameras. The system may further include a camera housing comprising at least an upper surface having an aperture positioned to allow light representing the scene to pass through the aperture to the central reflective element. The upper surface may be positioned orthogonal to the virtual optical axis at or above the apex of the central reflective element. The housing may further include a lower surface positioned substantially parallel to the upper surface and positioned at or below a lower surface of the central reflective element. Each of the plurality of cameras may be positioned between the upper surface and the lower surface of the camera housing. The system may further include a processor configured to receive image data comprising an image of a portion of the scene from each of the plurality of cameras and to perform a stitching operation to generate the target image. The system may further (for example, to correct tilt artifacts in the image data) include a processor that is further configured to perform a projective transform on the image data based at least partly a geometric relationship between the virtual center of projection and the center of projection of the lens assembly of each of the plurality of cameras and the location of the mirror for each of the plurality of cameras within a corresponding mirror plane. The system may further include a secondary light re-directing surface associated with each of the plurality of cameras, the secondary light re-directing surface positioned to direct light received from the lens assembly onto the image sensor. The system may further include a substrate having at least one aperture, the at least one aperture positioned to allow light representing the scene to pass through the at least one aperture to the mirror of each of the plurality of cameras, wherein the image sensor for each of the plurality of cameras is positioned on or within the substrate.
Another innovation includes a method of manufacturing an array camera substantially free of parallax artifacts in a captured image. In various embodiments, the method includes positioning a plurality of cameras in an array, each of the plurality of cameras configured to capture one of a plurality of portions of a target image scene, the plurality of cameras positioned to each capture image data from a location of a virtual camera having a virtual center of projection, each of the plurality of cameras including a lens assembly having at least one lens, a camera center of projection having a location determined at least partly by optics of the lens assembly, and a sensor positioned to receive light from the lens assembly. The method may further include providing, for each of the plurality of cameras, a mirror positioned on a mirror plane, the mirror plane positioned to intersect a point (for example, a midpoint) along a line connecting a center of projection of the camera and captured virtual center of projection. The mirror plane may be positioned orthogonally to the line.
The following are non-limiting examples of some features and embodiments of such method. For example, the method may further include providing the mirror for each of the plurality of cameras as a facet of a central reflective element, the plurality of cameras positioned around the central reflective element or reflective component. The method may include providing the central reflective element such that the mirror for each of the plurality of cameras intersects at an apex of the central reflective element. The method may include positioning each of the plurality of cameras such that an optical axis of each of the plurality of cameras intersects with the apex. The method may include providing a secondary light folding surface for each of the plurality of cameras. The method may include positioning the secondary light folding surface between the lens assembly and the sensor to direct light received from the lens assembly onto the sensor. The method may include providing a substrate having at least one aperture positioned to allow light representing the scene to pass through the aperture to the mirror for each of the plurality of cameras. The method may include mounting the sensor for each of the plurality of cameras on or within the substrate such that all sensors are positioned on (or within) the same plane. The method may include providing a processor in electronic communication with the sensor of each of the plurality of cameras, the processor configured to receive the plurality of portions of a target image scene and generate a complete image of the target image scene.
Another innovation includes a method of capturing an image free (or substantially free) of parallax, the method including splitting light representing a scene of the image into a plurality of portions using a plurality of mirrors, directing each of the plurality of portions toward a corresponding one of a plurality of cameras each positioned to capture image data from a location of a virtual camera having a virtual center of projection, each of the plurality of cameras having a lens assembly comprising at least one lens, a camera center of projection having a location determined at least partly by optics of the lens assembly, and a sensor positioned to receive light from the lens assembly, and assembling the plurality of portions into the image, where for each mirror of the plurality of mirrors, the mirror is positioned on a mirror plane. The mirror plane is positioned to intersect a point (for example, a midpoint) along a line connecting the camera center of projection of the corresponding camera of the plurality of cameras and the virtual center of projection. The mirror plane may further be orthogonal to the line.
The following are non-limiting examples of some features and embodiments of such a method. The method may include applying projective transform to captured image data to change an effective tilt of each of the plurality of cameras. The method may include, for each of the plurality of cameras, redirecting the light received from the lens assembly onto the image sensor using a secondary light folding surface.
Another innovation includes an apparatus for capturing a target image of a scene, the apparatus including an array of a plurality of cameras having a virtual center of projection, each of the plurality of cameras configured to capture one of a plurality of portions of the target image of the scene, and for each of the plurality of cameras, means for capturing an image, means for focusing light having a center of projection located along a line passing through the virtual center of projection, and means for redirecting light located at least partially within a primary light folding plane positioned to intersect a midpoint along the line, and at an angle orthogonal to the line.
The following are non-limiting examples of some features and embodiments of such an apparatus. In some embodiments, the light redirecting means includes one of a reflective surface positioned within the primary light folding plane and a refractive prism having a facet positioned within the primary light folding plane. The light focusing means may include a lens assembly including one or more lenses. each of the plurality of cameras may include an additional light folding means positioned to direct light received from the light focusing means onto the image capture means. The apparatus may include means for combining the plurality of portions into the target image. The apparatus may include means for compensating for tilt artifacts between the plurality of portions of the target image.
The disclosed aspects will hereinafter be described in conjunction with the appended drawings, provided to illustrate and not to limit the disclosed aspects, wherein like designations denote like elements.
Implementations disclosed herein provide systems, methods and apparatus for generating images substantially free of parallax artifacts using an array camera with folded optics. Aspects of the present disclosure relate to an array camera exhibiting little or no parallax artifacts in the captured images. For example, the planes of the central mirror surfaces of the array camera can be located at a midpoint along, and orthogonally to, a line between the corresponding camera location and the virtual camera location. Accordingly, the cones of all of the cameras in the array appear as if coming from the virtual camera location after folding by the mirrors. Each sensor in the array “sees” a portion of the image scene using a corresponding facet of the central mirror prism, and accordingly each individual sensor/mirror pair represents only a sub-aperture of the total array camera. The complete array camera has a synthetic aperture generated based on the sum of all individual aperture rays, that is, based on stitching together the images generated by the sub-apertures.
In the following description, specific details are given to provide a thorough understanding of the examples. However, the examples may be practiced without these specific details.
Referring now to
Referring to
The sensors 105, 125 may be mounted on the substrate 150 as shown in
In some embodiments, a central reflective surface 120 may be used to redirect light from a target image scene toward the sensors 105, 125. Central reflective surface 120 may be a mirror or a plurality of mirrors, and may be flat or shaped as needed to properly redirect incoming light to the image sensors 105, 125. For example, in some embodiments, central reflective surface 120 may be a mirror sized and shaped to reflect incoming light rays through the lens assemblies 115, 130 to sensors 105, 125. The central reflective surface 120 may split light comprising the target image into multiple portions and direct each portion at a different sensor. For example, a first side 122 of the central reflective surface 120 (also referred to as a primary light folding surface, as other embodiments may implement a refractive prism rather than a reflective surface) may send a portion of the light corresponding to a first field of view (FOV) 140 toward the left sensor 105 while a second side 124 sends a second portion of the light corresponding to a second FOV 145 toward the right sensor 125. It should be appreciated that together the fields of view 140, 145 of the image sensors cover at least the target image. The present example describes an embodiment comprising two sensors, but other embodiments may have greater than two sensors, for example 2, 3, 4, 8 or more (N) sensors.
In some embodiments in which the receiving sensors are each an array of a plurality of sensors, the central reflective surface may be made of multiple reflective surfaces angled relative to one another in order to send a different portion of the target image scene toward each of the sensors. Each sensor in the array may have a substantially different field of view, and in some embodiments the fields of view may overlap. Certain embodiments of the central reflective surface may have complicated non-planar surfaces to increase the degrees of freedom when designing the lens system. Further, although the central surface is discussed as being a reflective surface, in other embodiments the central surface may be refractive. For example, the central surface may be a prism configured with a plurality of facets, where each facet directs a portion of the light comprising the scene toward one of the sensors.
After being reflected off the central reflective surface 120, at least a portion of incoming light may propagate through each of the lens assemblies 115, 130. One or more lens assemblies 115, 130 may be provided between the central reflective surface 120 and the sensors 105, 125 and reflective surfaces 110, 135. The lens assemblies 115, 130 may be used to focus the portion of the target image which is directed toward each sensor.
In some embodiments, each lens assembly may comprise one or more lenses and an actuator for moving the lens among a plurality of different lens positions through a housing. The actuator may, for example, be a voice coil motor (VCM), micro-electronic mechanical system (MEMS), or a shape memory alloy (SMA). The lens assembly may further comprise a lens driver for controlling the actuator.
In some embodiments, traditional auto focus techniques may be implemented by changing the focal length between the lens 115, 130 and corresponding sensor 105, 125 of each camera. In some embodiments, this may be accomplished by moving a lens barrel. Other embodiments may adjust the focus by moving the central mirror up or down or by adjusting the angle of the mirror relative to the lens assembly. Certain embodiments may adjust the focus by moving the side mirrors over each sensor. Such embodiments may allow the assembly to adjust the focus of each sensor individually. Further, it is possible for some embodiments to change the focus of the entire assembly at once, for example by placing a lens, like a liquid lens, over the entire assembly. In certain implementations, computational photography may be used to change the focal point of the camera array.
Multiple side reflective surfaces, for example, reflective surfaces 110 and 135, can be provided around the central mirror 120 opposite the sensors. After passing through the lens assemblies, the side reflective surfaces 110, 135 (also referred to as a secondary light folding surface, as other embodiments may implement a refractive prism rather than a reflective surface) can reflect the light (downward, as depicted in the orientation of
Each sensor's FOV 140, 145 may be steered into the object space by the surface of the central mirror 120 associated with that sensor. Mechanical methods may be employed to tilt the mirrors and/or move the prisms in the array so that the FOV of each camera can be steered to different locations on the object field. This may be used, for example, to implement a high dynamic range camera, to increase the resolution of the camera system, or to implement a plenoptic camera system. Each sensor's (or each 3×1 array's) FOV may be projected into the object space, and each sensor may capture a partial image comprising a portion of the target scene according to that sensor's field of view. As illustrated in
The sensors 105, 125 may be mounted on the substrate 150 as shown in
Primary light folding surfaces 122, 124 may be prism surfaces as illustrated, or may be a mirror or a plurality of mirrors, and may be flat or shaped as needed to properly redirect incoming light to the image sensors 105, 125. In some embodiments the primary light folding surfaces 122, 124 may be formed as a central reflective element, as illustrated in
Other embodiments may combine the reflective and refractive elements illustrated by
As illustrated by
As used herein, the term “camera” refers to an image sensor, lens system, and a number of corresponding light folding surfaces, for example the primary light folding surface 124, lens assembly 130, secondary light folding surface 135, and sensor 125 as illustrated in
Device 200 may be a cell phone, digital camera, tablet computer, personal digital assistant, or the like. There are many portable computing devices in which a reduced thickness imaging system such as is described herein would provide advantages. Device 200 may also be a stationary computing device or any device in which a thin imaging system would be advantageous. A plurality of applications may be available to the user on device 200. These applications may include traditional photographic and video applications, high dynamic range imaging, panoramic photo and video, or stereoscopic imaging such as 3D images or 3D video.
The image capture device 200 includes the cameras 215a-n for capturing external images. The cameras 215a-n may each comprise a sensor, lens assembly, and a primary and secondary reflective or refractive surface for redirecting a portion of a target image to each sensor, as discussed above with respect to
The image processor 220 may be configured to perform various processing operations on received image data comprising N portions of the target image in order to output a high quality stitched image, as will be described in more detail below. Image processor 220 may be a general purpose processing unit or a processor specially designed for imaging applications. Examples of image processing operations include cropping, scaling (e.g., to a different resolution), image stitching, image format conversion, color interpolation, color processing, image filtering (for example, spatial image filtering), lens artifact or defect correction, etc. Image processor 220 may, in some embodiments, comprise a plurality of processors. Certain embodiments may have a processor dedicated to each image sensor. Image processor 220 may be one or more dedicated image signal processors (ISPs) or a software implementation of a processor.
As shown, the image processor 220 is connected to a memory 230 and a working memory 205. In the illustrated embodiment, the memory 230 stores capture control module 235, image stitching module 240, and operating system 245. These modules include instructions that configure the image processor 220 of device processor 250 to perform various image processing and device management tasks. Working memory 205 may be used by image processor 220 to store a working set of processor instructions contained in the modules of memory component 230. Alternatively, working memory 205 may also be used by image processor 220 to store dynamic data created during the operation of device 200.
As mentioned above, the image processor 220 is configured by several modules stored in the memories. The capture control module 235 may include instructions that configure the image processor 220 to adjust the focus position of cameras 215a-n. Capture control module 235 may further include instructions that control the overall image capture functions of the device 200. For example, capture control module 235 may include instructions that call subroutines to configure the image processor 220 to capture raw image data of a target image scene using the cameras 215a-n. Capture control module 235 may then call the image stitching module 240 to perform a stitching technique on the N partial images captured by the cameras 215a-n and output a stitched and cropped target image to imaging processor 220. Capture control module 235 may also call the image stitching module 240 to perform a stitching operation on raw image data in order to output a preview image of a scene to be captured, and to update the preview image at certain time intervals or when the scene in the raw image data changes.
Image stitching module 240 may comprise instructions that configure the image processor 220 to perform stitching and cropping techniques on captured image data. For example, each of the N sensors 215a-n may capture a partial image comprising a portion of the target image according to each sensor's field of view. The fields of view may share areas of overlap, as described above and below. In order to output a single target image, image stitching module 240 may configure the image processor 220 to combine the multiple N partial images to produce a high-resolution target image. Target image generation may occur through known image stitching techniques.
For instance, image stitching module 240 may include instructions to compare the areas of overlap along the edges of the N partial images for matching features in order to determine rotation and alignment of the N partial images relative to one another. Due to rotation of partial images and/or the shape of the FOV of each sensor, the combined image may form an irregular shape. Therefore, after aligning and combining the N partial images, the image stitching module 240 may call subroutines which configure image processor 220 to crop the combined image to a desired shape and aspect ratio, for example a 4:3 rectangle or 1:1 square. The cropped image may be sent to the device processor 250 for display on the display 225 or for saving in the storage 210.
Operating system module 245 configures the image processor 220 to manage the working memory 205 and the processing resources of device 200. For example, operating system module 245 may include device drivers to manage hardware resources such as the cameras 215a-n. Therefore, in some embodiments, instructions contained in the image processing modules discussed above may not interact with these hardware resources directly, but instead interact through standard subroutines or APIs located in operating system component 270. Instructions within operating system 245 may then interact directly with these hardware components. Operating system module 245 may further configure the image processor 220 to share information with device processor 250.
Device processor 250 may be configured to control the display 225 to display the captured image, or a preview of the captured image, to a user. The display 225 may be external to the imaging device 200 or may be part of the imaging device 200. The display 225 may also be configured to provide a view finder displaying a preview image for a use prior to capturing an image, or may be configured to display a captured image stored in memory or recently captured by the user. The display 225 may comprise an LCD or LED screen, and may implement touch sensitive technologies.
Device processor 250 may write data to storage module 210, for example data representing captured images. While storage module 210 is represented graphically as a traditional disk device, those with skill in the art would understand that the storage module 210 may be configured as any storage media device. For example, the storage module 210 may include a disk drive, such as a floppy disk drive, hard disk drive, optical disk drive or magneto-optical disk drive, or a solid state memory such as a FLASH memory, RAM, ROM, and/or EEPROM. The storage module 210 can also include multiple memory units, and any one of the memory units may be configured to be within the image capture device 200, or may be external to the image capture device 200. For example, the storage module 210 may include a ROM memory containing system program instructions stored within the image capture device 200. The storage module 210 may also include memory cards or high speed memories configured to store captured images which may be removable from the camera.
Although
Additionally, although
A virtual camera is the location from which an image captured by a camera, for example the image captured by 310A, appears to have been captured if no optical folding (e.g., reflection using mirrors) had been used. In a parallax free array physically formed by multiple folded optic cameras, all virtual cameras are merged into one single virtual camera having the same center of projection, shown in
In
Virtual camera V and cameras 310A, 310D are illustrated as a block diagram centered on the point representing the center of projection for each of virtual camera V and cameras 310A, 310D, respectively. As illustrated in the embodiment of
However, the direction in which a camera is looking—the alignment of its optical axis—can be a design parameter that may be varied while still complying with the parallax free spatial relationships described herein. There are tradeoffs that will be considered in specific designs, for example the direction of gaze of each camera and the location of each camera's center of projection (for example, its distance from the central point P and), each camera's focal length, number of cameras in the array, and whether the array should be symmetric. This generality of the parallax free design principles of
Accordingly, though optical axis 311A of
The parallax-free design principle as illustrated in
In the embodiment 300A of
The embodiment 300B of
As such, a reflecting reflective element for the wide FOV embodiment 300B is inverted in comparison with a reflecting reflective element for the embodiment 300A. Further, while the array 300A “sees” an image located along optical axis 321, the array 300B “sees” a 180 degree panoramic image of a space located circumferentially around axis 321 but does not see the central image scene that is seen by array 300A. In some embodiments, the designs 300A and 300B can be combined to form a camera that captures a full hemispheric field of view.
In order to design or construct an array camera free of parallax according to the parallax-free design principle illustrated in
In order to design or construct an array camera free of parallax according to the parallax-free design principle illustrated in
Although each plane 312A-312F is illustrated with a boundary, in theory a plane is infinite, and the designer can choose the appropriate size and location of a mirror within its corresponding plane according to the needs of the particular camera construction, for example to prevent overlapping mirrors, obstruction of a camera's field of view, etc. Reflections from those mirrors send each center of projection (in this case of cameras 310A-310F) to the same virtual canter of projection V of the thin camera 300A, 300B. This ensures no or substantially no parallax in the resulting stitched image between the fields of view of the cameras.
In some embodiments, the mirrors located within planes 312A-312C and 312D-312F can be arranged to form a reflective element. For example, planes 312A-312C can be extended outward from the illustrated boundaries until they intersect to form a reflective element with an apex at or near point P having mirrored facets sloping downward and outward. As another example, planes 312D-312F can be extended outward from the illustrated boundaries until they intersect to form a reflective element with an apex at or near point P having mirrored facets sloping upward and outward. As such, a reflecting reflective element for the embodiment 300B is inverted in comparison with a reflecting reflective element for the embodiment 300A. Accordingly, mirrors do not have to be positioned in the locations shown for planes 312A-312F and the actual mirrors do not have to be located along the lines 313A-313F or containing midpoints 314A-314F, but rather can be located anywhere in the infinitely extended plane.
The examples illustrated in
In some thin camera implementations, all cameras of the array may be positioned in the horizontal plane of (or physically inside the) thin camera module or camera housing. However, in other embodiments the cameras do not have to be positioned in the horizontal plane 305. For example, the cameras do not have to be positioned in the horizontal plane for a spherical camera, or for a free “3D cloud of centers” camera arrangement. In some array camera embodiments, the arrangement of the cameras may be symmetric, or substantially symmetric, so that all mirror planes have one common point of intersection. If such point exists, it is denoted P in the figures, as in
If symmetry is not sufficient, the array camera design, while producing parallax free images, may still show incorrect defocus blur artifacts. A defocus blur kernel can be created, in order to compensate at least partially for the defocus blur, from the common aperture that is synthesized from all camera apertures/mirrors. If there are gaps and irregularities at certain locations or directions in the view from the location of V, blur will appear unnatural. This is due to the fact that real individual cameras are not really points, but apertures each representing an area of a surface with orientation in 3D space. However, designs with insufficient symmetry may still be used for capturing good quality images of scenes that are almost completely in focus, or if there are no objects particularly close to the array camera, for example, from approximately 20 cm to approximately 50 cm from the camera, in some implementations.
The location of the mirror planes 312D-312F (illustrated in
The location of the mirror planes 312D-312F (illustrated in
As illustrated, there is overlap between the viewing areas 620A-620D due to skewing of the FOV of each camera 610A-610D. The illustrated four-camera array camera 600 was designed without significant overlap between the FOV of the cameras, however this can be achieved in one embodiment by rotating adjacent cameras about their optical axis, as the optical axes of the cameras reside in the same plane. For the adjacent cameras overlap exists in part due to skewing of the FOV of the cameras. This overlap can be larger than shown and can be skewed, as the overlapping area of the left camera is actually reflected from the mirror surface of the right camera and vice versa.
In some embodiments, overlap can be created between adjacent cameras by rotating them toward the virtual optical axis until they are back in an un-rotated position. However, placing the cameras parallel to each other can cause overlap to increase with increasing object distance while the non-overlapping area will not increase at increasing object distance (the percentage overlap will approach 100%), and therefore may not be a workable solution in certain embodiments. Rotating the cameras slightly toward the centerline to create some overlap can result in an overlap of approximately 10 degrees in each direction. However, depending on the size and relative positioning of each camera as well as the height constraints of the array, there may not be enough space available to place two rotated cameras side by side.
The process 900 then transitions to block 915, in which light comprising a target image of a scene is reflected off of the at least one reflective surface toward the imaging sensors. For example, a portion of the light may be refracted by or reflected off of each of a plurality of surfaces toward each of the plurality of sensors. This block may further comprise passing the light through a lens assembly associated with each sensor, and may also include reflecting the light off of a second surface onto a sensor. Block 915 may further comprise focusing the light using the lens assembly or through movement of any of the reflective surfaces.
The process 900 may then move to block 920, in which the sensors capture a plurality of images of the target image scene. For example, each sensor may capture an image of a portion of the scene corresponding to that sensor's field of view. Together, the fields of view of the plurality of sensors cover at least the target image in the object space. Though not illustrated, projective transforms can be applied to some or all of the captured images in order to digitally rotate the optical axis of the camera used to capture the image.
The process 900 then may transition to block 925 in which an image stitching method is performed to generate a single image from the plurality of images. In some embodiments, the image stitching module 240 of
Next, the process 900 transitions to block 930 in which the stitched image is cropped to a specified aspect ratio, for example 4:3 or 1:1. Finally, the process ends after storing the cropped image at block 935. For example, the image may be stored in storage 210 of
Implementations disclosed herein provide systems, methods and apparatus for multiple aperture array cameras free from parallax and tilt artifacts. One skilled in the art will recognize that these embodiments may be implemented in hardware, software, firmware, or any combination thereof.
In some embodiments, the circuits, processes, and systems discussed above may be utilized in a wireless communication device. The wireless communication device may be a kind of electronic device used to wirelessly communicate with other electronic devices. Examples of wireless communication devices include cellular telephones, smart phones, Personal Digital Assistants (PDAs), e-readers, gaming systems, music players, netbooks, wireless modems, laptop computers, tablet devices, etc.
The wireless communication device may include one or more image sensors, two or more image signal processors, a memory including instructions or modules for carrying out the processes discussed above. The device may also have data, a processor loading instructions and/or data from memory, one or more communication interfaces, one or more input devices, one or more output devices such as a display device and a power source/interface. The wireless communication device may additionally include a transmitter and a receiver. The transmitter and receiver may be jointly referred to as a transceiver. The transceiver may be coupled to one or more antennas for transmitting and/or receiving wireless signals.
The wireless communication device may wirelessly connect to another electronic device (e.g., base station). A wireless communication device may alternatively be referred to as a mobile device, a mobile station, a subscriber station, a user equipment (UE), a remote station, an access terminal, a mobile terminal, a terminal, a user terminal, a subscriber unit, etc. Examples of wireless communication devices include laptop or desktop computers, cellular phones, smart phones, wireless modems, e-readers, tablet devices, gaming systems, etc. Wireless communication devices may operate in accordance with one or more industry standards such as the 3rd Generation Partnership Project (3GPP). Thus, the general term “wireless communication device” may include wireless communication devices described with varying nomenclatures according to industry standards (e.g., access terminal, user equipment (UE), remote terminal, etc.).
The functions described herein may be stored as one or more instructions on a processor-readable or computer-readable medium. The term “computer-readable medium” refers to any available medium that can be accessed by a computer or processor. By way of example, and not limitation, such a medium may comprise RAM, ROM, EEPROM, flash memory, CD-ROM or other optical disk storage, magnetic disk storage or other magnetic storage devices, or any other medium that can be used to store desired program code in the form of instructions or data structures and that can be accessed by a computer. Disk and disc, as used herein, includes compact disc (CD), laser disc, optical disc, digital versatile disc (DVD), floppy disk and Blu-ray® disc where disks usually reproduce data magnetically, while discs reproduce data optically with lasers. It should be noted that a computer-readable medium may be tangible and non-transitory. The term “computer-program product” refers to a computing device or processor in combination with code or instructions (e.g., a “program”) that may be executed, processed or computed by the computing device or processor. As used herein, the term “code” may refer to software, instructions, code or data that is/are executable by a computing device or processor.
A transmission (or communication) means may be used to communicate between two devices. For example, if information is made from a website, server, or other remote source using a coaxial cable, fiber optic cable, twisted pair, digital subscriber line (DSL), or wireless technologies such as infrared, radio, and microwave, then the coaxial cable, fiber optic cable, twisted pair, DSL, or wireless technologies such as infrared, radio, and microwave are included in the definition of a transmission (or communication) means.
The methods disclosed herein comprise one or more steps or actions for achieving the described method. The method steps and/or actions may be interchanged with one another without departing from the scope of the claims. In other words, unless a specific order of steps or actions is required for proper operation of the method that is being described, the order and/or use of specific steps and/or actions may be modified without departing from the scope of the claims.
It should be noted that the terms “couple,” “coupling,” “coupled” or other variations of the word couple as used herein may indicate either an indirect connection or a direct connection. For example, if a first component is “coupled” to a second component, the first component may be either indirectly connected to the second component or directly connected to the second component. As used herein, the term “plurality” denotes two or more. For example, a plurality of components indicates two or more components.
The term “determining” encompasses a wide variety of actions and, therefore, “determining” can include calculating, computing, processing, deriving, investigating, looking up (e.g., looking up in a table, a database or another data structure), ascertaining and the like. Also, “determining” can include receiving (e.g., receiving information), accessing (e.g., accessing data in a memory) and the like. Also, “determining” can include resolving, selecting, choosing, establishing and the like.
The phrase “based on” does not mean “based only on,” unless expressly specified otherwise. In other words, the phrase “based on” describes both “based only on” and “based at least on.”
In the foregoing description, specific details are given to provide a thorough understanding of the examples. However, it will be understood by one of ordinary skill in the art that the examples may be practiced without these specific details. For example, electrical components/devices may be shown in block diagrams in order not to obscure the examples in unnecessary detail. In other instances, such components, other structures and techniques may be shown in detail to further explain the examples.
Headings are included herein for reference and to aid in locating various sections. These headings are not intended to limit the scope of the concepts described with respect thereto. Such concepts may have applicability throughout the entire specification.
It is also noted that the examples may be described as a process, which is depicted as a flowchart, a flow diagram, a finite state diagram, a structure diagram, or a block diagram. Although a flowchart may describe the operations as a sequential process, many of the operations can be performed in parallel, or concurrently, and the process can be repeated. In addition, the order of the operations may be re-arranged. A process is terminated when its operations are completed. A process may correspond to a method, a function, a procedure, a subroutine, a subprogram, etc. When a process corresponds to a software function, its termination corresponds to a return of the function to the calling function or the main function.
The previous description of the disclosed implementations is provided to enable any person skilled in the art to make or use the present disclosure. Various modifications to these implementations will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other implementations without departing from the spirit or scope of the disclosure. Thus, the present disclosure is not intended to be limited to the implementations shown herein but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.
The present application claims the benefit under 35 U.S.C. §119(e) of U.S. Provisional Patent Application No. 62/015,316, filed on Jun. 20, 2014, entitled “MULTI-CAMERA SYSTEM USING FOLDED OPTICS FREE FROM PARALLAX ARTIFACTS,” the contents of which is hereby incorporated by reference herein in its entirety.
Number | Name | Date | Kind |
---|---|---|---|
4114171 | Altman | Sep 1978 | A |
4437745 | Hajnal | Mar 1984 | A |
4639586 | Fender et al. | Jan 1987 | A |
4740780 | Brown et al. | Apr 1988 | A |
4751570 | Robinson | Jun 1988 | A |
5012273 | Nakamura et al. | Apr 1991 | A |
5016109 | Gaylord | May 1991 | A |
5063441 | Lipton et al. | Nov 1991 | A |
5142357 | Lipton et al. | Aug 1992 | A |
5194959 | Kaneko et al. | Mar 1993 | A |
5207000 | Chang et al. | May 1993 | A |
5231461 | Silvergate et al. | Jul 1993 | A |
5243413 | Gitlin et al. | Sep 1993 | A |
5313542 | Castonguay | May 1994 | A |
5475617 | Castonguay | Dec 1995 | A |
5539483 | Nalwa | Jul 1996 | A |
5606627 | Kuo | Feb 1997 | A |
5614941 | Hines | Mar 1997 | A |
5640222 | Paul | Jun 1997 | A |
5642299 | Hardin et al. | Jun 1997 | A |
5686960 | Sussman et al. | Nov 1997 | A |
5721585 | Keast et al. | Feb 1998 | A |
5734507 | Harvey | Mar 1998 | A |
5745305 | Nalwa | Apr 1998 | A |
5793527 | Nalwa | Aug 1998 | A |
5903306 | Heckendorn et al. | May 1999 | A |
5926411 | Russell | Jul 1999 | A |
5990934 | Nalwa | Nov 1999 | A |
6111702 | Nalwa | Aug 2000 | A |
6115176 | Nalwa | Sep 2000 | A |
6128143 | Nalwa | Oct 2000 | A |
6141145 | Nalwa | Oct 2000 | A |
6144501 | Nalwa | Nov 2000 | A |
6195204 | Nalwa | Feb 2001 | B1 |
6219090 | Nalwa | Apr 2001 | B1 |
6285365 | Nalwa | Sep 2001 | B1 |
6356397 | Nalwa | Mar 2002 | B1 |
6611289 | Yu et al. | Aug 2003 | B1 |
6628897 | Suzuki | Sep 2003 | B2 |
6650774 | Szeliski | Nov 2003 | B1 |
6700711 | Nalwa | Mar 2004 | B2 |
6701081 | Dwyer et al. | Mar 2004 | B1 |
6768509 | Bradski et al. | Jul 2004 | B1 |
6775437 | Kazarinov et al. | Aug 2004 | B2 |
6798406 | Jones et al. | Sep 2004 | B1 |
6809887 | Gao et al. | Oct 2004 | B1 |
6850279 | Scherling | Feb 2005 | B1 |
6855111 | Yokoi et al. | Feb 2005 | B2 |
6861633 | Osborn | Mar 2005 | B2 |
7006123 | Yoshikawa et al. | Feb 2006 | B2 |
7039292 | Breiholz | May 2006 | B1 |
7084904 | Liu et al. | Aug 2006 | B2 |
7116351 | Yoshikawa | Oct 2006 | B2 |
7215479 | Bakin | May 2007 | B1 |
7253394 | Kang | Aug 2007 | B2 |
7271803 | Ejiri et al. | Sep 2007 | B2 |
7336299 | Kostrzewski et al. | Feb 2008 | B2 |
7612953 | Nagai et al. | Nov 2009 | B2 |
7710463 | Foote | May 2010 | B2 |
7805071 | Mitani | Sep 2010 | B2 |
7817354 | Wilson | Oct 2010 | B2 |
7893957 | Peters et al. | Feb 2011 | B2 |
7961398 | Tocci | Jun 2011 | B2 |
8004557 | Pan | Aug 2011 | B2 |
8098276 | Chang et al. | Jan 2012 | B2 |
8115813 | Tang | Feb 2012 | B2 |
8139125 | Scherling | Mar 2012 | B2 |
8228417 | Georgiev et al. | Jul 2012 | B1 |
8267601 | Campbell et al. | Sep 2012 | B2 |
8284263 | Oohara et al. | Oct 2012 | B2 |
8294073 | Vance et al. | Oct 2012 | B1 |
8356035 | Baluja et al. | Jan 2013 | B1 |
8400555 | Georgiev et al. | Mar 2013 | B1 |
8482813 | Kawano et al. | Jul 2013 | B2 |
8791984 | Jones et al. | Jul 2014 | B2 |
8988564 | Webster et al. | Mar 2015 | B2 |
20010028482 | Nishioka | Oct 2001 | A1 |
20020070365 | Karellas | Jun 2002 | A1 |
20020136150 | Mihara et al. | Sep 2002 | A1 |
20030024987 | Zhu | Feb 2003 | A1 |
20030038814 | Blume | Feb 2003 | A1 |
20030214575 | Yoshikawa | Nov 2003 | A1 |
20040021767 | Endo et al. | Feb 2004 | A1 |
20040066449 | Givon | Apr 2004 | A1 |
20040105025 | Scherling | Jun 2004 | A1 |
20040183907 | Hovanky et al. | Sep 2004 | A1 |
20040246333 | Steuart et al. | Dec 2004 | A1 |
20040263611 | Cutler | Dec 2004 | A1 |
20050053274 | Mayer et al. | Mar 2005 | A1 |
20050057659 | Hasegawa | Mar 2005 | A1 |
20050081629 | Hoshal | Apr 2005 | A1 |
20050111106 | Matsumoto et al. | May 2005 | A1 |
20050185711 | Pfister et al. | Aug 2005 | A1 |
20050218297 | Suda et al. | Oct 2005 | A1 |
20060023074 | Cutler | Feb 2006 | A1 |
20060023106 | Yee et al. | Feb 2006 | A1 |
20060023278 | Nishioka | Feb 2006 | A1 |
20060140446 | Luo et al. | Jun 2006 | A1 |
20060193509 | Criminisi et al. | Aug 2006 | A1 |
20060215054 | Liang et al. | Sep 2006 | A1 |
20060215903 | Nishiyama | Sep 2006 | A1 |
20060238441 | Benjamin et al. | Oct 2006 | A1 |
20070024739 | Konno | Feb 2007 | A1 |
20070058961 | Kobayashi et al. | Mar 2007 | A1 |
20070064142 | Misawa et al. | Mar 2007 | A1 |
20070085903 | Zhang | Apr 2007 | A1 |
20070164202 | Wurz et al. | Jul 2007 | A1 |
20070216796 | Lenel et al. | Sep 2007 | A1 |
20070242152 | Chen | Oct 2007 | A1 |
20070263115 | Horidan et al. | Nov 2007 | A1 |
20070268983 | Elam | Nov 2007 | A1 |
20080088702 | Linsenmaier et al. | Apr 2008 | A1 |
20080117289 | Schowengerdt et al. | May 2008 | A1 |
20080218612 | Border et al. | Sep 2008 | A1 |
20080259172 | Tamaru | Oct 2008 | A1 |
20080266404 | Sato | Oct 2008 | A1 |
20080290435 | Oliver et al. | Nov 2008 | A1 |
20080291543 | Nomura et al. | Nov 2008 | A1 |
20080297612 | Yoshikawa | Dec 2008 | A1 |
20080316301 | Givon | Dec 2008 | A1 |
20090003646 | Au et al. | Jan 2009 | A1 |
20090005112 | Sorek et al. | Jan 2009 | A1 |
20090015812 | Schultz et al. | Jan 2009 | A1 |
20090051804 | Nomura et al. | Feb 2009 | A1 |
20090085846 | Cho et al. | Apr 2009 | A1 |
20090096994 | Smits | Apr 2009 | A1 |
20090153726 | Lim | Jun 2009 | A1 |
20090160931 | Pockett et al. | Jun 2009 | A1 |
20090268983 | Stone et al. | Oct 2009 | A1 |
20090296984 | Nijim et al. | Dec 2009 | A1 |
20090315808 | Ishii | Dec 2009 | A1 |
20100044555 | Ohara et al. | Feb 2010 | A1 |
20100045774 | Len et al. | Feb 2010 | A1 |
20100066812 | Kajihara et al. | Mar 2010 | A1 |
20100165155 | Chang | Jul 2010 | A1 |
20100215249 | Heitz et al. | Aug 2010 | A1 |
20100232681 | Fujieda et al. | Sep 2010 | A1 |
20100259655 | Takayama | Oct 2010 | A1 |
20100265363 | Kim | Oct 2010 | A1 |
20100278423 | Itoh et al. | Nov 2010 | A1 |
20100289878 | Sato et al. | Nov 2010 | A1 |
20100302396 | Golub et al. | Dec 2010 | A1 |
20100309286 | Chen et al. | Dec 2010 | A1 |
20100309333 | Smith et al. | Dec 2010 | A1 |
20110001789 | Wilson et al. | Jan 2011 | A1 |
20110007135 | Okada et al. | Jan 2011 | A1 |
20110009163 | Fletcher et al. | Jan 2011 | A1 |
20110012998 | Pan | Jan 2011 | A1 |
20110043623 | Fukuta et al. | Feb 2011 | A1 |
20110090575 | Mori | Apr 2011 | A1 |
20110096089 | Shenhav et al. | Apr 2011 | A1 |
20110096988 | Suen et al. | Apr 2011 | A1 |
20110128412 | Milnes et al. | Jun 2011 | A1 |
20110213664 | Osterhout et al. | Sep 2011 | A1 |
20110235899 | Tanaka | Sep 2011 | A1 |
20110249341 | Difrancesco et al. | Oct 2011 | A1 |
20110304764 | Shigemitsu et al. | Dec 2011 | A1 |
20120008148 | Pryce et al. | Jan 2012 | A1 |
20120033051 | Atanassov et al. | Feb 2012 | A1 |
20120056987 | Fedoroff | Mar 2012 | A1 |
20120075168 | Osterhout et al. | Mar 2012 | A1 |
20120249750 | Izzat et al. | Oct 2012 | A1 |
20120249815 | Bohn et al. | Oct 2012 | A1 |
20120269400 | Heyward | Oct 2012 | A1 |
20120281072 | Georgiev et al. | Nov 2012 | A1 |
20120293607 | Bhogal et al. | Nov 2012 | A1 |
20120293632 | Yukich | Nov 2012 | A1 |
20120327195 | Cheng | Dec 2012 | A1 |
20130003140 | Keniston et al. | Jan 2013 | A1 |
20130010084 | Hatano | Jan 2013 | A1 |
20130070055 | Atanassov et al. | Mar 2013 | A1 |
20130141802 | Yang | Jun 2013 | A1 |
20130222556 | Shimada | Aug 2013 | A1 |
20130229529 | Lablans | Sep 2013 | A1 |
20130260823 | Shukla et al. | Oct 2013 | A1 |
20130278785 | Nomura et al. | Oct 2013 | A1 |
20130286451 | Verhaegh | Oct 2013 | A1 |
20130335598 | Gustavsson et al. | Dec 2013 | A1 |
20130335600 | Gustavsson et al. | Dec 2013 | A1 |
20140104378 | Kauff et al. | Apr 2014 | A1 |
20140111650 | Georgiev et al. | Apr 2014 | A1 |
20140139623 | McCain et al. | May 2014 | A1 |
20140152852 | Ito et al. | Jun 2014 | A1 |
20140184749 | Hilliges et al. | Jul 2014 | A1 |
20140192253 | Laroia | Jul 2014 | A1 |
20150049172 | Ramachandra et al. | Feb 2015 | A1 |
20150177524 | Webster et al. | Jun 2015 | A1 |
20150244934 | Duparre et al. | Aug 2015 | A1 |
20150286033 | Osborne | Oct 2015 | A1 |
20150288865 | Osborne | Oct 2015 | A1 |
20150370040 | Georgiev | Dec 2015 | A1 |
20150371387 | Atanassov | Dec 2015 | A1 |
20150373252 | Georgiev | Dec 2015 | A1 |
20150373262 | Georgiev | Dec 2015 | A1 |
20150373268 | Osborne | Dec 2015 | A1 |
20150373269 | Osborne | Dec 2015 | A1 |
20150373279 | Osborne | Dec 2015 | A1 |
Number | Date | Country |
---|---|---|
101201459 | Jun 2008 | CN |
101581828 | Nov 2009 | CN |
0610605 | Aug 1994 | EP |
0751416 | Jan 1997 | EP |
1176812 | Jan 2002 | EP |
1383342 | Jan 2004 | EP |
1816514 | Aug 2007 | EP |
1832912 | Sep 2007 | EP |
2242252 | Oct 2010 | EP |
2354390 | Mar 2001 | GB |
2354391 | Mar 2001 | GB |
H089424 | Jan 1996 | JP |
H0847001 | Feb 1996 | JP |
H08125835 | May 1996 | JP |
8194274 | Jul 1996 | JP |
H08242453 | Sep 1996 | JP |
2001194114 | Jul 2001 | JP |
2003304561 | Oct 2003 | JP |
3791847 | Jun 2006 | JP |
2006279538 | Oct 2006 | JP |
2007147457 | Jun 2007 | JP |
2007323615 | Dec 2007 | JP |
2008009424 | Jan 2008 | JP |
2010041381 | Feb 2010 | JP |
2010067014 | Mar 2010 | JP |
2010128820 | Jun 2010 | JP |
2010524279 | Jul 2010 | JP |
WO-9321560 | Oct 1993 | WO |
WO-9847291 | Oct 1998 | WO |
WO-2006075528 | Jul 2006 | WO |
WO-2007129147 | Nov 2007 | WO |
WO-2008112054 | Sep 2008 | WO |
WO-2009047681 | Apr 2009 | WO |
WO-2009086330 | Jul 2009 | WO |
WO-2010019757 | Feb 2010 | WO |
WO-2012136388 | Oct 2012 | WO |
WO-2012164339 | Dec 2012 | WO |
WO-2013154433 | Oct 2013 | WO |
WO-2014012603 | Jan 2014 | WO |
WO-2014025588 | Feb 2014 | WO |
Entry |
---|
Arican, et al., “Intermediate View Generation for Perceived Depth Adjustment of Sterio Video”, Mitsubishi Electric Research Laboratories, http://www.merl.com, TR2009-052, Sep. 2009; 12 pages. |
Hoff, et al., “Surfaces from Stereo: Integrating Feature Matching, Disparity Estimation, and Contour Detection”, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 11, No. 2, pp. 121-136, Feb. 1989. |
Krotkov E., et al., “Active vision for reliable ranging: Cooperating focus, stereo, and vergence”, International Journal of Computer Vision. vol. 11, No. 2, Oct. 1, 1993, pp. 187-203, XP055149875, ISSN: 0920-5691. DOI: 10.1007/BF01469228. |
Murphy M., et al., “Lens Drivers Focus on Performance in High-Resolution Camera Modules,” Analog Dialogue, Nov. 2006, vol. 40, pp. 1-3. |
Narkhede, et al., “Stereoscopic Imaging: A Real-Time, In Depth Look,” IEEE Potentials, Feb./Mar. 2004, vol. 23, Issue 1, pp. 38-42. |
Sun W.S., et al., “Single-Lens Camera Based on a Pyramid Prism Array to Capture Four Images,” Optical Review, 2013, vol. 20 (2), pp. 145-152. |
Ricoh Imagine Change: “New Ricoh Theta Model, Capturing 360-degree Images in One Shot, is on Sale Soon—Spherical Video Function, API and SDK (Beta Version)”, News Release, Oct. 28, 2014, 3 pages. |
Han Y., et al., “Removing Illumination from Image Pair for Stereo Matching”, Audio, Language and Image Processing (ICALIP), 2012 International Conference on, IEEE, Jul. 16, 2012, XP032278010, pp. 508-512. |
Shuchun Y., et al., “Preprocessing for stereo vision based on LOG filter”, Proceedings of 2011 6th International Forum on Strategic Technology, Aug. 2011, XP055211077, pp. 1074-1077. |
Hao M., et al., “Object Location Technique for Binocular Stereo Vision Based on Scale Invariant Feature Transform Feature Points”, SIFT, Journal of Harbin Engineering University, Jun. 2009, vol. 30, No. 6 pp. 649-653. |
International Search Report and Written Opinion—PCT/US2015/033176—ISA/EPO—Dec. 18, 2015. |
Kawanishi T., et al., “Generation of High-Resolution Stereo Panoramic Images by Omnidirectional Imaging Sensor Using Hexagonal Pyramidal Mirrors”, Patiern Recognition, 1998, Proceedings, Fourteenth International Conference on Brisbane, QLD., Australia Aug. 16-20, 1998, Los Alamitos, CA, USA,IEEE Comput. Soc, US, Jan. 1, 1998, pp. 485-489, vol. 1, XP031098377, ISBN: 978-0-8186-8512-5. |
Partial International Search Report—PCT/US2015/033176—ISA/EPO—Aug. 13, 2015. |
Tan K-H., et al., “Multiview Panoramic Cameras Using a Pyramid”, Omnidirectional Vision, 2002, Proceedings, Third Workshop on Jun. 2, 2002, Piscataway, NJ, USA,IEEE, Jan. 1, 2002, pp. 87-93, XP010611080, ISBN: 978-0-7695-1629-5. |
Number | Date | Country | |
---|---|---|---|
20150373263 A1 | Dec 2015 | US |
Number | Date | Country | |
---|---|---|---|
62015316 | Jun 2014 | US |