CN110708532A - Universal light field unit image generation method and system - Google Patents
Universal light field unit image generation method and system Download PDFInfo
- Publication number
- CN110708532A CN110708532A CN201910981967.5A CN201910981967A CN110708532A CN 110708532 A CN110708532 A CN 110708532A CN 201910981967 A CN201910981967 A CN 201910981967A CN 110708532 A CN110708532 A CN 110708532A
- Authority
- CN
- China
- Prior art keywords
- image
- camera
- display
- pixel
- determining
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 40
- 238000010586 diagram Methods 0.000 description 3
- 230000000694 effects Effects 0.000 description 3
- 238000003384 imaging method Methods 0.000 description 2
- 230000000750 progressive effect Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/207—Image signal generators using stereoscopic image cameras using a single 2D image sensor
- H04N13/236—Image signal generators using stereoscopic image cameras using a single 2D image sensor using varifocal lenses or mirrors
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Image Processing (AREA)
Abstract
The invention discloses a method and a system for generating a universal light field unit image. The method comprises the following steps: determining the position of the intersection point of a display ray emitted by a display lens in the display lens array and the camera array; determining a first camera according to the position of the intersection point; respectively acquiring a first image and a second image; determining the pixel position difference degree of the first image and the second image; acquiring nearest neighbor shooting light; determining the distance delta between the display light and the nearest neighbor shooting light; according to the pixel difference degree and the first position offset amount delta idDetermining a second position offset Δ iu(ii) a Determining pixel values of an image displayed by the display lensThe method and the system for generating the universal light field unit image can avoid the problem that the pixel at the corresponding position cannot be accurately acquired due to the introduction of the position error of the reference surface, so as to improve the accuracy of generating the unit image.
Description
Technical Field
The invention relates to the technical field of image generation, in particular to a method and a system for generating a universal light field unit image.
Background
The unit image is based on integrated imaging light field display, the accuracy of the unit image determines the accuracy of the light field display, the existing method has matching requirements on a display system, one or more reference surfaces are required to be arranged in the process of generating the image, and the position error of the reference surface can be introduced by introducing the reference surface, so that the accuracy of the obtained unit image is poor, and the three-dimensional effect reconstructed by using the unit image is poor. Therefore, the invention provides a method capable of remarkably improving the imaging accuracy of an image, and is a technical problem to be solved in the field.
Disclosure of Invention
The invention aims to provide a method and a system for generating a universal light field unit image, which can accurately generate the unit image.
In order to achieve the purpose, the invention provides the following scheme:
a universal light field unit image generation method is applied to a light field image acquisition device based on a camera array; the light field image acquisition device based on the camera array comprises: a camera array and a display lens array; the camera array and the display lens array are arranged in parallel; an object to be photographed is placed between the camera array and the display lens array;
the method comprises the following steps:
determining the position of the intersection point of a display ray emitted by a display lens in the display lens array and the camera array;
determining a camera at the intersection point of the display ray and the camera array according to the position of the intersection point, and recording the camera as a first camera;
respectively acquiring a first image and a second image; the first image is an image shot by the first camera; the second image is an image shot by a camera which is closest to the display light ray in the adjacent cameras of the first camera;
determining the pixel position difference degree of the first image and the second image;
acquiring all the shooting light rays parallel to the display light rays, and correspondingly determining one of the shooting light rays emitted by the first camera which is closest to the display light rays, and recording the shooting light ray as a nearest neighbor shooting light ray;
determining a distance delta between the display ray and the nearest neighbor photographing ray,wherein x is the position of the intersection point of the display light emitted by the display lens array and the camera array, n is the serial number of the first camera,dP is the center-to-center distance between adjacent cameras;
determining a first position deviation amount delta i according to the pixel position difference degree and the distanced,Wherein k is the pixel position difference degree;
according to the pixel difference degree and the first position offset delta idDetermining said second amount of position offset Δ iu,Δiu=k-Δid;
Wherein,pP is the image of the camera at the intersectionPixel size, f is focal length of camera at intersection point, dE is center distance between display lenses, pE is pixel size of display lens emitting display light, g is distance between display lens emitting display light and display screen, j is j-th pixel serial number of display lens emitting display light, ceil is minimum integer of return greater than or equal to calculated result, a is weight value, a is delta/dP,is the i + Δ i th image in the first imagedThe value of each of the pixels is calculated,is the i- Δ i in the second imageuAnd n 'is the serial number of the camera which is closest to the display light ray in the cameras adjacent to the first camera, and n' is n +/-1.
Optionally, the method further includes: acquiring the center distance between every two adjacent cameras in the camera array, the pixel size of every camera, the focal length of every camera, the center distance between every two adjacent display lenses in the display lens array, the pixel size of every display lens, the distance between every display lens and a display screen and the distance between the display lens array and the camera array.
Optionally, the intersection point of the display light emitted by the display lens array and the camera array is x:
and D is the distance between the display lens array and the camera array.
Optionally, the determining the difference degree between the pixel positions of the first image and the second image includes:
According to the sum of squared errors Rj(k ') determining a degree of difference in pixel position between said first image and said second image, k' where Rj(k');
Where th is the set pixel search range, k' is the center position of the start of the pixel search range,for the p-th pixel value in the first image,for the t-th pixel value, k' while R, in the second imagej(k') is when Rj(k ') the value of k ' corresponding to the minimum value is obtained, where k ' is the center position of the start of the pixel search range.
A pervasive light field unit image generation system, comprising:
the position determining module is used for determining the position of the intersection point of a display ray emitted by a display lens in the display lens array and the camera array;
the first camera determining module is used for determining a camera at the intersection point of the display ray and the camera array according to the position of the intersection point, and recording the camera as a first camera;
the image acquisition module is used for respectively acquiring a first image and a second image; the first image is an image shot by the first camera; the second image is an image shot by a camera which is closest to the display light ray in the adjacent cameras of the first camera;
a pixel position difference degree determining module, configured to determine a pixel position difference degree of the first image and the second image;
a nearest neighbor shooting light ray obtaining module, configured to obtain all the shooting light rays emitted by the first camera and parallel to the display light ray, and correspondingly determine one of the shooting light rays closest to the display light ray, which is recorded as a nearest neighbor shooting light ray;
a distance determination module for determining the distanceThe distance delta between the display ray and the nearest neighbor photographing ray,wherein x is the position of the intersection point of the display light rays emitted by the display lens array and the camera array, n is the serial number of the first camera, and dP is the center distance between adjacent cameras;
a first position offset determination module, configured to determine a first position offset Δ i according to the pixel position difference and the distanced,Wherein k is the pixel position difference degree;
a second position offset determination module for determining the first position offset Δ i according to the pixel differencedDetermining said second amount of position offset Δ iu,Δiu=k-Δid;
A pixel value determination module for determining pixel values of an image displayed by the display lens
Wherein,pP is the pixel size of the camera at the intersection, f is the focal length of the camera at the intersection, dE is the center distance between the display lenses, pE is the pixel size of the display lens emitting the display light, g is the distance between the display lens emitting the display light and the display screen, j is the j-th pixel serial number of the display lens emitting the display light, ceil (x) is the minimum integer which returns the calculation result which is greater than or equal to x, a is the weight, a is delta/dP,is the firstI + Δ i in the imagedThe value of each of the pixels is calculated,is the i- Δ i in the second imageuAnd n 'is the serial number of the camera which is closest to the display light ray in the cameras adjacent to the first camera, and n' is n +/-1.
Optionally, the system further includes: and the initialization parameter acquisition module is used for respectively acquiring the center distance between every two adjacent cameras in the camera array, the pixel size of every camera, the focal length of every camera, the center distance between every two adjacent display lenses in the display lens array, the pixel size of every display lens, the distance between every display lens and a display screen and the distance between the display lens array and the camera array.
Optionally, the pixel position difference determining module includes:
a sum of squared errors acquisition unit for acquiring a sum of squared errors R between the first image and the second imagej(k'),
A pixel position difference degree determining unit for determining the difference degree according to the sum of squared errors Rj(k ') determining a degree of difference in pixel position between said first image and said second image, k' where Rj(k');
Where th is the set pixel search range, k' is the center position of the start of the pixel search range,for the p-th pixel value in the first image,for the t-th pixel value, k' while R, in the second imagej(k') is when Rj(k ') the value of k ' corresponding to the minimum value is obtained, where k ' is the center position of the start of the pixel search range.
Optionally, the intersection point of the display light emitted by the display lens array and the camera array is x:
and D is the distance between the display lens array and the camera array.
According to the specific embodiment provided by the invention, the invention discloses the following technical effects: according to the method and the system for generating the universal light field unit image, the pixel position obtained is corrected by adopting the pixel position difference degree and the position offset, so that the position of the obtained pixel has no error in the process of forming the unit image by adopting the pixel, the problem that the pixel at the corresponding position cannot be accurately obtained due to the introduction of the position error of the reference surface is solved, and the accuracy of generating the unit image is improved.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings needed in the embodiments will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art to obtain other drawings without creative efforts.
FIG. 1 is a schematic diagram of a camera array and a display lens array in a generalized light field unit image generating method disclosed in an embodiment of the present invention;
FIG. 2 is a flowchart of a generalized light field unit image generation method according to an embodiment of the present invention;
fig. 3 is a schematic structural diagram of a generalized light field unit image generation system according to an embodiment of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
The invention aims to provide a method and a system for generating a universal light field unit image, which can accurately generate the unit image.
In order to make the aforementioned objects, features and advantages of the present invention comprehensible, embodiments accompanied with figures are described in further detail below.
Fig. 1 is a schematic diagram of the operation of a camera array and a display lens array in the method for generating a ubiquitous light field unit image disclosed in the embodiment of the present invention, and fig. 2 is a flowchart of the method for generating a ubiquitous light field unit image disclosed in the embodiment of the present invention.
As shown in fig. 1 and 2, a generalized light field unit image generation method is mainly applied to a light field image acquisition device based on a camera array. The light field image acquisition device based on the camera array comprises: a camera array 1-1 and a display lens array 1-2. The camera array 1-1 and the display lens array 1-2 are arranged in parallel with each other. An object 1-3 to be photographed is placed between the camera array 1-1 and the display lens array 1-2.
The method comprises the following steps:
s100, determining the position of the intersection point of the display light ray emitted by one display lens in the display lens array 2 and the camera array 1.
S101, determining a camera at the position of the intersection point of the display light and the camera array according to the position of the intersection point, and recording the camera as a first camera.
And S102, respectively acquiring a first image and a second image. The first image is an image captured by the first camera. The second image is an image shot by a camera closest to the display light ray in cameras adjacent to the first camera.
S103, determining the pixel position difference degree of the first image and the second image.
S104, all the shooting light rays which are emitted by the first camera and are parallel to the display light rays are obtained, and one of the shooting light rays which is closest to the display light rays is correspondingly determined and recorded as a nearest neighbor shooting light ray.
S105, determining the distance delta between the display light ray and the nearest neighbor shooting light ray,wherein x is the position of the intersection point of the display light emitted by the display lens array and the camera array, n is the serial number of the first camera,dP is the center-to-center spacing between adjacent cameras.
S106, determining a first position deviation amount delta i according to the pixel position difference degree and the distanced,And k is the pixel position difference degree.
S107, according to the pixel difference degree and the first position offset delta idDetermining said second amount of position offset Δ iu,Δiu=k-Δid。
Wherein,pP is the pixel size of the camera at the intersection, f is the focal length of the camera at the intersection, dE is the center distance between the display lenses, pE is the pixel size of the display lens emitting the display light, g is the distance between the display lens emitting the display light and the display screen, j is the j-th pixel serial number of the display lens emitting the display light, ceil (x) is the minimum integer which returns the calculation result which is greater than or equal to x, a is the weight, a is delta/dP,is the i + Δ i th image in the first imagedThe value of each of the pixels is calculated,is the i- Δ i in the second imageuAnd n 'is the serial number of the camera which is closest to the display light ray in the cameras adjacent to the first camera, and n' is n +/-1.
Since the displayed unit image is composed of individual pixels, the final display image can be obtained by calculating the pixel at any position in the display lens array.
The second position offset Δ i in the above method of the present disclosureuAnd the essence of the position deviation is the position deviation of the shooting light which is parallel to the display light of the display lens and is closest to the display light in the shooting light emitted by the camera which is closest to the display light in the camera adjacent to the first camera.
The method may further comprise: respectively acquiring the center distance between every two adjacent cameras in the camera array, the pixel size of every camera, the focal length of every camera, the center distance between every two adjacent display lenses in the display lens array, the pixel size of every display lens, the distance between every display lens and a display screen and the distance between the display lens array and the camera array.
The position of the intersection point of the display light emitted by the display lens array and the camera array is x:
and D is the distance between the display lens array and the camera array.
The determining the pixel position difference degree of the first image and the second image comprises:
According to the sum of squared errors Rj(k ') determining a degree of difference in pixel position between said first image and said second image, k' where Rj(k')。
Where th is the set pixel search range, k' is the center position of the start of the pixel search range,for the p-th pixel value in the first image,for the t-th pixel value, k' while R, in the second imagej(k') is when Rj(k ') the value of k ' corresponding to the minimum value is obtained, where k ' is the center position of the start of the pixel search range.
The method disclosed above is mainly a method for determining pixels in the vertical direction of a camera array and a display lens array, and the method for determining pixels in the horizontal direction is the same as the method for determining pixels in the vertical direction, and is not repeated here.
The present invention also provides a system for generating a generic light field unit image, as shown in fig. 3, the system includes: the device comprises a position determining module 1, a first camera determining module 2, an image acquiring module 3, a pixel position difference degree determining module 4, a nearest neighbor shooting ray acquiring module 5, a distance determining module 6, a first position offset amount determining module 7, a second position offset amount determining module 8 and a pixel value determining module 9.
The position determining module 1 is used for determining the position of the intersection point of a display ray emitted by a display lens in the display lens array and the camera array.
And determining a camera at the intersection point of the display ray and the camera array according to the position of the intersection point by using a first camera determination module 2, and recording the camera as a first camera.
The image acquisition module 3 is used for acquiring a first image and a second image respectively. The first image is an image captured by the first camera. The second image is an image shot by a camera closest to the display light ray in cameras adjacent to the first camera.
And determining the pixel position difference degree of the first image and the second image by adopting a pixel position difference degree determining module 4.
And acquiring all the shooting light rays which are parallel to the display light rays and are emitted by the first camera by adopting a nearest neighbor shooting light ray acquisition module 5, correspondingly determining one of the shooting light rays which is closest to the display light rays, and recording the shooting light ray as the nearest neighbor shooting light ray.
The distance delta between the display ray and the nearest neighbor shooting ray is determined by a distance determining module 6,wherein x is the position of the intersection point of the display light emitted by the display lens array and the camera array, n is the serial number of the first camera,dP is the center-to-center spacing between adjacent cameras.
Determining a first position offset delta i by adopting a first position offset determination module 7 according to the pixel position difference degree and the distanced,And k is the pixel position difference degree.
A second position offset determining module 8 is adopted to determine the first position offset delta i according to the pixel difference degreedDetermining said second amount of position offset Δ iu,Δiu=k-Δid。
Determining pixel values of an image displayed by the display lens using a pixel value determining module 9
Wherein,pP is the pixel size of the camera at the intersection, f is the focal length of the camera at the intersection, dE is the center distance between the display lenses, pE is the pixel size of the display lens emitting the display light, g is the distance between the display lens emitting the display light and the display screen, j is the j-th pixel serial number of the display lens emitting the display light, ceil (x) is the minimum integer which returns the calculation result which is greater than or equal to x, a is the weight, a is delta/dP,is the i + Δ i th image in the first imagedThe value of each of the pixels is calculated,is the i- Δ i in the second imageuAnd n 'is the serial number of the camera which is closest to the display light ray in the cameras adjacent to the first camera, and n' is n +/-1.
The system also comprises an initialization parameter acquisition module. The initialization parameter acquisition module respectively acquires the center distance between every two adjacent cameras in the camera array, the pixel size of every camera, the focal length of every camera, the center distance between every two adjacent display lenses in the display lens array, the pixel size of every display lens, the distance between every display lens and a display screen and the distance between the display lens array and the camera array.
The pixel position difference degree determination module may further include: an error square sum acquisition unit and a pixel position difference degree determination unit.
Wherein the error square sum acquisition unit acquires the error square sum R between the first image and the second imagej(k'),
The pixel position difference degree determining unit determines the difference degree according to the sum of squares of errors Rj(k ') determining a degree of difference in pixel position between said first image and said second image, k' where Rj(k')。
Where th is the set pixel search range, k' is the center position of the start of the pixel search range,for the p-th pixel value in the first image,for the t-th pixel value, k' while R, in the second imagej(k') is when Rj(k ') the value of k ' corresponding to the minimum value is obtained, where k ' is the center position of the start of the pixel search range.
According to the specific embodiment provided by the invention, the invention discloses the following technical effects: according to the method and the system for generating the universal light field unit image, the pixel position obtained is corrected by adopting the pixel position difference degree and the position offset, so that the position of the obtained pixel has no error in the process of forming the unit image by adopting the pixel, the problem that the pixel at the corresponding position cannot be accurately obtained due to the introduction of the position error of the reference surface is solved, and the accuracy of generating the unit image is improved.
In addition, in the method and the system for generating the universal light field unit image, which are provided by the invention, weighting operation is carried out on various pixels in the process of forming the required pixels, so that the continuity and consistency of the unit image generated by the pixels can be obviously improved.
The embodiments in the present description are described in a progressive manner, each embodiment focuses on differences from other embodiments, and the same and similar parts among the embodiments are referred to each other. For the system disclosed by the embodiment, the description is relatively simple because the system corresponds to the method disclosed by the embodiment, and the relevant points can be referred to the method part for description.
The principles and embodiments of the present invention are explained herein using specific examples, which are presented only to assist in understanding the method and its core concepts. Meanwhile, for a person skilled in the art, according to the idea of the present invention, the specific embodiments and the application range may be changed. In view of the above, the present disclosure should not be construed as limiting the invention.
Claims (8)
1. A universal light field unit image generation method is applied to a light field image acquisition device based on a camera array; the light field image acquisition device based on the camera array comprises: a camera array and a display lens array; the camera array and the display lens array are arranged in parallel; an object to be photographed is placed between the camera array and the display lens array; characterized in that the method comprises:
determining the position of the intersection point of a display ray emitted by a display lens in the display lens array and the camera array;
determining a camera at the intersection point of the display ray and the camera array according to the position of the intersection point, and recording the camera as a first camera;
respectively acquiring a first image and a second image; the first image is an image shot by the first camera; the second image is an image shot by a camera which is closest to the display light ray in the adjacent cameras of the first camera;
determining the pixel position difference degree of the first image and the second image;
acquiring all the shooting light rays parallel to the display light rays, and correspondingly determining one of the shooting light rays emitted by the first camera which is closest to the display light rays, and recording the shooting light ray as a nearest neighbor shooting light ray;
determining a distance delta between the display ray and the nearest neighbor photographing ray,wherein x is emitted from the display lens arrayThe position of the intersection point of the display light and the camera array is shown, n is the serial number of the first camera,dP is the center-to-center distance between adjacent cameras;
determining a first position deviation amount delta i according to the pixel position difference degree and the distanced,Wherein k is the pixel position difference degree;
according to the pixel difference degree and the first position offset delta idDetermining said second amount of position offset Δ iu,Δiu=k-Δid;
Determining pixel values of an image displayed by the display lens
Wherein,pP is the pixel size of the camera at the intersection, f is the focal length of the camera at the intersection, dE is the center distance between the display lenses, pE is the pixel size of the display lens emitting the display light, g is the distance between the display lens emitting the display light and the display screen, j is the j-th pixel serial number of the display lens emitting the display light, ceil (x) is the minimum integer which returns the calculation result which is greater than or equal to x, a is the weight, a is delta/dP,is the i + Δ i th image in the first imagedThe value of each of the pixels is calculated,is the i- Δ i in the second imageuA pixel value n' is adjacent to the first cameraAnd n ═ n +/-1, of the serial number of the camera closest to the display ray.
2. A method for generating a generic light field unit image according to claim 1, further comprising: acquiring the center distance between every two adjacent cameras in the camera array, the pixel size of every camera, the focal length of every camera, the center distance between every two adjacent display lenses in the display lens array, the pixel size of every display lens, the distance between every display lens and a display screen and the distance between the display lens array and the camera array.
4. The method of claim 1, wherein the determining the difference degree of pixel positions of the first image and the second image comprises:
obtaining the sum of the squares of errors R between the first image and the second imagej(k'),
According to the sum of squared errors Rj(k ') determining a degree of difference in pixel position between said first image and said second image, k' where Rj(k');
Where th is the set pixel search range, k' is the center position of the start of the pixel search range,for the p-th pixel value in the first image,for the t-th pixel value, k' while R, in the second imagej(k') is when Rj(k ') the value of k ' corresponding to the minimum value is obtained, where k ' is the center position of the start of the pixel search range.
5. A pervasive light field unit image generation system, comprising:
the position determining module is used for determining the position of the intersection point of a display ray emitted by a display lens in the display lens array and the camera array;
the first camera determining module is used for determining a camera at the intersection point of the display ray and the camera array according to the position of the intersection point, and recording the camera as a first camera;
the image acquisition module is used for respectively acquiring a first image and a second image; the first image is an image shot by the first camera; the second image is an image shot by a camera which is closest to the display light ray in the adjacent cameras of the first camera;
a pixel position difference degree determining module, configured to determine a pixel position difference degree of the first image and the second image;
a nearest neighbor shooting light ray obtaining module, configured to obtain all the shooting light rays emitted by the first camera and parallel to the display light ray, and correspondingly determine one of the shooting light rays closest to the display light ray, which is recorded as a nearest neighbor shooting light ray;
a distance determining module for determining a distance delta between the display ray and the nearest neighbor photographing ray,wherein x is the position of the intersection point of the display light emitted by the display lens array and the camera arrayN is the serial number of the first camera,dP is the center-to-center distance between adjacent cameras;
a first position offset determination module, configured to determine a first position offset Δ i according to the pixel position difference and the distanced,Wherein k is the pixel position difference degree;
a second position offset determination module for determining the first position offset Δ i according to the pixel differencedDetermining said second amount of position offset Δ iu,Δiu=k-Δid;
A pixel value determination module for determining pixel values of an image displayed by the display lens
Wherein,pP is the pixel size of the camera at the intersection, f is the focal length of the camera at the intersection, dE is the center distance between the display lenses, pE is the pixel size of the display lens emitting the display light, g is the distance between the display lens emitting the display light and the display screen, j is the j-th pixel serial number of the display lens emitting the display light, ceil (x) is the minimum integer which returns the calculation result which is greater than or equal to x, a is the weight, a is delta/dP,is the i + Δ i th image in the first imagedThe value of each of the pixels is calculated,is the i- Δ i in the second imageuAnd n 'is the serial number of the camera which is closest to the display light ray in the cameras adjacent to the first camera, and n' is n +/-1.
6. The ubiquitous light field unit image generating system of claim 5, wherein said system further comprises: and the initialization parameter acquisition module is used for respectively acquiring the center distance between every two adjacent cameras in the camera array, the pixel size of every camera, the focal length of every camera, the center distance between every two adjacent display lenses in the display lens array, the pixel size of every display lens, the distance between every display lens and a display screen and the distance between the display lens array and the camera array.
7. The system of claim 5, wherein the pixel position disparity determination module comprises:
a sum of squared errors acquisition unit for acquiring a sum of squared errors R between the first image and the second imagej(k'),
A pixel position difference degree determining unit for determining the difference degree according to the sum of squared errors Rj(k ') determining a degree of difference in pixel position between said first image and said second image, k' where Rj(k');
Where th is the set pixel search range, k' is the center position of the start of the pixel search range,for the p-th pixel value in the first image,for the t-th pixel value, k' while R, in the second imagej(k') is when Rj(k ') the value of k ' corresponding to the minimum value is obtained, where k ' is the center position of the start of the pixel search range.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910981967.5A CN110708532B (en) | 2019-10-16 | 2019-10-16 | Universal light field unit image generation method and system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910981967.5A CN110708532B (en) | 2019-10-16 | 2019-10-16 | Universal light field unit image generation method and system |
Publications (2)
Publication Number | Publication Date |
---|---|
CN110708532A true CN110708532A (en) | 2020-01-17 |
CN110708532B CN110708532B (en) | 2021-03-23 |
Family
ID=69199885
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910981967.5A Active CN110708532B (en) | 2019-10-16 | 2019-10-16 | Universal light field unit image generation method and system |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110708532B (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111225221A (en) * | 2020-01-15 | 2020-06-02 | 未来新视界文化科技(嘉善)有限公司 | Panoramic video image processing method and device |
CN111595254A (en) * | 2020-06-04 | 2020-08-28 | 中国人民解放军陆军装甲兵学院 | Method and system for measuring axial distance between lens array and LCD display screen |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103475889A (en) * | 2013-09-01 | 2013-12-25 | 西安电子科技大学 | Method for obtaining micro cell image array in real time based on double cameras |
CN104469343A (en) * | 2014-11-26 | 2015-03-25 | 北京智谷技术服务有限公司 | Optical field display control and device and optical field display device |
CN104735360A (en) * | 2013-12-18 | 2015-06-24 | 华为技术有限公司 | Method and device for optical field image processing |
CN107121786A (en) * | 2017-05-18 | 2017-09-01 | 中国人民解放军装甲兵工程学院 | Integration imaging light field shows the calibration method of system lenses array |
CN107580207A (en) * | 2017-10-31 | 2018-01-12 | 武汉华星光电技术有限公司 | The generation method and generating means of light field 3D display cell picture |
US10061062B2 (en) * | 2015-10-25 | 2018-08-28 | Oculus Vr, Llc | Microlens array system with multiple discrete magnification |
-
2019
- 2019-10-16 CN CN201910981967.5A patent/CN110708532B/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103475889A (en) * | 2013-09-01 | 2013-12-25 | 西安电子科技大学 | Method for obtaining micro cell image array in real time based on double cameras |
CN104735360A (en) * | 2013-12-18 | 2015-06-24 | 华为技术有限公司 | Method and device for optical field image processing |
CN104469343A (en) * | 2014-11-26 | 2015-03-25 | 北京智谷技术服务有限公司 | Optical field display control and device and optical field display device |
US10061062B2 (en) * | 2015-10-25 | 2018-08-28 | Oculus Vr, Llc | Microlens array system with multiple discrete magnification |
CN107121786A (en) * | 2017-05-18 | 2017-09-01 | 中国人民解放军装甲兵工程学院 | Integration imaging light field shows the calibration method of system lenses array |
CN107580207A (en) * | 2017-10-31 | 2018-01-12 | 武汉华星光电技术有限公司 | The generation method and generating means of light field 3D display cell picture |
Non-Patent Citations (1)
Title |
---|
姚凯凯: "《集成成像三维显示微单元图像阵列快速生成方法》", 《中国优秀硕士学位论文全文数据库(电子期刊)》 * |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111225221A (en) * | 2020-01-15 | 2020-06-02 | 未来新视界文化科技(嘉善)有限公司 | Panoramic video image processing method and device |
CN111595254A (en) * | 2020-06-04 | 2020-08-28 | 中国人民解放军陆军装甲兵学院 | Method and system for measuring axial distance between lens array and LCD display screen |
CN111595254B (en) * | 2020-06-04 | 2021-09-21 | 中国人民解放军陆军装甲兵学院 | Method and system for measuring axial distance between lens array and LCD display screen |
Also Published As
Publication number | Publication date |
---|---|
CN110708532B (en) | 2021-03-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20230362344A1 (en) | System and Methods for Calibration of an Array Camera | |
JP7043085B2 (en) | Devices and methods for acquiring distance information from a viewpoint | |
JP5635218B1 (en) | Pattern alignment method and system for spatially encoded slide images | |
CN109813251B (en) | Method, device and system for three-dimensional measurement | |
CN107077743B (en) | System and method for dynamic calibration of an array camera | |
Genovese et al. | Stereo-camera calibration for large-scale DIC measurements with active phase targets and planar mirrors | |
CN109859272B (en) | Automatic focusing binocular camera calibration method and device | |
CN103776419B (en) | A kind of binocular distance measurement method improving measurement range | |
CN109325981B (en) | Geometric parameter calibration method for micro-lens array type optical field camera based on focusing image points | |
CN111080705B (en) | Calibration method and device for automatic focusing binocular camera | |
CN106612390B (en) | Camera module with double imaging modules and optical axis parallelism adjusting method thereof | |
US8749652B2 (en) | Imaging module having plural optical units in which each of at least two optical units include a polarization filter and at least one optical unit includes no polarization filter and image processing method and apparatus thereof | |
CN201662682U (en) | Stereo pick-up device | |
US8817121B2 (en) | Image apparatus and imaging method | |
JP2014182299A (en) | Microlens array unit and solid-state image pickup device | |
JP2019082680A (en) | Method, device, and method for calibration of three-dimensional display device | |
JP2016024052A (en) | Three-dimensional measurement system, three-dimensional measurement method and program | |
US20200221064A1 (en) | Combined stereoscopic and phase detection depth mapping in a dual aperture camera | |
JP7489253B2 (en) | Depth map generating device and program thereof, and depth map generating system | |
CN110708532B (en) | Universal light field unit image generation method and system | |
CN106888344A (en) | Camera module and its inclined acquisition methods of image planes and method of adjustment | |
CN112950727B (en) | Large-view-field multi-target simultaneous ranging method based on bionic curved compound eye | |
CN111127379B (en) | Rendering method of light field camera 2.0 and electronic equipment | |
KR102031485B1 (en) | Multi-view capturing apparatus and method using single 360-degree camera and planar mirrors | |
JP2018133064A (en) | Image processing apparatus, imaging apparatus, image processing method, and image processing program |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |