[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

CN116071237B - Video hyperspectral imaging method, system and medium based on filter sampling fusion - Google Patents

Video hyperspectral imaging method, system and medium based on filter sampling fusion Download PDF

Info

Publication number
CN116071237B
CN116071237B CN202310179342.3A CN202310179342A CN116071237B CN 116071237 B CN116071237 B CN 116071237B CN 202310179342 A CN202310179342 A CN 202310179342A CN 116071237 B CN116071237 B CN 116071237B
Authority
CN
China
Prior art keywords
hyperspectral
video
filter
image frame
video stream
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202310179342.3A
Other languages
Chinese (zh)
Other versions
CN116071237A (en
Inventor
李树涛
佃仁伟
单天赐
郭安静
韦晓辉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hunan University
Original Assignee
Hunan University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hunan University filed Critical Hunan University
Priority to CN202310179342.3A priority Critical patent/CN116071237B/en
Publication of CN116071237A publication Critical patent/CN116071237A/en
Application granted granted Critical
Publication of CN116071237B publication Critical patent/CN116071237B/en
Priority to PCT/CN2023/102544 priority patent/WO2024178879A1/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4053Scaling of whole images or parts thereof, e.g. expanding or contracting based on super-resolution, i.e. the output image resolution being higher than the sensor resolution
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01JMEASUREMENT OF INTENSITY, VELOCITY, SPECTRAL CONTENT, POLARISATION, PHASE OR PULSE CHARACTERISTICS OF INFRARED, VISIBLE OR ULTRAVIOLET LIGHT; COLORIMETRY; RADIATION PYROMETRY
    • G01J3/00Spectrometry; Spectrophotometry; Monochromators; Measuring colours
    • G01J3/28Investigating the spectrum
    • G01J3/2823Imaging spectrometer
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4046Scaling of whole images or parts thereof, e.g. expanding or contracting using neural networks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/262Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
    • H04N5/265Mixing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/01Conversion of standards, e.g. involving analogue television standards or digital television standards processed at pixel level
    • H04N7/0135Conversion of standards, e.g. involving analogue television standards or digital television standards processed at pixel level involving interpolation processes
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01JMEASUREMENT OF INTENSITY, VELOCITY, SPECTRAL CONTENT, POLARISATION, PHASE OR PULSE CHARACTERISTICS OF INFRARED, VISIBLE OR ULTRAVIOLET LIGHT; COLORIMETRY; RADIATION PYROMETRY
    • G01J3/00Spectrometry; Spectrophotometry; Monochromators; Measuring colours
    • G01J3/28Investigating the spectrum
    • G01J3/2823Imaging spectrometer
    • G01J2003/2826Multispectral imaging, e.g. filter imaging
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02ATECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
    • Y02A40/00Adaptation technologies in agriculture, forestry, livestock or agroalimentary production
    • Y02A40/10Adaptation technologies in agriculture, forestry, livestock or agroalimentary production in agriculture

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • Signal Processing (AREA)
  • Multimedia (AREA)
  • Spectroscopy & Molecular Physics (AREA)
  • Computational Linguistics (AREA)
  • Mathematical Physics (AREA)
  • Data Mining & Analysis (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Biophysics (AREA)
  • Software Systems (AREA)
  • Biomedical Technology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Image Processing (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a video hyperspectral imaging method, a system and a medium based on filter sampling fusion, which comprise the steps of obtaining RGB video stream and filter video stream obtained by a filter, extracting RGB image frame I and filter image frame F; upsampling RGB image frame I into shallow spectral feature map X 0 Reconstructing the spectrum into a grain refining characteristic diagram V; optimizing the filter curve of the filter image frame F to obtain a hyperspectral image X 1 The method comprises the steps of carrying out a first treatment on the surface of the The grain refining characteristic diagram V and the hyperspectral image X 1 Fused into hyperspectral image frames X that make up a hyperspectral video stream. The invention uses RGB video stream and filter video stream obtained by filter to process video hyperspectral fusion imaging, combines advanced deep learning technique to fully mine the characteristics among RGB image frame, filter image frame and hyperspectral image frame, can realize rapid imaging of hyperspectral video stream, and can be used for realizing small-volume high-spatial resolution video hyperspectral imaging equipment.

Description

Video hyperspectral imaging method, system and medium based on filter sampling fusion
Technical Field
The invention relates to the technical field of video image processing, in particular to a video hyperspectral imaging method, a system and a medium based on optical filter sampling fusion.
Background
Compared with the traditional RGB image, the hyperspectral image has higher spectral resolution and rich spectral information, and has irreplaceable effects in the fields of aerospace, medical treatment, industry and the like. However, due to limitations of imaging principles and imaging sensors, the time-space-spectral resolution of the existing hyperspectral imaging methods are mutually restricted, and thus it is difficult to acquire high-resolution hyperspectral videos. On the other hand, spectral imaging equipment is expensive to manufacture, and the application of hyperspectral is limited to a large extent. Currently, there are great advances and developments in the technology of obtaining multispectral images by means of a specific-response multi-filter combination. These filters are often narrow-band pass high-transmission filters, and the correlation between the filters is weak, so that it is difficult to recover the spectral information of the multispectral image. Meanwhile, the existing imaging apparatus can rapidly obtain an RGB image or an RGB video of high spatial resolution. Unfortunately, RGB images may lack much spectral information than hyperspectral images. The realization of video hyperspectral imaging by recovering spectra from RGB video has become a key technical problem to be solved urgently.
Disclosure of Invention
The invention aims to solve the technical problems: aiming at the problems in the prior art, the invention provides a video hyperspectral imaging method, a system and a medium based on filter sampling fusion, which are used for carrying out video hyperspectral fusion imaging by using RGB video stream and filter video stream obtained by a filter, fully excavating the characteristics among RGB image frames, filter image frames and hyperspectral image frames by combining with advanced deep learning technology, can realize rapid imaging of hyperspectral video stream, and can be used for realizing small-volume high-spatial-resolution video hyperspectral imaging equipment.
In order to solve the technical problems, the invention adopts the following technical scheme:
a video hyperspectral imaging method based on filter sampling fusion comprises the following steps:
s101, acquiring an RGB video stream and a filter video stream obtained through a filter, extracting an RGB image frame I from the RGB video stream in frames, and extracting a filter image frame F from the filter video stream in frames;
s102, up-sampling the RGB image frame I to obtain a shallow spectral feature map X 0 Performing spectrum reconstruction to obtain a fine grain characteristic diagram V; the filtering curve optimization is carried out on the filtering image frame F to obtain a hyperspectral image X 1
S103, the grain refining characteristic diagram V and the hyperspectral image X 1 Fused into hyperspectral image frames X that make up a hyperspectral video stream.
Optionally, in step S102, the RGB image frame I is up-sampled to obtain a shallow spectral feature map X 0 Comprising the following steps:
s201, subtracting a G wave band from an R wave band of an RGB image frame I to obtain RG wave band characteristics, subtracting a B wave band from a G wave band of the RGB image frame I to obtain GB wave band characteristics, and respectively taking the R wave band, the G wave band, the B wave band, the RG wave band characteristics and the GB wave band characteristics of the RGB image frame I as a group of characteristics in a combined characteristic diagram H to obtain a combined characteristic diagram H;
s202, performing spectral dimension up-sampling on the combined feature map H by adopting a trained spectral self-similarity guiding network layer to obtain a shallow spectral feature map X 0
Optionally, the functional expression of the trained spectral self-similarity guided network layer in step S202 is:
M i =Conv 3×3 (Conv 3×3 (Conv 3×3 (H i ))),
X 0 =Conv 3×3 ((ReLu(CAT[M 1 ,M 2 ,M 3 ,M 4 ,M 5 )),
in the above, M i Is the firstiGroup characteristic H i Conv 3×3 For convolution with a convolution kernel size of 3×3, reLu represents the ReLu activation function, CAT represents stacking in the channel dimension, M 1 ~M 5 Respectively 1 st to 5 th group of characteristics H i And shallow spectral feature map X 0 The same number of spectral bands as the hyperspectral image frame X.
Optionally, performing spectral reconstruction in step S102 to obtain a fine-grained feature map refers to subjecting the shallow spectral feature map X to 0 And inputting a trained spectrum reconstruction layer to perform spectrum reconstruction to obtain a grain refining characteristic diagram V, wherein the function expression of the spectrum reconstruction layer is as follows:
V=Res ×N (MAB(Res ×M (X 0 ))),
in the above, res ×M M residual network blocks representing a spectral reconstruction layer, MAB representing a hybrid attention module, res, of the spectral reconstruction layer ×N Representing N residual network blocks of the spectral reconstruction layer.
Optionally, in step S102, the filtered image frame F is subjected to filter curve optimization to obtain a hyperspectral imageX 1 The method is that the filtered image frame F is input into a trained filtering curve optimization layer to perform filtering curve optimization to obtain a hyperspectral image X 1 The function expression of the filter curve optimization layer is as follows:
X 1 = Conv 3×3 (Res ×K (Conv 3×3 (F))),
in the above, conv 3×3 For increasing convolution with a convolution kernel size of 3×3 for maintaining consistency of spectral channel numbers, res ×K For K residual network blocks to increase nonlinearity, F is the filtered image frame.
Optionally, the grain refinement feature map V and the hyperspectral image X are refined in step S103 1 The hyperspectral image frame X fused into the hyperspectral video stream is the grain-refining characteristic image V and hyperspectral image X 1 Inputting a trained fusion layer to obtain a hyperspectral image frame X forming a hyperspectral video stream, wherein the fusion layer has a function expression as follows:
X= Res ×K (V+X 1 )
in the above, res ×K For K residual network blocks.
Optionally, step S101 further includes: and acquiring RGB video stream samples and filtering video stream samples obtained through a filter, performing end-to-end supervised training on a video hyperspectral imaging network formed by a spectrum self-similarity guiding network layer, a spectrum reconstruction layer and a filtering curve optimization layer by utilizing training samples formed by the RGB video stream samples and the filtering video stream samples, wherein a loss function adopted during training is an average absolute error, and the average absolute error is the sum of absolute differences between a label value of the supervised training and an output value of the video hyperspectral imaging network.
In addition, the invention also provides a video hyperspectral imaging system based on the optical filter sampling fusion, which comprises a first video acquisition device, a second video acquisition device and a video hyperspectral imaging processing device, wherein the first video acquisition device is used for acquiring RGB video streams, the second video acquisition device is used for acquiring filtered video streams after passing through the optical filter, the output ends of the first video acquisition device and the second video acquisition device are respectively connected with the video hyperspectral imaging processing device, the video hyperspectral imaging processing device comprises a microprocessor and a memory which are connected with each other, and the microprocessor is programmed or configured to execute the video hyperspectral imaging method based on the optical filter sampling fusion.
In addition, the invention also provides a video hyperspectral imaging system based on the filter sampling fusion, which comprises a microprocessor and a memory which are connected with each other, wherein the microprocessor is programmed or configured to execute the video hyperspectral imaging method based on the filter sampling fusion.
Furthermore, the present invention provides a computer readable storage medium having stored therein a computer program for programming or configuring by a microprocessor to perform a filter sample fusion based video hyperspectral imaging method.
Compared with the prior art, the invention has the following advantages:
1. the invention uses RGB video stream and filter video stream obtained by filter to process video hyperspectral fusion imaging, based on the mapping relation between RGB video stream and filter video stream obtained by filter and hyperspectral image, and can effectively simulate the response parameter of filter set, and fully uses the internal correlation characteristic of the two, with interpretability (see the imaging principle part description below, in particular), the invention combines advanced deep learning technology to fully mine the characteristics of RGB image frame, filter image frame and hyperspectral image frame, and can realize rapid imaging of hyperspectral video stream, and can be used for realizing small-volume high-space resolution video hyperspectral imaging equipment.
2. The invention has better universality and robustness without changing the structure and parameters of the network when the invention is used for different types of filter slice groups and RGB images.
Drawings
FIG. 1 is a schematic diagram of a basic flow of a method according to an embodiment of the present invention.
Fig. 2 is a schematic diagram of a network structure of a video hyperspectral imaging network according to an embodiment of the present invention.
Fig. 3 is a schematic diagram of a network structure of a spectral self-similarity guiding network layer according to an embodiment of the present invention.
FIG. 4 is a comparison of the results of four imaging methods on a CAVE dataset hyperspectral image in an embodiment of the present invention.
Detailed Description
As shown in fig. 1, the video hyperspectral imaging method based on filter sampling fusion in this embodiment includes:
s101, acquiring an RGB video stream and a filter video stream obtained through a filter, extracting an RGB image frame I from the RGB video stream in frames, and extracting a filter image frame F from the filter video stream in frames;
s102, up-sampling the RGB image frame I to obtain a shallow spectral feature map X 0 Performing spectrum reconstruction to obtain a fine grain characteristic diagram V; the filtering curve optimization is carried out on the filtering image frame F to obtain a hyperspectral image X 1
S103, the grain refining characteristic diagram V and the hyperspectral image X 1 Fused into hyperspectral image frames X that make up a hyperspectral video stream.
The imaging principle of the video hyperspectral imaging method based on the filter sampling fusion in the embodiment is as follows:
first, for the relationship between the hyperspectral image frame X and the RGB image frame I, the following mapping relationship can be established:
I c (u,v)=Λ P c (λ)X(u,v,λ),(1)
in the above-mentioned method, the step of,I c (u,v) The channel c position for RGB image frame I is%u,v) Wherein lambda is the wavelength interval,P c (λ) As a function of the spectral response of channel c at wavelength lambda,X(u,v,λ) The position is [ ] under the wavelength lambdau,v) Spectral radiation at that time; channel c satisfies c epsilon { R, G, B }, wherein R, G, B are three channels (wave bands) of RGB image respectively, (-), the following are allu,v) Representing a certain position in space of the RGB image/hyperspectral image. In practical applications the wavelength interval is sampled discretely, so (1) can be rewritten as:
I c (u,v)=∑ n P c (λ)X(u,v,λ n ),(2)
in the above-mentioned method, the step of,nrepresenting the number of spectral bands,X(u,v,λ n ) Is of discrete wavelengthλ n The lower position is%u,v) Spectral radiation at that time.
Spectral response functionP c (λ) Is determined by the camera sensor and is related to the physical parameters. That is, in the case where the light source is necessarily noiseless, the hyperspectral image frame X and the RGB image frame I are in a linear relationship. In practice, however, when an RGB image is photographed, noise increases with an increase in exposure time and temperature, so that the hyperspectral image frame X and the RGB image frame I are not in a linear relationship in a real case.
Secondly, for the relation between the hyperspectral image frame X and the filtered image frame F, the following mapping relation can be established:
F=PBX
in the above formula, F is a filtered image, the size of the filtered image is w×h, and the number of wave bands is s; x is a hyperspectral image, the size of the hyperspectral image is w multiplied by h, and the wave band is S; b is a filter response matrix, the specification size of which is S×S, which can be essentially a band-pass filter, P is a spectral response matrix, the dimensions of which are s×S, w, h are the width and height of the hyperspectral image or the filtered image, respectively.
And thirdly, based on the relation between the hyperspectral image frame X and the RGB image frame I and the relation between the hyperspectral image frame X and the filter image frame F, the RGB image frame I can be extracted from the RGB video stream in a framing way, and the filter image frame F can be extracted from the filter video stream in a framing way, so that the hyperspectral image frame X of the corresponding imaging hyperspectral video stream can be realized.
In this embodiment, in step S102, the RGB image frame I is up-sampled to obtain a shallow spectral feature map X 0 Comprising the following steps:
s201, subtracting a G band from an R band of an RGB image frame I to obtain RG band characteristics, subtracting a B band from a G band of the RGB image frame I to obtain GB band characteristics, and respectively using the R band, the G band, the B band, the RG band characteristics and the GB band characteristics of the RGB image frame I as a group of characteristics in a combined characteristic diagram H to obtain a combined characteristic diagram H:
H =[R, RG, G,GB, B];
s202, performing spectral dimension up-sampling on the combined feature map H by adopting a trained spectral self-similarity guiding network layer to obtain a shallow spectral feature map X 0
In this embodiment, the functional expression of the trained spectral self-similarity guiding network layer in step S202 is:
M i =Conv 3×3 (Conv 3×3 (Conv 3×3 (H i ))),
X 0 =Conv 3×3 ((ReLu(CAT[M 1 ,M 2 ,M 3 ,M 4 ,M 5 )),
in the above, M i Is the firstiGroup characteristic H i Conv 3×3 For convolution with a convolution kernel size of 3×3, reLu represents the ReLu activation function, CAT represents stacking in the channel dimension, M 1 ~M 5 Respectively 1 st to 5 th group of characteristics H i And shallow spectral feature map X 0 The same number of spectral bands as the hyperspectral image frame X. As shown in fig. 3, the spectral self-similarity guiding network layer includes five convolution channels, each of which includes three concatenated convolution Conv with a convolution kernel size of 3×3 3×3 For extracting the firstiGroup characteristic H i Intermediate feature M of (2) i The method comprises the steps of carrying out a first treatment on the surface of the Then sequentially passing through a channel dimension stacking module to obtain results M of five convolution channels 1 ~M 5 Stacked in the channel dimension, then activated by a ReLu activation layer, and finally by a convolution kernel of size 3×Convolution Conv of 3 3×3 Outputting a shallow spectral feature map X 0
In this embodiment, the spectrum reconstruction in step S102 to obtain the fine grained feature map means that the shallow spectrum feature map X 0 And inputting a trained spectrum reconstruction layer to perform spectrum reconstruction to obtain a fine-grained feature map V, wherein the function expression of the spectrum reconstruction layer is as follows:
V= Conv 3×3 (Res ×N (MAB(Res ×M (Conv 3×3 (X 0 ))))),
in the above, conv 3×3 For convolutions with a convolution kernel size of 3 x 3 for dimension reduction or increase, res ×M M residual network blocks representing a spectral reconstruction layer, MAB representing a hybrid attention module, res, of the spectral reconstruction layer ×N Representing N residual network blocks of the spectral reconstruction layer. Referring to fig. 2, the spectrum reconstruction layer includes M residual network blocks (pre-layer), a mixed attention module MAB and N residual network blocks (post-layer) which are sequentially cascaded, so that hyperspectral features can be better learned. Where M and N may be selected as desired, for example m=4 and n=3 in the present embodiment. The spectrum reconstruction layer is combined with the residual error network block, so that the network has good convergence, a mixed attention mechanism is provided, the spatial spectrum characteristics of the hyperspectral image can be better learned, the parameter quantity is less, the video hyperspectral imaging speed is high, the precision is high, and better mobility is achieved.
In this embodiment, in step S102, the filtered image frame F is subjected to filter curve optimization to obtain the hyperspectral image X 1 The method is that the filtered image frame F is input into a trained filtering curve optimization layer to perform filtering curve optimization to obtain a hyperspectral image X 1 The function expression of the filter curve optimization layer is as follows:
X 1 = Conv 3×3 (Res ×K (Conv 3×3 (F))),
in the above, conv 3×3 For increasing convolution with a convolution kernel size of 3×3 for maintaining consistency of spectral channel numbers, res ×K In order to increase the nonlinear K residual network blocks, F is a filtered image frame, and the filtered image frame can be better simulated through multi-layer nonlinear convolution OR operationAnd (5) filtering the curve. Where K may be selected as desired, for example, in this embodiment, K has a value of k=3.
In the present embodiment, the grain refining feature map V and the hyperspectral image X are performed in step S103 1 The hyperspectral image frame X fused into the hyperspectral video stream is the grain-refining characteristic image V and hyperspectral image X 1 Inputting a trained fusion layer to obtain a hyperspectral image frame X forming a hyperspectral video stream, wherein the fusion layer has a function expression as follows:
X= Res ×T (V+X 1 )
in the above, res ×T For K residual network blocks. Wherein T may be selected as required, for example, in this embodiment, T takes a value of t=3.
In this embodiment, the step S101 further includes: and acquiring RGB video stream samples and filtering video stream samples obtained through a filter, performing end-to-end supervised training on a video hyperspectral imaging network formed by a spectrum self-similarity guiding network layer, a spectrum reconstruction layer and a filtering curve optimization layer by utilizing training samples formed by the RGB video stream samples and the filtering video stream samples, wherein a loss function adopted during training is an average absolute error, and the average absolute error is the sum of absolute differences between a label value of the supervised training and an output value of the video hyperspectral imaging network.
Because hyperspectral video can be divided into a plurality of frames of hyperspectral images, 32 disclosed by CAVE is utilized to carry out verification experiments on data sets in the embodiment, and in order to carry out validity verification on the video hyperspectral imaging method based on filter sampling fusion in the embodiment, a single frame of hyperspectral image is selected to carry out verification. We selected a hyperspectral public dataset CAVE dataset with a hyperspectral image band number of 31 and a spatial dimension of 512 x 512. In experiments, the hyperspectral image in the dataset was treated as a high resolution hyperspectral image, and a set of RGB images was downsampled as input images using a known spectral response function, while the filter response function was simulated according to the characteristics of the filtered images, and a set of filtered images was downsampled as input images. In the actual process, 20 pairs of data in the CAVE data set are used as training sets, 2 pairs of data are used as verification sets, 10 pairs of data are used as test sets, and 4 typical single RGB hyperspectral imaging methods are compared. The evaluation indexes of the fusion image are 4, namely a Spectrum Angle (SAM), a Root Mean Square Error (RMSE), a Unified Image Quality Index (UIQI) and a Structural Similarity (SSIM). Wherein a larger value of UIQI and SSIM indicates a better quality of the high resolution image and a larger value of SAM and RMSE indicates a worse quality of the high resolution image. Table 1 shows objective evaluation indices of imaging experiments on CAVE data sets for 4 typical imaging methods (Arad, HSCNN-R, AWAN+, HSRnet) and the method proposed in this example (TRFS), with the best numerical results being blackened.
Table 1: the method of this embodiment and four typical hyperspectral imaging methods provide objective performance metrics on the CAVE dataset.
Figure SMS_1
FIG. 4 shows the results of three exemplary imaging methods HSCNN-R, AWAN+, HSRnet and the method proposed in this example (FRFN) hyperspectral image imaging in a CAVE dataset, wherein the first row is the case hyperspectral image artwork in the CAVE dataset, wherein (a 1) is the case hyperspectral image artwork in the CAVE dataset, and (a 2) is the hyperspectral image error result plot of the HSCNN-R method in the CAVE dataset; wherein (b 1) is the original hyperspectral image of the case in the CAVE data set, and (b 2) is the error result graph of the hyperspectral image of the AWAN+ method in the CAVE data set; wherein (c 1) is the original hyperspectral image of the case in the CAVE data set, and (c 2) is the error result graph of the hyperspectral image of the HSRnet method in the CAVE data set; wherein (d 1) is the case hyperspectral image artwork in the CAVE dataset, and (d 2) is the hyperspectral image error result graph result of the method (FRFN) proposed by the embodiment in the CAVE dataset; wherein (e 1) is the original image of the hyperspectral image of the case in the CAVE data set, and (e 2) is the ideal error result image. As can be seen from table 1 and fig. 4, all objective evaluation indexes of the method (FRFN) proposed in this embodiment are superior to those of other methods, because the FRFN introduces the filtered image obtained by sampling the optical filter to form a fused image with the RGB image, so that the advantages of the two images can be combined, and a hyperspectral image with better effect can be reconstructed. More importantly, the adopted mixed attention mechanism can better learn the spatial spectrum characteristics of the hyperspectral image and save the spatial and spectral details of the image.
In summary, the video hyperspectral imaging method based on the filter sampling fusion in the method of the embodiment includes inputting an RGB video stream sampled by the filter and an RGB video stream photographed by a sensor; and then the hyperspectral video X after fusion is obtained through a filter response curve optimization layer, a spectral feature extraction layer, a reconstruction network layer, a fusion layer and the like frame by frame, so that hyperspectral video X with high spatial resolution can be effectively obtained based on filter sampling fusion, filter response parameter simulation and hyperspectral video reconstruction are realized, and the hyperspectral video X has strong universality and fewer network parameters. The video hyperspectral imaging method based on the filter sampling fusion utilizes the strong learning capability of the deep neural convolution network, and takes the filtered video obtained through the filter sampling as an auxiliary condition to realize hyperspectral video imaging. Firstly, carrying out spectrum up-sampling on a single frame image in an RGB video stream to obtain a shallow spectrum characteristic image, then sending the shallow spectrum characteristic image into a trained hyperspectral reconstruction network layer to further excavate potential association of RGB and hyperspectrum, and simultaneously designing a mixed attention mechanism to better extract the empty spectrum characteristic of the hyperspectral image. And simultaneously, carrying out spectral response curve optimization on a single frame image in the filtered video stream to obtain a hyperspectral image mapped with the filtered image, then sending the hyperspectral image and a shallow spectral feature image obtained through RGB image frames into a fusion layer, finally obtaining a high-resolution hyperspectral image frame, and then obtaining a multi-frame hyperspectral image and combining to obtain a hyperspectral video. The method has the advantages that extra hyperspectral data are not needed for training, no priori knowledge is needed, training is only needed on an RGB video and a filtered video data set which are easier to obtain, a filter response curve is simulated through a deep learning module, the method can be suitable for hyperspectral data of different types, the noise interference resistance is high, compared with other high-performance single RGB video hyperspectral imaging methods, hyperspectral videos obtained by the video hyperspectral imaging method based on the filter sampling fusion have better quality, and compared with the other high-performance single RGB video hyperspectral imaging method, hyperspectral videos obtained by multi-frame combination have better quality, have stronger noise interference resistance, and have strong universality and robustness for filtered videos and RGB videos which pass through different types of filter sets of different types, and the network structure and parameters do not need to be changed too much.
In addition, the embodiment also provides a video hyperspectral imaging system based on optical filter sampling fusion, which comprises a first video acquisition device, a second video acquisition device and a video hyperspectral imaging processing device, wherein the first video acquisition device is used for acquiring RGB video streams, the second video acquisition device is used for acquiring filtered video streams after passing through optical filters, the output ends of the first video acquisition device and the second video acquisition device are respectively connected with the video hyperspectral imaging processing device, the video hyperspectral imaging processing device comprises a microprocessor and a memory which are connected with each other, and the microprocessor is programmed or configured to execute the video hyperspectral imaging method based on the optical filter sampling fusion.
In addition, the embodiment also provides a video hyperspectral imaging system based on the filter sampling fusion, which comprises a microprocessor and a memory which are connected with each other, wherein the microprocessor is programmed or configured to execute the video hyperspectral imaging method based on the filter sampling fusion. Furthermore, the present embodiment also provides a computer readable storage medium having a computer program stored therein, the computer program being configured or programmed by a microprocessor to perform the aforementioned video hyperspectral imaging method based on filter sample fusion.
It will be appreciated by those skilled in the art that embodiments of the present application may be provided as a method, system, or computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment, or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-readable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein. The present application is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the application. It will be understood that each flow and/or block of the flowchart illustrations and/or block diagrams, and combinations of flows and/or blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks. These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks. These computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
The above description is only a preferred embodiment of the present invention, and the protection scope of the present invention is not limited to the above examples, and all technical solutions belonging to the concept of the present invention belong to the protection scope of the present invention. It should be noted that modifications and adaptations to the present invention may occur to one skilled in the art without departing from the principles of the present invention and are intended to be within the scope of the present invention.

Claims (6)

1. The video hyperspectral imaging method based on the optical filter sampling fusion is characterized by comprising the following steps of:
s101, acquiring an RGB video stream and a filter video stream obtained through a filter, extracting an RGB image frame I from the RGB video stream in frames, and extracting a filter image frame F from the filter video stream in frames;
s102, up-sampling the RGB image frame I to obtain a shallow spectral feature map X 0 Performing spectrum reconstruction to obtain a fine grain characteristic diagram V; the filtering curve optimization is carried out on the filtering image frame F to obtain a hyperspectral image X 1 The method comprises the steps of carrying out a first treatment on the surface of the The spectrum reconstruction to obtain a fine grain characteristic image refers to the shallow spectrum characteristic image X 0 And inputting a trained spectrum reconstruction layer to perform spectrum reconstruction to obtain a grain refining characteristic diagram V, wherein the function expression of the spectrum reconstruction layer is as follows:
V=Res ×N (MAB(Res ×M (X 0 ))),
in the above, res ×M M residual network blocks representing a spectral reconstruction layer, MAB representing a hybrid attention module, res, of the spectral reconstruction layer ×N N residual network blocks representing a spectral reconstruction layer; the filtering curve optimization is carried out on the filtering image frame F to obtain a hyperspectral image X 1 The method is that the filtered image frame F is input into a trained filtering curve optimization layer to perform filtering curve optimization to obtain a hyperspectral image X 1 The function expression of the filter curve optimization layer is as follows:
X 1 = Conv 3×3 (Res ×K (Conv 3×3 (F))),
in the above, conv 3×3 For increasing convolution with a convolution kernel size of 3×3 for maintaining consistency of spectral channel numbers, res ×K For K residual network blocks to increase nonlinearity, F is the filtered image frame;
s103, the grain refining characteristic diagram V and the hyperspectral image X 1 Fusing into hyperspectral image frames X forming a hyperspectral video stream;
in step S102, the RGB image frame I is up-sampled to obtain a shallow spectral feature map X 0 Comprising the following steps:
s201, subtracting a G wave band from an R wave band of an RGB image frame I to obtain RG wave band characteristics, subtracting a B wave band from a G wave band of the RGB image frame I to obtain GB wave band characteristics, and respectively taking the R wave band, the G wave band, the B wave band, the RG wave band characteristics and the GB wave band characteristics of the RGB image frame I as a group of characteristics in a combined characteristic diagram H to obtain a combined characteristic diagram H;
s202, performing spectral dimension up-sampling on the combined feature map H by adopting a trained spectral self-similarity guiding network layer to obtain a shallow spectral feature map X 0 The method comprises the steps of carrying out a first treatment on the surface of the Wherein the functional expression of the spectral self-similarity guiding network layer is as follows:
M i = Conv 3×3 (Conv 3×3 (Conv 3×3 (H i ))),
X 0 = Conv 3×3 ((ReLu(CAT[M 1 ,M 2 ,M 3 ,M 4 ,M 5 )),
in the above, M i Is the firstiGroup characteristic H i Conv 3×3 For convolution with a convolution kernel size of 3×3, reLu represents the ReLu activation function, CAT represents stacking in the channel dimension, M 1 ~M 5 Respectively 1 st to 5 th group of characteristics H i And shallow spectral feature map X 0 The same number of spectral bands as the hyperspectral image frame X.
2. The method for hyperspectral imaging of video based on filter sample fusion as claimed in claim 1, wherein the fine-grained feature map V, hyperspectral image X are processed in step S103 1 The hyperspectral image frame X fused into the hyperspectral video stream is the grain-refining characteristic image V and hyperspectral image X 1 Inputting a trained fusion layer to obtain a hyperspectral image frame X forming a hyperspectral video stream, wherein the fusion layer has a function expression as follows:
X= Res ×K (V+X 1 )
in the above, res ×K For K residual network blocks.
3. The method of video hyperspectral imaging based on filter sample fusion as claimed in claim 1, further comprising, prior to step S101: and acquiring RGB video stream samples and filtering video stream samples obtained through a filter, performing end-to-end supervised training on a video hyperspectral imaging network formed by a spectrum self-similarity guiding network layer, a spectrum reconstruction layer and a filtering curve optimization layer by utilizing training samples formed by the RGB video stream samples and the filtering video stream samples, wherein a loss function adopted during training is an average absolute error, and the average absolute error is the sum of absolute differences between a label value of the supervised training and an output value of the video hyperspectral imaging network.
4. The utility model provides a video hyperspectral imaging system based on light filter sampling fuses, its characterized in that includes first video acquisition device, second video acquisition device and video hyperspectral imaging processing equipment, first video acquisition device is used for gathering RGB video stream, second video acquisition device is used for gathering the filtering video stream after the light filter, the output of first video acquisition device, second video acquisition device links to each other with video hyperspectral imaging processing equipment respectively, video hyperspectral imaging processing equipment includes interconnect's microprocessor and memory, microprocessor is programmed or is configured in order to carry out the video hyperspectral imaging method based on light filter sampling fuses of any one of claims 1 ~ 3.
5. A filter sample fusion based video hyperspectral imaging system comprising a microprocessor and a memory connected to each other, wherein the microprocessor is programmed or configured to perform the filter sample fusion based video hyperspectral imaging method of any one of claims 1 to 3.
6. A computer readable storage medium having a computer program stored therein, wherein the computer program is for programming or configuring by a microprocessor to perform the filter sample fusion based video hyperspectral imaging method of any one of claims 1 to 3.
CN202310179342.3A 2023-03-01 2023-03-01 Video hyperspectral imaging method, system and medium based on filter sampling fusion Active CN116071237B (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN202310179342.3A CN116071237B (en) 2023-03-01 2023-03-01 Video hyperspectral imaging method, system and medium based on filter sampling fusion
PCT/CN2023/102544 WO2024178879A1 (en) 2023-03-01 2023-06-27 Optical filter sampling and fusion-based video hyperspectral imaging method and system, and medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310179342.3A CN116071237B (en) 2023-03-01 2023-03-01 Video hyperspectral imaging method, system and medium based on filter sampling fusion

Publications (2)

Publication Number Publication Date
CN116071237A CN116071237A (en) 2023-05-05
CN116071237B true CN116071237B (en) 2023-06-20

Family

ID=86182012

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310179342.3A Active CN116071237B (en) 2023-03-01 2023-03-01 Video hyperspectral imaging method, system and medium based on filter sampling fusion

Country Status (2)

Country Link
CN (1) CN116071237B (en)
WO (1) WO2024178879A1 (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116071237B (en) * 2023-03-01 2023-06-20 湖南大学 Video hyperspectral imaging method, system and medium based on filter sampling fusion

Family Cites Families (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2018165605A1 (en) * 2017-03-09 2018-09-13 Northwestern University Hyperspectral imaging sensor
WO2020199205A1 (en) * 2019-04-04 2020-10-08 合刃科技(深圳)有限公司 Hybrid hyperspectral image reconstruction method and system
US11860034B2 (en) * 2020-05-08 2024-01-02 Raytheon BBN Technologies, Corp. Systems, devices, and methods for hyperspectral imaging
CN113222823B (en) * 2021-06-02 2022-04-15 国网湖南省电力有限公司 Hyperspectral image super-resolution method based on mixed attention network fusion
CN113743001B (en) * 2021-08-13 2023-12-12 湖南大学 Spectral super-resolution reconstruction-oriented optical filter design method, optical filter and system
CN114757831B (en) * 2022-06-13 2022-09-06 湖南大学 High-resolution video hyperspectral imaging method, device and medium based on intelligent space-spectrum fusion
CN114998109B (en) * 2022-08-03 2022-10-25 湖南大学 Hyperspectral imaging method, system and medium based on dual RGB image fusion
CN115345909B (en) * 2022-10-18 2023-01-24 无锡学院 Hyperspectral target tracking method based on depth space spectrum convolution fusion characteristics
CN115700727A (en) * 2023-01-03 2023-02-07 湖南大学 Spectral super-resolution reconstruction method and system based on self-attention mechanism
CN116071237B (en) * 2023-03-01 2023-06-20 湖南大学 Video hyperspectral imaging method, system and medium based on filter sampling fusion

Also Published As

Publication number Publication date
WO2024178879A1 (en) 2024-09-06
CN116071237A (en) 2023-05-05

Similar Documents

Publication Publication Date Title
Anwar et al. Densely residual laplacian super-resolution
CN109064396B (en) Single image super-resolution reconstruction method based on deep component learning network
CN110009590B (en) High-quality color image demosaicing method based on convolutional neural network
CN113139898B (en) Light field image super-resolution reconstruction method based on frequency domain analysis and deep learning
Luo et al. Lattice network for lightweight image restoration
WO2024027095A1 (en) Hyperspectral imaging method and system based on double rgb image fusion, and medium
CN112801881B (en) High-resolution hyperspectral calculation imaging method, system and medium
CN112561799A (en) Infrared image super-resolution reconstruction method
CN116152120A (en) Low-light image enhancement method and device integrating high-low frequency characteristic information
CN111951164A (en) Image super-resolution reconstruction network structure and image reconstruction effect analysis method
CN116071237B (en) Video hyperspectral imaging method, system and medium based on filter sampling fusion
CN110717947B (en) High-quality spectrum reconstruction method based on external and internal training
CN112184553A (en) Hyperspectral image super-resolution method based on depth prior information
CN114841856A (en) Image super-pixel reconstruction method of dense connection network based on depth residual channel space attention
CN114266957A (en) Hyperspectral image super-resolution restoration method based on multi-degradation mode data augmentation
CN114926337A (en) Single image super-resolution reconstruction method and system based on CNN and Transformer hybrid network
Yang et al. RSAMSR: A deep neural network based on residual self-encoding and attention mechanism for image super-resolution
CN107203968A (en) Single image super resolution ratio reconstruction method based on improved subspace tracing algorithm
CN113743001B (en) Spectral super-resolution reconstruction-oriented optical filter design method, optical filter and system
CN110895790A (en) Scene image super-resolution method based on posterior degradation information estimation
CN116433548A (en) Hyperspectral and panchromatic image fusion method based on multistage information extraction
CN116823662A (en) Image denoising and deblurring method fused with original features
CN116205809A (en) Image restoration model and method for progressive guided decoding network
CN116612009A (en) Multi-scale connection generation countermeasure network medical image super-resolution reconstruction method
CN113793262A (en) Image demosaicing method based on residual error feature aggregation attention block

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant