SYSTEMS AND METHODS FOR HYPERSPECTRAL IMAGING OF PLANTS

Information

  • Patent Application
  • 20240060883
  • Publication Number
    20240060883
  • Date Filed
    December 17, 2021
    2 years ago
  • Date Published
    February 22, 2024
    2 months ago
Abstract
Systems and methods for hyperspectral imaging of plants are provided. Multispectral images of plants are transformed, e.g. by interpolation along a spectral axis, to generate hyperspectral images of plants. The transformation can be based on spectral bases formed from hyperspectral sample images including images of plant matter. Plant characteristics, such as plant health, may be predicted based on the hyperspectral image. Plant health may be predicted by comparing derivatives of reflectance values with respect to wavelength for a plant of a given image relative to a reference derivative based on a reference hyperspectral image. The derivatives may be compared by determining a regression loss.
Description
TECHNICAL FIELD

The present disclosure relates generally to hyperspectral imaging, and in particular to hyperspectral imaging of plants.


BACKGROUND

Hyperspectral imaging involves capturing light across many spectral bands. Hyperspectral images are sometimes referred to as “hyperspectral cubes” to reflect their (typically) two spatial dimensions (e.g. corresponding to a two-dimensional array of pixels) and their spectral dimension (corresponding to various wavelengths, sometimes binned as “channels”). A hyperspectral image may have tens, hundreds, or even thousands of channels. Hyperspectral imaging has found applications in agriculture and other plant-related disciplines; such applications may make use of plants' spectral characteristics which are not readily distinguishable in a conventional RGB image.


For example, hyperspectral images have been used to determine characteristics of plants, such as detecting the presence of plant matter and/or distinguishing between healthy and unhealthy plants. Such approaches typically make use of the spectral characteristics of healthy plants (and/or portions of plants), which tend to have more reflectance intensity in certain wavelengths relative to unhealthy plants and/or non-plant objects. See, for example, U.S. Pat. No. 7,715,013. A variety of measures of relative reflectance are used, of which the most common is normalized differential vegetation index (NDVI), which can be expressed as (NIR−red)/(NIR+red), where NIR is reflectance measured in near-infrared wavelengths (where healthy plants tend to have high reflectance) and red is reflectance in red wavelengths (where healthy plants tend to have limited reflectance). Plants with a high NDVI value can be predicted to be more likely to be healthy than those with lower NDVI values. See, for example, US Patent Publication No. 2019/0236358.


Hyperspectral images often require complex and costly imaging devices and may require significant computational resources to store and process the resulting images. It can be preferable to use multispectral images, which generally provide fewer spectral resolution than hyperspectral images (e.g. on the order of 3-15 channels) but can provide greater spatial resolution. Although multispectral images will generally provide less dense spectral information, the foregoing measures require only a few channels and so can be amenable to multispectral imaging contexts. See, for example, US Patent Publication No. 2019/0236358.


There is a general desire for improved hyperspectral imaging techniques for plants and techniques for making use of such hyperspectral images to determine characteristics of plants.


The foregoing examples of the related art and limitations related thereto are intended to be illustrative and not exclusive. Other limitations of the related art will become apparent to those of skill in the art upon a reading of the specification and a study of the drawings.


SUMMARY

The following embodiments and aspects thereof are described and illustrated in conjunction with systems, tools and methods which are meant to be exemplary and illustrative, not limiting in scope. In various embodiments, one or more of the above-described problems have been reduced or eliminated, while other embodiments are directed to other improvements.


One aspect of the invention provides a system and method for hyperspectral imaging of plants. The system comprises one or more processors and a memory storing instructions which cause the one or more processors to perform operations according to the method. The method comprises receiving a multispectral image comprising a number m of multispectral channels, at least one multispectral channel comprising an infrared wavelength, the multispectral image representing at least a portion of at least one plant; generating a hyperspectral image comprising a number n of hyperspectral channels based on the multispectral image and a plurality of spectral bases, the number n of hyperspectral channels greater than the number m of multispectral channels; and generating a determination for the at least the portion of at least one plant based on the hyperspectral image.


In some embodiments, the determination comprises a prediction of plant health and generating the determination comprises generating the prediction based on a plurality of reflectance values of the hyperspectral image.


In some embodiments, determining the prediction of plant health comprises determining a derivative of the plurality of reflectance values with respect to wavelength. In some embodiments, determining the prediction of plant health comprises determining the prediction of plant health based on the derivative of the plurality of reflectance values and a plurality of reference reflectance values.


In some embodiments, the method further comprises generating the plurality of reference reflectance values based on a reference hyperspectral image representing at least a healthy portion of a reference plant. In some embodiments, generating the plurality of reference reflectance values comprises determining an average of reflectance values for a plurality of spatial locations of the at least the healthy portion of the reference plant for each of a plurality of the n hyperspectral channels.


In some embodiments, determining the prediction of plant health comprises determining a difference between the derivative of the plurality of reflectance values and a derivative of the plurality of reference reflectance values with respect to wavelength. In some embodiments, determining the difference comprises determining a regression loss metric based on the derivative of the plurality of reflectance values and a derivative of the plurality of reference reflectance values. In some embodiments, the regression loss comprises at least one of: a mean square error, a mean absolute error, a Huber loss, a log-cosh loss, and a quantile loss.


In some embodiments, the plurality of reference reflectance values comprise a first plurality of reference reflectance values corresponding to at least a first portion of at least a first plant and a second plurality of reference reflectance values corresponding to at least a second portion of at least a second plant, the first and second portions differing in at least one of: species of plant, type of disease, type of damage, and degree of damage.


In some embodiments, determining the prediction of plant health comprises: determining a first prediction of plant health based on the derivative of the plurality of reflectance values and the first plurality of reference reflectance values; determining a second prediction of plant health based on the derivative of the plurality of reflectance values and the second plurality of reference reflectance values; and selecting the first prediction based on the first prediction corresponding to a greater likelihood of health than the second prediction.


In some embodiments, the spectral bases having been generated from one or more images comprising at least one image representing at least a further portion of at least one further plant. In some embodiments, the plurality of spectral bases comprises at least four spectral bases.


In some embodiments, generating the hyperspectral image comprises interpolating at least one hyperspectral reflectance value for a wavelength of at least one of the n hyperspectral channels outside of the m multispectral channels.


In some embodiments, the method further comprises segmenting the multispectral image into plant and non-plant regions; wherein generating the hyperspectral image comprises generating the hyperspectral image for the plant regions.


In some embodiments, the method further comprises: receiving a calibration multispectral image representing at least a portion of a calibration subject, the at least the portion of the calibration subject substantially non-reflective in one or more multispectral channels of the m multispectral channels; and determining, for at least one of the one or more multispectral channels, a corresponding calibration reflectance of at least a portion of the multispectral image representing at least the portion of the calibration subject; wherein generating the hyperspectral image comprises, for the at least one of the one or more multispectral channels, subtracting the corresponding calibration reflectance.


In some embodiments, at least one of the m multispectral channels comprises at least one wavelength in a range of about 525 nm to 575 nm. In some embodiments, at least one of the m multispectral channels comprises at least one wavelength in a range of about 600 nm to 700 nm. In some embodiments, at least one of the m multispectral channels comprises at least one wavelength in a range of about 400 nm to 500 nm. In some embodiments, the m multispectral channels comprise at least four multispectral channels. In some embodiments, the m multispectral channels comprise no more than ten multispectral channels.


In some embodiments, receiving the multispectral image comprises causing an imaging sensor having infrared sensitivity to capture one or more frames through one or more optical filters.


In some embodiments, the imaging sensor comprises at least one of: an RGB imaging sensor with NIR sensitivity and a monochrome imaging sensor; the one or more optical filters comprise a plurality of optical filters; and causing the imaging sensor to capture one or more frames comprises causing the imaging sensor to capture a plurality of frames by capturing at least one frame through each of the plurality of optical filters.


In some embodiments, causing the imaging sensor to capture the plurality of frames by capturing at least one frame through each of the plurality of optical filters comprises causing the plurality of optical filters to revolve through a field of view of the imaging sensor while causing the imaging sensor to capture frames.


One aspect of the invention provides a system and method for hyperspectral characterization of plants. The system comprises one or more processors and a memory storing instructions which cause the one or more processors to perform operations according to the method. The method comprises receiving a hyperspectral image comprising a number m of hyperspectral channels, at least one hyperspectral channel comprising an infrared wavelength, the hyperspectral image representing at least a portion of at least one plant; generating a determination for the at least the portion of at least one plant based on the hyperspectral image based on a derivative of a plurality of reflectance values of the hyperspectral image with respect to wavelength.


In some embodiments, the determination comprises a prediction of plant health and generating the determination comprises determining the prediction of plant health based on the derivative of the plurality of reflectance values and a plurality of reference reflectance values.


In some embodiments, the method comprises generating the plurality of reference reflectance values based on a reference hyperspectral image representing at least a healthy portion of a reference plant. In some embodiments, generating the plurality of reference reflectance values comprises determining an average of reflectance values for a plurality of spatial locations of the at least the healthy portion of the reference plant for each of a plurality of the n hyperspectral channels.


In some embodiments, determining the prediction of plant health comprises determining a difference between the derivative of the plurality of reflectance values and a derivative of the plurality of reference reflectance values with respect to wavelength. In some embodiments, determining the difference comprises determining a regression loss metric based on the derivative of the plurality of reflectance values and a derivative of the plurality of reference reflectance values. In some embodiments, the regression loss comprises at least one of: a mean square error, a mean absolute error, a Huber loss, a log-cosh loss, and a quantile loss.


In some embodiments, the plurality of reference reflectance values comprise a first plurality of reference reflectance values corresponding to at least a first portion of at least a first plant and a second plurality of reference reflectance values corresponding to at least a second portion of at least a second plant, the first and second portions differing in at least one of: species of plant, type of disease, type of damage, and degree of damage.


In some embodiments, determining the prediction of plant health comprises: determining a first prediction of plant health based on the derivative of the plurality of reflectance values and the first plurality of reference reflectance values; determining a second prediction of plant health based on the derivative of the plurality of reflectance values and the second plurality of reference reflectance values; and selecting the first prediction based on the first prediction corresponding to a greater likelihood of health than the second prediction.


In some embodiments, the method comprises one or more acts of the method of hyperspectral imaging of plants described above.


In addition to the exemplary aspects and embodiments described above, further aspects and embodiments will become apparent by reference to the drawings and by study of the following detailed descriptions.





BRIEF DESCRIPTION OF THE DRAWINGS

Exemplary embodiments are illustrated in referenced figures of the drawings. It is intended that the embodiments and figures disclosed herein are to be considered illustrative rather than restrictive.



FIG. 1 is a flowchart illustrating an example method for generating hyperspectral images based on multispectral images.



FIG. 2 is a flowchart illustrating an example method for characterizing plants based on reflectance values of a hyperspectral image, such as a hyperspectral image generated according to the method of FIG. 1 (or otherwise obtained).



FIG. 3 is a perspective view schematic diagram illustrating an example apparatus for imaging plants, which images may be used by the methods of FIGS. 1 and 2.



FIG. 4 is a side elevation view schematic diagram of the example apparatus of FIG. 3.



FIG. 5 is a detail view schematic diagram of a portion of the example apparatus of FIG. 3 shown in a perspective orientation generally from below, illustrating generally an example planter, illumination source, and camera.



FIG. 6 shows a first exemplary operating environment that includes at least one computing system for performing methods described herein, such as the methods of FIGS. 1 and 2.





DESCRIPTION

Throughout the following description specific details are set forth in order to provide a more thorough understanding to persons skilled in the art. However, well known elements may not have been shown or described in detail to avoid unnecessarily obscuring the disclosure. Accordingly, the description and drawings are to be regarded in an illustrative, rather than a restrictive, sense.


The present disclosure provides systems and methods for hyperspectral imaging of plants. One aspect of the present disclosure relates to techniques for generating hyperspectral images of plants from multispectral images of plants. Such generating can take advantage of the distinctive spectral characteristics of healthy plants through the use of spectral bases generated from sample hyperspectral images of plants. Spectral information can be interpolated from the multispectral image based on the spectral bases to generate the hyperspectral image. Such techniques may, in suitable circumstances, allow for analysis of plants based on images which are more spectrally dense, more spectrally accurate, more spatially dense, and/or more efficient to generate (e.g. in terms of reduced imaging and/or computational resources required) than might otherwise be provided by conventional multispectral or hyperspectral imaging techniques. This aspect includes an example apparatus for capturing multispectral images as part of a system for generating hyperspectral images therefrom.


Another aspect of the present disclosure relates to generating determinations about plants from multispectral images (which may comprise hyperspectral images). For instance, plant health can be predicted by comparing derivatives of reflectance values with respect to wavelength for a plant of a given hyperspectral image relative to a reference derivative based on a reference hyperspectral image (e.g. of a healthy plant). The derivatives may be compared, for example, based on a difference between the derivatives, e.g. by determining a regression loss. Such techniques may, in suitable circumstances, provide more accurate characterization of healthy vs. unhealthy plants (by making use of more complete spectral information to characterize plants' spectral response), and/or provide improved consistency between illumination intensities.


For the purposes of this disclosure and the appended claims, “multispectral” and “hyperspectral” include pairs of images where one image (the multispectral image) represents fewer spectral channels and/or wavelengths than another image (the hyperspectral image), where the hyperspectral image is generated from the multispectral image. Thus, for a pair of images where a first one has 100 spectral channels and a second one (generated from the first one) has 200 spectral channels, the first image is “multispectral” in that context and the second image is “hyperspectral” in that context. Similarly, for a pair of images where a first one has 3 spectral channels and a second one (generated from the first one) has 5 spectral channels, the first image is “multispectral” in that context and the second image is “hyperspectral” in that context. The terms “multispectral” and “hyperspectral” are used, at least in part, for convenience to distinguish between such images, and not to require a specific number of spectral channels.


A Method for Generating Hyperspectral Images of Plants from Multispectral Images



FIG. 1 is a flowchart illustrating an example method 100 for generating hyperspectral images based on multispectral images. Method 100 is performed by a computing system, as described elsewhere herein (e.g. with reference to FIG. 6). Method 100 may, optionally, comprise calibration acts 110 (e.g. based on one or more reference and/or calibration images). Method 100 acquires a multispectral image at act 122, generates a hyperspectral image therefrom at acts 130, and determines one or more plant characteristics based on the hyperspectral image at act 142. Each of these acts is discussed in greater detail below.


Method 100 involves interpolating spectral information based on a plurality of spectral bases. Such spectral bases may be predetermined, generated as part of method 100, and/or otherwise obtained. In some embodiments, including the illustrated example of FIG. 1, method 100 generates a plurality of spectral bases based on one or more reference images at act 112. Reference images may comprise hyperspectral images, which may have no spatial dimension (e.g. hyperspectral readings produced by a spectrometer), one or two spatial dimensions, and/or any other suitable number of spatial dimensions.


Act 112 may comprise determining characteristic spectra of the one or more reference images and describing the characteristic spectra as spectral bases comprising a set of basis vectors. For instance, given a set of p reference images, each image comprising spectral intensity s(λ)=[s(λ1), s(λ2), . . . , s(λn)]T where s(λi) is the spectral intensity for the ith wavelength (or channel) λi, the computing system may generate a correlation matrix R=Σi=1psi(λ)si(λ)T and determine therefrom the eigenvectors {σj} of R. Each eigenvector is a potential basis vector; the computing system may generate the spectral bases by selecting a plurality of the eigenvectors (which may comprise some or all of the eigenvectors σj). The computing system may optionally transform the basis vectors, e.g. by normalizing them. Further details on generation of spectral bases is provided by Parkkinen et al., Characteristic spectra of munsell colors, Journal of the Optical Society of America A 6 (1989) 318-322, which is incorporated by reference.


In some embodiments, the computing system generates (and/or otherwise receives) one or more spectral bases based on one or more reference images representing at least a portion of a plant. For example, the one or more reference images may comprise one or more hyperspectral images of healthy plants. In at least one embodiment, the one or more reference images comprise a plurality of hyperspectral images of non-plant matter such as Munsell chips (substantially as described by Parkkinen et al., referenced above) and further comprise hyperspectral images of healthy plants and/or portions thereof, such as healthy leaves. Such hyperspectral reference images of healthy plants may comprise measurements of plants' reflectance intensity in infrared spectral wavelengths, such as in near-IR spectral wavelengths (e.g. approx. 700 nm to 800 nm). In some embodiments, the computing system generates a number of eigenvalues as described above based on the reference images and selects from them a number of spectral bases. In at least one embodiment, the computing system selects four spectral bases. Experimentation with such example embodiments has demonstrated that the accuracy of hyperspectral interpolation of method 100 can be improved by the inclusion of such plant spectral information in the spectral bases, in at least some circumstances.


Optionally, at act 116 the computing system determines one or more intensities of reflectance of a calibration image (called calibration reflectance values herein), e.g. for optional use in calibrating images at act 132. The calibration image may comprise, for example, a multispectral image (e.g. having the same or similar multispectral channels to those received at act 122, described elsewhere herein) representing a calibration subject which is substantially non-reflective in one or more of the calibration image's multispectral channels. In some embodiments the calibration subject comprises a black patch positioned in the field of view of a multispectral imaging apparatus, the black patch being substantially non-reflective in visible and near-IR spectra. The computing system may measure an intensity of reflectance at one or more spatial locations (e.g. pixels of the multispectral image) representing the calibration subject for at least one of the spectral channels of the calibration image and determine from these a calibration reflectance (e.g. by using the value of the spectral intensity as provided by the multispectral image, by averaging the intensities in a channel at multiple spatial locations, and/or in any other suitable way). For example, in at least one embodiment, the computing system calculates an average calibration reflectance intensity c(λi) for a plurality of pixels representing a substantially non-reflective black patch in each channel λi.


At act 122, the computing system receives a multispectral image representing at least a portion of a plant. The multispectral image comprises m spectral channels (called multispectral channels herein for convenience), at least one of which comprises an infrared wavelength. For instance, the multispectral image may comprise at least one channel comprising a wavelength in the near-IR spectrum, e.g. in approx. 700 nm to 800 nm and/or 700 nm to 1000 nm. The multispectral image may comprise at least one channel in the green spectrum, e.g. in approx. 525 nm to 575 nm. The multispectral image may comprise one or more channels with wavelengths shorter than the green spectrum (e.g. in the 400 nm to 500 nm range), between the green and infrared spectra (e.g. in the 600 nm to 700 nm range), and/or longer than the near-IR spectrum (e.g. longer than 1000 nm). In some embodiments, the multispectral image comprises at least four spectral channels, to aid in interpolation.


In some example embodiments, the multispectral image comprises seven channels with center wavelengths of 450 nm, 500 nm, 550 nm, 600 nm, 650 nm, 700 nm, and 750 nm, respectively. In one example embodiment, each of these seven channels has a FWHM bandwidth of 50 nm. In another example embodiment, each such channel has a FWHM bandwidth of 25 nm. These example embodiments further comprise an eighth channel covering the visible spectrum (roughly 400 nm to 700 nm).


In some embodiments, act 122 comprises generating the multispectral image from frames generated by an imaging device. For example (e.g. as described in greater detail with reference to FIG. 6), the computing system may receive a plurality of frames captured by an imaging device, each frame captured through one of a plurality of optical filters (which may, e.g., revolve through a field of view of the imaging device as it captures frames). The imaging device may comprise, for example, an RGB camera with near-IR sensitivity (e.g. in the blue channel) generating 3-channel RGB frames, a monochrome CCD sensor generating single-channel frames, and/or any other suitable imaging device. The computing system may combine the plurality of frames into a multispectral image, e.g. with each multispectral channel corresponding to an optical filter. In at least some embodiments, the computing system receives frames as raw sensor data—e.g. without automatic white balancing, adjusted exposure, or other common adjustments (often intended to improve the appearance of images to the human eye), which modify spectral characteristics and can interfere with certain applications, such as those which require accurate readings of relative reflectance.


In some embodiments the imaging device generates multichannel frames (e.g. as with an RGB camera). In such embodiments, act 122 may comprise combining the multiple channels of a frame into a single channel (and/or a reduced number of channels) and/or combining frames' channels non-destructively (e.g. via concatenation). For example, act 122 may comprise generating a multispectral image having a channel for each optical filter through which the imaging device generates an image by summing the intensities of each channel for a given frame to form one intensity for the multispectral channel corresponding to the optical filter through which the frame was captured. In some embodiments, summing the intensities of a frame's channels may comprise performing a weighted sum and/or integration of the intensities of the channels, weighted according to the relative sensitivities of the imaging device's sensors for each of the imaging device's channels.


Optionally, at act 132, the computing system calibrates the multispectral image. Calibration may comprise, for example, correcting one or more channels of the multispectral image based on the calibration reflectance values generated at act 116 (e.g. based on a black patch). For instance, the computing system may subtract, from each of one or more channels (and optionally all channels) of the multispectral image, the calibration reflectance corresponding to that channel. This can reduce the effect of optical imperfections in the imaging system, unexpected/leaky illumination in the environment, and/or other miscalibrations. Act 132 may alternatively or additionally comprise any suitable calibration technique, such as adding a highly reflective calibration subject (e.g. a diffuse white board), such as is described in greater detail by Han et al. (2011) Fast Spectral Reflectance Recovery Using DLP Projector, Computer Vision—ACCV 2010. Lecture Notes in Computer Science, vol 6492. doi:10.1007/978-3-642-19315-6_25, which is incorporated herein by reference.


Act 132 may occur as part of act 122, act 136, and/or separately (e.g. afterwards). For example, suppose the calibration image comprises a plurality of RGB frames, e.g. in an exemplary embodiment where the imaging device comprises an RGB camera as described above with reference to act 122. Suppose also that the calibration image represents a black patch which is substantially non-reflective in each of the channels of the calibration image (and/or the multispectral image). If the computing system generated at act 116 a calibration reflectance of [16,0,0] for an RGB frame of the calibration image captured through a 650 nm optical filter (e.g. with 50 nm FWHM bandwidth), thereby indicating an intensity of 16 in the “red” channel, the computing system may subtract that calibration reflectance from an RGB frame received at act 122 which is also associated with the 650 nm optical filter. For instance, if that RGB frame has intensities [32, 105, 12], the computing system may correct those intensities based on the calibration reflectance to determine a corrected reflectance of [16, 105, 12]. The computing system may proceed to combine the corrected RGB frame with other frames as described in act 122. Alternatively, or in addition, the computing system may combine frames for the calibration image to generate a single, scalar calibration reflectance for a given wavelength and may correct the multispectral image of act 122 based on such scalar calibration reflectance. In some embodiments, the computing system alternatively or additionally corrects reflectance values of the hyperspectral image of act 136 based on the calibration reflectance values of act 116.


Optionally, at act 134 the computing system segments the multispectral image to classify regions of the multispectral image as representing plant or non-plant (e.g. background) objects. Any suitable segmenting method may be used; for example, the computing system may classify plants as foreground and non-plants as background based on Otsu thresholding. In some embodiments, the computing system performs segmentation based on one frame of the multispectral image (e.g. a frame captured through a shortpass filter covering some or all of the visible spectrum, and/or a frame comprising a conventional RGB image to facilitate segmentation by available segmentation models) and may apply that segmentation mask to all frames of the multispectral image. Subsequent acts based on the hyperspectral image, such as interpolation at act 136, may be limited to portions of the multispectral image classified as plant. This may, for example, reduce the computational resources required to generate the hyperspectral image at act 136 and/or generate determinations about plant characteristics at act 142. In some embodiments, the computing system alternatively or additionally performs segmentation on the hyperspectral image generated from the multispectral image.


In some embodiments, the computing system applies morphological adjustments to reduce the likelihood of including non-plant objects in plant-labelled regions, such as by applying binary closing and/or binary erosion. Although such adjustments are not always desirable, for at least some applications of the present techniques it can be desirable to make such adjustments to reduce the likelihood that non-plant objects will be included in regions classified as plant. For example, where the computing system will use the resulting hyperspectral image to assess plant health based on the spectral characteristics of the plant, such adjustments may be desirable in suitable circumstances.


At act 136, the computing system generates a hyperspectral image based on the multispectral image and a plurality of spectral bases (e.g. the spectral bases generated at act 112 and/or otherwise obtained). The hyperspectral image at least partially represents the (at least a portion of a) plant represented in the multispectral image. The computing system generates the hyperspectral image to comprise a greater number n of spectral channels (called hyperspectral channels herein for convenience) than the number m of multispectral channels of the multispectral image. Such generating may comprise interpolating at least one reflectance value for a spatial location (e.g. a pixel) of the hyperspectral image and for a given wavelength (e.g. corresponding to one of the hyperspectral channels) outside of the m multispectral channels. Such interpolation may be based on the m multispectral channels.


For example, supposing the imaging device generating the multispectral image has a linear intensity response and generates frames having one or more spectral channels, the intensity Im,n of a spatial location (e.g. a pixel) in the hyperspectral image may be determined based on:






I
m,n
=∫s(λ)cm(λ)In(λ)


where λ is the wavelength (and/or channel), s(λ) is the spectral reflectance at the spatial location for λ, cm(λ) is the spectral response function of the imaging device at the mth colour channel, and ln(λ) is the spectrum of the nth frame.


Spectral reflectance s(λ) can be recovered from such a linear model based on the spectral bases. In particular, the spectral reflectance for a given spatial location may be determined based on:







s

(
λ
)

=




j
=
1

i



α
j




b
j

(
λ
)







where bj(λ) is the jth spectral basis and αj is a corresponding coefficient which may be estimated based on any suitable technique. An example technique for estimating aj is provided, for example, by Han et al. (2011) Fast Spectral Reflectance Recovery Using DLP Projector, Computer Vision—ACCV 2010. Lecture Notes in Computer Science, vol 6492. doi:10.1007/978-3-642-19315-6_25, which is incorporated herein by reference.


In some embodiments, act 136 comprises generating (e.g. interpolating) spectral reflectance values at spatial locations labelled as plant at act 134, without necessarily doing so at other spatial locations.


At act 142, the computing system generating a determination for the (at least the portion of a) plant based on the hyperspectral image. For example, the computing system may predict a measure of plant health based on the hyperspectral image, e.g. as described in greater detail with reference to FIG. 2.


A Method for Generating Determinations from Spectral Characteristics of Plants



FIG. 2 is a flowchart illustrating an example method 200 for characterizing plants based on their spectral characteristics. Method 200 is performed by a computing system, as described elsewhere herein (e.g. with reference to FIG. 6). The method involves determining a derivative of the plant's reflectance with respect to wavelength. Plant spectral characteristics in multispectral images tend to be strongly affected by illumination and other factors, whereas plant spectral characteristics tend to be less affected by such factors in the first derivative (and higher-order derivatives). Characterization of plants may comprise, for example, predicting plant health based on such a derivative of plant spectral characteristics.


Method 200 may be performed based on an image representing at least a portion of a plant having sufficient spectral channels to determine a derivative with respect to wavelength (e.g. with respect to channels). Such number of wavelengths/channels may be 3, 4, 5, 6, 7, 8, 9, 10, 20, 50, 100, 200, 500, 1000, any number therebetween, and/or any greater number. The image may thus conventionally be thought to be hyperspectral or multispectral. For convenience, the following disclosure may occasionally refer to such an image processed according to method 200 as a “multispectral image” without the intent of limiting such images to non-hyperspectral embodiments. The image may have any suitable number of spatial dimensions, such as zero (e.g. as may be the case for images produced by a spectrometer), one, or two. The multispectral image may be obtained in any suitable way, e.g. it may be predetermined, acquired from an imaging system (e.g. a multi- and/or hyperspectral imaging system), generated according to the method of FIG. 1, and/or otherwise obtained.


Method 200 involves determining a derivative of reflectance values with respect to wavelength (which may comprise, for example, determining a derivative of reflectance values with respect to the image's spectral channels) to characterize the plant. The computer system may, for example, compare that derivative to a reference, such as a derivative of a reference reflectance generated from a reference multispectral image of a (healthy) reference plant, and may generate a prediction based on that comparison. The reference multispectral image may comprise the same spectral channels as multispectral images processed by method 200.


The reference may be predetermined, generated by the computer system as part of method 200, and/or otherwise obtained. For example, in at least the depicted embodiment, at act 202, the computing system generates a reference reflectance based on a reference multispectral image representing (at least a portion of) a reference plant. The reference plant (and/or a portion thereof) may be a healthy plant, thereby providing a reference for the spectral characteristics of a healthy plant. Generating the reference reflectance may comprise, for example, determining a reflectance value for each of a plurality of wavelengths (and/or channels) of the reference multispectral image. Such determining may comprise, for example, selecting the reflectance value from a predetermined spatial location (e.g. a center of the image, a center of mass of the representation of the plant, etc.), determining a measure of a plurality of reflectance values (e.g. the maximum, minimum, median, average, or other measure of reflectance values), and/or any other suitable determination.


In some embodiments, act 202 comprises determining an average reflectance for a plurality of spatial locations (e.g. pixels) of the reference multispectral image representing the plant. For instance, the computing system may optionally segment the multispectral image to classify regions of the multispectral image as representing plant or non-plant (e.g. background) objects. (Such segmentation may be predetermined, e.g. at act 134 of method 100.) Any suitable segmenting method may be used; for example, the computing system may classify plants as foreground and non-plants as background based on Otsu thresholding. The computing system may determine an average reflectance of all spatial locations (e.g. pixels) classified as “plant”. In some embodiments, the computing system applies morphological adjustments to reduce the likelihood of including non-plant objects in plant-labelled regions, such as by applying binary closing and/or binary erosion.


In some embodiments where the computing system determines a measure of a plurality of reflectance values, the computing system excludes from its determination (e.g. excludes from the average) one or more spatial locations based on specularity and/or non-illumination. For example, experimentation has shown that prediction of plant health can be unreliable in areas with significant specularity, such as in the case of the highly reflective leaves of cabbage plants, and/or in areas covered by shadow where a plant's natural reflectance may not be visible due to a lack of light. In some embodiments, the computing system excludes spatial locations (e.g. pixels) with an average reflectance across one or more (e.g. all) channels of a multispectral image which is greater than a specularity threshold. In some embodiments, the computing system excludes spatial locations with an individual measure of reflectance (e.g. an average reflectance) across one or more (e.g. all) channels of a multispectral image which is less than a non-illumination threshold. For example, the computing system may determine an average reflectance value for a given channel based only on pixels which have an average reflectance across all channels which do not exceed the specularity threshold and which do exceed the non-illumination threshold. Such an approach may be particularly advantageous in hyperspectral imaging embodiments by making use of relatively dense spectral information available to draw inferences about the spectral characteristics of specific spatial locations, but may also be advantageously applied in at least some multispectral imaging embodiments, in suitable circumstances.


In some embodiments, act 202 comprises generating a reference reflectance based on a plurality of reference multispectral images. For example, the computing system may generate a reflectance for each channel of each image (e.g. as described above) and may average or otherwise combine such reflectance values to generate the reference reflectance. For example, the computing system may average the reflectance values of a plurality of images of healthy plants to generate the reference reflectance.


In some embodiments, act 202 comprises generating a plurality of reference reflectance values. For example, act 202 may generate a first reference reflectance for a plant (or plants, and/or portions thereof) of a first species, and may generate a second reference reflectance for another plant (or plants, and/or portions thereof) of a second species. Alternatively, or in addition, the computing system may generate different reference reflectance values for images of different organs of plants (e.g. for leaves and for stems, optionally for the same species), for different health statuses (e.g. for healthy plants and unhealthy plants), for different types of disease (e.g. for Sclerotinia and for powdery mildew), for different types of damage (e.g. for disease, for breakage, and/or for malnourishment), for different degrees of damage (e.g. for severe disease and for mild disease), and/or for other distinctions between plants and/or portions thereof.


In some embodiments, method 200 involves comparing a derivative of reflectance for a multispectral image with a derivative of reflectance for a reference multispectral image. The derivative of reflectance for the reference multispectral image may be predetermined, generated by the computer system as part of method 200, and/or otherwise obtained. For example, method 200 may comprise act 204, which comprises determining a derivative of the reference reflectance with respect to wavelength. This may include, for example, determining a derivative with respect to the spectral channels of the reference multispectral image and/or with respect to the channels/wavelengths of another multispectral image (e.g. by estimating, interpolating, or otherwise generating spectral values corresponding to the channels/wavelengths of the multichannel image received at act 206). The derivative may be calculated in any suitable way, including by discrete methods (such as those provided by the numpy scientific library), by continuous methods (e.g. by fitting discrete reference reflectance values to a curve and determining the derivative of the curve), and/or by any other suitable method.


At act 206, the computing system receives a multispectral image representing at least a portion of at least one plant. The multispectral image comprises an image representing at least a portion of a plant having sufficient spectral channels to determine a derivative with respect to wavelength (e.g. with respect to channels). Such number of wavelengths/channels may be 3, 4, 5, 6, 7, 8, 9, 10, 20, 50, 100, 200, 500, 1000, any number therebetween, and/or any greater number. Act 206 may comprise receiving the multispectral image from a user, accessing a predetermined multispectral image, generating the multispectral image by the computing system (e.g. by generating a hyperspectral image according to method 100), and/or otherwise obtaining the multispectral image.


At act 208, the computing system determines a derivative of reflectance with respect to wavelength for the multispectral image received at act 206. This may include, for example, obtaining reflectance values for the multispectral image (and/or of a plant and/or of a portion thereof represented by the image) substantially as described with reference to act 202 and determining a derivative of the reflectance values with respect to the spectral channels of the multispectral image substantially as described with reference to act 204. In some embodiments, the computing system also or alternatively determines the derivative with respect to channels/wavelengths corresponding to another multispectral image, for instance by estimating, interpolating, and/or otherwise generating spectral values corresponding to the channels/wavelengths of the reference multichannel image. The derivative may be calculated in any suitable way, including by discrete methods (such as those provided by the numpy scientific library), by continuous methods (e.g. by fitting discrete reference reflectance values to a curve and determining the derivative of the curve), and/or by any other suitable method.


In some embodiments, at act 210, the computing system compares the reflectance values for the plant of the multispectral image with the reference reflectance values. In some embodiments, such comparison comprises comparing the derivative of reflectance values of act 208 (called the “target derivative” or ds/dλ herein) with the derivative of reference reflectance values of act 204 (called the “reference derivative” or dr/dλ herein). Such comparisons may comprise, for example, determining a difference between the target derivative and the reference derivative. In at least one example embodiment, act 210 comprises determining a regression loss metric based on the target derivative and the reference derivative. For instance, act 210 may comprise determining a mean square error of the target derivative relative to the reference derivative over n channels λi, which may be expressed as:







MSE

[



ds



d

λ


,

dr

d

λ



]

=


1
n






i
=
1

n



(




ds



d

λ




(

λ
i

)


-


dr

d

λ




(

λ
i

)



)

2







Alternatively, or in addition, act 210 may comprise determining a mean absolute error between the target derivative and the reference derivative, a Huber loss between the target derivative and the reference derivative, a log-cosh loss between the target derivative and the reference derivative, a quantile loss between the target derivative and the reference derivative, and/or any other suitable regression loss metric between the target derivative and the reference derivative.


In some embodiments, act 210 comprises a plurality of comparisons. For example, in an embodiment where the computing system has a plurality of reference reflectance values (e.g. having generated such reference reflectance values at act 204 and/or otherwise obtained such reference reflectance values), the computing system may perform comparisons as described above between the target derivative and a derivative for each of the reference reflectance values with respect to wavelength.


At act 212, the computing system generates a determination for the (at least a portion of a) plant represented in the multispectral image of act 206 based on the target derivative. In at least some embodiments, the computing system generates the determination based on the comparison of act 212. In some embodiments, the determination comprises a prediction of plant health. For example, the computing system may determine that a regression loss metric value exceeds a threshold and, based on such determination, may predict that the plant is not healthy. Alternatively, or in addition, the computing system may bin regression loss metric values into various categorical bins (e.g. “healthy”, “partially healthy”, “unhealthy”). Alternatively, or in addition, the computing system may provide a predicted healthiness score based on the regression loss metric (e.g. as a heatmap of regression metric values). For instance, the computing system may provide the regression loss metric as a healthiness score for regression loss metric values below the threshold and may group regression loss metric values above the threshold as 100% unhealthy.


The computing system may generate predictions for specific spatial locations (e.g. pixels), for regions of a multispectral image (and thus regions of the represented plant), and/or for an entire multispectral image and/or plant. For example, where a multispectral image comprises representations of multiple plants, the computing system may segment the plants (or otherwise identify each plant) and generate a plurality of predictions for each spatial location (e.g. pixels) representing portions of that plant.


The computing system may optionally generate a combined prediction for the plant by combining the plurality of predictions for each spatial location representing portions of that plant. For instance, the computing system may generate an average regression loss metric, and/or may score each spatial location for the given plant characteristic (e.g. plant health) as described herein and provide an area-based measure for that characteristic. For example, for a leaf which has one healthy half and one diseased half, the computing system may generate a prediction for each spatial location representing that leaf (e.g. as heatmap), and/or may generate a prediction comprising an average regression loss metric for the leaf (and may, e.g., generate a prediction based on such average regression loss metric as described herein), and/or may generate a prediction comprising measure of how much of the leaf is “healthy” and/or “diseased” (e.g. based on one or more thresholds for healthy and/or diseased predictions) as a proportion of the visible area.


In some embodiments where the computing system has received a plurality of reference reflectance values (e.g. varying by species, disease, or other factors, as described elsewhere herein), the computing system generates a first determination for the (at least a portion of a) plant based on the target derivative and a first reference derivative, and generates a second determination for the (at least a portion of a) plant based on the target derivative and a second reference derivative. (It will be understood that more than two determinations may be generated.) The computing system may select one of the predictions based on which of the predictions corresponds to the highest confidence, smallest regression loss, and/or the greatest likelihood of health. As one example, if the first prediction corresponds to a “healthy” prediction and the second prediction corresponds to an “unhealthy” prediction, the computing system may select the first prediction. As another example, if the first prediction corresponds to a healthy reference plant and the second prediction corresponds to an unhealthy reference plant (e.g. one displaying significant disease), a low regression loss metric value for the second prediction may indicate a high confidence that the plant is unhealthy and the computing system may select the second prediction.


In some embodiments, the computing system receives an indication associating a set of one or more plants with one of a plurality of references. For example, the computing system may receive a multispectral image at 206 representing a corn field and may receive an indication that the multispectral image comprises representations of corn plants. Such indication may be predetermined, provided by a user, generated by the computing system (e.g. according to a classification machine learning model executed by the computing system and trained over images of various species of plants), and/or otherwise suitably obtained. The computing system may select a reference derivative based on the indication (e.g., in the foregoing example, the computing system may select a reference derivative based on a reference multispectral image of healthy corn) and may generate a prediction based on target derivative and the selected reference derivative as described elsewhere herein without necessarily performing a comparison between the target derivative and one or more unselected reference derivatives.


In some embodiments, the computing system alternatively or additionally generates a prediction of plant species, type of disease, type of damage, and/or degree of damage based on a comparison between the target derivative and a reference derivative, and in particular based on the plant species, type of disease, type of damage, and/or degree of damage (as appropriate) of the reference plant corresponding to the most-similar (e.g. lowest-loss) reference derivative. For instance, if the first prediction corresponds to a reference plant comprising wheat and the second prediction corresponds to a reference plant comprising corn, a low regression loss metric value (e.g. lower than a threshold, and/or lower than a regression loss metric value for the first prediction) may indicate a high confidence that the target plant (i.e. the plant corresponding to the target derivative) is corn, and the computing system may generate such a prediction on that basis.


A System for Hyper spectral Imaging of Plants



FIGS. 3, 4, and 5 are schematic diagrams of an example apparatus 300 for imaging plants. Such images may be multispectral and/or hyperspectral and may optionally be used by methods 100 and/or 200. FIG. 3 provides a perspective view for context, FIG. 4 provides a side elevation view, and FIG. 5 provides an enlarged detail view shown in a perspective orientation generally from below to show certain elements of a camera and associated elements. FIGS. 3, 4, and 5 are discussed together.


Apparatus 300 comprising an illumination source 312 and an imaging device 320. Illumination source 312 illuminates a region 314 which substantially aligns with a field of view of imaging device 320. In the example embodiment of FIGS. 3-5, apparatus 300 comprises a frame 302 for supporting illumination source 312 and imaging device 320. Apparatus 300 may further comprise a conveyor 304 (e.g. supported by frame 302) for conveying plants and/or other imaging subjects through region 314. Plants may, for example, be supported by a planter 310 conveyed by conveyor 304.


In some embodiments, apparatus 302 comprises a shroud (not shown) for blocking and/or otherwise reducing external illumination in region 314. For example, the shroud may comprise walls and/or other substantially opaque surfaces around (and optionally supported by) frame 302. In some embodiments, apparatus 300 comprises a movable barrier along conveyor 304 (and optionally two movable barriers on opposing sides of region 314) which are openable to admit a plant and/or planter 310 to region 314 for imaging and/or to allow the plant and/or planter 310 to exit region 314 subsequent to imaging. Such movable barriers may be opaque and may be controlled by apparatus 300 (e.g. via a controller, not shown) to shut, thereby blocking and/or otherwise reducing external illumination in region 314 during imaging by imaging device 320.


An exemplary embodiment of imaging device 320 is shown in greater detail in FIG. 5. Imaging device 320 comprises an imaging sensor 322. For example, imaging device 320 may comprise an RGB camera with near-IR sensitivity. For instance, in an exemplary embodiment, imaging device 320 comprises a Sony™ IMX219 module (comprising an imaging sensor 320) for a Raspberry Pi™ controller (not shown). Other elements of apparatus 300, such as illumination source 312, conveyor 304, movable barriers, may optionally be controlled by such controller and/or by one or more other controllers (e.g. such as by a computing system, as described with reference to FIG. 6).


Imaging device 320 may further comprise one or more optical filters 326. In some embodiments, imaging device 320 comprises a plurality of optical filters 326. For example, imaging device 320 may comprise a filter wheel 324 comprising a plurality of optical filters 326. Imaging device 320 may drive filter wheel 324 (e.g. via a rotor 328) to revolve optical filters 326 through a field of view of imaging sensor 322 to allow imaging sensor 322 to capture images through such optical filters 326. Rotor 328 may drive filter wheel 324 continuously and/or rotor 328 may drive filter wheel 324 intermittently, e.g. by pausing during an exposure of imaging sensor 322 through a given optical filter 326.


In at least the depicted exemplary embodiment, filter wheel 324 comprises eight optical filters: seven bandpass filters with center wavelengths of 450 nm, 500 nm, 550 nm, 600 nm, 650 nm, 700 nm, and 750 nm, respectively, with each a FWHM bandwidth of 25 nm, and one shortpass filter blocking wavelengths longer than approx. 700 nm. Such a shortpass filter may, in suitable embodiments (e.g. those comprising a near-IR sensitive RGB and/or monochrome camera) facilitating the capture of conventional, visible-spectrum-only images by imaging sensor 322. Such images may be used for segmentation, display, and/or any other purpose.


Example System Implementation


FIG. 6 illustrates a first exemplary operating environment 600 that includes at least one computing system 602 for performing methods described herein. System 602 may be any suitable type of electronic device, such as, without limitation, a mobile device, a personal digital assistant, a mobile computing device, a smart phone, a cellular telephone, a handheld computer, a server, a server array or server farm, a web server, a network server, a blade server, an Internet server, a work station, a mini-computer, a mainframe computer, a supercomputer, a network appliance, a web appliance, a distributed computing system, multiprocessor systems, or combination thereof. System 602 may be configured in a network environment, a distributed environment, a multi-processor environment, and/or a stand-alone computing device having access to remote or local storage devices.


A computing system 602 may include one or more processors 604, a communication interface 606, one or more storage devices 608, one or more input and output devices 612, and a memory 610. A processor 604 may be any commercially available or customized processor and may include dual microprocessors and multi-processor architectures. The communication interface 606 facilitates wired or wireless communications between the computing system 602 and other devices. A storage device 608 may be a computer-readable medium that does not contain propagating signals, such as modulated data signals transmitted through a carrier wave. Examples of a storage device 608 include without limitation RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disks (DVD), or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage. In at least some embodiments such embodiments of storage device 608 do not contain propagating signals, such as modulated data signals transmitted through a carrier wave. There may be multiple storage devices 608 in the computing system 602. The input/output devices 612 may include a keyboard, mouse, pen, voice input device, touch input device, display, speakers, printers, etc., and any combination thereof.


The memory 610 may be any non-transitory computer-readable storage media that may store executable procedures, applications, and data. The computer-readable storage media does not pertain to propagated signals, such as modulated data signals transmitted through a carrier wave. It may be any type of non-transitory memory device (e.g., random access memory, read-only memory, etc.), magnetic storage, volatile storage, non-volatile storage, optical storage, DVD, CD, floppy disk drive, etc. that does not pertain to propagated signals, such as modulated data signals transmitted through a carrier wave. The memory 610 may also include one or more external storage devices or remotely located storage devices that do not pertain to propagated signals, such as modulated data signals transmitted through a carrier wave.


The memory 610 may contain instructions, components, and data. A component is a software program that performs a specific function and is otherwise known as a module, program, engine, and/or application. The memory 610 may include an operating system 614, a multispectral engine 616, an interpolation engine 618, a prediction engine 620, spectral bases 622, calibration settings 624, one or more images 626 (e.g. multispectral images and/or hyperspectral images, which may comprise reference images), and other applications and data 630. Depending on the embodiment, some such elements may be wholly or partially omitted. For example, an embodiment intended for prediction based on received multispectral and/or hyperspectral images may exclude interpolation engine 618. As another example, memory 610 may include no images 626 prior to performing a method described herein and may receive such images via an input device 612 and/or from a storage device 608 and/or generate such images as described elsewhere herein.


While a number of exemplary aspects and embodiments have been discussed above, those of skill in the art will recognize certain modifications, permutations, additions and sub-combinations thereof. It is therefore intended that the following appended claims and claims hereafter introduced are interpreted to include all such modifications, permutations, additions and sub-combinations as are within their true spirit and scope.

Claims
  • 1. A method for hyperspectral imaging of plants, the method performed by a processor and comprising: receiving a multispectral image comprising a number m of multispectral channels, at least one multispectral channel comprising an infrared wavelength, the multispectral image representing at least a portion of at least one plant;generating a hyperspectral image comprising a number n of hyperspectral channels based on the multispectral image and a plurality of spectral bases, the number n of hyperspectral channels greater than the number m of multispectral channels; andgenerating a determination for the at least the portion of at least one plant based on the hyperspectral image.
  • 2. The method according to claim 1 wherein the determination comprises a prediction of plant health and generating the determination comprises generating the prediction based on a plurality of reflectance values of the hyperspectral image.
  • 3. The method according to claim 2 wherein determining the prediction of plant health comprises determining a derivative of the plurality of reflectance values with respect to wavelength.
  • 4. The method according to claim 3 wherein determining the prediction of plant health comprises determining the prediction of plant health based on the derivative of the plurality of reflectance values and a plurality of reference reflectance values.
  • 5. The method according to claim 4 further comprising generating the plurality of reference reflectance values based on a reference hyperspectral image representing at least a healthy portion of a reference plant.
  • 6. The method according to claim 5 wherein generating the plurality of reference reflectance values comprises determining an average of reflectance values for a plurality of spatial locations of the at least the healthy portion of the reference plant for each of a plurality of the n hyperspectral channels.
  • 7. The method according to claim 4 wherein determining the prediction of plant health comprises determining a difference between the derivative of the plurality of reflectance values and a derivative of the plurality of reference reflectance values with respect to wavelength.
  • 8. The method according to claim 7 wherein determining the difference comprises determining a regression loss metric based on the derivative of the plurality of reflectance values and a derivative of the plurality of reference reflectance values, wherein the regression loss optionally comprises at least one of: a mean square error, a mean absolute error, a Huber loss, a log-cosh loss, and a quantile loss.
  • 9. (canceled)
  • 10. The method according to claim 4 wherein the plurality of reference reflectance values comprise a first plurality of reference reflectance values corresponding to at least a first portion of at least a first plant and a second plurality of reference reflectance values corresponding to at least a second portion of at least a second plant, the first and second portions differing in at least one of: species of plant, organ of plant, type of disease, type of damage, and degree of damage.
  • 11. The method according to claim 10 wherein determining the prediction of plant health comprises: determining a first prediction of plant health based on the derivative of the plurality of reflectance values and the first plurality of reference reflectance values;determining a second prediction of plant health based on the derivative of the plurality of reflectance values and the second plurality of reference reflectance values; andselecting the first prediction based on the first prediction corresponding to a greater likelihood of health than the second prediction.
  • 12. The method according to claim 1 wherein the spectral bases having been generated from one or more images comprising at least one image representing at least a further portion of at least one further plant, wherein optionally the plurality of spectral bases comprises at least four spectral bases.
  • 13. (canceled)
  • 14. The method according to claim 1 wherein generating the hyperspectral image comprises interpolating at least one hyperspectral reflectance value for a wavelength of at least one of the n hyperspectral channels outside of the m multispectral channels.
  • 15. The method according to claim 1 comprising segmenting the multispectral image into plant and non-plant regions; wherein generating the hyperspectral image comprises generating the hyperspectral image for the plant regions.
  • 16. The method according to claim 1 comprising: receiving a calibration multispectral image representing at least a portion of a calibration subject, the at least the portion of the calibration subject substantially non-reflective in one or more multispectral channels of the m multispectral channels; anddetermining, for at least one of the one or more multispectral channels, a corresponding calibration reflectance of at least a portion of the multispectral image representing at least the portion of the calibration subject;wherein generating the hyperspectral image comprises, for the at least one of the one or more multispectral channels, subtracting the corresponding calibration reflectance.
  • 17. The method according to claim 1 wherein at least one of the m multispectral channels comprises at least one wavelength in a range of about 525 nm to about 575 nm, at least one wavelength in a range of about 600 nm to about 700 nm, and/or at least one wavelength in a range of about 400 nm to about 500 nm.
  • 18. (canceled)
  • 19. (canceled)
  • 20. The method according to claim 17 wherein the m multispectral channels comprise at least four multispectral channels, and/or wherein the m multispectral channels comprise no more than ten multispectral channels.
  • 21. (canceled)
  • 22. The method according to claim 17 wherein receiving the multispectral image comprises causing an imaging sensor having infrared sensitivity to capture one or more frames through one or more optical filters.
  • 23. The method according to claim 22 wherein: the imaging sensor comprises at least one of: an RGB imaging sensor with NIR sensitivity and a monochrome imaging sensor;the one or more optical filters comprise a plurality of optical filters; andcausing the imaging sensor to capture one or more frames comprises causing the imaging sensor to capture a plurality of frames by capturing at least one frame through each of the plurality of optical filters.
  • 24. The method according to claim 23 wherein causing the imaging sensor to capture the plurality of frames by capturing at least one frame through each of the plurality of optical filters comprises causing the plurality of optical filters to revolve through a field of view of the imaging sensor while causing the imaging sensor to capture frames.
  • 25. A computer system comprising: one or more processors; anda memory storing instructions which cause the one or more processors to perform operations comprising: performing the acts of the method according to claim 1.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application claims priority to, and the benefit of, U.S. provisional patent application No. 63/127,813 filed 18 Dec. 2020 entitled Systems and Methods for Hyperspectral Imaging of Plants, the entirety of which is incorporated by reference herein for all purposes.

PCT Information
Filing Document Filing Date Country Kind
PCT/CA2021/051829 12/17/2021 WO
Provisional Applications (1)
Number Date Country
63127813 Dec 2020 US