High Intensity Waterproof LED Spotlights | Car & Home - led red spotlight
LED emergency lights for vehicles
Our prototype was designed to house 8 different types of LEDs, each having a unique bandwidth of our choice. The optical characteristics of each LED model is listed in Table.1, the selection of bandwidths will be further discussed in next section. We mount 4 LEDs per color, and as is shown in Fig.1(b), the LEDs with same color are mounted in a rectangular pattern in order to illuminate the subject as evenly as possible. Such arrangement minimizes the difference of light distribution across different colors.
Hyper-spectral imaging (HSI) have had a significant impact in agriculture and remote sensing. For example, classification allows for separating diseased from healthy plants using additional color band information [13, 14, 15, 16, 17]. In addition, detecting the maturity of crops and finding stressed crops are important HSI tasks [18, 19, 20, 21, 22, 23].
Although the model showed certain capabilities of providing extended spectral information based on its 8-band input, it is worth mentioning that we do not expect the same model to work as well for unknown plants or unknown variants. The reason we are able to use only 8 bands to reconstruct the entire spectrum is the network learns the types and states of a pixel in 8 different bands and internally maps it back to the same known state of a high resolution spectrum distribution. This process is purely based on the previous "knowledge" stored inside the model. It should not be able to be transferred to unknown plant species nor different soil types.
Another challenge is the bright LED spots found in our camera’s results. These light spots (reflections of the LED lights) are supposed to be ignored and filled up by the model. In order to help our U-Net model learn to do this, we implemented Otsu’s thresholding as an optional add-on to the data augmentation pipeline. The Otsu method searches for the threshold that minimizes the intra-class variation for both sides of the threshold. This method applies well on our case, which has a cluster of very bright pixels (the light spots) and another cluster of dim pixels (the soil and root pixels). As is shown in Fig.9, this method creates an accurate mask separating light spot pixels from useful data. Based on the mask, our framework then performs an interpolation based on other bandwidths of the same pixel which are not affected by the light spot and uses the interpolated value in place of the over-exposed value.
We propose a type of camera that uses LEDs of different bands instead of color filters to capture spectral information. In our setup, the camera sensor does not need to be equipped with any type of optical filter, not even the Bayer filter for an RGB sensor or IR filter commonly seen on the back of lenses. This setup works perfectly in the scenarios where the on-camera LEDs are the only source of light. Moreover, it can also deal with static scenes with a moderate amount of ambient light by treating the ambient light as "dark field" and only tracking the incremental light upon firing each LED.
We performed analysis on data gathered by the reference HSI camera, using the same rhizobox samples that will be used by our camera (analysis was performed by T. M. Sazzad). Out of all 8 bandwidths, three of our selections matched exactly with the optimal bands (blue, green and red). The other bandwidths are limited by stock availability when producing the prototype, and thus did not match exactly with calculated optimal values.
This is a custom loss function that measures how different a pixel of a band is against its neighboring bands. The total error is the mean of all bands’ max absolute difference against their neighboring bands. This loss function rewards smoothness of the band plot for each pixel.
Construction light barkit
For the next step, we will capture data from real minirhizotrons and validate our model’s ability to transfer its knowledge to those data. This is expected to be more challenging because the plants might be of a different species, and the type of soil might be drastically different than the rhizoboxes from which we obtained data for training.
This is a custom loss function that measures how different a pixel is against its neighboring pixels (in both width and height direction). The total error is the mean of all pixels’ max absolute difference against their neighboring pixels. This loss function smooths out the noise introduced by the model.
Constructiontruck strobe lights kits
As is shown in the figure, the reconstruction model was able to bring out the details of the spectrum of a root pixel, especially the curved region near 700700700700nm. The model also did well in reconstructing average soil pixels, which are expected to look like a nice Gaussian distribution because it is composed by a mixture of many different ingredients.
The challenge lies in the geometry of the minirhizotron. The diameter of a typical minirhizotron is between 5 to 10 centimeters. And the one used by our collaborating research group is only 5cm in diameter. Even for our camera, fitting itself into such a narrow tube is a significant engineering challenge. Most of the existing hyper-spectral cameras in the market feature a much larger dimension, making them impossible to be deployed in such environments.
With the help of the data processing and training techniques shown in in Sec.6.3, our model performed well in handling light spots. Although the output is relatively blurry in spatial dimensions, a significant amount of plant root features were conserved in the spectral dimension and showed a high relevance towards the ground truth.
Hyper-spectral imaging has gained increasing interest in recent years. The ability to capture rich spectral information beyond traditional RGB channels has been proven to help reveal more information from the subject of interest, including agricultural and plant researches [1, 2, 3, 4, 5]. Compared to RGB cameras, hyper-spectral imaging provides both richer information inside the visible bandwidth (380 - 700nm) and additional information outside of the visible bandwidths which would be unavailable otherwise. However, hyper-spectral cameras also tend to cost significantly more than traditional cameras and to have a large physical size. Its shortcomings limited the chance of integrating hyper-spectral imaging into a larger variety of industrial applications or research projects.
LEDLight Barfor truck roof
In this work, we demostrated the feasibility of a cost-efficient approach to build a compact sized hyper-spectral camera. We performed experiments on spectrum reconstruction from a reduced spectral image back into a full-sized hyper-spectral image. The reconstruction results showed higher relevance for plant root pixels, but did not perform as well on soil pixels. Such a result indicates that additional, useful data can indeed be obtained by our active illumination camera setup.
We collected a dataset to train the U-Net based model for spectrum reconstruction. Since the ground truth was required to train the model, and our reference data needed to be captured by a reference HSI device which does not fit into a minirhizotron, the initial version of camera we developed (shown in Fig.1(a)) was especially designed to match the results from the reference camera. Both cameras would take hyper-spectral images of plant roots in rhizoboxes. The rhizobox is a container that works similarly to a minirhizotron, the difference being the rhizobox features a flat transparent surface so it’s easier to take photos with all types of cameras. In addition, the compact size of a rhizobox makes it easy to be moved around.
This per-pixel error quantization treats pixels as high dimensional vectors (299 dimensions in our case), and calculates per-pixel angular difference for each spatial location (x,y)𝑥𝑦(x,y)( italic_x , italic_y ) between ground truth p→1subscript→𝑝1\vec{p}_{1}over→ start_ARG italic_p end_ARG start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT and model prediction p→2subscript→𝑝2\vec{p}_{2}over→ start_ARG italic_p end_ARG start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT. In this way, the error compares the shape of spectral dimension and ignores difference in pixel intensity.
As shown in Sec.6.4, our model did not perform as great then handling soil pixels. This is partially caused by the difficulty to accurately match up high spatial frequency details of fine-grained soil pixels (see Fig.6.D). In addition to the error introduced by template matching algorithms, we also noticed that some soil grains did not stay in the same location before and after we retrieved the rhizobox from the cabinet for the reference HSI camera. Although done carefully, the transportation inevitably caused some soil grains to move around, causing a mismatch between ground truth and our camera’s data. This might also be a significant contributing factor for the blurry output of our model.
In order to fully leverage the resolution of our camera sensor, we reformed the model into an asymmetric structure. The left side of the network (encoder) descends faster on spatial resolution, while the right side (decoder) descends slower on spectral dimension. In this setup, the dimensions of the corresponding layers on each side no longer match. A resizer had to be implemented within each feed-forward shortcut connection.
In the first stage, the model was fed with ground truth as input (pre-train). The ground truth (299 bands) data-points were first projected into the 8 bands corresponding to our camera’s results, and then reshaped to mimic the dimensions of actual input data. This stage is designed to teach the network to learn to extend low dimension data into high dimensions without any interference. Random affine augmentation was enabled in this stage.
Construction light barled
Due to the distortion and uneven intensity distribution introduced by hardware, the raw data (Fig.4(a)) captured from our camera has to be pre-processed before proceeding. In our calibration pipeline, the raw image is firstly mapped by a "reference white" calibration image, eliminating uneven intensity distribution introduced by light sources and lens projection. The reference white was captured and applied separately for each color. Geometric correction was then applied based on a chess board image, which is captured by the same camera. The brightness distribution calibrated in the previous step will be retained during this process (calibration pipeline shown in Fig.3, sample calibrated image shown in Fig.4(b)).
A model based on U-Net was implemented and trained using the data we collected. The model was intended for both removing the bright LED light spots introduced by our custom light module and expanding the number of bands from 8 to 299.
Light Barfor truck roof
Supervised and Semi-supervised labelling: Supervised approaches require labelled samples for the selection of most favourable bands during training and learning [67] where similarity measure metric is used among the class labels. These techniques require a number of assessment conditions which can be categorised as: information divergence [68], maximum ellipsoid volume [69], Euclidean distance [70] etc. For semi-supervised approaches graph based models are used for labelled and unlabelled data samples for the selection of appropriate spectral bands [71, 72] but suffers from providing contextual information.
U-Net is a convolutional neural network architecture that is commonly used for segmentation tasks in computer vision [74]. However, instead of performing typical segmentation tasks, we explored the feasibility of using such architecture to perform spectral reconstruction tasks. This is based on the assumption that the U-Net could learn the inherent spectrum structure of different soil and root components based on the training set, and then transfer its prior knowledge to new data while maintaining contextual awareness.
The model parameters from the first stage of training were then loaded back for the second (main) stage of training. In this stage, 3 out of 5 samples (each batch) were projected from the ground truth, just like what we did in the first stage. And the rest 2 samples are loaded from raw samples from our camera. Affine augmentation was also enabled in this process. This stage was added in order to make the network aware of LED light spots. The spots can appear in any shape, size and form factor with the help of affine transformation, they could also be completely absent since 60%percent6060\%60 % of the inputs used in this stage were converted directly from ground truth. This stage helps the network to gradually learn to deal with LED spots in our input dataset.
There exist attempts to integrate multispectral camera into minirhizotron[73]. This work implements an automatically operated multi-spectral camera that can actuate by itself inside a minirhizotron. Their camera was also equipped with multi-bandwidth light sources. However, instead of exploring the information contained in data acquired from such device, this work primarily focused on autonomous operation and remote deployment.
Our design addressed these challenges by both utilizing small monochrome sensor modules and a custom designed printed circuit board (PCB) to integrate the LEDs. The camera module and the LED module were stacked upon each other with only 5mm distance. The LED module has an opening in its center to let through a wide angle lens that sits on the camera module’s PCB. Each module has an upstream USB port to be connected with the host computer (a raspberry pi), and the USB port handles both power and communication. The first camera was designed to work with the flat surface of a rhizobox, but the same core components can be easily adapted into a 5cm diameter minirhizotron with a different case (Fig.2).
It is worth mentioning that, weights for each component above were manually configured in each stage of training. In pre-train phase, all above loss functions are used to derive the total loss. The delta-pixel and delta-bands error each has a weight of 4, the MSE and the MAE each has a weight of 1. In main training phase, we only use smooth L1 loss.
Given the potential benefits of enabling hyper-spectral imaging inside narrow spaces such as a minirhizotron, we investigated a new approach of obtaining spectral information. Instead of performing a full-spectral scan, our approach focuses on a few discrete bands that were selected to carry the most potential information. Our device utilizes readily available parts from the consumer market, and our design features an ultra compact size so it can fit into applications with very narrow operating space for imaging. In addition to compact size, our device also has the advantages of low cost and ease of use. Such advantages makes it possible to be massively deployed and autonomously operated.
We used a combination of several different loss function to train the model. The effective loss function is the weighted sum of the following functions:
In addition to the above qualitative analysis, thanks to the segmentation model[5] provided by Spencer et al., the prediction error was categorized into plant root pixels and soil pixels respectively. As is shown in Fig.11. derived prediction error from 4 prediction images are each divided into root pixels and soil pixels. The figure indicates that our model performed slightly better when handling plant root pixels.
Since all parts of our camera are standard market-ready parts, we can achieve an extremely low unit cost compared to hyper-spectral cameras currently in the market. In addition, the constraints of the environment in which the camera will be operated is another crucial aspect we should take care of. In our case, we are targeting the minirhizotrons broadly used to study plants. Minirhizotron is a specialized tool used in biology and agriculture researches to study roots in their natural environment. It is a transparent tube inserted into the soil, making it possible to observe roots as they grow over time.
In the following analysis and "Prediction Error" graphs shown in Fig.10, we used the spectral angular difference to quantize prediction. It is defined as:
The mixture of ground truth inputs and raw inputs also ensured that the model "respects" the details of its input. In our previous trials, the model over-processed the input and removed the details entirely due to pixel shifts caused by alignment errors. This small misalignment confused the network and issued a punishment when the model generated sharp and detailed predictions. This misalignment can be found by comparing point B in sample (3) of Fig.10. In ground truth, point B sits on the left side of the root, while in the raw input, point B sits on the right side. This mismatch can also be spotted from the prediction error, the blue curve and red curve concentrate on different sides of the sample point.
Hyper-spectral imaging has recently gained increasing attention for use in different applications, including agricultural investigation, ground tracking, remote sensing and many other. However, the high cost, large physical size and complicated operation process stop hyperspectral cameras from being employed for various applications and research fields. In this paper, we introduce a cost-efficient, compact and easy to use active illumination camera that may benefit many applications. We developed a fully functional prototype of such camera. With the hope of helping with agricultural research, we tested our camera for plant root imaging. In addition, a U-Net model for spectral reconstruction was trained by using a reference hyperspectral camera’s data as ground truth and our camera’s data as input. We demonstrated our camera’s ability to obtain additional information over a typical RGB camera. In addition, the ability to reconstruct hyperspectral data from multi-spectral input makes our device compatible to models and algorithms developed for hyperspectral applications with no modifications required.
EmergencyLight Bar
Dimensionality reduction of HS samples: A range of band-selection approaches have been proposed in the existing literature. Typically, to decrease the dimensionality of a hyper-spectral image, these band selection approaches can be categorised into two specific groups that include feature extraction [59, 60] and feature selection (band selection) [61, 62, 63, 64, 46, 65]. Interestingly, the above mentioned two specific methods extract or select data from all HSI bands to correspond to the whole spectral cube, and the outcomes are almost or roughly equivalent to the full HSI bands. Currently, traditional feature selection approaches are considered to be the most commonly applied techniques which include Principal Component Analysis (PCA) [53], Maximum Noise Fraction (MNF) [66], Genetic Algorithm, and FICA (Fast Independent Component Analysis) [59]. For feature extraction approaches, high dimensional spatial data are mapped into low-dimensional space considering specific criteria, thus extracting a complete new sub-set of features which represents the original HS data. Unfortunately, during spatial transformation, the physical denotation of the original HS data cannot be found same as well as it is also possible that some of the key or main information may be lost forever. On the other hand, for the band selection approaches a distinct and representative sub-set is selected from the unique hyper-spectral data which preserves the physical denotation and information without loss. Additionally, it also preserves the intrinsic characteristics of the HS data. In this article, we have focused on feature (band) selection rather than feature (band) extraction.
The ever-increasing accessibility of HSI has fueled widespread impact[24, 25, 26, 27, 28, 29], with studies on hundreds of spectral bands [30, 31, 32, 33]. These bands provide rich information that can be analyzed and identified to differentiate objects in the scene [34, 35]. HSI have impacts on fields as diverse as agriculture (plant disease detection and classification, etc), tissue analysis in medical science, forest land detection, mining, and mineral study, vegetation estimation, protection of the environment, biological analysis, etc [36, 37, 38, 39, 40, 41].
However, the large data footprint of the spectral bands has disadvantages, such as high computational time complexity, transmission, storage and analysis [32, 42, 43, 44, 45]. Therefore, redundant information and time complexity minimization is important to reconstruct hyper-spectral data efficiently [43, 46, 47, 48]. Techniques for compressing the measurements include clustering [49, 50], Ranking approaches [51, 52, 53, 54], greedy approaches [55, 56] and evolutionary approaches [57, 58].
Our processed dataset has 95 pairs of working samples, each of them consists of a reference matrix of shape 286×286×299286286299286\times 286\times 299286 × 286 × 299 pixels (the ground truth) and an input matrix of shape 1024×1024×81024102481024\times 1024\times 81024 × 1024 × 8 pixels. These samples (95 total) were divided into 85 samples as training set and 10 samples as test set. The list of samples used as test set was randomly generated upon initialization of a training. No manual selection was involved in the process of splitting training set and test set. However, it is worth noticing that there might be a minor overlap between nearby samples. The percentage of overlap is generally less than 10%, and our splitting algorithm does not take care of the overlap.
Due to the high volume of data expected in both spatial and spectral dimensions, it is not straightforward to demonstrate the overall performance in generalized metrics. Instead, we randomly picked a set of test samples and the spectrum of select pixels in these samples to demonstrate the training result (shown in Fig.10).
Bestconstruction light bar
As an example, minirhizotron is a commonly used tool for research of plant roots [6, 7, 8, 9]. It is a tube shaped inspection tool made of transparent materials. Its inner diameter varies from as small as 5555cm up to 20−30203020-3020 - 30cm. It reaches deep into soil near the plant to be inspected, enabling researchers to visually inspect plant roots in a nondestructive way. A lot of studies have been carried out with the help of minirhizotron and compact monochrome or RGB cameras [10, 11, 12]. Although these studies could potentially be significantly benefited from hyper-spectral imaging, up to our best knowledge, there does not exist a feasible way to provide hyperspectral imaging inside such narrow space.
Random affine augmentation was employed in the training process. The affine transformation is a combination of five different transformations. Each transformation is controlled by a separate set of parameters. The list of applicable transformations and their probability distributions are listed in Table.2, all parameters follow normal distribution within the given range. This data augmentation technique is used to compensate for the limited amount of samples we are able to collect. It also helps to prevent the model from over fitting to scale and/or distortion. If the transformed image cannot fill up the entire canvas, we have two different methods to fill it up. The simpler approach is to use a uniform average color for filling, another is mirroring the original image so the content looks consistent and meaningful. In practice, the mirroring option would consume too much resources (8 times more memory consumption) and would consequently limit the batch size to 1. Therefore, we adopted uniform solid color fill throughout all stages of training. In addition, the mask of valid pixels was preserved for each affine transform and re-applied for the loss function.
The final step of data processing is matching our camera’s result with the reference camera’s result. The sizes of (1) the rhizobox, (2) the reference camera’s imaging area, and (3) our camera’s imaging area are shown in Fig.5. We first manually downscale the size of our camera’s output (1024×1024102410241024\times 10241024 × 1024 after calibration) to match the pixel density of the reference camera (286×286286286286\times 286286 × 286). Then we are able to run template matching on the reference image (using the down-scaled image as template). Fig.6 shows a sample of a successful match.
Data availability Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.