Abstract
In contemporary biomedical research, the accurate automatic detection of cells within intricate microscopic imagery stands as a cornerstone for scientific advancement. Leveraging state-of-the-art deep learning techniques, this study introduces a novel amalgamation of Fuzzy Automatic Contrast Enhancement (FACE) and the You Only Look Once (YOLO) framework to address this critical challenge of automatic cell detection. Yeast cells, representing a vital component of the fungi family, hold profound significance in elucidating the intricacies of eukaryotic cells and human biology. The proposed methodology introduces a paradigm shift in cell detection by optimizing image contrast through optimal fuzzy clustering within the FACE approach. This advancement mitigates the shortcomings of conventional contrast enhancement techniques, minimizing artifacts and suboptimal outcomes. Further enhancing contrast, a universal contrast enhancement variable is ingeniously introduced, enriching image clarity with automatic precision. Experimental validation encompasses a diverse range of yeast cell images subjected to rigorous quantitative assessment via Root-Mean-Square Contrast and Root-Mean-Square Deviation (RMSD). Comparative analyses against conventional enhancement methods showcase the superior performance of the FACE-enhanced images. Notably, the integration of the innovative You Only Look Once (YOLOv5) facilitates automatic cell detection within a finely partitioned grid system. This leads to the development of two models—one operating on pristine raw images, the other harnessing the enriched landscape of FACE-enhanced imagery. Strikingly, the FACE enhancement achieves exceptional accuracy in automatic yeast cell detection by YOLOv5 across both raw and enhanced images. Comprehensive performance evaluations encompassing tenfold accuracy assessments and confidence scoring substantiate the robustness of the FACE-YOLO model. Notably, the integration of FACE-enhanced images serves as a catalyst, significantly elevating the performance of YOLOv5 detection. Complementing these efforts, OpenCV lends computational acumen to delineate precise yeast cell contours and coordinates, augmenting the precision of cell detection.
Similar content being viewed by others
Introduction
In the realm of medical and biological research, the utilization of microscopic imagery capturing cells, tissues, and entire organisms has become a prevailing practice1. The rapid development of single-cell sequencing technologies makes it possible to explore cell heterogeneity of genome, epigenome, transcriptome, and cell interaction/communication in the context of a specific environment in a tissue. However, the accurate detection of individual cells within complex microscopic imagery presents significant challenges. Due to various technical noises such as dropout rate, accurately measuring the expression level in a single cell is pretty challenging. Human analysis can no longer sustain the high throughput image data2. In the era of artificial intelligence and machine learning, these cutting-edge approaches are now widely applied in biological image processing, such as cell type identification, cell count, cell structure analysis, and cell reconstruction3. On the other hand, the effectiveness of image classification algorithms and the quality of the extracted image characteristics limit the accuracy and robustness of most of the methods now in use.
The yeast cell, with its precise developmental requirements and genetic tractability, serves as a valuable model organism in the field of biology. Optical investigations of yeast cell structure play a crucial role across various microbiological research areas, encompassing cell cycle modelling and ageing dynamics studies4,5,6. The current landscape of image processing and segmentation techniques employed for yeast images predominantly relies on traditional methodologies7, including approaches such as thresholding, edge detection, contour fitting, and watershed algorithms. However, the segmentations8 produced by these techniques often necessitate frequent human interventions, as demonstrated in multiple tests. Challenges inherent in yeast cell picture segmentation include issues such as cell crowding, irregular shapes, transparent inclusions, anomalies in visible structures, budding events, and inconsistent imaging focus. Complicating matters further, yeast cell images obtained from diverse imaging devices, such as microscopes, exhibit low contrast9 due to various factors, including the nature of the specimen, image-capturing device calibration settings, and suboptimal ambient lighting conditions.
In recent years, the field of automatic yeast cell detection has been revolutionized by the emergence of deep learning techniques, particularly convolutional neural networks (CNNs)10. These modern approaches leverage large amounts of labelled data to train highly complex models capable of directly learning intricate patterns and features from images. Automatic yeast cell detection within images characterized by low contrast presents a substantial challenge in biological image analysis, where intricate details often become obscured, hampering accurate detection and subsequent analysis. Contrast enhancement is one of the methods of enhancing image quality and highlighting the intricacies of hidden data in the low-contrast regions of an image. Different intensity adjustments were applied to boost the visual contrast11. While enhancing the contrast depends on the selection of parameters of function coefficients that users have artificially determined. Most transformation techniques change the original intensity of the image to a new value, resulting in a distinct pattern. No single transformation function is ideal for improving contrast in each image12.
The task of automatic yeast cell detection presents inherent challenges within image analysis, often exacerbated by factors such as low contrast and intricate backgrounds. To overcome these challenges, this study proposes the utilization of image enhancement techniques emerges as a pivotal strategy to enhance yeast cell visibility and accentuate their distinctive attributes. Notably, the integration of Fuzzy Automatic Contrast Enhancement (FACE)13 stands out as an effective solution, optimizing contrast dynamics while preserving the inherent natural colour distribution. Moreover, the incorporation of the You Only Look Once (YOLO)14 framework, a sophisticated deep learning-based object detection methodology, serves to elevate the precision and efficiency of yeast cell identification. This combination of image enhancement and YOLO enables the precise identification and localization of yeast cells within complex biological images, contributing to the advancement of research in microbiology and biology by facilitating more accurate and insightful analyses of yeast cells in their complex environmental contexts. However, some limitations of the study include challenges tied to labelling, manpower, and time constraints during dataset preparation.
Related work
Image enhancement
Image enhancement plays a significant role in augmenting image quality by emphasizing crucial features, mitigating secondary ones, enriching information, and refining details are undeniable. This process can be broadly categorized into two groups15: spatial ___domain processing, which operates on individual pixels and then combines neighbouring ones; and frequency ___domain processing, which processes the entire image collectively. Recent advancements have introduced efficient algorithms for image enhancement, with many traditional techniques being supplanted by neural network-based models16. This study explores the utilization of various image enhancement models, such as Fuzzy Automatic Contrast Enhancement (FACE), Retinex Algorithm, and Histogram Equalization (HE), to enhance the image quality of cell images.
Histogram equalization, a contrast adjustment technique using image histograms, aids in enhancing global contrast in images17,18. Despite its effectiveness in addressing images with extreme brightness levels, it may not precisely adjust localized areas, resulting in overexposed or underexposed images with amplified noise19. The Retinex algorithm20, founded on retinex theory21, captures the interaction of illumination and reflection to determine object color22,23. Single Scale Retinex (SSR) and Multiple Scale Retinex (MSR) methods exhibit different enhancement outcomes, with MSR integrating SSR outputs using varying weights to address dynamic range limitations24. To overcome colour distortion issues, Multi-Scale Retinex with Color Restoration (MSRCR) emerges as a solution, surpassing MSR in imaging quality and enhancement. The Fuzzy Automatic Contrast Enhancement (FACE)13 technique employs Fuzzy C-Means (FCM)25 clustering for pixel classification, leveraging a universal contrast enhancement variable (UCEV) to automatically maximize image entropy. This approach ensures enhanced contrast without introducing artifacts or noise, benefiting from both accuracy and smoothness in color distribution.
This paper presents the FACE as a contrast enhancement technique for enhancing the microscopic images of yeast cells, changing the original image pixels to a more uncongested pixel distribution. The fuzziness will maintain the smoothness of colour distribution to prevent artifacts in yeast cell images. The maximization approach using a universal contrast enhancement variable was presented to have better contrast enhancement automatically without manually fed parameters. Root-Mean-Square Deviation (RMSD) and Root-Mean-Square Contrast (RMSC) are used to evaluate the contrast of images and the difference between before and after enhancement26,27.
Cell detection
Biological research has substantially adopted Machine Learning (ML) and Deep Learning (DL), notably in creating automated computer vision tools for recognizing and classifying microscopic images, with a focus on species-level identification28,29. Various computer vision techniques have emerged in this context for cell recognition and detection. In cell detection images, challenges arise due to the visual similarity, close proximity, and overlapping nature of cells in the 2D representation of 3D structures30. Uneven illumination, low contrast, low resolution, and diverse foreground/background intensities further complicate the task. Object detection, especially in cell detection, revolves around precise segmentation for local and global differentiation31. Given the multitude of cells in these images, precise detection is essential, particularly in overlap scenarios.
Li et al.32 conducted a comprehensive review of methods for microorganism image analysis, encompassing preparation, feature extraction, classification, and evaluation. Rea et al.33 introduced a GPU-based software utilizing the simplex method for efficient yeast cell detection in time-lapse microscopy, offering real-time analysis capabilities. An advanced RIC-Unet neural network was developed for accurate nuclei segmentation in histology images, integrating residual blocks, multi-scale, and channel attention. Van Valen et al.34 employed the DeepCell software with CNNs for cell segmentation in diverse microscopy images, sharing training insights. Hilsenbeck et al.34 introduced fastER, an efficient cell segmentation tool using SVM to estimate cell likelihood from candidate regions with texture and shape features, adopting a divide and conquer approach for optimal region selection.
Li et al.32 extensively reviewed methods for analyzing microscope images of microorganisms, where he focused on techniques for classifying microorganisms using different image processing steps like preparation, feature extraction, classification, and evaluation. Rea et al.33develops a GPU-based software using the simplex method for the efficient detection of single yeast cells in time-lapse microscopy images. This approach offers significant speedup compared to CPU processing, enabling real-time analysis of microfluidics chip data. An advanced RIC-Unet neural network introduces accurate nuclei segmentation in histology images, incorporating techniques like residual blocks, multi-scale, and channel attention34. Van Valen et al. in34,35 use DeepCell software with CNNs for cell segmentation across diverse microscopy images, sharing training insights. Hilsenbeck et al.34,36 introduce fastER, an efficient cell segmentation tool using SVM to estimate cell likelihood from candidate regions with texture and shape features, using divide and conquer for optimal region selection.
Worldwide, researchers have been attracted to the fundamental and crucial object detection task in computer vision37. Deep learning, particularly deep convolutional neural networks (CNNs), can automatically classify input datasets, such as images. CNNs directly learn from raw pixel data and class labels through end-to-end learning. There are primarily three families of detectors for object detection using deep learning models. These models are Region-CNN(R-CNN)38,39, the Single Shot Detector (SSD)40 and the Yolo series14. However, it’s worth noting that despite the remarkable capabilities of deep learning models like R-CNN, SSD, and YOLO in object detection, they often encounter challenges when dealing with low-light images. The reduced visibility in such conditions can hinder the efficiency and accuracy of object detection processes using these models. So the effect of contrast enhancement on YOLOv5 object detection was studied in this research, and in conclusion, the object detection capabilities of yeast cell image patterns are compared.
Materials and methods
Dataset
To curate the dataset, we captured raw images using a commercially available 8-megapixel camera featuring a 1/1.8-inch sensor size and the SONY imx334 sensor (Pixel size: 2.0 × 2.0 μm). The camera was paired with a 14X-90X optical zoom achromatic field objective, seamlessly integrated into a custom-designed positioning system. To initiate the experimentation phase, a few drops of a solution containing yeast cells (Saccharomyces cerevisiae) were delicately placed onto a petri dish strategically positioned beneath the camera’s objective. This meticulous setup enabled the acquisition of high-resolution images, effectively capturing a diverse spectrum of yeast cells spanning various growth stages. After image capture, a meticulous post-processing regimen ensued, involving resizing these images to a standardized dimension of 500 pixels × 500 pixels. This resizing operation facilitated the creation of images depicting a reduced quantity of cells, thereby rendering them optimally suited for subsequent analytical procedures.
The resultant dataset encompassed a comprehensive compilation of 1000 raw microscopic images of yeast cells, all meticulously formatted to the dimensions of 500 pixels × 500 pixels. This dataset served as the foundational substrate for the FACE image enhancement model. In the initial phases, the FACE methodology was applied to enhance the contrast of these 1000 raw microscopic images, culminating in creation of an enriched and enhanced image dataset. Notably, this dual dataset, consisting of original raw images and their corresponding FACE-enhanced counterparts, formed the bedrock for the subsequent development and validation of the YOLOv5 cell detection model.
Fuzzy automation contrast enhancement (FACE)
Fuzzy Automatic Contrast Enhancement (FACE)13 is a unique method based on fuzzy clustering and a fuzzy expansion of pixel dispersion whose schematic diagram is shown in Fig. 1.
Initially, FACE utilizes a Fuzzy C-Means (FCM) clustering technique to segment an image while categorizing CIELAB colour space pixels with similar colours into tiny image clusters with appropriate attributes. Each group’s RGB colour space pixels are dispersed away from the respective cluster’s centre to enhance the image’s contrast, which maintains the consistency of pixel colours within each cluster. To automatically improve picture contrast, a Universal Contrast Enhancement Variable (UCEV) is created and adjusted to maximize the entropy of the image. A less cluttered distribution of the image pixels guaranteed more excellent picture contrast. The suggested entropy-maximization approach can enhance image quality without manually setting control parameters.
Fuzzy C-Means clustering
Assuming there are \(n\) image pixels \(x_{i}\) for \(i = 1...n\), and arranged into \(k\) groups whose center is denoted by \(j = 1...k\). The inclusion of the \(i{\text{th}}\) pixel in \(j{\text{th}}\) cluster denotes \(u_{ij}\). K-means as an exact clustering method41,42, the value of \(u_{ij}\) is either 0 or 1, where 0 stands for inclusion and 1 stand for non-inclusion. The Fuzzy C-Means (FCM) algorithm allows for imperfect pixel inclusiveness in each cluster. As it states, that \(u_{ij}\) will be a real number with an interval of [0, 1] and always satisfy the Eq. (1).
The definition of total in-group variance is shown in the Eq. (2).
Different researchers43 suggest the strength of inclusion parameters varies from 1 to 5, so for this study, the inclusion parameter is considered as \(m = 3\). For optimal image segmentation, pixels \(J\) must be minimized due to smaller in-group variations leading to larger ones44. The following formulation of a Lagrangian equation is as follows:
where \(\lambda_{i}\) is a Lagrangian multiplier, which was evaluated by the L minimization problem. The two equalities in the below-given Eqs. (4), and (5) can be satisfied to get the optimal fuzzy inclusiveness of each image pixel.
A parameterized method based on the Eqs. (1), (4), and (5) are demonstrated below, to evaluate the ideal parameters:
Equation (6) states the center of the \(j{\text{th}}\) cluster based on given inclusiveness, whereas the Eq. (7) states the inclusiveness parameter \(u_{ij}\) based on the given cluster center points. The value of L converges until the parameterized process continues. In our approach, the parameterized technique begins with arbitrary assumptions given by \(u_{ij}\), and the applied approach can produce consistent results across trials. The FCM generally produces the best identification findings in the CIELAB color space. The Commission International d’Eclairage established the CIELAB color space, where L stands for lightness and AB represents the color component dimensions.
Entropy maximization concerning the universal contrast enhancement variable (UCEV)
The Universal Contrast Enhancement Variable (UCEV) is a pivotal concept within the framework of contrast enhancement, serving as a dynamic control parameter that orchestrates the transformation of pixel values in images to achieve enhanced contrast and improved visual perception. UCEV introduces an adaptive mechanism to adjust pixel values, broadening or contracting their distribution across the color space, thereby intensifying contrast between different features or addressing issues like overly bright regions. This concept hinges on dynamically manipulating pixel distributions while maintaining image integrity, offering adaptability and versatility across various image types and contexts. UCEV’s automated approach eliminates the need for manual parameter adjustment, optimizing image quality by achieving a balance between contrast enhancement and preservation of image details through interaction with entropy maximization, resulting in enhanced visual interpretation and improved image quality.
UCEV introduces an adaptive mechanism, denoted as \(\alpha\) influencing each new pixel value \(x^{\prime}_{i}\), as articulated through the following Eq. (8).
where \(\alpha\) symbolizes the step size in a line search, leading away from the central cluster points through a fusion of fuzzy directions. A positive value of \(\alpha\) widens the image dispersion, resulting in heightened contrast enhancement. Conversely, a negative \(\alpha\) value signifies the contraction of pixel dispersion, potentially aiding in the adjustment of excessively bright photographs. Within the context of contrast enhancement, the Eq. (8) serves as an indispensable tool for harmonizing pixel colours. This study is specifically centred on assessing performance under the condition of K = 3. However, it is imperative to acknowledge the need for future investigations to ascertain the optimal clustering parameter. Furthermore, it becomes evident that achieving a brighter photograph with enhanced contrast and harmonious colour distribution is attainable by utilising real number values within the interval of [0, 1] for \(u_{ij}\).
A global measurement, entropy J45 is maximized concerning \(\alpha\), which is the final step of contrast enhancement shown by the Eq. (9).
where \(p\) stands for probability density and \(x^{\prime}\) a set of image pixels which was obtained from the Eq. (8) for measuring the unpredictability of distribution of pixels. The analysis is conducted through the application of a gridding set \(\Omega\), which is employed for the computation of image entropy. A higher entropy value represents an increased degree of randomness within pixel distribution, showcasing substantial variance between the pixels. This results in enhancing image contrast by the optimization process stated in the Eq. (9).
Within the analyzed colour space, the expansion of the pixel distribution is curtailed. In this study, the repositioned pixel points that exceed the boundaries of the colour space are constrained to remain within those limits. As a consequence, the escalation in image entropy ceases once specific pixel points converge upon the boundary edges. A higher value of \(\alpha\) will lead to a greater accumulation of pixel points clustering along the bounds, resulting in reduced variation among them and potentially yielding a lower level of entropy. Consequently, the application of the maximization technique yields maximal entropy. An illustration of the FACE process as applied to yeast cells is depicted in Fig. 2.
Demonstration of the FACE enhancement for Yeast Cells
An insightful observation has been presented, highlighting the significance of determining the optimal UCEV value (\(\alpha\)). This value intricately ties in with the accumulation of pixel points along the image boundaries, subsequently influencing the entropy. When UCEV provides a higher value, it corresponds to the maximization of entropy. The entire process of enhancement with the relation of \(\alpha\) and entropy are illustrated in Table 1. The first row within the table visually maps out the FACE enhancement procedure, encompassing alterations in the \(\alpha\) value. The accompanying graph elegantly portrays the resulting variation in entropy as a function of the \(\alpha\) value, providing a comprehensive visual representation of this process. Therefore, by employing the maximization process embedded in FACE, the optimal value of UCEV can be identified, leading to the attainment of maximum entropy. This endeavour significantly enhances image quality, as visually depicted in Fig. 3.
Figure 4a shows the original raw microscopic image of yeast cells to demonstrate the contrast enhancement. The suggested FACE approach uses fuzzy clustering and fuzzy entropy maximization to determine the optimal conditions for the contrast enhancement of the image. The smooth enhancement is accomplished by the fuzziness in the clustering process, which assures the belongingness of the pixel point’s accuracy. With just one control parameter, the specification of UCEV enables the contrast enhancement process could efficiently produce a global contrast enhancement effect. The entropy maximization technique assures that the pixel distributions are greatly enlarged and do not gather on the color space’s boundaries. An enhanced image is optimized by the proposed FACE method, as shown in Fig. 4b, and it is determined without the need for human feed control parameters.
Enhancement performance measurement
Two statistical methods are used to measure the qualitative index of contrast enhancement performed by FACE and other techniques. The first method is the Root-Mean-Square Deviation (RMSD), which was performed between the original and enhanced images. The Eq. (10) shows the calculation for the RMSD value.
where \(x_{i}\) and \(x^{\prime}_{i}\) are the \(i{\text{th}}\) pixel point of the original image and enhanced image, respectively, in case of a large number of RMSD values shows the large variation in pixel color caused by the method used for enhancement. If the original image has low contrast, it is desired to have an enhanced image with a greater RMSD value. If the case is reversed, then the value of RMSD would be desirably small.
Another method known as Root-Mean-Square Contrast (RMSC) is used for quantifying the contrast enhancement process, where the difference between each pixel point is calculated as shown in the Eq. (11).
Spatial distribution of contrast does not rely on the measurement of the RMSC, which means that it cannot provide the existence of artifacts. This method provides an overall impression of the image feature and its background. An image with a higher RMSC is desirable, but if its value is too high, it will be over-enhanced.
Automatic cell detection using YOLOv5
Yolov5
YOLOv5 is a single-stage object detector containing three crucial components like another single-stage detector. The components of YOLOv5 are the model backbone, model neck, and model head. Model backbone is the initial stage, whose primary purpose is to extract significant features from an input image. In YOLOv5, the cross-stage partial networks (CSP) are applied as a backbone to extract rich, informative features from an input image. With deeper networks, CSPNet has demonstrated a considerable reduction in processing time.
YOLOv5 has a second component known as the model neck, which generates feature pyramids. Feature pyramids assist models in developing decent object scaling generalizations. Model neck helps to recognize the same object in various sizes and scales. Models that employ feature pyramids perform well on unseen data. Different feature pyramid methods are available such as BiFPN, PANet, FPN, etc. PANet is a feature pyramid technique that is used in YOLOv5. The model head performs the final detection part. Anchor boxes are applied to features through which final output vectors are generated along with class probabilities, objectness, and bounding boxes. The model head is the same as in previous YOLO versions. The final model architecture for YOLOv5 is shown in Fig. 5.
YOLOv5 was used to recognize the yeast cells in the investigated images. YOLOv5 is an object detection method based on CNN structure. The confidence level of each recognized image pattern based on CNN is evaluated, and the optimal bounding box of a recognized image pattern is then determined. The yeast cell image patterns learning database in this paper is built on the tool Labellmg and Roboflow.
Automatic cell detection using FACE-YOLO model
To comprehensively investigate the impact of contrast enhancement in automatic yeast cell detection, our study meticulously created two sets of image databases. The first encompassed the original raw yeast cell images, while the second was composed of meticulously enhanced images utilizing the FACE approach. This dual-pronged approach enabled us to intricately probe the effect of contrast enhancement on the efficacy of the yeast cell detection methodology, seamlessly integrated with the cutting-edge YOLOv5 framework. The process of automatic cell detection using this FACE-YOLO model was shown by a schematic diagram shown in Fig. 6.
The yeast cell detection process commences with the meticulous labelling and organization of the dataset. To facilitate this, two distinct datasets are meticulously curated: the initial dataset comprising raw cell images and an additional dataset comprising images of yeast cells subjected to enhancement via the FACE methodology. Subsequently, the labelled training images are allocated to the backbone, neck, and head of the architecture, culminating in the generation of the desired output. The weights acquired through the training process are then applied to the testing phase, yielding conclusive results. Detection is executed through systematic manipulation of the input images and datasets. Initially, the raw training dataset is employed for cell detection, involving both raw images and FACE-enhanced images. Following this, the training dataset enriched with FACE-enhanced images is deployed for cell detection, encompassing both the original raw images and their enhanced counterparts. The computational parameter used for the FACE-YOLO model is detailed in Table 2.
A comparative analysis presented in Table 3 scrutinises the efficiency of training and testing interchange utilizing original raw images and FACE-enhanced images for yeast cell detection. The results illuminate that training with FACE-enhanced images consistently yields superior detection performance automatic cell detection with YOLOv5 for both raw and enhanced image datasets.
For the task of finding the real-time position of the cell, the primary objective is to get the contour and pinpoint the precise coordinates of the yeast cell. This dynamic real-time process for detecting yeast cell contour and centre of the cell is illustrated in Fig. 7. The process unfolds with the utilization of YOLOv5 to train the FACE dataset, followed by employing the trained FACE model to detect yeast cells in the raw image. Subsequently, the application of OpenCV comes into play to identify the contour and ___location of each yeast cell. The procedure commences by extracting the bounding box from the YOLOv5 image’s outcome to verify the yeast cell’s position. Subsequent steps involve transforming the image into binary form, eliminating the bounding box and redundant portions, and employing the OpenCV function ‘cv2’ to locate the cell’s contour and centre. Finally, the identified contour and centre are overlaid onto the original raw image.
Results analysis
The microscopic images of yeast cells are taken from the laboratory, creating a database of 1000 raw images. This paper focuses on the performance of fuzzy clustering at \(k = 2\). Comparative analysis is done using six original images from the database using three different techniques: FACE, SSR, and HE. The comparison results of contrast enhancement of the original image with other methods as shown in Table 4.
The qualitative index was measured using RMSD and RMSC values for original microscopic and enhanced images, shown in Table 4. RMSC values of original and enhanced images were calculated, shown in Table 5. RMSC values have increased after the image enhancement using different techniques. The RMSC of all the enhanced images is higher than the RMSC of the Original image. But when compared, HE has a very high RMSC value which provides over-contrast enhancement. While we can observe that features are apparent using FACE. Using an SSR image has generated a greater amount of color change than the original image. Many artifacts were ground in the background when SSR and HE were used. Hence fuzzy clustering in FACE keeps the image free of artifacts and provides good features that keep the image’s originality.
While comparing the RMSD values, Table 6 shows that FACE provides a lower RMSD value than HE and SSR.FACE does not overchange the image’s contrast to the original image, while the RMSD value is small compared to the other two methods, which shows that the other techniques have changed the colour tones and features lacking issues compared with the original image. Hence FACE delivers the proper amount of contrast enhancement for the microscopic images.
Next, YOLOv5 was used to recognize the yeast cells from the original microscopic images. In this study, the two models using a training dataset of the yeast cell image patterns are generated, and the effect of image enhancement on object detection by YOLOv5 is analyzed. The first model using the dataset was original raw yeast cell images, and the second model the dataset of enhanced images by FACE. Each model uses a different dataset to build a YOLO object detection system. When the performance compared for the detection of yeast cells between the original raw image model and the FACE model, it is observed that when the original raw image model is considered for YOLO detection, many cells are not identified due to being small in nature or blurred image, as shown in Fig. 8a. On the other hand, the FACE model could effectively increase the visual quality of the cells in the image, as shown in Fig. 8b, and improve object detection accuracy. As observed from the two models of cell detection using YOLOv5 for comparative analysis, the FACE model provides good results with higher accuracy of yeast cell detection using YOLOv5, and it is also capable of detecting the yeast cells with original raw input image having low contrast features with high accuracy.
The performance of the FACE training model was analyzed using tenfold cross-validation accuracy and confidence score. As we can see, in the first fold and the second fold of the accuracy, the detection of the FACE training model had an accuracy above 90%, as compared with the detection of the raw image training model accuracy the FACE training model is more suitable for the blurry image or boundary of the cells are not clear. The confidence scores of the FACE training model are much higher than the raw training model, although the third and the fifth fold are lower, and these two confidence scores do not differ much. So the confidence scores of the FACE model are nearly the highest among the two models. The tenfold cross-validation accuracy and confidence scores are shown in Table 7.
The image shown in Fig. 9 is the performance metrics of the FACE training model, resulting in the loss value of train and validation convergence gradually and then reaching a stable level, and the precision, recall, and mAP value rise to high accuracy and reach almost 100%. The training leads to a satisfactory result that shows our model is trained well.
The recognition accuracy is demonstrated using different raw, enhanced, and synthetic training datasets. The synthetic dataset is a computer-generated yeast cell image from Blender3D46 representing all the object classes. It is noted that the boundaries of the yeast cells in the enhanced images are more precise and accessible to recognize than those in the original images. At first, the model is trained with 100% raw images to set a reference result. Then the next model is trained repeatedly, and for each new session, 10% of the dataset is switched out to enhanced images until the training dataset is 100% enhanced images. The same was done with the synthetic dataset. Then the following models are trained with different enhanced: synthetic dataset ratios. For determining the quality of the results, we used metrics such as mAP_0.5. The recognition accuracy demonstrated by the dark blue curve, which is found to be low, was based on the original microscopic images of yeast cells. When the accuracy is performed using 40% synthetic, and 60% enhanced images, it provides good detection accuracy, shown by the curve in grey color. The other curves are based on different research settings and can be seen in Fig. 10.
In the pursuit of enhancing the efficiency of the proposed model, a comparative study was undertaken involving two state-of-the-art deep learning models: FACE-YOLO and Mask R-CNN, for the task of automated yeast cell detection. The tenfold accuracy results of both models are presented in Table 8, revealing that the FACE-YOLO model outperforms Mask R-CNN in terms of accuracy. The comparitive outcomes of automatic cell detection by FACE-YOLO and Mask R-CNN are visually depicted in the accompanying Fig. 11.
Conclusion
This work proposes Fuzzy Automatic Contrast Enhancement (FACE) as an image quality enhancement technique to enhance the microscopic images of yeast cells. FACE employs fuzzy clustering to prevent visible pixel separation, while a universal contrast enhancement variable (UCEV) optimizes entropy for enhanced image quality. In this work, the image enhancement of the microscopic image of yeast cells was done automatically without any feed of human-defined parameters. Finally, FACE created a less congested image pixel distribution with a significantly larger picture contrast while effectively enhancing the yeast cell image’s contrast level without producing local flaws such as a change in colour, visual artifacts, and over-strengthening. When the qualitative index was performed, and compared to other methods, FACE performed better enhancement results. The integration of YOLOv5 for automatic yeast cell detection further elevates the study's contributions. Notably, the FACE-enhanced model within YOLOv5 improves yeast cell detection accuracy. Impressively, the FACE-YOLO model outperforms Mask R-CNN with a remarkable 94.2% accuracy. This study underscores the promising synergy between FACE-enhanced imagery and advanced detection techniques, offering a pathway for heightened precision and effectiveness in cell analysis across a spectrum of biomedical applications. Additionally, the localization of cell boundaries and centres through OpenCV enhances cell positioning accuracy. As a future direction, integrating the automatic FACE-YOLO cell detection with a microgripper robot holds the potential for facilitating cell manipulation and handling.
Data availability
The datasets used during the current study are available from the corresponding author upon request.
References
Wang, H. et al. Biological image analysis using deep learning-based methods: Literature review. Digit. Med. 4, 157. https://doi.org/10.4103/digm.digm_16_18 (2018).
Eliceiri, K. W. et al. Biological imaging software tools. Nat. Methods 9, 697–710. https://doi.org/10.1038/nmeth.2084 (2012).
Liu, J., Fan, Z., Zhao, W. & Zhou, X. Machine intelligence in single-cell data analysis: advances and new challenges. Front. Genet. https://doi.org/10.3389/fgene.2021.655536 (2021).
Simon, I. et al. Serial regulation of transcriptional regulators in the yeast cell cycle. Cell 106, 697–708. https://doi.org/10.1016/s0092-8674(01)00494-9 (2001).
Iyer, V. R. et al. Genomic binding sites of the yeast cell-cycle transcription factors SBF and MBF. Nature 409, 533–538. https://doi.org/10.1038/35054095 (2001).
Rustici, G. et al. Periodic gene expression program of the fission yeast cell cycle. Nat. Genet. 36, 809–817. https://doi.org/10.1038/ng1377 (2004).
Kulwa, F. et al. A state-of-the-art survey for microorganism image segmentation methods and future potential. IEEE Access 7, 100243–100269. https://doi.org/10.1109/ACCESS.2019.2930111 (2019).
Bredies, K. & Wolinski, H. An active-contour based algorithm for the automated segmentation of dense yeast populations on transmission microscopy images. Comput. Visual Sci. 14, 341–352. https://doi.org/10.1007/s00791-012-0178-8 (2011).
Casacio, C. A. et al. Quantum-enhanced nonlinear microscopy. Nature 594, 201–206. https://doi.org/10.1038/s41586-021-03528-w (2021).
Rani, P., Kotwal, S., Manhas, J., Sharma, V. & Sharma, S. Machine learning and deep learning based computational approaches in automatic microorganisms image recognition: Methodologies, challenges, and developments. Arch. Comput. Methods Eng. 29, 1801–1837. https://doi.org/10.1007/s11831-021-09639-x (2022).
Gonzalez, R. C. Digital image processing. (Pearson education INDIA, 2009).
Demirel, H., Ozcinar, C. & Anbarjafari, G. Satellite image contrast enhancement using discrete wavelet transform and singular value decomposition. IEEE Geosci. Remote Sens. Lett. 7, 333–337. https://doi.org/10.1109/LGRS.2009.2034873 (2009).
Lin, P. T. & Lin, B. R. in 2016 12th IEEE/ASME International Conference on Mechatronic and Embedded Systems and Applications (MESA). 1–10 (IEEE).
Redmon, J., Divvala, S., Girshick, R. & Farhadi, A. in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 779–788.
Chaker, A., Mlika, A., Laribi, M., Romdhane, L. & Zeghloul, S. Clearance and manufacturing error’s effects on the accuracy of the 3-RCC Spherical Parallel Manipulator. Eur. J. Mech. A/Solids 37, 86–95. https://doi.org/10.1016/j.euromechsol.2012.05.009 (2013).
Dietler, N. et al. A convolutional neural network segments yeast microscopy images with high accuracy. Nat. Commun. 11, 5723. https://doi.org/10.1038/s41467-020-19557-4 (2020).
Pizer, S. M. et al. Adaptive histogram equalization and its variations. Comput. Gr. Image Process. 39, 355–368. https://doi.org/10.1016/S0734-189X(87)80186-X (1987).
Abdullah-Al-Wadud, M., Kabir, M. H., Dewan, M. A. A. & Chae, O. A dynamic histogram equalization for image contrast enhancement. IEEE T Consum. Electr. 53, 593–600. https://doi.org/10.1109/TCE.2007.381734 (2007).
Pisano, E. D. et al. Contrast limited adaptive histogram equalization image processing to improve the detection of simulated spiculations in dense mammograms. J. Digit. Imaging 11, 193–200. https://doi.org/10.1007/BF03178082 (1998).
Land, E. H. & McCann, J. J. Lightness and retinex theory. J. Opt. Soc. Am. 61, 1–11. https://doi.org/10.1364/JOSA.61.000001 (1971).
Jobson, D. J., Rahman, Z.-U. & Woodell, G. A. Properties and performance of a center/surround retinex. IEEE Trans. Image Process. 6, 451–462. https://doi.org/10.1109/83.557356 (1997).
Fu, X., Zeng, D., Huang, Y., Zhang, X.-P. & Ding, X. in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 2782–2790.
Guo, X., Li, Y. & Ling, H. LIME: Low-light image enhancement via illumination map estimation. IEEE Trans. Image Process. 26, 982–993. https://doi.org/10.1109/TIP.2016.2639450 (2016).
Rahman, Z.-u., Jobson, D. J. & Woodell, G. A. in Proceedings of 3rd IEEE International Conference on Image Processing. 1003–1006 (IEEE).
Bezdek, J. C., Ehrlich, R. & Full, W. FCM: The fuzzy c-means clustering algorithm. Comput. Geosci. 10, 191–203. https://doi.org/10.1016/0098-3004(84)90020-7 (1984).
Dileep, M. & Murthy, A. S. in 2011 International Conference on Emerging Trends in Electrical and Computer Technology. 708–712 (IEEE).
Ionescu, C., Fosalau, C. & Petrisor, D. in 2014 International Conference and Exposition on Electrical and Power Engineering (EPE). 100–104 (IEEE).
Alanazi, A. Using machine learning for healthcare challenges and opportunities. Inform. Med. Unlocked 30, 100924 (2022).
Kaul, D., Raju, H. & Tripathy, B. Deep learning in healthcare. Deep Learning in Data Analytics: Recent Techniques, Practices and Applications, 97–115 (2022).
Ghafari, M. et al. in 2022 International Conference on Artificial Intelligence in Information and Communication (ICAIIC). 204–209 (IEEE).
Maddalena, L., Antonelli, L., Albu, A., Hada, A. & Guarracino, M. R. Artificial intelligence for cell segmentation, event detection, and tracking for label-free microscopy imaging. Algorithms 15, 313. https://doi.org/10.3390/a15090313 (2022).
Li, C., Wang, K. & Xu, N. A survey for the applications of content-based microscopic image analysis in microorganism classification domains. Artif. Intell. Rev. 51, 577–646. https://doi.org/10.1007/s10462-017-9572-4 (2019).
Rea, D., Perrino, G., di Bernardo, D., Marcellino, L. & Romano, D. A GPU algorithm for tracking yeast cells in phase-contrast microscopy images. Int. J. High Perform. Comput. Appl. 33, 651–659. https://doi.org/10.1177/10943420188014 (2019).
Zeng, Z., Xie, W., Zhang, Y. & Lu, Y. RIC-Unet: An improved neural network based on Unet for nuclei segmentation in histology images. IEEE Access 7, 21420–21428. https://doi.org/10.1109/ACCESS.2019.2896920 (2019).
Moen, E. et al. Deep learning for cellular image analysis. Nat. Methods 16, 1233–1246. https://doi.org/10.1038/s41592-019-0403-1 (2019).
Hilsenbeck, O. et al. fastER: a user-friendly tool for ultrafast and robust cell segmentation in large-scale microscopy. Bioinformatics 33, 2020–2028. https://doi.org/10.1093/bioinformatics/btx107 (2017).
Wang, B., Cao, G., Zhou, L., Zhang, Y. & Shang, Y. Task differentiation: Constructing robust branches for precise object detection. Comput. Vis. Image Underst. 199, 103030. https://doi.org/10.1016/j.cviu.2020.103030 (2020).
Yang, S. et al. in 2017 International Conference on Security, Pattern Analysis, and Cybernetics (SPAC). 345–350 (IEEE).
Fujita, S. & Han, X. -H. in Proceedings of the Asian Conference on Computer Vision.
Bresilla, K. et al. Single-shot convolution neural networks for real-time fruit detection within the tree. Front. Plant Sci. 10, 611. https://doi.org/10.3389/fpls.2019.00611 (2019).
Jain, A. K. Data clustering: 50 years beyond K-means. Pattern Recognit. Lett. 31, 651–666. https://doi.org/10.1016/j.patrec.2009.09.011 (2010).
Wagstaff, K., Cardie, C., Rogers, S. & Schrödl, S. in Proceedings of the Eighteenth International Conference on Machine Learning. 577–584 (Morgan Kaufmann Publishers Inc.).
Cannon, R. L., Dave, J. V. & Bezdek, J. C. Efficient implementation of the fuzzy c-means clustering algorithms. IEEE Trans. Pattern Anal. Mach. Intell. https://doi.org/10.1109/TPAMI.1986.4767778 (1986).
Otsu, N. A threshold selection method from gray-level histograms. IEEE T Syst. Man. Cyb 9, 62–66. https://doi.org/10.1109/TSMC.1979.4310076 (1979).
Narayan, R. & Nityananda, R. Maximum entropy image restoration in astronomy. Annu. Rev. Astronomy Astrophys. 24, 127–170. https://doi.org/10.1146/annurev.aa.24.090186.001015 (1986).
Community, B. Blender–A 3D Modelling and Rendering Package (Stichting Blender Foundation, Amsterdam, 2018).
Acknowledgements
This paper was supported by the National Science and Technology Council (NSTC), Taiwan (grant numbers MOST 108-2221-E-011-129-MY3 and MOST 110-2923-E-011-006-MY3), Research Council of Lithuania (LMTLT), (Grant Number P-LLT-21-6), State Education Development Agency of Latvia (Grant Number LV-LT-TW/2021/8), Taipei Medical University-National Taiwan University of Science and Technology Joint Research Program (Grant Number TMU-NTUST-108-08), and Intelligent Manufacturing Innovation Center (IMIC) at National Taiwan University of Science and Technology (NTUST), which is a featured areas research center in Higher Education Sprout Project of Ministry of Education (MOE), Taiwan.
Author information
Authors and Affiliations
Contributions
V.B., M.G., Y.W.W., and P.T.L. design the conceptual ideas. Data collection and analysis were done by Z.J.H., T.Y.Y., and W.C.T. The first draft of the manuscript was written by B.P., and all authors commented on previous versions of the manuscript. All authors read and approved the final manuscript.
Corresponding authors
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Huang, ZJ., Patel, B., Lu, WH. et al. Yeast cell detection using fuzzy automatic contrast enhancement (FACE) and you only look once (YOLO). Sci Rep 13, 16222 (2023). https://doi.org/10.1038/s41598-023-43452-9
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-023-43452-9
This article is cited by
-
Research on X-ray Image Fusion Algorithm for Food Foreign Object Detection
Journal of Nondestructive Evaluation (2025)
-
Defect detection of photovoltaic modules based on improved VarifocalNet
Scientific Reports (2024)