A Color and Shape-Aware TSR Model Enhanced by Morphological Filtering and Fuzzy Logic
Abstract:
Real-time traffic sign recognition (TSR) plays a crucial role in intelligent transportation systems (ITS) and autonomous driving technologies. It enhances road safety, ensures efficient traffic rule enforcement, and supports the seamless operation of both autonomous and driver-assist systems. This paper proposes a hybrid TSR model that integrates mathematical morphology, edge detection, and fuzzy logic to accurately identify and classify traffic signs across diverse environmental conditions. The preprocessing stage applies contrast enhancement and Gaussian filtering to improve the visibility of key features. Next, shape- and color-based segmentation using mathematical morphology extracts regions of interest that are likely to contain traffic signs. These regions are then analyzed using a fuzzy inference system (FIS) that evaluates features such as color intensity, geometric shape ratios, and edge sharpness. The fuzzy system handles the inherent ambiguity in visual patterns, enabling robust decision-making. The entire model is developed in MATLAB R2015a, ensuring both computational efficiency and real-time performance. The integration of classical mathematical techniques with fuzzy reasoning allows the system to maintain high accuracy and reliability across a wide variety of traffic scenes. The proposed approach demonstrates significant potential for practical deployment in ITS applications, including smart vehicles and automated road safety systems.
1. Introduction
TSR plays a vital role in the development of ITS and autonomous driving technologies. The ability to accurately detect and interpret traffic signs in real time not only enhances driver assistance systems but also significantly contributes to road safety. With increasing vehicle automation, reliable TSR models are essential to ensure that vehicles can correctly understand and react to regulatory and warning signs under diverse environmental conditions. Recent studies have explored a wide range of methodologies for improving TSR accuracy, including traditional image processing, machine learning, and deep learning frameworks [1], [2], [3], [4], [5]. These approaches emphasize robustness in challenging scenarios such as poor lighting, partial occlusion, and cluttered backgrounds, laying a solid foundation for the advancement of intelligent and context-aware driving systems.
Researchers have been studying deep learning techniques extensively to improve the accuracy and efficiency of traffic sign detection and recognition systems. Wang et al. [6] proposed an enhanced YOLOv5-based model that achieves a balance between detection performance and computational speed. In another approach, Barodi et al. [7] developed an explainable artificial intelligence (XAI) model to ensure transparency and trustworthiness in TSR—a critical aspect in safety-critical domains like ITS. Nabou et al. [8] applied the YOLOv8 model and demonstrated its robustness across varied road conditions, showcasing its adaptability to real-world scenarios. Meanwhile, Zhu and Yan [9] highlighted the effectiveness of deep convolutional neural networks in achieving high recognition accuracy for traffic signs. Zhan and Temirgaziyeva [10] further contributed by optimizing convolutional neural network architectures for more efficient road sign detection. These contributions collectively reflect a growing trend toward deploying deep learning-driven solutions that outperform traditional methods in terms of scalability, robustness, and real-time performance.
Researchers have also focused on enhancing the preprocessing and detection stages of TSR to further optimize system performance. Liu and Yan [11] introduced an improved histogram equalization technique combined with bilinear interpolation to preprocess traffic sign images, resulting in enhanced image contrast and resolution. However, their approach demonstrated limited adaptability in handling varying illumination and complex backgrounds. Mani et al. [12] proposed a deep learning-based traffic sign detection and recognition framework tailored for autonomous vehicles. Their model achieved high accuracy and real-time performance, yet the reliance on large annotated datasets remains a challenge, particularly in resource-constrained settings. In another effort, Jenifer and Balamanigandan [13] developed an advanced image processing technique for traffic sign detection, which showed promising results in terms of sign localization, though its robustness under noisy and occluded conditions was not fully addressed. Furthermore, Suresha et al. [14] reviewed recent advancements in small traffic sign detection and highlighted state-of-the-art approaches and datasets. While their survey provides a valuable overview, it also emphasized the persistent issue of small sign misdetection due to scale variation and background clutter. These studies collectively reveal significant progress in TSR while also underlining the existing gaps that necessitate further research and refinement.
Considering the limitations identified in existing approaches—such as sensitivity to lighting variations in histogram-based methods, high computational demands in deep learning-based models, and limited generalizability of handcrafted techniques, this study proposes a novel real-time TSR model that integrates classical mathematical tools with fuzzy logic to achieve enhanced accuracy and robustness under diverse environmental conditions. Figure 1 illustrates the workflow from image acquisition and preprocessing to feature extraction and fuzzy logic-based classification, highlighting key stages such as edge detection, Hough transform-based shape isolation, and multi-criteria decision-making. Unlike previous works that rely heavily on either deep learning or simple thresholding methods, our model begins with a hybrid preprocessing stage involving histogram equalization and Gaussian filtering to improve contrast and suppress noise. It then applies shape- and color-based segmentation using mathematical morphology, tailored to reduce false positives in complex backgrounds. An FIS is employed for final classification, incorporating features such as color intensity, edge clarity, and geometric proportions—an approach aimed at overcoming the rigidity of traditional rule-based or SVM-based classifiers. To ensure computational feasibility, the entire framework is implemented in MATLAB R2015a with optimized routines for image filtering, candidate region extraction, and fuzzy classification. Experimental evaluations confirm that the proposed model delivers improved recognition accuracy and stability across diverse traffic and weather conditions, validating its effectiveness for real-time deployment in ITS.

2. Related Work
Researchers have extensively studied various techniques for traffic sign and gesture recognition, focusing on improving accuracy, real-time performance, and adaptability to complex environments. These studies encompass a range of approaches, including fuzzy logic, color and shape-based detection, and advanced image processing frameworks. The literature reveals significant progress in addressing challenges related to environmental variability, small object detection, and integration into ITS. This section high-lights key contributions and limitations of recent advancements in traffic sign and sign recognition.
Broadly, existing approaches can be categorized into three groups: (i) traditional image processing methods, (ii) deep learning-based methods, and (iii) fuzzy logic and hybrid methods. Each category has unique strengths but also notable weaknesses, which highlight the research gap addressed by our proposed model.
Ampadu and Huebner [15] proposed a novel approach to autonomous traffic gesture integrity monitoring by employing fuzzy logic to interpret complex road gestures in real-time. The model emphasizes the use of fuzzy decision-making to improve system flexibility and reduce dependency on rigid, rule-based frameworks. A significant achievement of their study is the ability to handle gesture ambiguity in dynamic traffic scenarios, thereby enhancing system reliability and safety. Their methodology also allows for adaptability in the presence of occlusions and varying gesture patterns.
Despite its strengths, the study is limited by its constrained dataset and lack of large-scale validation. The fuzzy logic-based approach, while robust to uncertainty, may suffer from performance drops in cases where gesture transitions are too subtle or vary significantly across regions. Furthermore, the implementation details remain limited, making replication and benchmarking against other models difficult.
Li and Liu [16] investigated image processing technologies specific to traffic sign detection for autonomous vehicles. Their work integrates image denoising, color enhancement, and shape extraction techniques to accurately detect and classify traffic signs. One of the key achievements of their method is its robustness against motion blur and lighting variations, which are common in autonomous driving environments. The authors also demonstrated that their sign recognition algorithm could maintain high accuracy during both day and night conditions.
However, the research primarily focuses on static video frames and does not extend to real-time applications. While the preprocessing pipeline shows promising results, it lacks integration with end-to-end vehicle control systems. Additionally, the dataset used was not fully disclosed, raising concerns about the generalizability of their findings to diverse real-world scenarios.
Gan [17] presented a fast recognition method for pedestrian-related signs using a combination of color feature extraction and support vector machines (SVM). The model was particularly effective in detecting red-bordered triangular and circular signs that are common in pedestrian zones. A notable achievement of this approach is its balance between computational efficiency and classification accuracy, making it suitable for embedded vision systems in smart vehicles. The use of color segmentation significantly reduced the search space, while the SVM provided robust discrimination across sign types.
Nonetheless, the system exhibits certain limitations in recognizing faded or heavily occluded signs, where color cues become unreliable. Additionally, the method may not scale well when introduced to multilingual or highly cluttered traffic environments. The reliance on handcrafted features also limits adaptability compared to deep learning-based methods, which could offer improved performance on complex datasets.
Table 1 summarizes the strengths and weaknesses of traditional, deep learning, and fuzzy logic-based methods. While traditional methods are computationally efficient, they struggle with occlusion and environmental variability. Deep learning approaches achieve high accuracy but demand large annotated datasets and high computational power. Fuzzy logic-based methods effectively handle uncertainty and interpretability but often lack large-scale benchmarking. Our proposed fuzzy logic-based road sign recognition model is designed to combine robustness against uncertainty with improved adaptability to real-world conditions, thereby addressing these gaps.
Category | Techniques | Strengths | Limitations |
Traditional Image Processing | Color thresholding, shape detection, SVMs | Fast, low computational demand, suitable for embedded systems | Sensitive to lighting, noise, and occlusion; limited adaptability |
Deep Learning | CNNs, transfer learning, large datasets | High accuracy, automatic feature extraction, scalable to complex datasets | Requires large annotated datasets, high computational cost, less interpretable |
Fuzzy Logic / Hybrid | Rule-based fuzzy systems, neuro-fuzzy methods | Handles uncertainty, interpretable, adaptable to variations | Limited benchmarking, sometimes less scalable for real-time large-scale use |
3. Methodology
The proposed methodology for real-time TSR is structured into a sequence of processing stages, each designed to enhance accuracy, speed, and robustness under real-world driving conditions. The system begins with image acquisition, preprocessing is then applied to remove noise and improve contrast, using Gaussian filtering and histogram equalization. Following this, color and shape-based segmentation is performed to isolate regions of interest that may contain traffic signs. Feature extraction is subsequently carried out, where each candidate region is analyzed for attributes such as shape descriptors, edge density, and color distribution. These features are then fed into an FIS, which evaluates the degree of membership of each region to different traffic sign classes. This rule-based fuzzy classifier enables the model to handle uncertainty and imprecise data, making it particularly effective in challenging environments such as fog, low light, or occlusion. Finally, the recognized traffic sign is highlighted and annotated on the original image in real-time.
The proposed system processes video frames captured at a rate of 25 frames per second (fps), which ensures a smooth and near real-time performance suitable for TSR. Each frame is then subjected to a preprocessing pipeline aimed at enhancing color-based features critical for distinguishing traffic signs.
Color Transformation: Since RGB is highly sensitive to illumination, frames are transformed into the Hue, Saturation, Value (HSV) space [18], [19], which separates chromatic information from brightness and is more robust for traffic sign detection. The transformation highlights hue ($H$) for dominant color, saturation ($S$) for color purity, and value ($V$) for brightness:
$ S=\frac{C}{\max \left(R^{\prime}, G^{\prime}, B^{\prime}\right)}, \quad V=\max \left(R^{\prime}, G^{\prime}, B^{\prime}\right) $
where, $C=\max \left(R^{\prime}, G^{\prime}, B^{\prime}\right)-\min \left(R^{\prime}, G^{\prime}, B^{\prime}\right)$. The hue is determined based on the relative dominance of RGB channels.
Color Range Modeling: Thresholds in HSV space are defined to capture traffic sign colors:
- Red: $H \in\left[ 0^{\circ}, 10^{\circ}\right] \cup\left[ 170^{\circ}, 180^{\circ}\right]$
- Blue: $H \in\left[ 100^{\circ}, 130^{\circ}\right]$ (critical for circular/rectangular signs)
- Yellow:$H \in\left[ 20^{\circ}, 40^{\circ}\right]$
- White: low $S$, high $V$
Multi-colored signs are segmented by combining masks (e.g., red border $+$ white interior $+$ blue icon). This approach follows standard practices in traffic sign detection.
To accurately isolate regions corresponding to traffic signs, edge detection followed by morphological operations is employed. These techniques are fundamental in identifying object boundaries and refining segmented regions.
Canny Edge Detector: The Canny operator is used for edge detection due to its robustness against noise and ability to preserve fine details. The process involves Gaussian smoothing, gradient computation, non-maximum suppression, and hysteresis thresholding [20]. This ensures reliable extraction of traffic sign contours under varying image conditions.
Morphological Operations: Basic morphological operations, including erosion and dilation, are applied with a disk-shaped structuring element to remove noise and enhance segmented sign regions. More complex operations such as opening and closing are further used to refine object boundaries and maintain shape consistency. These steps strengthen the binary masks before feature extraction, ensuring clearer and more accurate sign representation.
To robustly localize traffic signs, the proposed model employs both shape detection and color thresholding. This combined strategy leverages the inherent circular structure of traffic signs and their distinctive color features in the HSV color space.
Hough Circle Transform: The Hough Circle Transform is utilized to detect circular objects in the edge-detected image. The standard form of the circle equation is:
$ x^2+y^2=r^2 $
In the Hough space, the presence of a circle is determined by accumulating votes in a three-dimensional parameter space $(x, y, r)$, where $x$ and $y$ denote the circle’s center, and $r$ is its radius. This is mathematically represented as:
$ H(x, y, r)=\sum_{i, j} f(i, j) \delta\left((i-x)^2+(j-y)^2-r^2\right) $
here, $f (i, j)$ represents the binary edge image (typically from a Canny detector), and $\delta$ is the Dirac delta function, which acts as an accumulator. A peak in $H(x, y, r)$ indicates a potential circle in the image. This transform is effective for detecting the circular lights in traffic signs, regardless of minor variations in scale or position.
Color Thresholding (HSV space): Color thresholding in the HSV color space is performed to filter regions based on hue ($H$), saturation ($S$), and value ($V$). The red sign is identified by the condition $H \in[ 0,10] \cup[ 160,180]$, with $S>0.5$ and $V>0.5$, ensuring that only vivid and bright red colors are selected. Similarly, the blue sign is detected using the range $H \in[ 100,130]$, again with $S>0.5$ and $V>0.5$. These ranges are selected to improve robustness against shadows and illumination variations. This segmentation isolates colored regions that potentially represent traffic sign for further analysis.
After segmentation, feature extraction is applied to characterize the candidate traffic sign regions for classification.
Hu Moments: Hu moments are shape descriptors derived from image moments and are invariant to translation, rotation, and scaling. They are computed using normalized central moments $\eta_{p q}$, which are defined as:
$ \eta_{p q}=\frac{\mu_{p q}}{\mu_{00}^{1+\frac{p+q}{2}}} $
here, $\mu_{p q}$ are the central moments of the image region. The seven Hu moment invariants $\phi_i$ are then calculated as nonlinear combinations of these normalized moments:
$ \phi_i=f\left(\eta_{p q}\right) $
These features capture the geometric structure of traffic signs and help differentiate circular lights from other objects.
Color Histogram: To capture the color distribution within a segmented region, a normalized color histogram is computed. The image is typically quantized into 64 bins by dividing the HSV color space into levels. The histogram for bin $k$ is given by:
$ H_k=\frac{n_k}{\sum_{j=1}^{64} n_j}, \quad k=1,2, \ldots, 64 $
here, $n_k$ denotes the number of pixels falling into the $k^{\text {th }}$ color bin. Normalization ensures that the feature is invariant to region size and lighting changes. The histogram helps classify the sign by distinguishing between red, green, and blue intensities.
Symmetry Ratio: Most traffic signs exhibit vertical symmetry. To quantify this property, the symmetry ratio $R_s$ is defined as:
$ R_s=\frac{\left|A_{\text {left }}-A_{\text {right }}\right|}{A_{\text {left }}+A_{\text {right }}} $
where, $A_{\text {left }}$ and $A_{\text {right }}$ denote the pixel areas on the left and right halves of the segmented object, respectively. A low symmetry ratio (i.e., $R_s \approx 0$ ) indicates a highly symmetrical object, which is a strong indicator of a traffic sign. This feature helps to eliminate asymmetrical artifacts and improves recognition reliability.
To robustly classify the detected traffic sign candidates, an FIS is utilized. The FIS models human-like reasoning by incorporating uncertainty and imprecision in input data, making it well-suited for real-world visual environments where sensor readings may not be crisp.
Inputs: The FIS takes three inputs: Shape Confidence ($SC$), which measures the degree of similarity between the detected region and an ideal geometric shape (e.g., circle or triangle); Color Match ($CM$), which represents the degree of similarity between the region’s color histogram and the expected color profile of a traffic sign; Symmetry Ratio ($SR$), which measures the vertical symmetry of the region, as traffic signs are typically symmetric in shape.
Fuzzification: Fuzzification maps crisp numerical input values into fuzzy sets using predefined membership functions. For instance, the membership function for the fuzzy set “High” in Shape Confidence is defined as:
$ \mu_{S C}(x)= \begin{cases}1, & x>0.9 \\ \frac{x-0.7}{0.2}, & 0.7 \leq x \leq 0.9 \\ 0, & x<0.7\end{cases} $
This piecewise linear function means that if the shape confidence is greater than 0.9, it is fully considered “high.” If it lies between 0.7 and 0.9, it partially belongs to the “high” category, and below 0.7, it is not considered high at all. Similar fuzzification functions are defined for other inputs like $CM$ and $SR$, with sets such as “low,” “medium,” and “high.”
Rule Base: Once the inputs are fuzzified, fuzzy rules are applied to derive conclusions.
These rules follow the IF–THEN format and are derived from expert knowledge or empirical observations. For example:
$ \text { IF } S C \text { is High AND } C M \text { is High AND } S R \text { is Low THEN Class }=\text { Circular Prohibitory } $
This rule indicates that if the detected shape closely matches a circle (high shape confidence), has strong color similarity (high color match), and is symmetric (low asymmetry), then it is likely a circular prohibitory sign.
Another example is:
$ \text { IF } S C \text { is Medium AND } C M \text { is High THEN Class = Warning Triangle } $
This rule captures cases where the shape does not strongly resemble a circle but has a high color match, possibly indicating a triangular warning sign.
Defuzzification: After evaluating the fuzzy rules, the fuzzy output needs to be converted into a crisp decision. This is achieved through defuzzification. The most commonly used method is the **centroid method**, which calculates the center of gravity of the output fuzzy set. It is given by:
$ \text { Output }=\frac{\int z \mu(z) d z}{\int \mu(z) d z} $
In this equation, $z$ is the output variable (e.g., class score), and $\mu(z)$ is the aggregated membership function derived from the rule evaluations. The result is a crisp value used to classify the traffic sign into specific categories such as “Circular Prohibitory,” “Warning Triangle,” or others, depending on the output scale.
The FIS thus allows for smooth transitions between classes and handles uncertainty in the visual feature values, making it particularly effective in outdoor environments with noise, varying lighting, and occlusions.
4. Results and Discussion
To validate the effectiveness of the proposed TSR model, a series of experiments were carried out using the German Traffic Sign Recognition Benchmark (GTSRB) dataset. This benchmark dataset is widely used in TSR research and contains over 50,000 images of German traffic signs belonging to more than 40 categories. The images vary in terms of lighting, weather conditions, occlusions, and viewing angles, providing a comprehensive and realistic evaluation environment.
The dataset was divided into training and testing subsets, ensuring a balanced distribution of sign classes for model evaluation. The proposed method was implemented in MATLAB R2015a, following a structured pipeline involving preprocessing, segmentation, feature extraction, and classification. Performance metrics such as accuracy and recognition rate were calculated to objectively assess the model’s ability to identify traffic signs under diverse real-world scenarios.
To ensure reproducibility and transparency, the complete FIS rule set used in this study is fully documented, detailing the input membership functions, rule base, and output defuzzification strategy. While the full code implementation is not publicly hosted, it is available for research purposes upon request via email. Additionally, the dataset was divided into training and testing subsets according to standard GTSRB protocols, and all random initializations were controlled using fixed seeds to maintain consistent results across experiments. These measures collectively support replicability while maintaining accessibility for interested researchers.
The proposed TSR model was configured with carefully chosen parameters to ensure optimal performance across all processing stages. In the HSV color space transformation, hue thresholds were set to $[ 0-10] \cup[ 160-180]$ for red, and $[ 100-130]$ for blue, with both saturation and value thresholds greater than 0.5 to ensure vivid and bright color segmentation. These ranges were selected based on established traffic sign color standards in prior studies and verified empirically to remain robust under varying illumination conditions. For edge detection, a Gaussian kernel size of $5 \times 5$ with a standard deviation of $\sigma=1.4$ was used prior to applying the Canny edge detector, with low and high thresholds set at 0.1 and 0.3, respectively, to balance edge sensitivity and noise suppression. These values were tuned experimentally to minimize false edges while retaining sign boundaries across different environmental scenarios. In morphological operations, a disk-shaped structuring element with a radius of 3 pixels was used. Erosion and dilation were iteratively applied two times each to clean and enhance the binary masks. This configuration was tested against different radii and iterations, and the chosen values consistently produced optimal mask refinement without excessive loss of shape information. For shape detection, the Hough Circle Transform was tuned with a radius range of 10–30 pixels and an accumulator threshold of 20, allowing reliable detection of circular lights. These parameters were determined by observing the size distribution of target traffic signs in the dataset and adjusted to avoid false positives. The feature extraction phase computed 7 Hu moments, a 64-bin HSV color histogram, and a symmetry ratio, where a threshold of $R_s<0.2$ was considered indicative of high vertical symmetry. This threshold was empirically chosen after testing different values, where 0.2 provided the best trade-off between capturing sign symmetry and minimizing misclassification.
The proposed sign recognition model follows a systematic pipeline designed to robustly detect and classify traffic signs under varying conditions in Figure 2. The process begins with image acquisition, where input images are captured for analysis. Next, preprocessing techniques are applied, including Gaussian filtering for noise reduction and histogram-based contrast enhancement to improve feature visibility. The model then transforms the image into a suitable color space (e.g., HSV) to facilitate color-based segmentation, enabling the isolation of sign regions based on hue and saturation.

For feature extraction, the model employs edge detection combined with morphological operations (e.g., dilation and erosion) to refine shape boundaries. Geometric shapes are isolated using Hough transform approximations, allowing the detection of primitives such as circles, triangles, and rectangles. These steps ensure accurate localization of potential signs within the image.
In the classification phase, an FIS evaluates two key metrics: shape confidence (measuring geometric fit) and color match score (assessing hue consistency). By combining these scores probabilistically, the model achieves robust sign recognition even in challenging scenarios (e.g., occlusions or lighting variations). The final output classifies the detected sign into predefined categories (e.g., stop signs, speed limits) with high accuracy.
This modular approach ensures adaptability across diverse environments, making the model suitable for real-world applications such as autonomous driving and traffic management systems. The integration of traditional computer vision techniques with fuzzy logic enhances both precision and interpretability, addressing common challenges in sign recognition.
To validate the performance of the proposed TSR model, we conduct a statistical analysis using standard evaluation metrics: Accuracy, Precision, Recall, and F1- score. These metrics are derived from the confusion matrix elements: True Positives (TP), True Negatives (TN), False Positives (FP), and False Negatives (FN). The formulations are provided as follows:
$\bullet$ Accuracy: Measures the overall correctness of the model’s predictions.
$ \text { Accuracy }=\frac{T P+T N}{T P+T N+F P+F N} $
$\bullet$ Precision: Measures the proportion of correctly predicted positive observations to the total predicted positives.
$ \text { Precision }=\frac{T P}{T P+F P} $
$\bullet$ Recall (Sensitivity): Measures the proportion of correctly predicted positive observations
to all actual positives.
$ \text { Recall }=\frac{T P}{T P+F N} $
$\bullet$ F1-Score: Harmonic mean of Precision and Recall, offering a balance between the two.
$ \text { F1-Score }=2 \times \frac{\text { Precision } \times \text { Recall }}{\text { Precision }+ \text { Recall }} $
Using the above metrics, we evaluate the model’s ability to correctly recognize and classify traffic signs under various conditions. High values across all metrics (as shown in Table 2) confirm the reliability, generalizability, and robustness of the proposed recognition system. These quantitative results validate that the model effectively minimizes false classifications while maintaining high sensitivity to true signs, making it suitable for deployment in real-time ITS.
Condition | Accuracy | Precision | Recall | F1-Score | FPS | Latency |
Normal | 98.7 | 98.5 | 98.6 | 98.55 | 32 | 31.2 |
Poor Lighting | 96.2 | 95.8 | 96.0 | 95.9 | 30 | 33.5 |
Partial Occlusion | 95.5 | 95.1 | 95.3 | 95.2 | 29 | 34.7 |
Rotated Signs | 96.9 | 96.5 | 96.7 | 96.6 | 31 | 32.1 |
Blurred Images | 94.7 | 94.2 | 94.5 | 94.3 | 28 | 36.0 |
Average | 96.4 | 96.0 | 96.2 | 96.1 | 30 | 33.5 |
Table 2 presents the performance of the proposed TSR model under varying real-world conditions. To evaluate robustness under diverse conditions, additional test subsets were constructed from the GTSRB dataset. Specifically, poor lighting scenarios were simulated by reducing brightness and contrast levels in MATLAB, while partial occlusion was introduced by masking random portions of the traffic sign area with rectangular patches. Rotated signs were generated by applying rotation transformations within the range of $\left[-25^{\circ},+25^{\circ}\right]$, and blurred images were obtained using a Gaussian blur with $\sigma=2.0$. These subsets were created to systematically mimic real-world challenges while maintaining consistency with the original dataset distribution. Approximately 20% of the total test images were assigned to each condition, ensuring balanced evaluation across scenarios. The model demonstrates exceptional accuracy across all scenarios, achieving a peak accuracy of 98.7% in well-lit, clear conditions. Even under challenging environments such as poor lighting and partial occlusion, the model maintains high accuracy levels of 96.2% and 95.5%, respectively, indicating strong generalization and robustness.
Precision, recall, and F1-score metrics also remain consistently high, with average values of 96.0%, 96.2%, and 96.1%, respectively. Notably, the model performs reliably even with rotated signs (96.9% accuracy) and blurred inputs (94.7% accuracy), reflecting its effectiveness in real-world, unpredictable traffic conditions.
The inclusion of real-time performance metrics further validates the efficiency of the proposed model. As shown in Table 2, the system achieves an average of 30 frames per second (FPS), which aligns with the typical requirement for real-time vision systems in intelligent transportation. Latency per frame is consistently maintained around 33.5 ms, even under challenging conditions such as poor lighting, occlusion, and motion blur. These results indicate that the model can process traffic signs at video frame rates without perceptible delays, ensuring smooth integration into real-world autonomous driving and traffic monitoring applications.
In summary, the proposed TSR model demonstrates high accuracy and robustness across various road conditions and sign types, effectively handling illumination changes, occlusions, and background complexities. The consistent performance observed in all evaluated scenarios confirms the reliability of the model in real-world applications. The use of integrated feature extraction, fuzzy logic inference, and optimized classification enhances both precision and computational efficiency. These results collectively validate the model’s suitability for deployment in ITS and real-time driver assistance frameworks, marking a significant advancement in the domain of TSR.
5. Conclusion
This paper presented a hybrid real-time TSR framework that integrates mathematical morphology, color and shape-based modeling, and fuzzy logic inference. The framework can effectively process road scene images to detect and recognize traffic signs by first isolating candidate regions through morphological operations, followed by verification using shape and color features. The FIS enhanced the robustness of classification by handling uncertainties such as partial occlusion, poor lighting, or faded signs.
The proposed model demonstrated reliable performance in diverse road conditions, maintaining real-time responsiveness and low computational complexity. Its rule-based structure ensured interpretability, which is vital for applications in driver assistance systems and autonomous vehicles. Furthermore, the integration of fuzzy logic added flexibility to decision-making, making the system more tolerant of variations in input data.
However, the system has some limitations. The manually defined fuzzy rule base, while interpretable, lacks adaptability to dynamic or unseen environments. In addition, although the framework is real-time on standard systems, its performance on low-power embedded platforms remains to be fully optimized.
Future work will address these limitations by incorporating adaptive fuzzy rule learning techniques such as reinforcement learning or data-driven rule generation to improve model adaptability. Additionally, further research will explore lightweight hardware-friendly implementations, such as deploying the model on FPGA or Jetson platforms. The system will also be extended to support multilingual and region-specific signs to improve its usability in diverse geographic contexts.
The data used to support the findings of this study are available from the corresponding author upon request.
The authors declare that there are no conflicts of interest.
