Abstract
Regulating traffic in cities plays a crucial role in addressing climate change. The rapid advancement of artificial intelligence technologies offers new opportunities for managing urban traffic within the framework of Smart Cities. One common method for classifying traffic is the level of service (LoS) criterion, which evaluates traffic quality based on factors like density, speed, and location-specific characteristics. Because of these variations, LoS must be assessed individually for each location, often with expert assistance. In this article, we propose and compare several approaches for LoS classification using neural networks, fuzzy sets, and high-dimensional random vectors. Our goal is to reduce the reliance of LoS determination on local conditions, making the methodology adaptable across different locations. The results show that all three used methods achieved sufficient accuracy, which supports their potential integration with meteorological and pollution data for further applications.
1 Introduction
Transport is an integral part of contemporary civilization. Although efforts are being made to mitigate its environmental impacts, it is not negligible. This is particularly the case in densely built-up and populated areas, where transport emissions can significantly impact the quality of the environment and, consequently, the health of urban residents.
Efforts to introduce smart technologies in urban planning can help to reduce the production and concentration of pollution in cities. To do this, data collection and processing technologies must be deployed and integrated within a smart city framework.
Beneš et al. [1] proposed a situation model of the transport, transport emissions, and meteorological conditions, based on knowledge graphs [2] in the context of smart cities. The main objective is to integrate data from different providers such as Technical Manager of Roads, Operator ICT, Czech Hydrometeorological Institute, and Institute of Planning and Development to provide real-time urban actions, such as traffic management based on the weather forecasts, aiming for efficient city planning and improved environmental conditions.
According to Beneš and Svítek [3], it is essential to integrate and categorize data from multiple sources. Such a system for automated data categorization was described in the article [1] where the Takagi-Sugeno fuzzy inference system (FIS) was used to model a level of service degree. The authors propose to take into account data on:
wind direction and wind speed,
categorization of meteorological conditions such as temperature and precipitation,
level of service of traffic flow.
In this article, we focus on comparing the effectiveness of automated categorization of the level of service using different approaches such as neural networks, fuzzy sets, and high-dimensional random vectors. The level of service was chosen to test the approaches because of the least number of input parameters. After testing, the model can be extended with additional input parameters concerning weather and emissions.
2 Level of service categorization
In this section, we focus on description of the level of service, as well as the location for data collection.
2.1 Level of service
Level of service (LoS) in transport is a criterion that can be used to quantitatively describe the extent of traffic conditions within a traffic stream. This criterion takes into account measures such as speed, travel time, maneuverability, traffic interruption, comfort, and convenience.
Since the data used in this thesis were collected in the city of Prague, we will follow the Czech standards [6], which categorize traffic flow data into six categories of LoS. The definitions of each category are described in words as follows:
LoS 1 – The traffic flow is free.
LoS 2 – The traffic flow is almost continuous.
LoS 3 – The traffic situation is stable.
LoS 4 – The traffic situation is still stable.
LoS 5 – The lane capacity is full.
LoS 6 – The section is congested.
As each location may differ in conditions, such as the number of lanes in a given direction, the maximum speed limit, and traffic light control, a traffic expert must categorize each location individually. We aim to develop a system that automatically categorizes locations based on their types.
2.2 Location
For testing purposes, traffic data (speed of traffic flow and traffic flow) on Legerova Street in Prague collected during the year 2020 were used. It is a three-lane street as shown in Figure 1, with the position of the traffic sensors marked.

Map of the Legerova Street.
In Figure 2, the data are categorized based on the Greenshield fundamental model into six categories according to the Czech standards. Based on the experts’ evaluation, the areas, shown by colored rectangles, which represent the individual LoS are determined. Individual points in Figure 2 represent real measured data.

The LoS categories for Legerova Street in Prague according to traffic and speed of vehicles.
2.3 Data preparation
We checked the traffic data for possible errors. The collected traffic data includes information on time, cumulative number of vehicles passed, and average speed of vehicles per 5 minutes. All values are stored separately for three vehicle categories (cars, trucks, unrecognized). We have prepared hourly aggregated data regardless of vehicle category. We excluded incomplete records in which some data were missing and visualized the data to identify any anomalous values.
3 Methods used
The area of machine learning has experienced a major boom in recent years and is finding applications in many areas. Since the majority of available data is currently unlabeled, it is challenging to use supervised learning-based methods.
One solution is to use a classical unsupervised learning-based clustering method like
Better results can be achieved using supervised machine learning methods that incorporate information from an expert in the learning process. These methods require labeled data. At the same time, however, it should be noted that many of the algorithms used are often computationally intensive and have high hardware performance requirements.
In the next sections, we introduce each of the three compared methods used for automated LoS categorization.
3.1 Deep neural networks (DNNs)
DNNs are used today in many scientific fields as well as in a number of applications used on a daily basis, such as voice and image recognition, assistive services, or classification.
A DNN by default consists of more than three layers including an input layer, an output layer, and several hidden layers. Each layer consists of interconnected nodes. In DNNs, each layer of nodes trains on a different set of features based on the previous layer’s output. So, the more you go into the net, the more you can advance into it, and the more complex features the nodes can recognize. These nets allow to identify complex nonlinear relationships. The major advantage that DNNs hold is the ability to deal with unstructured and unlabeled data, which is most of the world’s data [8].
Numerous use cases can be found ranging from classification, making predictions, signal processing to examples of using generative artificial intelligence (AI) models. Michalíková and Pažický [9] used AI to determine the position of a tire in the processed image for traceological purposes. In the study of Vagač et al. [10], DNN was used to extract features based on which it was possible to look for similarity between pairs of images. Povinský et al. [11] deal with the use of DNN to create a chatbot compatible with TJBot. Some papers deal with explainability [12] and error detection in neural networks [13]. Another application of DNN for prediction of molecular docking score can be found in the previous studies [14,15].
Usually, it requires significant computational resources to train a DNN model, which means access to high-performance resources including GPGPUs. Another challenge is the acquisition of data with sufficient quantity and quality for DNN training process. The drawback of the DNN is the lack of interpretability, especially those DNNs with many layers used. This can make it difficult to understand how the model is making predictions and to identify any errors or biases in the model. DNN’s strength is in the automatic feature training process, which can find even nonlinear relationships in data, as well as that they can process both structured and unstructured data of different types. With sufficient data and proper optimization, DNNs can achieve very high accuracy and robustness in predictions.
3.2 Fuzzy inference systems (FISs)
Fuzzy sets were proposed in 1965 [16]. They were designed to deal with uncertainty and to create rules in human language. Therefore, their results are easy to explain. FIS derived from fuzzy sets can capture expert knowledge and represent it in the form of IF-THEN rules. Let us give an example of such a rule for our problem:
“IF Traffic Flow is Very_Low AND Speed is High, THEN Level_Of_Service is equal to 1.”
FISs use simple operations, such as max, min, weighted average, to obtain results from the IF-THEN rules. Several applications of FISs can be found, for instance in the study by Mardani et al. [17], where the outputs of more adaptive network-based fuzzy inference systems (ANFISs) were combined and used to predict and analyze the mutual relationship between the selected variables, which affect carbon dioxide emissions. Tashayo and Alimohammadi [18] used a Hierarchical FIS to model urban air pollution.
Intuitionistic fuzzy sets, which represent an extension of fuzzy sets, are also suitable for modeling various situations and problems. Poryazov et al. [19] used two approaches to the intuitionistic fuzzy estimation of the uncertainty of compositions of traffic quality services. Intuitionistic fuzzy sets are also a suitable tool for classification, as in the study by Michalíková and Dudáš [20], where the authors present the design, implementation, and experimental evaluation of parallel algorithms for classifying tire tread pattern images based on intuitionistic fuzzy set methods that incorporate similarity measure, distance function, and selected intuitionistic fuzzy negations. In the study by Michalíková et al. [21], a fuzzy interference system of the Takagi-Sugeno type is used for the identification of wood-decaying basidiomycetous macrofungi.
In general, there are two basic approaches to creating FISs: to obtain rules analyzing large amounts of data or to use the knowledge of experts who describe in human language terms certain properties of the input variables. Often a combination of these two approaches is used. A drawback of the FISs is considered to be the difficulty of parameter tuning, which has a large impact on the result. This problem arises especially in the case of small amounts of training data, where parameter optimization using known data may not be sufficiently accurate. However, with precise information from an expert, high accuracy can be achieved even without a large amount of known data. The main advantage of the FIS is generation of human-readable and implementable rules.
3.3 High-dimensional computing (HDC)
Commonly used algorithms in machine learning are often computationally intensive and thus require significant hardware performance. On the other hand, solutions that could be applied on less powerful hardware are desirable, allowing them to be executed at the edge, e.g., embedded systems or IoT devices. High-dimensional computing (HDC) is emerging as a promising solution to this situation. The basic idea behind HDC is brain-inspired information storage and processing and is based on a very simple neural architecture [22].
The principle of HDC is to project input data onto vectors in high-dimensional space. For this, a uniform static mapping to randomly generated high-dimensional vectors (HVs) containing numbers is usually used. The dimension of the HV is usually larger than 10,000, which provides sufficient space for information representations. In addition, HVs are usually initialized randomly, so they can be assumed to be orthogonal to each other. Various embeddings can be used to represent HVs, for instance, from bipolar vectors (values
Consequently, we can do three operations with the data embedded as HVs: addition, multiplication, and permutation of elements, each of which has a specific physical meaning.
Addition – takes two HVs as operands and performs element-wise addition on the numbers at the same positions. It is usually used to aggregate the information of two HVs from the same modality and create a superposition of them.
Multiplication – takes two HVs as operands and performs element-wise multiplication on the numbers at the same positions. It is usually used to combine information from different modalities and create new information of another modality based on these two.
Permutation – takes only one HV as an operand and performs cyclic rotation of elements over the HV. It is usually used to reflect temporal or spatial pattern of information.
Two types of HDC memory are used: item memory and associative memory. Item memory stores information about input data. Associative memory is used to represent the output model. Measuring the similarity measure of two HVs, i.e., the similarity of two pieces of information, is mostly done using Hamming distance or cosine similarity metrics. The entire setup is elementary and lends itself to fast, low-power hardware realization [24].
Several applications of HDC can be found. For example, determining the belonging of a text to a particular language based on clustering [25], speach recognition [26], human activity recognition [27,28], and anomaly detection [29]. Ni et al. [30] deal with uncertainty quantification of a predictive regression model. Many works showed that HDC is an energy-efficient, low-latency, and noise-resilient alternative to conventional machine learning tools.
The major challenge of HDC is the encoding process, which is application-specific. The appropriate encoding process has to be manually designed and evaluated. In HDC, only the associative memory is trainable, but the elements inside each class HV is not individually trainable due to their holographic representation of HVs [23]. Since the information is represented in high-dimensional vectors, the process can require significant memory, particularly when handling large data sets. Like DNNs, HDCs can also be difficult to interpret. The main advantage of the HDC is its speed and better energy efficiency, smaller model size, and acceleration on heterogeneous platforms. HDC can handle incomplete or noisy data, which is beneficial when data quality is low.
4 Implementation and results
We use the same traffic input data for all implementations of the tested methods. The data contains 3,825 records of hourly aggregated information on average speed and number of vehicles passed (traffic flow). Each record has been assigned a LoS value as described above.
In the first step, we can focus on comparison of the LoS assigned by the expert and the LoS classification based on

Classification of points under consideration using
We can see that the
For this reason, in the following, we will only consider methods that can categorize the data with respect to its nature. The entire input dataset is divided into training and test data, as described for each method in the following sections, if training process is involved. Finally, each method is evaluated in terms of the number of correctly categorized points on the entire input dataset. We focused on number of correctly assigned LoS values to input data and elapsed time required to the LoS value assignment.
4.1 DNNs
The first approach used is based on DNNs, which are now commonly used for a number of classification tasks. Our DNN implementation is based on the Python libraries Tensorflow and Keras together with the Sklearn and Numpy libraries. Before the actual training of the neural network, all the input data are proportionally split into training, testing, and validation data in the ratio of 75, 15, and 10%, respectively. Figure 2 shows that number of points belonging to each LoS category is not evenly distributed across all categories, e.g., LoS 6 contains a small number of points. Therefore, we aim to have all LoS values reasonably represented in each data set after splitting.
The network topology is set to 8 dense layers, the number of neurons in individual layers is 20, 60, 80, 120, 100, 80, 60, and 6 in the last output layer. Each layer except the last one uses the ReLu activation function, and the dropout rate is set to 0.002. The initial weights in the neural networks were set according to the standard He uniform Keras function, with default settings.
The use of He initialization in combination with the ReLU activation function can significantly reduce the problem of vanishing and/or exploding gradients at the beginning of training. The batch normalization proposed by Ioffe and Szegedy [31] can be employed to address this problem during training. It also leads to a lower sensitivity to the initialization of the DNN weights and allows the use of larger learning rates, which significantly speeds up the learning process. Adding normalization before each hidden DNN layer makes each epoch take much longer, but this is balanced by faster convergence and fewer epochs needed to achieve the same performance.
The model was trained for a fixed number of 200 epochs total using the stochastic gradient descent optimizer, and a learning rate of 0.001 and momentum of 0.9. Other parameters were kept at their default values. The general idea of methods based on gradient descent approach is to tweak parameters iteratively to minimize a cost function. It measures the local gradient of the error function with regard to the parameter vector, and it goes in the direction of descending gradient until the gradient is equal to zero. The learning rate parameter determines the size of step in each iteration. If it is too big, optimizer may jump over the minimum. If it is too small, optimization takes too much time to find minimum.
Another problem of gradient descent method is convergence to local minimum. The stochastic gradient descent picks a random instance in the training set at every step and computes the gradients based only on that single instance. Working on a single instance at a time makes the algorithm much faster because it has very little data to manipulate at every iteration. It also makes it possible to train on huge training sets, since only one instance needs to be in memory at each iteration. This approach can help to jump out of local minima, but it also cause that it never settle at the minimum due to random selection of instance in each iteration. We can observe this behavior in Figure 4. In the later stage of the optimization, the loss as well as accuracy functions fluctuate.

Loss function and accuracy of DNN model.
A validation set was used to mitigate the trained model’s possible overfitting and ensure the best model selection. Figure 4 depicts evolution of the loss function and accuracy during the training process.
After training, we use the trained model to assign an output LoS value to each input record (speed and traffic flow). The model outputs six probability values, one for each LoS category. To compare the results, the category with the highest probability value is used for each input pair and compared to the real output provided by the expert. We obtained 53 incorrectly determined values, which represents a 98.61% accuracy of input evaluation.
We can look at the obtained results (Figure 5). Incorrectly classified data pairs are marked in squares. The color of the circle represents the correct LoS value and the color of the square represents the incorrectly assigned LoS value. Most of the misclassified data pairs are located at the boundaries between LoS categories. This is evident especially in areas with low density of training data points. Better results can be achieved mainly by adding more training data in this areas.

Classification of points under consideration using DNN. Points marked in squares represent misclassified values.
4.2 FISs
To model the data we used also fuzzy approach. There are several types of FISs. We chose the Takagi-Sugeno FIS [32] because this system was designed to approximate data that can be described by linear functions on certain parts of the domain and that need to be approximated on the remaining parts of the domain to obtain a continuous function. When we plot our data into the 3D graph (Figure 6), we see that the presented FIS is perfectly suited to solving the mentioned problem. Takagi-Sugeno FIS consists of fuzzy IF-THEN rules. The inputs and outputs of these rules are specific. Input variables are represented by fuzzy set membership functions. Several types of fuzzy set membership functions have been defined [33]. In this research, we used the so-called trapezoidal and Gaussian membership functions. Output variables are described by polynomial functions (most often constant or linear functions).

3D graph of LoS on Legerova street in Prague.
In the first step, the rule base consists of fuzzy IF-THEN rules need to be designed. Subsequently, the data can be classified. There are two ways to design a rule base. If we have enough amount of data, we can obtain rules by analyzing them. If the solved problem is well described by expert knowledge, we can translate this knowledge directly into fuzzy IF-THEN rules. In our research, we used both approaches. We process these rules in the MATLAB software, in the Fuzzy Logic Designer toolbox.
First, we analyze data automatically using specific methods. One of them, called Grid partition method, was designed such way that it divide the domain of each input variable into the required number of equal parts. Let the required number of domain division be
To optimize parameters, we chose the ANFIS method [34], which is based on a combination of neural networks and fuzzy sets. Therefore, we divided the data into training, testing, and validation sets with the ratio of 70, 15 and 15%, and we used ANFIS method to achieve the highest possible classification accuracy.
We used different number of domain divisions, from 4 to 7. The reason for choosing these values was that we wanted to obtain a small number of rules. We also used different sets of training, testing, and validation data, which were randomly generated. The best results were obtain when we divided each domain into the six parts (i.e., 36 rules were created), and each variable was described by trapezoidal membership functions. From 3,825 input data pairs, 3,789 pairs were classified correctly, that represent 99.06% accuracy of classification. We were not satisfied with obtained results, and therefore, we used also another methods.
The second used fuzzy method, subtractive clustering method, also analyze data automatically. This approach is based on a specific clustering method, where the points with the largest number of neighbors are chosen as cluster centers [35]. Then the rules are created in such a way that the fuzzy membership functions of input variables are computed from the values of cluster centers. On the other hand, the output functions are computed as approximations of the output values of cluster centers and their neighbors. Depending on the parameters set by the user, the method determines the number of cluster centers. By changing the method parameters, it is possible to achieve a smaller/larger number of cluster centers and thus also the smaller/larger number of rules, which leads to possibility of optimization of the number of rules in the system. In the next step, the parameters of input and output variables need to be optimized. To optimize parameters, we used the ANFIS method.
To achieve different number of cluster centers (and also different number of rules), we change the value of parameter range of influence of the cluster center. This parameter can reach any value from the unit interval. When we set up the parameter to value 0.25, we obtained six rules, which seems appropriate given the six classes we classify into. Using this FIS only 3,746 pairs from 3,825 considered pairs were classified correctly. The best results we obtained when we set up this parameter to value 0.15. The system created 13 rules and correctly classified 3,753 pairs. These results represent the lowest obtained accuracy of classification, just 98.11%.
Because the investigated problem can be well described in human language, we decided to design the rules also in the form of expert knowledge. We used so-called trapezoidal membership functions as input functions and constant functions as output functions. The input Traffic Flow was described by six membership functions Very_Low – Low – Middle – High – Very_High – Extremely_High Flow. The input Speed was described by five membership functions Very_Low – Low – Middle – High – Very_High Speed. As an example, trapezoidal membership functions of fuzzy values for the variable Speed are displayed in Figure 7.

Example of membership functions of input variable Speed.
Constant functions were used as output variables. We decided to consider as a domain of output variable the interval [0, 6] and as the constant output values, we considered values one, two, …, and six. Considering the solved problem, we created 27 fuzzy IF-THEN rules. The output of Takagi-Sugeno FIS was the function that represented the approximation of the considered mutual relations (Figure 8).

3D graph of LoS obtained using Takagi-Sugeno FIS.
Figure 8 shows that for some input values, the approximation function reached a value equal to zero (gray parts of the function graph). Since the domain of the considered function is a rectangle, some areas are not used as input values in the description of the LoS. These data values represent anomalies, and using of these settings for the approximation function, we could detect them very quickly.
By using the function created by the proposed FIS, we can assign the output LoS value to each input pair of speed and traffic flow values. As an output of the approximation function, we mostly obtain a natural number, while at the boundaries of two levels of LoS, we usually obtain the decimal number. This is also one of the advantages of the proposed system that will be used in the future. The decimal output value indicates that the considered input pair represents the output that does not strictly belong to one type of LoS.
When we want to compare the obtained results (the obtained values of LoS) first we have to round the obtained decimal numbers to natural numbers. In our research, we have 3,825 input data pairs. After we processed them using an approximation function, we obtained only 28 incorrectly determined values of LoS. This represents a 99.27% accuracy of classification.
It is very interesting to look at the obtained results (Figure 9). Most of the incorrectly classified data pairs are located on the line, which represents the speed value equal to 30. When the values of traffic flow are between 1,750 and 3,500, then this line represent the border between LoS 2 and LoS 4. From our results, it follows that the incorrectly classified points are classified into the LoS 3. If we look at the output FIS function (Figure 8), we can see that the function is continuous in the given area and creates a junction between output value 2 and output value 4 (which, of course, gives an output value equal to 3). Similar situation is when the values of traffic flow are between 0 and 1,750. Then the line, which represents the speed value equal to 30, represents the border between LoS 2 and LoS 5. In this area, the incorrectly classified points are classified into the LoS 4 or LoS 3. It follows from the same reason as in the previous case.

Classification of points under consideration using FIS. Points marked in squares represent misclassified values.
Despite achieving satisfactory classification accuracy, we decided to focus on improving the classification of points in the aforementioned area. It is necessary to change the function curve around the input speed value 30 so, that the resulting curve is steeper. We change the parameters of membership functions low and middle of input variable speed (Figure 7) as close to value 30, as is possible (Figure 10).

Improved membership functions of input variable speed.
With this adjustment, we achieved an improvement in the number of correctly classified points. We obtain only five incorrectly determined values of LoS. These results represent the 99.87% accuracy of classification. Incorrectly classified points are displayed in Figure 11.

Improved classification of points under consideration using FIS. Points marked in squares represent misclassified values.
4.3 HDC
The HDC implementation is based on the Python libraries Torch and Torchhd together with the Sklearn and Numpy libraries. Similarly to the DNN, all the input data are column-wise normalized and proportionally split into training and testing data in the ratio of 80 and 20%, respectively.
Embedding to high-dimensional random vector (or HV) for input data is generated according to the intRVFL model [36] by embeddings.Density routine from the Torchhd library. Each row of input data is encoded into dense bipolar MAPTensor HV [37] with a dimension of 10,000 elements from
Since this is a classification task, six models need to be created and trained, one for each LoS value. The new model is usually represented as an HV with the same dimension as the embeddings HVs and is initialized with 0 value. Training each of the six models involves element-wise addition of the HV encoding the input data to the corresponding model HV. It means that all HVs from one LoS category are aggregated to one model HV.
The model is then tested to determine its accuracy. Each row of test data is assigned a LoS value based on the maximum cosine similarity of input HV with the six trained model HVs. The higher similarity between input data HV and model HV is, the higher is probability it belongs to the LoS category represented by the model HV. These models achieve classification accuracy of 76.34%.
To achieve higher classification accuracy using HDC, the generated models can be iteratively re-trained in several epochs. The re-training process consists of going through all the input data again, and in the case of an incorrect assignment of a LoS value to a given input line, its HV is subtracted from the model to which it was incorrectly assigned and added to the model to which it should belong according to equation (1). This procedure leads to decrease the similarity of the input HV to the incorrect model assignment and, in turn, increase its similarity to the correct model. The model accuracy evolution within iterations is shown in Figure 12.

Accuracy of HDC model.
The highest HDC classification accuracy was achieved after 99 epochs, with 8 out of 3,825 values being incorrectly determined, representing 99.79% accuracy of input evaluation. Incorrectly classified points are displayed in Figure 13.

Classification of points under consideration using HDC. Points marked in squares represent misclassified values.
We can observe that the majority of misclassified points lie on the border between two LoS categories, especially in areas with low density of points form correct LoS category. Due to this fact evaluated input point is in hyperspace more similar to the closest point from incorrect neighboring LoS category.
Further improvement can be done on the HDC approach with respect to faster convergence of models training. We will add variable
This modification speeds up the HDC training process. In Figure 14, we can see progress in accuracy during the training process with

Accuracy of HDC model after modification.
In this case, convergence to the highest models accuracy occurred after 62 epochs, with only 4 out of 3,825 values being incorrectly determined, which represents 99.95% accuracy of input evaluation. Incorrectly classified points are displayed in Figure 15.

Classification of points under consideration using modified HDC. Points marked in squares represent misclassified values.
5 Discussion
In this article, we focused on the evaluation of three methods used to classify the input data into six LoS categories. All three methods provide a high level of accuracy, but each has a different computational complexity. Since in our work, we consider only a small number of input parameters (traffic flow and speed) to determine LoS. It is possible to design rules using FIS without the need for further optimization, in case of quality expert knowledge availability. This provides some advantage over DNN and HDC-based methods, which need to first train models on the training dataset. Standard datasets used with DNN in common consist of at least 50,000 entries.
Another aspect to consider is the training process itself. For the DNN topology used, 41,850 parameters had to be trained over 200 epochs, which is quite a computationally intensive problem. In contrast, with HDC, it was necessary to iteratively refine six models for each LoS, which consists of the addition or subtraction of HVs with 10,000 elements. Both DNN and HDC implementations can take advantage of parallel processing in the training process.
A final aspect to consider in the evaluation is the computational complexity of using the trained model to classify the new input data. Classification using FIS requires evaluation of continuous nonpolynomial function for each input point. In the case of a DNN, the computation is more complicated because it goes through the entire neural network structure, including normalization and checking near-zero weights for dropout. Finally, the HDC implementation uses the cosine similarity measure of the two HVs for classification, computing this value 6 times, once with each LoS category model. Table 1 presents real computation run time required by each of three methods to classify all 3,825 input records. Measurements were done on PC with Intel i7-7500U CPU at 2.70 GHz frequency with 8 GB of RAM.
Comparison of classification run times
| Method | DNN | FIS | HDC |
|---|---|---|---|
| Time (s) | 0.23 | 0.07 | 0.91 |
According to results of measurement, we can conclude that in our study, the method based on the FIS is faster than other two methods based on DNN and high-dimensional random vectors. This is because obtaining a result using FIS only requires the evaluation of one function for each input record. Obtaining the result using DNN is primarily dependent on the complexity of the neural network, while the main rate-determining factor for HDC is the dimension of the HVs. The larger the dimension we choose, the more accuracy we can obtain; however, we increase the computation time.
6 Conclusion
The results in this article point to different possibilities for classifying traffic data. Furthermore, these data can be combined with other data on meteorological conditions as well as pollution data, on the basis of which recommendations for traffic management in the smart city concept can be proposed. We compared the three chosen approaches (DNN, FIS, HDC) in terms of their classification accuracy and also pointed out the different computational demands of the compared approaches.
Methods based on FIS use IF-THEN rules. According to these rules and using fuzzy sets, it is possible to determine the resulting LoS value, but also to identify borderline cases where a given point may partially belong to more than one category. This approach provides the possibility to understand the relationships between the different input values and the results and thus contributes to the explainability of the model.
To effectively manage traffic in the city based on the acquired sensor data, it is necessary to process the data in real-time. For this, it is often necessary to process the data close to the place of its origin within the concept of edge computing and not centrally using powerful computing resources. Thus, these methods must provide accurate results while keeping computational complexity low. As we have shown, all three compared methods are fast enough and can be used for real-time classification, assuming that we have a trained model for DNN and HDC.
The proposed solution can be extended in the future with additional input parameters from other sensors to provide more sophisticated solutions and predictions based on analyzed data. The LoS criterion is directly related to the number of vehicles traveled at a certain time in the monitored location. Given this, a change in the amount of emissions produced at a given location, whether from exhaust fumes or dust particles from brakes and so on, can be expected. Increased pollution levels have a significant impact on the quality of the environment. Once the traffic data and meteorological data are integrated, vulnerabilities can be identified, on the basis of which traffic can be managed, thereby reducing the risk of emission limits being exceeded in the locality and/or warnings can be issued to vulnerable populations.
All three methods provide high accuracy results for the studied locality. In future work, these models can be tested and further tuned with data from different localities. Integration of traffic data with meteorological and pollution data in the sense of mentioned knowledge graphs is also desired.
Acknowledgments
We would like to express our gratitude to Viktor Beneš for the preparation and preprocessing of the data used in this paper.
-
Funding information: Part of the research results was obtained using the computational resources procured in the national project National competence center for high performance computing (project code: 311070AKF2) funded by European Regional Development Fund, EU Structural Funds Informatization of society, and Operational Program Integrated Infrastructure. This work was co-funded by the Slovak Grant Agency VEGA (contract No. 1/0779/25) and the High Performance Computing Center of the Matej Bel University in Banská Bystrica using the supercomputing infrastructure acquired in project ITMS 26230120002 and 26210120002 (Slovak infrastructure for high-performance computing) and supported by the Research & Development Operational Programme funded by the ERDF.
-
Author contributions: All authors have accepted responsibility for the entire content of this manuscript and approved its submission.
-
Conflict of interest: The authors declare that there is no conflict of interest with respect to the publication of this article.
-
Data availability statement: The datasets mentioned in this article are not readily available due to privacy restrictions of the Technical Manager of Roads and Operator ICT.
References
[1] V. Beneš, M. Svítek, A. Michalíková, and M. Melicherčík, “Situation model of the transport, transport emissions and meteorological conditions,” Neural Network World, vol. 34, no. 1, pp. 7–36, 2024. 10.14311/NNW.2024.34.002Search in Google Scholar
[2] V. Beneš, and M. Svítek, “Knowledge graphs for Smart Cities,” Smart City Symposium Prague, Czech Republic, Prague, 2022. 10.1109/SCSP54748.2022.9792541Search in Google Scholar
[3] V. Beneš, and M. Svítek, “Knowledge graphs for transport emissions concerning meteorological conditions,” Smart City Symposium Prague, Czech Republic, Prague, 2023. 10.1109/SCSP58044.2023.10146219Search in Google Scholar
[4] V. Beneš, M. Svítek, A. Michalíková, and M. Melicherčík, “Investigating the impact of meteorological and traffic flow conditions on emissions,” in 2024 IEEE 17th International Scientific Conference on Informatics, IEEE, 2024, pp. 33–38. 10.1109/Informatics62280.2024.10900910Search in Google Scholar
[5] M. Antl, E. Pietriková, and P. Poór, “Intelligent Urban traffic management systems - Takeways from Martin City,” 2024 IEEE 17th International Scientific Conference on Informatics, IEEE, 2024, pp. 21–26. 10.1109/Informatics62280.2024.10900814Search in Google Scholar
[6] S. Prokeš, Projektování místních komunikací: kometář k ČSN 73 6110: komentované příklady řešení. Stavebnictví (komunikace, silnice), Český normalizační institut, Praha, 2007, 134 p. ISBN 978-80-7283-216-3. Search in Google Scholar
[7] S. Lloyd, “Least squares quantization in PCM,” IEEE Trans. Inform. Theory, vol. 28, no. 2, pp. 129–137, 1982. 10.1109/TIT.1982.1056489Search in Google Scholar
[8] R. Mohanasundaram, A. S. Malhotra, R. Arun, and P. S. Periasamy, “Deep learning and semi-supervised and transfer learning algorithms for medical imaging,” in Deep learning and parallel computing environment for bioengineering systems, Academic Press - Elsevier, Cambridge, Massachusetts, 2019. pp. 139–151. 10.1016/B978-0-12-816718-2.00015-4Search in Google Scholar
[9] A. Michalíková, and B. Pažický, “Classification of tire tread images by using neural networks,” 2019 IEEE 15th International Scientific Conference on Informatics, IEEE, 2019, pp. 000107–000112. 10.1109/Informatics47936.2019.9119306Search in Google Scholar
[10] M. Vagač, M. Povinský, and M. Melicherčík, “Detection of shoe sole features using DNN,” 2017 IEEE 14th International Scientific Conference on Informatics, IEEE, 2017. pp. 416–419. 10.1109/INFORMATICS.2017.8327285Search in Google Scholar
[11] M. Povinský, M. Melicherčík, and V. Siládi, “A Chatbot based on deep neural network and public cloud services with TJBot interface,” in 2019 IEEE 15th International Scientific Conference on Informatics, IEEE, 2019. pp. 000101–000106. 10.1109/Informatics47936.2019.9119304Search in Google Scholar
[12] M. Klimo, J. Kopčan, and K. Králik, “Explainability as a method for learning from computers,” IEEE Access, vol. 11, 2023, pp. 35853–35865. 10.1109/ACCESS.2023.3265582Search in Google Scholar
[13] M. Klimo, P. Lukáč, and P. Tarábek, “Deep neural networks classification via binary error-detecting output codes,” Appl. Sci., vol. 11, no. 8, p. 3563, 2021. 10.3390/app11083563Search in Google Scholar
[14] L. Bučinský, B. Bortňák, M. Gall, J. Matuuuška, V. Milata, M. Pitoňák, and D. Zajaček, “Machine learning prediction of 3CLpro SARS-CoV-2 docking scores,” Comput. Biol. Chem., vol. 98, pp. 107656, 2022. 10.1016/j.compbiolchem.2022.107656Search in Google Scholar PubMed PubMed Central
[15] L. Bučinský, M. Gall, J. Matúška, M. Pitoňák, and M. Štekláč, “Advances and critical assessment of machine learning techniques for prediction of docking scores,” Int. J. Quantum. Chem., vol. 123, no. 24, p. e27110, 2023. 10.1002/qua.27110Search in Google Scholar
[16] L. A. Zadeh, “Fuzzy sets,” Inform. Control, vol. 8, no. 3, pp. 338–353, 1965. 10.1016/S0019-9958(65)90241-XSearch in Google Scholar
[17] A. Mardani, Y. Van Fan, M. Nilashi, R. E. Hooker, S. Ozkul, D. Streimikiene, et al., “A two-stage methodology based on ensemble adaptive Neuro-Fuzzy inference system to predict carbon dioxide emissions,” J. Cleaner Production, vol. 231, pp. 446–461, 2019. 10.1016/j.jclepro.2019.05.153Search in Google Scholar
[18] B. Tashayo, and A. Alimohammadi, “Modeling urban air pollution with optimized hierarchical fuzzy inference system,” Environ. Sci. Pollution Res., vol. 23, pp. 19417–19431, 2016. 10.1007/s11356-016-7059-5Search in Google Scholar PubMed
[19] S. Poryazov, V. Andonov, E. Saranova, and K. Atanassov, “Two approaches to the traffic quality intuitionistic fuzzy estimation of service compositions,” Mathematics, vol. 10, no. 23, p. 4439, 2022. 10.3390/math10234439Search in Google Scholar
[20] A. Michalíková, and A. Dudáš, “Use of intuitionistic fuzzy set methods in parallel classification of image data,” in 2022 IEEE 16th International Scientific Conference on Informatics (Informatics), IEEE, 2022, pp. 214–219. 10.1109/Informatics57926.2022.10083463Search in Google Scholar
[21] A. Michalíková, T. Beck, J. Gáper, P. Pristaš, and S. Gáperová, “Can wood-decaying urban macrofungi be identified by using fuzzy interference system? An example in Central European Ganoderma species,” Scientific Reports, vol. 11, no. 1, p. 13222, 2021. 10.1038/s41598-021-92237-5Search in Google Scholar PubMed PubMed Central
[22] P. Kanerva, “Hyperdimensional computing: An introduction to computing in distributed representation with high-dimensional random vectors,” Cognitive Comput., vol. 1, pp. 139–159, 2009. 10.1007/s12559-009-9009-8Search in Google Scholar
[23] D. Ma, C. Hao, and X. Jiao, “Hyperdimensional computing vs. neural networks: Comparing architecture and learning process,” 2024 25th International Symposium on Quality Electronic Design (ISQED), IEEE, 2024, pp. 1–5. 10.1109/ISQED60706.2024.10528698Search in Google Scholar
[24] A. Thomas, S. Dasgupta, and T. Rosing, “A theoretical perspective on hyperdimensional computing,” J. Artif. Intel. Res., vol. 72, pp. 215–249, 2021. 10.1613/jair.1.12664Search in Google Scholar
[25] A. Joshi, J. T. Halseth, and P. Kanerva, “Language geometry using random indexing,” in: Quantum Interaction: 10th International Conference, QI 2016, San Francisco, CA, USA, July 20–22, 2016, Revised Selected Papers 10, Springer International Publishing, 2017, pp. 265–274. 10.1007/978-3-319-52289-0_21Search in Google Scholar
[26] M. Imani, D. Kong, A. Rahimi, and T. Rosing, “Voicehd: Hyperdimensional computing for efficient speech recognition,” in: 2017 IEEE International Conference on Rebooting Computing (ICRC), IEEE, 2017, pp. 1–8. 10.1109/ICRC.2017.8123650Search in Google Scholar
[27] D. Ma, R. Thapa, and X. Jiao, MoleHD: Ultra-low-cost drug discovery using hyperdimensional computing, 2021, arXiv: http://arXiv.org/abs/arXiv:2106.02894. 10.1109/BIBM55620.2022.9995708Search in Google Scholar
[28] Y. Kim, M. Imani, N. Moshiri, and T. Rosing, “Geniehd: Efficient dna pattern matching accelerator using hyperdimensional computing,” in 2020 Design, Automation & Test in Europe Conference & Exhibition (DATE), IEEE, 2020, pp 115–120. 10.23919/DATE48585.2020.9116397Search in Google Scholar
[29] R. Wang, F. Kong, H. Sudler, and X. Jiao, “Brief industry paper: Hdad: Hyperdimensional computing-based anomaly detection for automotive sensor attacks,” in 2021 IEEE 27th Real-Time and Embedded Technology and Applications Symposium (RTAS), IEEE, 2021, pp. 461–464. 10.1109/RTAS52030.2021.00052Search in Google Scholar
[30] Y. Ní, H. Chen, P. Poduval, Z. Zou, P. Mercati, and M. Imani, “Brain-inspired trustworthy hyperdimensional computing with efficient uncertainty quantification,” in 2023 IEEE/ACM International Conference on Computer Aided Design (ICCAD). IEEE, 2023, pp. 01–09. 10.1109/ICCAD57390.2023.10323657Search in Google Scholar
[31] S. Ioffe, and C. Szegedy, “Batch normalization: Accelerating deep network training by reducing internal covariate shift,” in International Conference on Machine Learning, pmlr, 2015, pp. 448–456. Search in Google Scholar
[32] T. Takagi, and M. Sugeno, “Fuzzy identifications of fuzzy systems and its applications to modelling and control,” IEEE Trans. Syst. Man Cybernet., vol. 15, no. 1, pp. 116–132, 1985. 10.1109/TSMC.1985.6313399Search in Google Scholar
[33] M. Jezewski, R. Czabanski, and J. Leski, “Introduction to fuzzy sets,” in Theory and Applications of Ordered Fuzzy Numbers: A Tribute to Professor Witold Kosiński, 2017, pp. 3–22. 10.1007/978-3-319-59614-3_1Search in Google Scholar
[34] J. S. Jang, “ANFIS: adaptive-network-based fuzzy inference system,” IEEE Trans. Syst. Man Cybernet., vol. 23, no. 3, pp. 665–685, 1993. 10.1109/21.256541Search in Google Scholar
[35] S. L. Chiu, “Fuzzy model identification based on cluster estimation,” J. Intel. Fuzzy Syst., vol. 2, no. 3, pp. 267–278, 1994. 10.3233/IFS-1994-2306Search in Google Scholar
[36] D. Kleyko, M. Kheffache, E. P. Frady, U. Wiklund, and E. Osipov, “Density encoding enables resource-efficient randomly connected neural networks,” IEEE Trans. Neural Netw. Learn. Syst., vol 32, no. 8, pp. 3777–3783, 2020. 10.1109/TNNLS.2020.3015971Search in Google Scholar PubMed
[37] R. W. Gayler, Multiplicative binding, representation operators & analogy, (workshop poster), 1998. Search in Google Scholar
© 2025 the author(s), published by De Gruyter
This work is licensed under the Creative Commons Attribution 4.0 International License.
Articles in the same Issue
- Review Article
- Enhancing IoT network security: a literature review of intrusion detection systems and their adaptability to emerging threats
- Research Articles
- Intelligent data collection algorithm research for WSNs
- A novel behavioral health care dataset creation from multiple drug review datasets and drugs prescription using EDA
- Speech emotion recognition using long-term average spectrum
- PLASMA-Privacy-Preserved Lightweight and Secure Multi-level Authentication scheme for IoMT-based smart healthcare
- Basketball action recognition by fusing video recognition techniques with an SSD target detection algorithm
- Evaluating impact of different factors on electric vehicle charging demand
- An in-depth exploration of supervised and semi-supervised learning on face recognition
- The reform of the teaching mode of aesthetic education for university students based on digital media technology
- QCI-WSC: Estimation and prediction of QoS confidence interval for web service composition based on Bootstrap
- Line segment using displacement prior
- 3D reconstruction study of motion blur non-coded targets based on the iterative relaxation method
- Overcoming the cold-start challenge in recommender systems: A novel two-stage framework
- Optimization of multi-objective recognition based on video tracking technology
- An ADMM-based heuristic algorithm for optimization problems over nonconvex second-order cone
- A multiscale and dual-loss network for pulmonary nodule classification
- Artificial intelligence enabled microgrid power generation prediction
- Special Issue on AI based Techniques in Wireless Sensor Networks
- Blended teaching design of UMU interactive learning platform for cultivating students’ cultural literacy
- Special Issue on Informatics 2024
- Analysis of different IDS-based machine learning models for secure data transmission in IoT networks
- Using artificial intelligence tools for level of service classifications within the smart city concept
- Applying metaheuristic methods for staffing in railway depots
- Interacting with vector databases by means of domain-specific language
- Data analysis for efficient dynamic IoT task scheduling in a simulated edge cloud environment
- Analysis of the resilience of open source smart home platforms to DDoS attacks
- Comparison of various in-order iterator implementations in C++
Articles in the same Issue
- Review Article
- Enhancing IoT network security: a literature review of intrusion detection systems and their adaptability to emerging threats
- Research Articles
- Intelligent data collection algorithm research for WSNs
- A novel behavioral health care dataset creation from multiple drug review datasets and drugs prescription using EDA
- Speech emotion recognition using long-term average spectrum
- PLASMA-Privacy-Preserved Lightweight and Secure Multi-level Authentication scheme for IoMT-based smart healthcare
- Basketball action recognition by fusing video recognition techniques with an SSD target detection algorithm
- Evaluating impact of different factors on electric vehicle charging demand
- An in-depth exploration of supervised and semi-supervised learning on face recognition
- The reform of the teaching mode of aesthetic education for university students based on digital media technology
- QCI-WSC: Estimation and prediction of QoS confidence interval for web service composition based on Bootstrap
- Line segment using displacement prior
- 3D reconstruction study of motion blur non-coded targets based on the iterative relaxation method
- Overcoming the cold-start challenge in recommender systems: A novel two-stage framework
- Optimization of multi-objective recognition based on video tracking technology
- An ADMM-based heuristic algorithm for optimization problems over nonconvex second-order cone
- A multiscale and dual-loss network for pulmonary nodule classification
- Artificial intelligence enabled microgrid power generation prediction
- Special Issue on AI based Techniques in Wireless Sensor Networks
- Blended teaching design of UMU interactive learning platform for cultivating students’ cultural literacy
- Special Issue on Informatics 2024
- Analysis of different IDS-based machine learning models for secure data transmission in IoT networks
- Using artificial intelligence tools for level of service classifications within the smart city concept
- Applying metaheuristic methods for staffing in railway depots
- Interacting with vector databases by means of domain-specific language
- Data analysis for efficient dynamic IoT task scheduling in a simulated edge cloud environment
- Analysis of the resilience of open source smart home platforms to DDoS attacks
- Comparison of various in-order iterator implementations in C++