Processing math: 100%
A journal of IEEE and CAA , publishes high-quality papers in English on original theoretical/experimental research and development in all areas of automation
T. Y. K. Zhang, J. X. Zhan, J. M. Shi, J. M. Xin, and  N. N. Zheng,  “Human-like decision-making of autonomous vehicles in dynamic traffic scenarios,” IEEE/CAA J. Autom. Sinica, vol. 10, no. 10, pp. 1905–1917, Oct. 2023. doi: 10.1109/JAS.2023.123696
Citation: T. Y. K. Zhang, J. X. Zhan, J. M. Shi, J. M. Xin, and  N. N. Zheng,  “Human-like decision-making of autonomous vehicles in dynamic traffic scenarios,” IEEE/CAA J. Autom. Sinica, vol. 10, no. 10, pp. 1905–1917, Oct. 2023. doi: 10.1109/JAS.2023.123696

Human-Like Decision-Making of Autonomous Vehicles in Dynamic Traffic Scenarios

doi: 10.1109/JAS.2023.123696
Funds:  This work was supported by the National Key R&D Program of China (2022YFB2502900) and the National Natural Science Foundation of China (62088102, 61790563)
More Information
  • With the maturation of autonomous driving technology, the use of autonomous vehicles in a socially acceptable manner has become a growing demand of the public. Human-like autonomous driving is expected due to the impact of the differences between autonomous vehicles and human drivers on safety. Although human-like decision-making has become a research hotspot, a unified theory has not yet been formed, and there are significant differences in the implementation and performance of existing methods. This paper provides a comprehensive overview of human-like decision-making for autonomous vehicles. The following issues are discussed: 1) The intelligence level of most autonomous driving decision-making algorithms; 2) The driving datasets and simulation platforms for testing and verifying human-like decision-making; 3) The evaluation metrics of human-likeness; personalized driving; the application of decision-making in real traffic scenarios; and 4) The potential research direction of human-like driving. These research results are significant for creating interpretable human-like driving models and applying them in dynamic traffic scenarios. In the future, the combination of intuitive logical reasoning and hierarchical structure will be an important topic for further research. It is expected to meet the needs of human-like driving.

     

  • WITH the development of artificial intelligence, autonomous driving has a significant influence on human lifestyle and travel modes. It brings about a revolution in intelligent transportation [1]. Researchers are no longer limited to low-level autonomous driving assistance systems (ADAS) and have begun to explore the realization of Level 4 and Level 5 autonomous driving as defined by the society of automotive engineers (SAE) [2]. This means that the running process of autonomous vehicles will be completely independent of human drivers.

    While there are already mature solutions available for simple driving scenarios, autonomous driving applications must inevitably face the challenge of dynamic and complex scenarios. Unexpected events such as sudden acceleration or deceleration of other vehicles, pedestrians crossing the street, sudden weather changes, and vehicle malfunctions leading to road congestion, can occur. Given the unpredictability of these situations, the autonomous driving system must be able to make appropriate judgments and decisions. It is necessary for autonomous vehicles to approach or even exceed the performance of human drivers.

    Human-like decision-making is a key concept in autonomous driving technology. It enables autonomous driving systems to make correct judgments and decisions in complex traffic environments. Achieving human-like decision-making requires dealing with various uncertain factors in dynamic traffic scenes, meeting the needs of passengers and other road users, and ensuring efficiency and safety.

    Autonomous vehicles with human-like characteristics have the potential to outperform human drivers due to their faster reaction times, highly accurate decision-making abilities, and greater adaptability. The high-performance computing unit can process vast amounts of data and make real-time decisions based on it, reducing the impact of human error. What is more, autonomous vehicles have the ability to continuously learn and enhance their performance. As time progresses, they will surpass human drivers in terms of performance, leading to improved efficiency and safety on the road.

    Although there are differences in implementation, researchers have reached a consensus on the advantages of human-like driving. The specific advantages of human-like driving strategies can be summarized as follows: 1) to help drivers of surrounding vehicles predict the behavior of autonomous vehicles based on existing experience, thus improving overall driving safety; 2) to improve the performance of the human-machine cooperative driving system, because if autonomous vehicles make decisions in a predictable way, other human drivers will be easier to manipulate; and 3) to personalize the driving style of autonomous vehicles to provide services for different types of passengers and vehicles.

    Implementing human-like decision-making requires significant algorithmic and technical support. For example, imitation learning can extract driving experience, reinforcement learning algorithms can formulate optimal driving strategies, and Bayesian networks can handle uncertain factors. It is also necessary to consider multiple factors, such as road safety, passenger needs, and road user expectations, to ensure that the decisions made by the autonomous driving system are safe, efficient, and reasonable.

    In order to explore how to develop decision-making algorithms to meet public needs, researchers from different academic backgrounds study human-like driving from different perspectives [3]–[5]. However, these efforts have not yet formed a unified theoretical framework. The decision-making ability of some early autonomous driving systems can only support the completion of simple driving tasks [6], [7]. It is easy to make mistakes in complex and dynamic traffic scenarios, which is far from human capabilities. With the development of artificial neural network technology and the improvement of computing power, the study of learning-based decision-making has gradually increased. Learning-based approaches are outcome-oriented by using large amounts of instructional data or environmental interactions to train models, and they can reach the level of human drivers in terms of results and guide the vehicle to complete driving tasks [8]. However, current learning-based methods have some common problems, such as poor interpretation. When faced with complex scenarios, even the best current system will make mistakes, such as being unable to avoid obstacles or even cause serious traffic accidents.

    Therefore, it is necessary to build a system with both performance and mechanism in one framework that combines reasoning and learning. First of all, the concept of human drivers’ driving behavior should be established, and the driving intentions of oneself and surrounding vehicles should be understood. Then autonomous vehicles should have driving styles and characteristics that can be adjusted as needed.

    To promote research of human-like driving, many other factors should be considered in addition to the method itself, such as the design of driving simulators, driving datasets, and evaluation metrics. Due to the danger of unmanned testing in real traffic scenarios, driving simulation and the collection of driving datasets are essential for high-level decision-making methods. High-quality driving datasets and driving simulations are required to ensure the usability of algorithms in real scenarios. Hence it is necessary to summarize key elements and provide some help for the research in related fields.

    This paper focuses on the above aspects of human-like decision-making. Some key questions are raised to help readers fully understand the current research. In particular, the following issues are discussed in detail: 1) the intelligence level of decision-making algorithms for autonomous driving and their imitation of human drivers; 2) the elements required for driving datasets, simulation platforms, and testing scenarios to verify human-like decision-making; 3) the evaluation metrics of human likeness, personalized driving, and the application of decision-making algorithms in real traffic scenarios; and 4) the potential research topics of human-like driving systems. In fact, each type of existing decision-making algorithm has a certain degree of human-likeness in mechanism or performance, but it cannot completely reproduce the driving style of human drivers. Although the designs of existing driving simulators, driving datasets, and evaluation indicators are relatively complete, a unified specification has not yet been formed. Driving simulations should collect driving data of autonomous vehicles in a virtual traffic environment to train decision-making algorithms and compare autonomous vehicles with humans in a risk-free environment. In the future, the application of artificial intelligence will result in breakthroughs in the decision-making of autonomous driving. It is expected that the hierarchical structure combined with intuitive-logical reasoning can meet the needs of human-like driving.

    This paper is organized as follows. The background and the basic knowledge of human-like decision-making are introduced in Section II, while existing work on the decisionmaking of autonomous vehicles is reviewed and the challenges faced by current decision-making algorithms are discussed in Section III. The development of driving datasets and simulators is presented in Section IV, and related topics on humanlike decision-making are discussed in Section V. Finally, some potential future research directions are considered in Section VI.

    Human driving decisions are affected by many factors including knowledge, experience, personality, cultural background, and environmental factors. Some of these factors contribute to the safety of autonomous driving systems, while other factors such as emotional disorders may lead to negative outcomes. It was pointed out that professional knowledge and driving experience can effectively reduce driving violations [9], [10], which is an important guide. The impact of driving risk on driving behavior should also be considered. If algorithms can simulate the human perception of driving risk in a real environment, it will be of great significance for building an intelligent human-like driving system.

    The similarity between artificial intelligence systems and humans is a research hotspot. As artificial intelligence becomes increasingly important in areas focused on humans, there is a growing understanding that both reasoning and emotions must be considered. Even social responsibility is something that AI systems need to possess [11]. Although there are some debates about whether autonomous driving strategies need to be consistent with that of human beings, the potential advantages of human-like strategies cannot be ignored, such as comfort and driving safety. Therefore, researchers study human-like intelligence, such as hybrid augmented intelligence [12], which introduces human roles or cognitive models into intelligent systems. This is a feasible and important growth model of artificial intelligence, which provides further inspiration for the design of human-like decision-making systems [13].

    Cognitive science is the theoretical basis for the development of human-like driving decision-making systems, and it has important guiding significance. Since most readers of this paper are developers of autonomous driving systems and may not be familiar with cognitive science, some basic concepts and existing research progress are provided herein.

    The cognitive architecture model demonstrates and simulates the functions of human brains, rather than their physics and structure. For current research on automatic driving, understanding the cognitive architecture model is more important than understanding the physiological model of human brains. The related research can be divided into symbolist, connectionism architecture, and hybrid architecture, among which the representative examples are Soar [15], adaptive resonance theory (ART) [16], and the adaptive characteristics of thought-rational (ACT-R) [4]. In particular, ACT-R specifically points out that the human cognitive process requires the participation of four different modules: goal, vision, action, and commonly used descriptive knowledge modules. Many models describing driving behavior are based on ACT-R [17], [18].

    Due to the complexity of cognitive processes, there is no single cognitive framework that can explain them all. However, some researchers have suggested that a general cognitive architecture model should include perception, movement, working memory, descriptive memory, and long-term memory [19]. This is an important reference for establishing a general cognitive architecture model. Furthermore, cognitive architecture models can be combined with other physiological models to expand their applicability and coverage [20]. In addition to the basic elements of cognitive architecture such as memory, perception, and execution modules, a complete cognitive framework should include long-term memory modules for storing prior knowledge and a target module for processing goal-oriented driving tasks.

    In order to simulate parallel activities more effectively, the queuing network-model human processor (QN-MHP) was adopted [21], [22]. In fact, the QN-MHP integrates two complementary methods: the queuing network method and the symbolic method (taking ACT-R as an example). Therefore, the QN-MHP provides a framework for modeling and real-time generation of concurrent activities. Such a model can intuitively reflect the ability of the human cognitive system, but each subunit of the system needs to be carefully designed.

    Dynamic traffic environments contain various factors of uncertainty that can impact the operation and efficiency of the entire transportation system. Here are some examples:

    1) Vehicle speed and direction are unpredictable. Some vehicles may be speeding, while others may be traveling slowly. This uncertainty can lead to traffic congestion and accidents.

    2) The position and movement direction of pedestrians and other obstacles are also uncertain. Pedestrians may suddenly run onto the road or cross the road. Other obstacles, such as bicycles or motorcycles, may also appear on the road. This uncertainty can lead to traffic accidents.

    3) Weather and road conditions also influence traffic. For instance, on rainy days, the road surface becomes slippery, making it harder for vehicles to travel. This uncertainty can cause traffic congestion and accidents.

    4) Vehicle breakdowns or stops are also uncertainties. Vehicles may suddenly malfunction or stop on the road, waiting for repairs. This uncertainty can lead to traffic congestion and jams.

    Therefore, it is crucial to understand and consider these uncertainties to ensure efficient operation and safety of the transportation system. The layered models are widely used to classify different information into different layers to describe dynamic traffic scenarios in autonomous driving applications. A typical layered model includes the following five layers [23]:

    Layer 1: Road-Level (road structure).

    Layer 2: Traffic Infrastructure (traffic lights, etc.).

    Layer 3: Temporary manipulation of Layers 1 and 2 (road maintenance, temporary closure).

    Layer 4: Objects (dynamic vehicles and pedestrians).

    Layer 5: Environment (weather, etc.).

    The first two layers (i.e., Layers 1 and 2) describe static information, while the last three layers (i.e., Layers 3–5) describe dynamic information. As shown in Fig. 1, the latest traffic scenario model of the PEGASUS project [14] includes a digital communication layer in addition to the traditional model. This layer describes the information that relies on communication technologies such as vehicle to everything (V2X) and high definition (HD) maps. There are a large number of traffic participants whose behavior will affect the safety of vehicles in the dynamic traffic scenario. In addition, temporary operations and complex environmental conditions should be considered. Therefore, it is necessary to thoroughly test representative scenario types or instances and pay special attention to the performance of algorithms in boundary scenarios and hazards [24].

    Figure  1.  Multi-layered model of a dynamic traffic scenario proposed by the PEGASUS project [14].

    Traditional decision-making methods for autonomous vehicles can be classified as either logic-based or statistics-based. Decisions defined by logical and statistical models are considered “rational”. In recent years, the application of artificial intelligence technology has allowed autonomous vehicles to break through these categories. Algorithms such as imitation learning and reinforcement learning can be used to directly learn driving strategies from driving data or environmental interactions. This section provides a comprehensive analysis of traditional and artificial intelligence-based decision-making methods, including their design and performance levels, which are compared to humans.

    1) Finite State Machine: The finite state machine (FSM) comprises a set of states and transition relationships between each state, where the internal rules between each state are executed through conditional judgments. The structure of the FSM is divided into three types: series, parallel, and hybrid. Among them, the hybrid structure is the most widely used in the decision-making of autonomous driving. As shown in Fig. 2, the final decision result is calculated through state estimation and target selection in the hybrid structure of the FSM. The results of different submodules can be arbitrated according to priority, which combines all aspects of risk theory to model human-like driving. In most cases, the hybrid structure of the FSM has satisfactory performance, but it is still not a perfect solution. For the rule base, it is a major challenge to fully cover all potential situations that may occur in complex traffic scenarios. If the status enters an area not covered by the rule base, the vehicle can no longer be guaranteed to operate safely.

    Figure  2.  Example of decision-making based on a finite state machine.

    2) Fuzzy Reasoning: Fuzzy reasoning refers to the imitation of human thinking. It can effectively express qualitative and empirical information that has unclear boundaries. The decision-making system based on fuzzy reasoning has been widely used in the fields of medicine [25], agriculture [26], and social services [27]. Fuzzy logic provides a higher level of explanation for decision-making [28]. It realizes flexible rule design and can better adapt to a fuzzy traffic environment [29]. The main challenge of fuzzy reasoning is the lack of logical traceability and the inability to accurately determine the state of vehicles based on their behavior. Combining fuzzy logic with other complementary algorithms can improve the performance of the decision-making system.

    1) Bayesian Network: As for data processing, the Bayesian network is good at classifying the probability of events and analyzing the reliability of events. The Bayesian network not only uses the probability description of the time-space relationship between vehicles but also incorporates the uncertainty of input data into the threat assessment of vehicles. It can efficiently represent uncertain events, such as estimating the probability of a vehicle collision [30]–[32].

    Bayesian network-based methods offer advantages in dealing with uncertainty, which can improve the performance and reliability of autonomous driving systems. However, in practical applications, the computational cost of this algorithm is typically high, leading to long computation times. Additionally, the complexity of the algorithm also increases due to the need to balance multiple uncertain factors.

    2) Markov Decision Process: Markov decision process (MDP) provides a framework for modeling decisions with partially random and partially controlled outcomes. In particular, partially observable MDP (POMDP) introduces perception uncertainty because driver intentions cannot be directly measured [33]. It has been found that POMDP is suitable for autonomous driving to make real-time decisions [34], [35]. However, obtaining a general solution to a specific problem is difficult with the original POMDP. Therefore, the multipolicy decision-making (MPDM) method has received extensive attention [36], [37].

    1) Game Theory Based Methods: Game theory describes the process of interaction in information structures. Compared to other methods, game theory emphasizes the interaction between drivers and more accurately reflects their driving behavior [38], [39]. Some researchers believe that the potential of game theory is underestimated [40]. With the development of sensor perception and dataset simulation, more accurate traffic participant intentions can be gradually obtained to achieve a complete game, and it is easy to combine with other lane-changing models. In fact, the core of game theory is utility. The application of game theory in lane changing was reviewed in [40], based on the difference between the utility revenue function and the convergence strategy.

    Compared to traditional decision-making methods, game theory makes more reasonable choices by considering the interaction between autonomous vehicles and the surrounding environment. Therefore, decision-making based on game theory has research potential.

    2) Reinforcement Learning: Reinforcement learning optimizes sequential driving strategies through feedback evaluation while interacting with the environment. The basic reinforcement learning is modeled as MDP [41], where the goal is to find the optimal strategy π, which can achieve the highest discount reward expectation after the implementation of the strategy as follows:

    π=argmaxπEπ{H1k=0γkrk+1|s0=s}. (1)

    Model-free methods, such as deep Q-networks (DQN) [42] and depth deterministic strategy gradient (DDPG) [43], are applied to generate advanced driving strategies without learning the dynamic model of the environment [44], [45]. Meanwhile, model-based methods make decisions by learning the dynamic model of the environment, thus reducing the time and cost required for actual interaction in the learning process [46]. Additionally, reinforcement learning can integrate recurrent neural networks to consider the POMDP for decision-making, making it more suitable for uncertain driving environments [47]. All of these works have effectively completed decision-making tasks under specific traffic scenarios.

    Indeed, reinforcement learning is a powerful technology of artificial intelligence, but some problems must be discussed and resolved. First, the maximization of the reward function is usually different in different environmental definitions. Only carefully constructed rewards can be used for training human-like intelligence [48]. In addition, it is quite difficult to improve the learning efficiency of reinforcement learning. Although the learning efficiency can be improved to some extent by adopting the exploration strategy, it is still necessary to develop more efficient reinforcement learning algorithms.

    In real-world applications, there is no doubt that auto- nomous vehicles must be absolutely accurate, which conflicts with the “trial and error” approach used in the training process. To deal with extreme scenarios in the real world and ensure driving safety, a complex architecture with safety mechanisms should be adopted.

    3) Imitation Learning: The purpose of imitation learning is to learn expected behavior and perform tasks through expert data. By assuming that the best action is given by the expert dataset D consisting of the state-action pair (s,a), the strategy πθ(s) is trained to be as close as possible to the expert strategy π(s) by

    argminθEsP(s|θ)L(π(s),πθ(s)) (2)

    where P(s|θ) is the state distribution of the trained strategy πθ, and L is the loss function.

    Behavior cloning treats the task of imitation learning as supervised learning. Fig. 3 depicts a convolutional neural network (CNN)-based “end-to-end” decision-making model designed by NVIDIA [49]. Research shows that the deep neural network can learn and obtain intermediate features that are meaningful for vehicle decision-making while ignoring a large amount of redundant information. Based on NVIDIA and other works [50], a full convolutional network with long and short-term memory (FCN-LSTM) was used to create a general model. The FCN-LSTM network extracts the temporal and spatial features of images and considers the semantics of visual features and their related time series to obtain decisions similar to human driver decisions. Then, based on the traditional imitation learning framework, Intel proposed conditional imitation learning to complete the internal “expert intention” modeling (including planning objectives, intentions, and prior knowledge) [51]. In Intel’s model, environment images and state observations during training are used, and the “expert intention” is introduced as a control input. Therefore, driving behavior can be separated from specific tasks, and the response to driving commands can handle the uncertainty of observed behavior.

    Figure  3.  “End-to-end” decision-making framework based on convolutional neural networks designed by NVIDIA [49].

    Inverse reinforcement learning is an “indirect” method that first learns the “intention” of human drivers through demonstration, and then generates a strategy based on the learned hidden goals. Compared with behavior cloning, the middle layer of inverse reinforcement learning can be designed to significantly improve the interpretability of the entire system. Several studies have been conducted to extract driving habits from human driving datasets. When designing the reward function, ride comfort [52], safety [53], and traffic efficiency [54] are emphasized. Furthermore, various attempts have been made to build decision-making systems, such as generating cost maps and combining model predictive controllers for decision-making [55].

    In particular, generative adversarial imitation learning (GAIL) has achieved remarkable results in imitating complex behaviors in high-dimensional environments [56]. It has verified its capabilities on the real vehicle dataset [57]. Compared with the behavior cloning method, GAIL can effectively reduce the cumulative error and make the behavior more robust to fluctuations in the trajectory, bringing it closer to real behavior. On this basis, the algorithm combines parameter sharing trust region policy optimization and GAIL (PS-GAIL), which is more stable and has fewer collisions in long-term interaction [58]. Considering the insufficient response of multi-agent imitation learning to sudden emergency driving, the customized reward further improves the local interaction performance of agents and teaches them emergency avoidance abilities in complex scenarios [59].

    In the above, different decision-making methods are reviewed. Now, their human-likeness can be summarized from the following aspects, as shown in Table I:

    Table  I.  Comparison of Adaptability of Decision-Making Methods in Practical Application
    MethodReferenceCombination of deterministic
    and fuzzy logic
    Adaptability to unknown
    environments
    Consideration of random
    factors
    Learning ability
    Finite state machines[60]–[62]++++
    Fuzzy reasoning[63]–[65]+++++++
    Bayesian network[30]–[32]+++++++
    Markov decision process[34]–[37]+++++++
    Game theory approach[38], [39]++++++++
    Reinforcement learning[44]–[47]+++++++
    Imitation learning[49]–[51]+++++++
    Note: +++: Good performance; ++: Average performance; +: Poor performance.
     | Show Table
    DownLoad: CSV

    a) Combination of deterministic and fuzzy logic: This is related to the human being’s ability to handle complex driving decisions with limited computing power.

    b) Adaptability to unknown environments: In order to accurately and comprehensively understand its surrounding environment, autonomous vehicles should have effective decision-making algorithms to model and process some of the observed states.

    c) Consideration of random factors: The adaptability of random factors is of great significance to the safety, intelligence, and reliability of autonomous driving.

    d) Learning ability: In order to improve decision-making capabilities, autonomous vehicles should learn from expert data or self-learning in environmental interactions.

    As shown in Table I, different methods focus on these four aspects respectively. Various decision-making algorithms have unique advantages and are applicable to different traffic situations. For instance, FSMs are suitable for discrete state and action spaces, such as lane changing, emergency braking, and steering. However, they cannot calculate continuous control outputs directly. Statistical-based methods are more suitable for scenarios with a large number of traffic participants in the environment, where predicting motion patterns is difficult, such as mixed traffic of motorized and non-motorized vehicles. Learning-based methods are applicable to almost all decision scenarios, depending on the specific algorithm selection and the overall design of the decision-making system. In conclusion, the autonomous driving decision-making system comprises multiple algorithms that work in combination, rather than a single type of algorithm. Therefore, the combination of different types of models is crucial for developing advanced human-like decision-making methods.

    It is of great significance to quantify and evaluate the decision-making ability of autonomous vehicles. The current evaluation of autonomous driving can be divided into three types: 1) dataset-based evaluation, 2) simulation-based evaluation, and 3) real-world-based evaluation [1]. Here, we review the relevant research on human driving datasets and autonomous driving simulation platforms.

    As more and more decision-making algorithms learn human driving habits from vehicle driving data, it is necessary to collect interactive vehicle data from the real world. Evaluating the performance of decision-making algorithms by comparing them with human driving data is also important. Table II lists the commonly used datasets for autonomous driving, and the main characteristics of these databases are as follows.

    Table  II.  Comparison of Some Representative Human Driving Datasets and Their Characteristics
    DatasetReferenceDescription
    NGSIM[66]One of the most popular driving datasets; raw video information; high precision and wide coverage.
    HighD[67]Accurate vehicle motions; dynamic urban scenarios are not included.
    Lyft[68]Detailed movements for traffic participants.
    INTERACTION[69]High-precision semantic maps, data on crisis situations.
    HDD[70]Record driving scenes with a long time span and a large spatial range.
    Argoverse[71]Provides relatively rich semantic map information.
    Commonroad[72]The first driving dataset for planning.
    nuPlan[73]The latest release dataset dedicated to ML-mased planning; including benchmarks and simulated traffic scenarios.
     | Show Table
    DownLoad: CSV

    1) NGSIM [66]: The “next generation simulation” project (NGSIM) includes a dataset of vehicle-road collaborations. The NGSIM dataset was collected from various regions, including structured road intersections, high-speed entrances and exits, and other hotspots of vehicle-road cooperation research. The original video information is used to generate track data for each vehicle in traffic flow. Many human-like driving algorithms have been tested and validated on the NGSIM dataset [59], [82].

    2) HighD [67]: Compared with the dataset of NGSIM, the large-scale naturalistic vehicle trajectory dataset from German highways (HighD) attempts to the accuracy of vehicle motion information is improved. However, urban scenes containing intensive and highly interactive behaviors, such as roundabouts and unsignalized intersections, are not included in NGSIM and HighD.

    3) Lyft [68]: Lyft includes sensing data and prediction data. It covers the movement logs of cars, cyclists, pedestrians, and other traffic agents encountered by their autonomous fleets.

    4) INTERACTION [69]: The INTERnational, Adversarial and Cooperative moTION (INTERACTION) dataset was proposed for semantic mapping of interactive driving scenarios with different driving cultures. It not only focuses on the integrity of high-precision semantic map information and interactive entities, but also contains crisis data, which can effectively promote research on extreme scenarios.

    5) HDD [70]: The Honda Research Institute Driving Dataset (HDD) is a public dataset based on onboard sensors, recording driving scenes with a longer time span and larger space range.

    6) Argoverse [71]: Argoverse is also a public dataset based on onboard sensors. It provides relatively rich semantic map information. However, due to the limitations of onboard sensors, the motion track of surrounding objects is incomplete. The inaccuracy of other vehicles’ positions around the environment makes it more difficult to confirm the aerial view.

    7) Commonroad [72]: Composable benchmarks for motion planning on roads (CommonRoad) are the first planning benchmarks with driving dataset for autonomous driving. In CommonRoad, real-world scenarios in the dataset can be easily converted into simulation scenarios to test the performance of decision-making algorithms offline.

    8) nuPlan [73]: nuPlan is a closed-loop ML-based planning benchmark. It includes the latest released dataset dedicated to machine learning-based planning methods. The dataset of nuPlan can effectively evaluate the performance of long-term planning and provides a closed-loop simulation framework and metrics for specific scenarios.

    As described above, the latest driving datasets, such as Commonroad [72] and nuPlan [73], already possess the basic elements for training and evaluating human-like driving decision-making methods. However, the data in complex scenarios should be further supplemented and improved, and data on extreme cases is still insufficient.

    1) Classic Driving Simulators: Simulation plays an important role in data collection and experimental verification [83]. Considering the risks involved in recording driving data in certain scenarios, driving simulators should be used to supplement extreme (even crash) data to reduce the security risks of testing in the real world. Therefore, researchers have higher requirements for driving simulation, which promotes the development of high-fidelity simulators. Table III lists the driving simulations commonly used in autonomous driving.

    Table  III.  Comparison of Some Representative Autonomous Driving Simulators and Their Characteristics
    SimulatorReferenceDescription
    TORCS[74]The open racing car simulator (TORCS) is an open-source racing simulator which is also widely used in verification for autonomous driving algorithms. However, the functions of TORCS are limited, and the motion characteristics of the simulated vehicle in TORCS are significantly different from those of a real vehicle.
    CARLA[75]Car learning to act (CARLA) is an open-source simulator that provides virtual scenes required to develop autonomous driving algorithms, including urban layouts, buildings, and vehicles. CARLA supports the free configuration of simulation scenarios and onboard sensors and controls both static and dynamic participants.
    CarSim[76]The CAR-following SIMulation model (CarSim) is specifically designed for vehicle dynamics that simulate the response of a vehicle to its driver, road surface, and aerodynamic input. Due to CarSim’s high-fidelity vehicle dynamics, it is usually combined with other simulation software to build autonomous driving simulation systems [77].
    Gazebo[78]Gazebo is a built-in simulation software of the robot operating system (ROS) [79], which offers the ability to accurately and efficiently simulate robots. Since ROS has been widely applied in the development of autonomous driving [80], Gazebo-based simulators have become increasingly popular.
    ADAPS[81]Autonomous driving via principled simulations (ADAPS) learns autonomous driving strategies from vehicle accidents, which includes a high-fidelity simulation platform for testing and a simulation platform for accident analysis. This framework provides a more effective online framework for learning that significantly reduces the number of iterations required to learn autonomous driving strategies.
     | Show Table
    DownLoad: CSV

    The complete system for evaluating and testing decision-making capabilities based on driving simulators includes modules for collecting driving data from human drivers, identifying driving behaviors, and evaluating these behaviors. In recent times, researchers have started testing self-driving algorithms in simulations before putting them into practice in the real world. Different solutions were compared to obtain more convincing results [84], [85]. Moreover, simulation testing has been combined with real-world testing to develop and validate decision algorithms.

    In practice, several necessary factors should be considered when building a simulation platform to evaluate and test decision-making capabilities. These factors include: 1) a dynamic vehicle model that accurately reflects the kinematics and dynamics characteristics of the actual vehicle; 2) various typical traffic scenarios; and 3) tests conducted with and without drivers under the same test conditions, as well as driving behavior evaluation based on vehicle movement. As shown in Fig. 4, the autonomous driving test simulation system should be a unified framework that includes offline simulation, real-time hardware-in-loop simulation, and driving simulators. The system not only needs to analyze driving behavior through collected driving data, but also needs to analyze the driving behavior of autonomous vehicles controlled by algorithms, evaluating their decision-making abilities under the same criteria.

    Figure  4.  An example of a simulation platform for autonomous driving.

    2) Mixed Reality Based Driving Simulators: Despite the interrelationship of virtual simulation, it is necessary to find a method to provide a more comprehensive and realistic representation of traffic scenarios, in order to convincingly verify the decision-making algorithms related to autonomous driving. Mixed reality (MR) is a combination of real and virtual elements, in which physical and digital objects coexist and interact in real-time. It retains the real-world scene restoration with security and element diversity. In recent years, many MR-based methods have been proposed for autonomous driving [86], and some platforms have been developed to verify the decision-making algorithm. Compared to traditional simulators, the results are more convincing.

    In addition, virtual reality (VR) technology is used to simulate the virtual scene, where pedestrians interact with autonomous vehicles. By wearing VR devices, pedestrians can explore their responses to autonomous vehicles in different scenarios [87]. Ride comfort can also be investigated by collecting information about users’ immersion in VR-based driving simulations [88].

    The main problems of the current MR-based driving simulator are as follows:

    a) Consistent with reality: When drivers or pedestrians use mixed reality equipment, the simulation system needs to provide sufficient support to display and interpret the scene. This ensures that the response of the human tester is consistent with the real-world situation.

    b) Usability: Compared to the combination of a traditional driving simulator and a large screen display, MR equipment is more difficult to use. Many users are not familiar with MR equipment and may deviate from the real world due to unskilled operation.

    c) Standardization: Most MR-based simulators lack a common framework and unified metrics, which is not conducive to comparing different works.

    The latest research shows that current technology effectively solves decision-making tasks for different levels of autonomous vehicles. However, many challenges remain unresolved. In this section, we will discuss some important issues.

    As one of the important purposes of human-like driving, personalized driving deserves further discussion. Driving style is defined as the driver’s relatively stable, long-term, and inherent behavioral tendencies. It integrates the driver’s psychological thinking and behavioral mode. The future research trend is to consider personal characteristics to improve vehicle safety and passenger acceptance [89].

    Driving styles can generally be divided into three categories:

    1) Radical driving style: This style takes some risks to achieve the driving goal, accelerates and decelerates violently and frequently, and shows more aggressive driving behavior.

    2) Careful driving style: This style places safety above achieving driving goals, showing more conservative acceleration and deceleration, or preferring to stay in the current lane or give way to other vehicles.

    3) Moderate driving style: This style is between radical and cautious.

    Obviously, such classification is rough and lacks unified standards. The driving styles of different individuals are significantly different. Thus, designing a reasonable scale to distinguish different driving styles is an important prerequisite for personalized driving. Recently, some works on driving styles have been conducted to define and simulate the driving behavior of humans or autonomous vehicles in specific environments or traffic scenarios, such as lane changing, car following, and intersections [90]–[92]. The features and scales selected in these works are also related to specific driving scenarios. In addition, the designed passenger preference measurement adopts parameters directly related to vehicle movement, including the preferred horizontal and vertical growth areas and the maximum allowable acceleration [93].

    In fact, the public’s preference for driving style tends to be personalized, but there are also some rules. For example, it was found that passengers tend to be cautious when seated in autonomous vehicles [94]. When people can not control their driving, they prefer a cautious style. In addition, autonomous vehicles can use collaboration to perceive and share information that human drivers cannot directly obtain. Therefore, autonomous vehicles may be more confident than humans. These behaviors may be seen as causing public discomfort, even if they do not pose a danger.

    The similarity between algorithms and human decisions has been studied from different perspectives. Therefore, a unified indicator is needed to evaluate the human similarity of the algorithm, and the principle and driving performance should be considered to accurately evaluate the human similarity of the algorithm.

    As described in Section III, the existing decision-making methods are evaluated for their human likeness through the following four aspects: 1) combination of deterministic and fuzzy logic, 2) adaptability to unknown environments, 3) consideration of random factors, and 4) learning ability. In addition, interpretability is also an important basis to evaluate the human similarity of the algorithm. It is the degree to which humans understand how to make decisions [95]. The higher the interpretability of the decision-making model, the easier it is for humans to understand why they make certain decisions. Only when the algorithm can be understood and interpreted by humans, can trust be established between humans and the model.

    Furthermore, the driving performance of human drivers is an important reference for evaluating human likeness from the perspective of driving performance. In order to demonstrate the potential of their models to be similar to human driving behavior, some researchers have explained the “human-like” nature of their methods by directly comparing them with human driving data [96]. The similarity between them shows the similarity between intelligent decision-making algorithms and human decision-making processes. In general, the evaluation of human-like driving performance includes the following indicators: 1) driving comfort, 2) driving safety, 3) similarity with human demonstration trajectories, and 4) characteristics relative to other traffic participants.

    There is no doubt that safety is the most important challenge and critical factor in autonomous driving. In the real world, traffic participants usually exhibit dynamic and uncertain behavior, so the autonomous vehicle must be able to handle all possible situations. Otherwise, decision-making will no longer be reliable and may even lead to serious traffic accidents. The data-driven method can improve the traversal depth of traffic scenarios, but it requires plenty of data to ensure safe decision-making. Therefore, it is necessary to design road tests to collect data covering environmental uncertainties. In addition, researchers should understand the interaction between autonomous vehicles and other elements in the traffic environment through road testing. Potential collision risks should be discovered in advance by designing some typical extreme scenarios.

    Due to the danger of real-world testing, it is important to develop and utilize high-fidelity driving simulators. Unfortunately, current simulated traffic scenarios are still unrealistic and cannot fully cover all possible situations in the real world. Although it cannot be guaranteed that the algorithms that meet safety requirements in the simulation environment will still be safe when transplanted to the real world, it can be predicted that driving simulators will play an increasingly important role in the development of decision-making algorithms.

    Human drivers face limitations that can impact their ability to navigate roads safely. These limitations include distractions, fatigue, emotions, and subjective behavior. These factors impair their judgment, decision-making abilities, and reaction times, leading to slower responses, delayed decisions, and an increased likelihood of accidents.

    The purpose of researchers developing human-like driving algorithms is to learn the driving style and good driving habits of human drivers, rather than learning the problems caused by the shortcomings of human drivers. In contrast, autonomous vehicles process vast amounts of data and make real-time decisions. They are not subject to the same limitations as human drivers, which means they can make more accurate and efficient decisions in dynamic traffic environments, taking into account various factors of uncertainty and ensuring the safety of all road users.

    It should be noted that there is currently no single type of humanoid driving decision algorithm that can completely overcome the limitations of human drivers. To achieve a more accurate, efficient, and reliable autonomous driving system that surpasses the performance of human drivers in all aspects, it is necessary to analyze the psychology, decision-making models, and decision-making processes of human driving and use a combination of various advanced algorithms.

    The decision-making based solely on logical reasoning is insufficient to deal with complex, dynamic traffic scenarios [97]. Intuitive reasoning enables human beings to effectively avoid risks. Therefore, it is important to combine intuition and logical reasoning in order to adapt to dynamic traffic scenarios and make them comparable to human capabilities. One of the representative methods of this combination is fuzzy reinforcement learning, which is crucial to the realization of general decision-making for autonomous vehicles.

    Fig. 5 shows the process of an adaptive cruise controller implemented through fuzzy reinforcement learning. Fuzzy reinforcement learning uses a membership function to transform continuous states into fuzzy semantic states; it then can generate complete rules after training. According to the rule base, the continuous output is obtained by weighted summation. In traditional fuzzy systems, the rule base is usually determined by a large amount of expert knowledge, and reinforcement learning can independently learn expert knowledge through the interaction between the agent and the environment, which plays a key role in the practical application of the system.

    Figure  5.  Decision-making framework based on fuzzy reinforcement learning for adapt cruise control.

    In order to verify the effectiveness of the proposed framework shown in Fig. 5, several test scenarios are designed on the driving simulator [98]. When the front vehicle accelerates violently, the human-like decision-making system will give appropriate acceleration to smoothly control the autonomous vehicle to catch up with the vehicle ahead. Although the speed will have a certain overshoot in this process, the acceleration is relatively stable, which ensures the safety of vehicle following tasks. On the contrary, when the vehicle ahead decelerates, the proposed method controls the autonomous vehicle so that it decelerates quickly without collision risk, so that the distance between the vehicles will never be less than the safe distance to ensure the safety of the autonomous vehicle. The driving behaviors learned from the proposed framework are not predefined but are fully learned through the process of interaction with the environment. It shows the good potential of combining logic and intuitive reasoning to learn human driving behavior.

    On the basis of traditional reinforcement learning, the rapid development of deep reinforcement learning in recent years is crucial to developing a decision-making method that combines intuition and logical reasoning. For example, the success of deep reinforcement learning in AlphaGo [99]–[101] and electronic games [102]–[104] has proven its potential to build intelligent human-like decision-making systems. It has been pointed out that rewards are enough to drive intelligent behaviors studied in nature and artificial intelligence, including knowledge, learning, perception, social intelligence, language, generalization, and imitation [48]. Therefore, the combination of deep reinforcement learning and traditional model-based methods is an important direction of current research on the human-like decision-making technology of autonomous driving.

    In the “end-to-end” model, only low-dimensional control commands or sparse rewards are used for learning. Due to this limitation, behavioral intelligence may not be well displayed. One possible way to improve this is through hierarchical decision-making, which can combine planning, decision-making, and path-tracking control into a complete module. The decision-making framework includes several levels, such as task, maneuver, and motions, where different levels refer to the decision-making process for human beings to achieve different goals. Furthermore, the hierarchical decision-making framework combines semantic and numerical methods to ensure safety and compliance with traffic rules, while providing common functions to solve new complex scenarios. In addition, the effective combination of rules and learning allows the network to focus on problems that do not exist in the rule base. For example, a hierarchical decision-making system was proposed to combine reinforcement learning with a sampling-based motion planner, which is a reasonable method [105].

    With the development of artificial intelligence, people have higher expectations for the possibility of autonomous driving. Exploring human-like decision-making technology is an important research direction for the development of autonomous driving. Thus, it is very important to improve the ability of the decision-making system to build a reasonable driving model so that the autonomous vehicle can learn human expert knowledge and driving habits.

    In this paper, the development of human-like driving technology was reviewed, where the advantages and disadvantages of various decision-making methods were analyzed, and related problems were discussed. In addition, the relevant research on driving datasets and simulation platforms was also reviewed for the development of advanced decision-making algorithms. Finally, several issues related to human-like driving were discussed, and potential research directions were presented as well.

    It is an arduous challenge to develop human-like driving so that autonomous vehicles have decision-making ability similar to human beings, but it has great potential value in academic and industrial applications. We firmly believe that the joint efforts of academia and industry will help promote the rapid development of automatic driving technology.

  • [1]
    Y. Ma, Z. Wang, H. Yang, and L. Yang, “Artificial intelligence applications in the development of autonomous vehicles: A survey,” IEEE/CAA J. Autom. Sinica, vol. 7, no. 2, pp. 315–329, 2020. doi: 10.1109/JAS.2020.1003021
    [2]
    Taxonomy and Definitions for Terms Related to On-Road Motor Vehicle Automated Driving Systems, SAE Int. Standard, vol. 3016, pp. 1–12, 2014.
    [3]
    S. Hecker, D. Dai, and L. V. Gool, “Learning accurate, comfortable and human-like driving,” arXiv preprint arXiv: 1903.10995, 2019.
    [4]
    F. E. Ritter, F. Tehranchi, and J. D. Oury, “Act-R: A cognitive architecture for modeling cognition,” Wiley Interdisciplinary Reviews: Cognitive Science, vol. 10, no. 3, p. e1488, 2019.
    [5]
    K. Sama, Y. Morales, H. Liu, N. Akai, A. Carballo, E. Takeuchi, and K. Takeda, “Extracting human-like driving behaviors from expert driver data using deep learning,” IEEE Trans. Vehicular Technology, vol. 69, no. 9, pp. 9315–9329, 2020. doi: 10.1109/TVT.2020.2980197
    [6]
    S. Thrun, M. Montemerlo, H. Dahlkamp, D. Stavens, A. Aron, J. Diebel, P. Fong, J. Gale, M. Halpenny, G. Hoffmann, et al., “Stanley: The robot that won the darpa grand challenge,” J. Field Robotics, vol. 23, no. 9, pp. 661–692, 2006. doi: 10.1002/rob.20147
    [7]
    M. Montemerlo, J. Becker, S. Bhat, H. Dahlkamp, D. Dolgov, S. Ettinger, D. Haehnel, T. Hilden, G. Hoffmann, B. Huhnke, et al., “Junior: The stanford entry in the urban challenge,” J. Field Robotics, vol. 25, no. 9, pp. 569–597, 2008. doi: 10.1002/rob.20258
    [8]
    A. Tampuu, T. Matiisen, M. Semikin, D. Fishman, and N. Muhammad, “A survey of end-to-end driving: Architectures and training methods,” IEEE Trans. Neural Networks and Learning Systems, vol. 33, no. 4, pp. 1364–1384, 2022. doi: 10.1109/TNNLS.2020.3043505
    [9]
    J. J. Rolison, S. Regev, S. Moutari, and A. Feeney, “What are the factors that contribute to road accidents? An assessment of law enforcement views, ordinary drivers’ opinions, and road accident records,” Accident Analysis &Prevention, vol. 115, pp. 11–24, 2018.
    [10]
    K. Bucsuházy, E. Matuchová, R. Zvala, P. Moravcová, M. Kostíková, and R. Mikulec, “Human factors contributing to the road traffic accident occurrence,” Transportation Research Procedia, vol. 45, pp. 555–561, 2020. doi: 10.1016/j.trpro.2020.03.057
    [11]
    A. Vetro, A. Santangelo, E. Beretta, and J. C. De Martin, “AI: From rational agents to socially responsible agents,” Digital Policy,Regulation and Governance, vol. 21, no. 3, pp. 291–304, 2019. doi: 10.1108/DPRG-08-2018-0049
    [12]
    N. Zheng, Z. Liu, P. Ren, Y. Ma, S. Chen, S. Yu, J. Xue, B. Chen, and F. Wang, “Hybrid-augmented intelligence: Collaboration and cognition,” Frontiers of Information Technology &Electronic Engineering, vol. 18, no. 2, pp. 153–179, 2017.
    [13]
    W. Wang, X. Na, D. Cao, J. Gong, J. Xi, Y. Xing, and F.-Y. Wang, “Decision-making in driver-automation shared control: A review and perspectives,” IEEE/CAA J. Autom. Sinica, vol. 7, no. 5, pp. 1289–1307, 2020.
    [14]
    M. Schiementz, K. Groh, S. Wagner, and T. Kühbeck, “Pegasus-test case variation and execution,” in Proc. PEGASUS Symp., 2019.
    [15]
    J. E. Laird, “Introduction to SOAR,” 2022.
    [16]
    L. E. B. da Silva, I. Elnabarawy, and D. C. Wunsch II, “A survey of adaptive resonance theory neural network models for engineering applications,” Neural Networks, vol. 120, pp. 167–203, 2019. doi: 10.1016/j.neunet.2019.09.012
    [17]
    M. Cina and A. B. Rad, “Categorized review of drive simulators and driver behavior analysis focusing on ACT-R architecture in autonomous vehicles,” Sustainable Energy Technologies and Assessments, vol. 56, p. 103044, 2023. doi: 10.1016/j.seta.2023.103044
    [18]
    A. Li, W. Zhao, X. Wang, and X. Qiu, “ACT-R cognitive model based trajectory planning method study for electric vehicle’s active obstacle avoidance system,” Energies, vol. 1, no. 1, p. 75, 2018.
    [19]
    J. E. Laird, C. Lebiere, and P. S. Rosenbloom, “A standard model of the mind: Toward a common computational framework across artificial intelligence, cognitive science, neuroscience, and robotics,” AI Magazine, vol. 38, no. 4, pp. 13–26, 2017. doi: 10.1609/aimag.v38i4.2744
    [20]
    C. L. Dancy, “A hybrid cognitive architecture with primal affect and physiology,” IEEE Trans. Affective Computing, vol. 12, no. 2, pp. 318–328, 2019.
    [21]
    S. Ko, Y. Zhang, and M. Jeon, “Modeling the effects of auditory display takeover requests on drivers’ behavior in autonomous vehicles,” in Proc. 11th Int. Conf. Automotive User Interfaces and Interactive Vehicular Applications: Adjunct Proceedings, 2019, pp. 392–398.
    [22]
    Y. Zhang, C. Wu, C. Qiao, A. Sadek, and K. F. Hulme, “A cognitive computational model of driver warning response performance in connected vehicle systems,” IEEE Trans. Intelligent Transportation Systems, vol. 23, no. 9, pp. 14790–14805, 2022. doi: 10.1109/TITS.2021.3134058
    [23]
    G. Bagschik, T. Menzel, and M. Maurer, “Ontology based scene creation for the development of automated vehicles,” in Proc. IEEE Intelligent Vehicles Symposium, 2018, pp. 1813–1820.
    [24]
    F. Hauer, T. Schmidt, B. Holzmüller, and A. Pretschner, “Did we test all scenarios for automated and autonomous driving systems?” in Proc. IEEE Intelligent Transportation Systems Conf., 2019, pp. 2950–2955.
    [25]
    D. U. Ozsahin, B. Uzun, I. Ozsahin, M. T. Mustapha, and M. S. Musa, “Fuzzy logic in medicine,” in Proc. Biomedical Signal Processing and Artificial Intelligence in Healthcare, 2020, pp. 153–182.
    [26]
    A. Indahingwati, M. Barid, N. Wajdi, D. Susilo, N. Kurniasih, and R. Rahim, “Comparison analysis of topsis and fuzzy logic methods on fertilizer selection,” Int. J. Eng. Technol, vol. 7, no. 2.3, pp. 109–114, 2018. doi: 10.14419/ijet.v7i2.3.12630
    [27]
    R. Jafari, M. A. Contreras, W. Yu, and A. Gegov, “Applications of fuzzy logic, artificial neural network and neuro-fuzzy in industrial engineering,” in Proc. Latin American Symp. Industrial and Robotic Systems, 2019, pp. 9–14.
    [28]
    J. M. Garibaldi, “The need for fuzzy AI,” IEEE/CAA J. Autom. Sinica, vol. 6, no. 3, pp. 610–622, 2019. doi: 10.1109/JAS.2019.1911465
    [29]
    X. Zhao, H. Mo, K. Yan, and L. Li, “Type-2 fuzzy control for driving state and behavioral decisions of unmanned vehicle,” IEEE/CAA J. Autom. Sinica, vol. 7, no. 1, pp. 178–186, 2019.
    [30]
    C. Chen, X. Liu, H.-H. Chen, M. Li, and L. Zhao, “A rear-end collision risk evaluation and control scheme using a bayesian network model,” IEEE Trans. Intelligent Transportation Systems, vol. 20, no. 1, pp. 264–284, 2018.
    [31]
    T. Makaba, W. Doorsamy, and B. S. Paul, “Bayesian network-based framework for cost-implication assessment of road traffic collisions,” Int. J. Intelligent Transportation Systems Research, vol. 19, pp. 240–253, 2021. doi: 10.1007/s13177-020-00242-1
    [32]
    S. Noh and K. An, “Decision-making framework for automated driving in highway environments,” IEEE Trans. Intelligent Transportation Systems, vol. 19, no. 1, pp. 58–71, 2017.
    [33]
    C. Hubmann, M. Becker, D. Althoff, D. Lenz, and C. Stiller, “Decision making for autonomous driving considering interaction and uncertain prediction of surrounding vehicles,” in Proc. IEEE Intelligent Vehicles Symp., 2017, pp. 1671–1678.
    [34]
    C. Zhang, S. Ma, M. Wang, G. Hinz, and A. Knoll, “Efficient pomdp behavior planning for autonomous driving in dense urban environments using multi-step occupancy grid maps,” in Proc. IEEE 25th Int. Conf. Intelligent Transportation Systems, 2022, pp. 2722–2729.
    [35]
    Z. Qiao, K. Muelling, J. Dolan, P. Palanisamy, and P. Mudalige, “Pomdp and hierarchical options MDP with continuous actions for autonomous driving at intersections,” in Proc. 21st Int. Conf. Intelligent Transportation Systems, 2018, pp. 2377–2382.
    [36]
    A. G. Cunningham, E. Galceran, D. Mehta, G. Ferrer, R. M. Eustice, and E. Olson, “MPDM: Multi-policy decision-making from autonomous driving to social robot navigation,” Control Strategies for Advanced Driver Assistance Systems and Autonomous Driving Functions: Development, Testing and Verification, pp. 201–223, 2019.
    [37]
    T. Nishi, P. Doshi, and D. Prokhorov, “Merging in congested freeway traffic using multipolicy decision making and passive actor-critic learning,” IEEE Trans. Intelligent Vehicles, vol. 4, no. 2, pp. 287–297, 2019. doi: 10.1109/TIV.2019.2904417
    [38]
    H. Yu, H. E. Tseng, and R. Langari, “A human-like game theory-based controller for automatic lane changing,” Transportation Research Part C: Emerging Technologies, vol. 88, pp. 140–158, 2018. doi: 10.1016/j.trc.2018.01.016
    [39]
    X. Na and D. J. Cole, “Modelling of a human driver’s interaction with vehicle automated steering using cooperative game theory,” IEEE/CAA J. Autom. Sinica, vol. 6, no. 5, pp. 1095–1107, 2019. doi: 10.1109/JAS.2019.1911675
    [40]
    A. Ji and D. Levinson, “A review of game theory models of lane changing,” Transportmetrica A: Transport Science, vol. 16, no. 3, pp. 1628–1647, 2020. doi: 10.1080/23249935.2020.1770368
    [41]
    B. R. Kiran, I. Sobh, V. Talpaert, P. Mannion, A. A. A. Sallab, S. Yogamani, and P. Pérez, “Deep reinforcement learning for autonomous driving: A survey,” IEEE Trans. Intelligent Transportation Systems, vol. 23, no. 6, pp. 4909–4926, 2022. doi: 10.1109/TITS.2021.3054625
    [42]
    V. Mnih, K. Kavukcuoglu, D. Silver, A. Graves, I. Antonoglou, D. Wierstra, and M. Riedmiller, “Playing atari with deep reinforcement learning,” 2013. [Online]. Available: https://arxiv.org/abs/1312.5602.
    [43]
    T. P. Lillicrap, J. J. Hunt, A. Pritzel, N. Heess, T. Erez, Y. Tassa, D. Silver, and D. Wierstra, “Continuous control with deep reinforcement learning,” 2015. [Online]. Available: https://arxiv.org/abs/1509.02971.
    [44]
    A. Kendall, J. Hawke, D. Janz, P. Mazur, D. Reda, J.-M. Allen, V.-D. Lam, A. Bewley, and A. Shah, “Learning to drive in a day,” in Proc. Int. Conf. Robotics and Automation, 2019, pp. 8248–8254.
    [45]
    Z. Huang, J. Zhang, R. Tian, and Y. Zhang, “End-to-end autonomous driving decision based on deep reinforcement learning,” in Proc. 5th Int. Conf. Control, Automation and Robotics, 2019, pp. 658–662.
    [46]
    T. M. Moerland, J. Broekens, A. Plaat, C. M. Jonker, et al., “Modelbased reinforcement learning: A survey,” Foundations and Trends® in Machine Learning, vol. 16, no. 1, pp. 1–118, 2023.
    [47]
    A. E. Sallab, M. Abdou, E. Perot, and S. Yogamani, “Deep reinforcement learning framework for autonomous driving,” Electronic Imaging, vol. 2017, no. 19, pp. 70–76, 2017.
    [48]
    D. Silver, S. Singh, D. Precup, and R. S. Sutton, “Reward is enough,” Artificial Intelligence, vol. 299, p. 103535, 2021. doi: 10.1016/j.artint.2021.103535
    [49]
    M. Bojarski, D. Del Testa, D. Dworakowski, B. Firner, B. Flepp, P. Goyal, L. D. Jackel, M. Monfort, U. Muller, J. Zhang, X. Zhang, J. Zhao, and K. Zieba, “End to end learning for self-driving cars,” 2016. [Online]. Available: https://arxiv.org/abs/1604.07316.
    [50]
    H. Xu, Y. Gao, F. Yu, and T. Darrell, “End-to-end learning of driving models from large-scale video datasets,” in Proc. IEEE Conf. Computer Vision and Pattern Recognition, 2017, pp. 2174–2182.
    [51]
    F. Codevilla, M. Miiller, A. López, V. Koltun, and A. Dosovitskiy, “End-to-end driving via conditional imitation learning,” in Proc. IEEE Int. Conf. Robotics and Automation, 2018, pp. 1–9.
    [52]
    D. Kishikawa and S. Arai, “Comfortable driving by using deep inverse reinforcement learning,” in Proc. IEEE Int. Conf. Agents, 2019, pp. 38–43.
    [53]
    Z. Wu, L. Sun, W. Zhan, C. Yang, and M. Tomizuka, “Efficient sampling-based maximum entropy inverse reinforcement learning with application to autonomous driving,” IEEE Robotics and Automation Letters, vol. 5, no. 4, pp. 5355–5362, 2020. doi: 10.1109/LRA.2020.3005126
    [54]
    Y. Jiang, W. Deng, J. Wang, and B. Zhu, “Studies on drivers’ driving styles based on inverse reinforcement learning,” SAE Technical Paper,Tech. Rep., 2018.
    [55]
    K. Lee, D. Isele, E. A. Theodorou, and S. Bae, “Spatiotemporal costmap inference for mpc via deep inverse reinforcement learning,” IEEE Robotics and Automation Letters, vol. 7, no. 2, pp. 3194–3201, 2022. doi: 10.1109/LRA.2022.3146635
    [56]
    J. Ho and S. Ermon, “Generative adversarial imitation learning,” in Advances in Neural Information Processing Systems, vol. 29, 2016.
    [57]
    A. Kuefler, J. Morton, T. Wheeler, and M. Kochenderfer, “Imitating driver behavior with generative adversarial networks,” in Proc. IEEE Intelligent Vehicles Symp., 2017, pp. 204–211.
    [58]
    R. P. Bhattacharyya, D. J. Phillips, B. Wulfe, J. Morton, A. Kuefler, and M. J. Kochenderfer, “Multi-agent imitation learning for driving simulation,” in Proc. IEEE/RSJ Int. Conf. Intelligent Robots and Systems, 2018, pp. 1534–1539.
    [59]
    R. P. Bhattacharyya, D. J. Phillips, C. Liu, J. K. Gupta, K. DriggsCampbell, and M. J. Kochenderfer, “Simulating emergent properties of human driving behavior using multi-agent reward augmented imitation learning,” in Proc. Int. Conf. Robotics and Automation, 2019, pp. 789–795.
    [60]
    S.-H. Bae, S.-H. Joo, J.-W. Pyo, J.-S. Yoon, K. Lee, and T.-Y. Kuc, “Finite state machine based vehicle system for autonomous driving in urban environments,” in Proc. 20th Int. Conf. Control, Automation and Systems, 2020, pp. 1181– 1186.
    [61]
    N. D. Van, M. Sualeh, D. Kim, and G.-W. Kim, “A hierarchical control system for autonomous driving towards urban challenges,” Applied Sciences, vol. 10, no. 10, p. 3543, 2020. doi: 10.3390/app10103543
    [62]
    Y. Hu, L. Yan, J. Zhan, F. Yan, Z. Yin, F. Peng, and Y. Wu, “Decisionmaking system based on finite state machine for low-speed autonomous vehicles in the park,” in Proc. IEEE Int. Conf. Realtime Computing and Robotics, 2022, pp. 721–726.
    [63]
    S. Coskun and R. Langari, “Predictive fuzzy Markov decision strategy for autonomous driving in highways,” in Proc. IEEE Conf. Control Technology and Applications, 2018, pp. 1032–1039.
    [64]
    L. Claussmann, M. O’Brien, S. Glaser, H. Najjaran, and D. Gruyer, “Multi-criteria decision making for autonomous vehicles using fuzzy dempster-shafer reasoning,” in Proc. IEEE Intelligent Vehicles Symposium, 2018, pp. 2195–2202.
    [65]
    Q. Wu, S. Cheng, L. Li, F. Yang, L. J. Meng, Z. X. Fan, and H. W. Liang, “A fuzzy-inference-based reinforcement learning method of overtaking decision making for automated vehicles,” Proc. Institution of Mechanical Engineers,Part D: J. Automobile Engineering, vol. 236, no. 1, pp. 75–83, 2022. doi: 10.1177/09544070211018099
    [66]
    US Highway 101 Dataset, Next Generation Simulation Program (NGSIM), 2007. [Online]. Available: https://www.fhwa.dot.gov/publications/research/operations/07030/index.cfm.
    [67]
    R. Krajewski, J. Bock, L. Kloeker, and L. Eckstein, “The highd dataset: A drone dataset of naturalistic vehicle trajectories on german highways for validation of highly automated driving systems,” in Proc. 21st Int. Conf. Intelligent Transportation Systems, 2018, pp. 2118–2125.
    [68]
    J. Houston, G. Zuidhof, L. Bergamini, Y. Ye, A. Jain, S. Omari, V. Iglovikov, and P. Ondruska, “One thousand and one hours: Self-driving motion prediction dataset.” [Online]. Available: https://level-5.global/level5/data/.
    [69]
    W. Zhan, L. Sun, D. Wang, H. Shi, A. Clausse, M. Naumann, J. Kummerle, H. Konigshof, C. Stiller, A. de La Fortelle, and M. Tomizuka, “Interaction dataset: An international, adversarial and cooperative motion dataset in interactive driving scenarios with semantic maps,” 2019. [Online]. Available: https://arxiv.org/abs/1910.03088.
    [70]
    V. Ramanishka, Y.-T. Chen, T. Misu, and K. Saenko, “Toward driving scene understanding: A dataset for learning driver behavior and causal reasoning,” in Proc. IEEE Conf. Computer Vision and Pattern Recognition, 2018, pp. 7699–7707.
    [71]
    M.-F. Chang, J. Lambert, P. Sangkloy, J. Singh, S. Bak, A. Hartnett, D. Wang, P. Carr, S. Lucey, D. Ramanan, et al., “Argoverse: 3d tracking and forecasting with rich maps,” in Proc. IEEE/CVF Conf. Computer Vision and Pattern Recognition, 2019, pp. 8748–8757.
    [72]
    M. Althoff, M. Koschi, and S. Manzinger, “Commonroad: Composable benchmarks for motion planning on roads,” in Proc. IEEE Intelligent Vehicles Symposium, 2017, pp. 719–726.
    [73]
    H. Caesar, J. Kabzan, K. S. Tan, W. K. Fong, E. Wolff, A. Lang, L. Fletcher, O. Beijbom, and S. Omari, “Nuplan: A closed-loop ML-based planning benchmark for autonomous vehicles,” 2021. [Online]. Available: https://arxiv.org/abs/2106.11810.
    [74]
    B. Wymann, E. Espié, C. Guionneau, C. Dimitrakakis, R. Coulom, and A. Sumner, “Torcs, the open racing car simulator,” 2000. [Online]. Available: http://torcs.sourceforge.net.
    [75]
    R. Bhattacharyya, B. Wulfe, D. Phillips, A. Kuefler, J. Morton, R. Senanayake, and M. Kochenderfer, “Modeling human driving behavior through generative adversarial imitation learning,” 2020. [Online]. Available: https://arxiv.org/abs/2006.06412.
    [76]
    A. Dosovitskiy, G. Ros, F. Codevilla, A. Lopez, and V. Koltun, “Carla: An open urban driving simulator,” in Proc. Conf. Robot Learning, 2017, pp. 1–16.
    [77]
    Mechanical Simulation Corporation. “Carsim software,” 1997. [Online]. Available: https://www.carsim.com/.
    [78]
    Y. Li, H. Deng, X. Xu, and W. Wang, “Modelling and testing of inwheel motor drive intelligent electric vehicles based on co-simulation with Carsim/Simulink,” IET Intelligent Transport Systems, vol. 13, no. 1, pp. 115–123, 2018.
    [79]
    N. Koenig and A. Howard, “Design and use paradigms for Gazebo, an open-source multi-robot simulator,” in Proc. IEEE/RSJ Int. Conf. Intelligent Robots and Systems, 2004, vol. 3, pp. 2149–2154.
    [80]
    M. Quigley, K. Conley, B. Gerkey, J. Faust, T. Foote, J. Leibs, R. Wheeler, A. Y. Ng, et al., “ROS: An open-source robot operating system,” in Proc. ICRA Workshop on Open Source Software, 2009.
    [81]
    Y. Pan, C.-A. Cheng, K. Saigol, K. Lee, X. Yan, E. Theodorou, and B. Boots, “Agile autonomous driving using end-to-end deep imitation learning,” 2017. [Online]. Available: https://arxiv.org/abs/1709.07174.
    [82]
    W. Li, D. Wolinski, and M. C. Lin, “Adaps: Autonomous driving via principled simulations,” in Proc. Int. Conf. Robotics and Automation, 2019, pp. 7625–7631.
    [83]
    J. Huang, Y. Chen, X. Peng, L. Hu, and D. Cao, “Study on the driving style adaptive vehicle longitudinal control strategy,” IEEE/CAA J. Autom. Sinica, vol. 7, no. 4, pp. 1107–1115, 2020. doi: 10.1109/JAS.2020.1003261
    [84]
    L. Li, K. Ota, and M. Dong, “Humanlike driving: Empirical decisionmaking system for autonomous vehicles,” IEEE Trans. Vehicular Technology, vol. 67, no. 8, pp. 6814–6823, 2018. doi: 10.1109/TVT.2018.2822762
    [85]
    S. Chen, S. Zhang, J. Shang, B. Chen, and N. Zheng, “Brain-inspired cognitive model with attention for self-driving cars,” IEEE Trans. Cognitive and Developmental Systems, vol. 11, no. 1, pp. 13–25, 2017.
    [86]
    A. Riegler, A. Riener, and C. Holzmann, “A research agenda for mixed reality in automated vehicles,” in Proc. 19th Int. Conf. Mobile and Ubiquitous Multimedia, 2020, pp. 119–131.
    [87]
    A. Pillai, “Virtual reality based study to analyse pedestrian attitude towards autonomous vehicles,” M.S. thesis, Aalto University. School of Science, 2017. [Online]. Available: http://urn.fi/URN:NBN:fi:aalto-201710307409.
    [88]
    D. Goedicke, A. W. Bremers, S. Lee, F. Bu, H. Yasuda, and W. Ju, “Xroom: Mixed reality driving simulation with real cars for research and design,” in Proc. CHI Conf. Human Factors in Computing Systems, 2022, pp. 1–13.
    [89]
    C. Gkartzonikas and K. Gkritza, “What have we learned? A review of stated preference and choice studies on autonomous vehicles,” Transportation Research Part C: Emerging Technologies, vol. 98, pp. 323–337, 2019. doi: 10.1016/j.trc.2018.12.003
    [90]
    C. Huang, C. Lv, P. Hang, and Y. Xing, “Toward safe and personalized autonomous driving: Decision-making and motion control with DPF and CDT techniques,” IEEE/ASME Trans. Mechatronics, vol. 26, no. 2, pp. 611–620, 2021. doi: 10.1109/TMECH.2021.3053248
    [91]
    Z. Deng, D. Chu, C. Wu, S. Liu, C. Sun, T. Liu, and D. Cao, “A probabilistic model for driving-style-recognition-enabled driver steering behaviors,” IEEE Trans. Systems,Man,and Cybernetics: Systems, vol. 52, no. 3, pp. 1838–1851, 2022. doi: 10.1109/TSMC.2020.3037229
    [92]
    B. Zhu, Y. Jiang, J. Zhao, R. He, N. Bian, and W. Deng, “Typical-driving-style-oriented personalized adaptive cruise control design based on human driving data,” Transportation Research Part C: Emerging Technologies, vol. 100, pp. 274–288, 2019. doi: 10.1016/j.trc.2019.01.025
    [93]
    I. Bae, J. Moon, J. Jhung, H. Suk, T. Kim, H. Park, J. Cha, J. Kim, D. Kim, and S. Kim, “Self-driving like a human driver instead of a robocar: Personalized comfortable driving experience for autonomous vehicles,” 2020. [Online]. Available: https://arxiv.org/abs/2001.03908.
    [94]
    L. Oliveira, K. Proctor, C. G. Burns, and S. Birrell, “Driving style: How should an automated vehicle behave?” Information, vol. 10, no. 6, p. 219, 2019. doi: 10.3390/info10060219
    [95]
    T. Miller, “Explanation in artificial intelligence: Insights from the social sciences,” Artificial Intelligence, vol. 267, pp. 1–38, 2019. doi: 10.1016/j.artint.2018.07.007
    [96]
    E. Rehder, J. Quehl, and C. Stiller, “Driving like a human: Imitation learning for path planning using convolutional neural networks,” in Proc. Int. Conf. Robotics and Automation Workshops, 2017.
    [97]
    S. Chen, Z. Jian, Y. Huang, Y. Chen, Z. Zhou, and N. Zheng, “Autonomous driving: Cognitive construction and situation understanding,” Science China Information Sciences, vol. 62, no. 8, pp. 1–27, 2019.
    [98]
    Y. Chen, S. Chen, T. Zhang, S. Zhang, and N. Zheng, “Autonomous vehicle testing and validation platform: Integrated simulation system with hardware in the loop,” in Proc. IEEE Intelligent Vehicles Symp., 2018, pp. 949–956.
    [99]
    D. Silver, A. Huang, C. J. Maddison, A. Guez, L. Sifre, G. Van Den Driessche, J. Schrittwieser, I. Antonoglou, V. Panneershelvam, M. Lanctot, et al., “Mastering the game of go with deep neural networks and tree search,” Nature, vol. 529, no. 7587, pp. 484–489, 2016. doi: 10.1038/nature16961
    [100]
    D. Silver, J. Schrittwieser, K. Simonyan, I. Antonoglou, A. Huang, A. Guez, T. Hubert, L. Baker, M. Lai, A. Bolton, et al., “Mastering the game of go without human knowledge,” Nature, vol. 550, no. 7676, pp. 354–359, 2017. doi: 10.1038/nature24270
    [101]
    D. Silver, T. Hubert, J. Schrittwieser, I. Antonoglou, M. Lai, A. Guez, M. Lanctot, L. Sifre, D. Kumaran, T. Graepel, et al., “A general reinforcement learning algorithm that masters chess, shogi, and go through self-play,” Science, vol. 362, no. 6419, pp. 1140–1144, 2018. doi: 10.1126/science.aar6404
    [102]
    K. Shao, Z. Tang, Y. Zhu, N. Li, and D. Zhao, “A survey of deep reinforcement learning in video games,” arXiv preprint arXiv: 1912.10944, 2019.
    [103]
    D. Ye, Z. Liu, M. Sun, B. Shi, P. Zhao, H. Wu, H. Yu, S. Yang, X. Wu, Q. Guo, et al., “Mastering complex control in moba games with deep reinforcement learning,” in Proc. AAAI Conf. Artificial Intelligence, 2020, vol. 34, no. 4, pp. 6672–6679.
    [104]
    E. Alonso, M. Peter, D. Goumard, and J. Romoff, “Deep reinforcement learning for navigation in AAA video games,” arXiv preprint arXiv: 2011.04764, 2020.
    [105]
    J. Wang, Y. Wang, D. Zhang, Y. Yang, and R. Xiong, “Learning hierarchical behavior and motion planning for autonomous driving,” in Proc. IEEE/RSJ Int. Conf. Intelligent Robots and Systems, 2020, pp. 2235–2242.
  • Related Articles

    [1]Luigi D’Alfonso, Francesco Giannini, Giuseppe Franzè, Giuseppe Fedele, Francesco Pupo, Giancarlo Fortino. Autonomous Vehicle Platoons In Urban Road Networks: A Joint Distributed Reinforcement Learning and Model Predictive Control Approach[J]. IEEE/CAA Journal of Automatica Sinica, 2024, 11(1): 141-156. doi: 10.1109/JAS.2023.123705
    [2]Zirui Chen, Jingchuan Tang, Zongyu Zuo. A Novel Prescribed-Performance Path-Following Problem for Non-Holonomic Vehicles[J]. IEEE/CAA Journal of Automatica Sinica, 2024, 11(6): 1476-1484. doi: 10.1109/JAS.2024.124311
    [3]Xuerao Wang, Qingling Wang, Yanxu Su, Yuncheng Ouyang, Changyin Sun. Adaptive Sensor-Fault Tolerant Control of Unmanned Underwater Vehicles With Input Saturation[J]. IEEE/CAA Journal of Automatica Sinica, 2024, 11(4): 907-918. doi: 10.1109/JAS.2023.123837
    [4]Shengnan Gao, Zhouhua Peng, Haoliang Wang, Lu Liu, Dan Wang. Long Duration Coverage Control of Multiple Robotic Surface Vehicles Under Battery Energy Constraints[J]. IEEE/CAA Journal of Automatica Sinica, 2024, 11(7): 1695-1698. doi: 10.1109/JAS.2023.123438
    [5]Li-Ying Hao, Gege Dong, Tieshan Li, Zhouhua Peng. Path-Following Control With Obstacle Avoidance of Autonomous Surface Vehicles Subject to Actuator Faults[J]. IEEE/CAA Journal of Automatica Sinica, 2024, 11(4): 956-964. doi: 10.1109/JAS.2023.123675
    [6]Zhiqiang Pu, Yi Pan, Shijie Wang, Boyin Liu, Min Chen, Hao Ma, Yixiong Cui. Orientation and Decision-Making for Soccer Based on Sports Analytics and AI: A Systematic Review[J]. IEEE/CAA Journal of Automatica Sinica, 2024, 11(1): 37-57. doi: 10.1109/JAS.2023.123807
    [7]Xiaohua Ge, Qing-Long Han, Qing Wu, Xian-Ming Zhang. Resilient and Safe Platooning Control of Connected Automated Vehicles Against Intermittent Denial-of-Service Attacks[J]. IEEE/CAA Journal of Automatica Sinica, 2023, 10(5): 1234-1251. doi: 10.1109/JAS.2022.105845
    [8]Wen Hu, Zejian Deng, Dongpu Cao, Bangji Zhang, Amir Khajepour, Lei Zeng, Yang Wu. Probabilistic Lane-Change Decision-Making and Planning for Autonomous Heavy Vehicles[J]. IEEE/CAA Journal of Automatica Sinica, 2022, 9(12): 2161-2173. doi: 10.1109/JAS.2022.106049
    [9]Jing Huang, Yimin Chen, Xiaoyan Peng, Lin Hu, Dongpu Cao. Study on the Driving Style Adaptive Vehicle Longitudinal Control Strategy[J]. IEEE/CAA Journal of Automatica Sinica, 2020, 7(4): 1107-1115. doi: 10.1109/JAS.2020.1003261
    [10]Parham M. Kebria, Abbas Khosravi, Syed Moshfeq Salaken, Saeid Nahavandi. Deep Imitation Learning for Autonomous Vehicles Based on Convolutional Neural Networks[J]. IEEE/CAA Journal of Automatica Sinica, 2020, 7(1): 82-95. doi: 10.1109/JAS.2019.1911825
    [11]Wenshuo Wang, Xiaoxiang Na, Dongpu Cao, Jianwei Gong, Junqiang Xi, Yang Xing, Fei-Yue Wang. Decision-Making in Driver-Automation Shared Control: A Review and Perspectives[J]. IEEE/CAA Journal of Automatica Sinica, 2020, 7(5): 1289-1307. doi: 10.1109/JAS.2020.1003294
    [12]Chen Sun, Jean M. Uwabeza Vianney, Ying Li, Long Chen, Li Li, Fei-Yue Wang, Amir Khajepour, Dongpu Cao. Proximity Based Automatic Data Annotation for Autonomous Driving[J]. IEEE/CAA Journal of Automatica Sinica, 2020, 7(2): 395-404. doi: 10.1109/JAS.2020.1003033
    [13]Long Chen, Xuemin Hu, Wei Tian, Hong Wang, Dongpu Cao, Fei-Yue Wang. Parallel Planning: A New Motion Planning Framework for Autonomous Driving[J]. IEEE/CAA Journal of Automatica Sinica, 2019, 6(1): 236-246. doi: 10.1109/JAS.2018.7511186
    [14]Wenchao Xu, Haibo Zhou, Nan Cheng, Feng Lyu, Weisen Shi, Jiayin Chen, Xuemin (Sherman) Shen. Internet of Vehicles in Big Data Era[J]. IEEE/CAA Journal of Automatica Sinica, 2018, 5(1): 19-35. doi: 10.1109/JAS.2017.7510736
    [15]Jiacai Huang, YangQuan Chen, Haibin Li, Xinxin Shi. Fractional Order Modeling of Human Operator Behavior with Second Order Controlled Plant and Experiment Research[J]. IEEE/CAA Journal of Automatica Sinica, 2016, 3(3): 271-280.
    [16]Zhaolei Wang, Qing Wang, Chaoyang Dong. Asynchronous H Control for Unmanned Aerial Vehicles: Switched Polytopic System Approach[J]. IEEE/CAA Journal of Automatica Sinica, 2015, 2(2): 207-216.
    [17]Chaoxu Mu, Qun Zong, Bailing Tian, Wei Xu. Continuous Sliding Mode Controller with Disturbance Observer for Hypersonic Vehicles[J]. IEEE/CAA Journal of Automatica Sinica, 2015, 2(1): 45-55.
    [18]Zixuan Liang, Zhang Ren, Xingyue Shao. Decoupling Trajectory Tracking for Gliding Reentry Vehicles[J]. IEEE/CAA Journal of Automatica Sinica, 2015, 2(1): 115-120.
    [19]Derong Liu, Changyin Sun, Bin Xian. Guest Editorial for Special Issue on Autonomous Control of Unmanned Aerial Vehicles[J]. IEEE/CAA Journal of Automatica Sinica, 2015, 2(1): 1-1.
    [20]Ruimin Zhang, Lu Dong, Changyin Sun. Adaptive Nonsingular Terminal Sliding Mode Control Design for Near Space Hypersonic Vehicles[J]. IEEE/CAA Journal of Automatica Sinica, 2014, 1(2): 155-161.

Catalog

    通讯作者: 陈斌, bchen63@163.com
    • 1. 

      沈阳化工大学材料科学与工程学院 沈阳 110142

    1. 本站搜索
    2. 百度学术搜索
    3. 万方数据库搜索
    4. CNKI搜索

    Figures(5)  / Tables(3)

    Article Metrics

    Article views (1577) PDF downloads(399) Cited by()

    Highlights

    • In recent years, there have been many works on autonomous driving decision-making. However, there hasn't been a comprehensive review from a "human-like" perspective. This paper is the first review that provides a comprehensive overview of human-like decision-making for autonomous vehicles
    • In this paper, some original issues are discussed: 1) The intelligence level of most autonomous driving decision-making algorithms; 2) The driving datasets and simulation platforms for testing and verifying human-like decision-making; 3) The evaluation metrics of human-likeness; personalized driving; the application of decision-making in real traffic scenarios; and 4) The potential research direction of human-like driving
    • It is very important to improve the ability of the decision-making system to build a reasonable driving model so that the autonomous vehicle can learn human expert knowledge and driving habits. The research results of this paper are significant for creating interpretable human-like driving models and applying them in dynamic traffic scenarios

    /

    DownLoad:  Full-Size Img  PowerPoint
    Return
    Return