PlumX Metrics
Embed PlumX Metrics

Attentive hybrid reinforcement learning-based eco-driving strategy for connected vehicles with hybrid action spaces and surrounding vehicles attention

Energy Conversion and Management, ISSN: 0196-8904, Vol: 321, Page: 119059
2024
  • 1
    Citations
  • 0
    Usage
  • 3
    Captures
  • 1
    Mentions
  • 0
    Social Media
Metric Options:   Counts1 Year3 Year

Metrics Details

  • Citations
    1
  • Captures
    3
  • Mentions
    1
    • News Mentions
      1
      • 1

Most Recent News

Researchers from Yanshan University Detail Findings in Energy Conversion and Management (Attentive Hybrid Reinforcement Learning-based Eco-driving Strategy for Connected Vehicles With Hybrid Action Spaces and Surrounding Vehicles Attention)

2024 DEC 02 (NewsRx) -- By a News Reporter-Staff News Editor at Energy Daily News -- Data detailed on Energy - Energy Conversion and Management

Article Description

In environments characterized by complex multi-source traffic information, the interaction between the ego vehicle and surrounding vehicles, along with behavioral interference, present fundamental challenges to the stability of eco-driving strategies. Therefore, an integrated action eco-driving strategy is proposed based on attention mechanism to explore strategy differences under varying interaction relationship with surrounding vehicles. Specifically, within the eco-driving strategy framework of the twin delayed deep deterministic policy gradient (TD3) algorithm, a multi-head self-attention module is introduced for feature extraction from traffic flow information. A hybrid action representation mechanism is assimilated to cooperate longitudinal acceleration control and lateral lane-change decisions, integrating attention output features to stabilize strategy shifts. Results indicate that the proposed strategy outperforms parameterized action deep deterministic policy gradient strategies (PADDPG) and discrete attention-based deep Q network eco-driving strategies (Attentive-DQN) under adversarial reward functions, achieving a 42.18% stability improvement over PADDPG and an 84.79% time stability improvement over Attentive-DQN. Compared to vehicles using the Krauss car-following model and LC2013 lane-changing model, the proposed strategy achieves a 30.25% energy optimization with only a 1.01% loss in time efficiency. Additionally, as behavioral interference factors are emphasized in strategy objectives, the parameter distribution of the attention module changes to enhance the ego vehicle’s interaction with interfering objects. This indicates that considering surrounding vehicle dynamics in strategy execution shapes different interaction relationships.

Provide Feedback

Have ideas for a new metric? Would you like to see something else here?Let us know