liu.seSearch for publications in DiVA
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
A Novel Energy Management Strategy for PHEV Considering Cabin Heat Demand Under Low Temperature Based on Reinforcement Learning
Jilin Univ, Peoples R China; Jilin Univ Campus Nanling, Peoples R China.
Jilin Univ, Peoples R China; Tongji Univ, Peoples R China.
Jilin Univ, Peoples R China; Jilin Univ Campus Nanling, Peoples R China.
Linköping University, Department of Electrical Engineering, Vehicular Systems. Linköping University, Faculty of Science & Engineering.ORCID iD: 0000-0001-8646-8998
Show others and affiliations
2025 (English)In: IEEE TRANSACTIONS ON TRANSPORTATION ELECTRIFICATION, ISSN 2332-7782, Vol. 11, no 1, p. 3062-3077Article in journal (Refereed) Published
Abstract [en]

The fuel economy of hybrid electric vehicles (HEVs) deteriorates due to high cabin heat demands and engine efficiency decay caused by temperature drops in low-temperature environments. This article proposes a novel energy management strategy based on deep reinforcement learning (RL) for plug-in HEVs (PHEVs) under cold environments. A series HEV model is first presented, and the coupling relationship between the engine-cabin-coupled thermal management system (CTMS) and the energy management system is analyzed. Considering the influence of the engine coolant temperature on the fuel consumption rate and cabin heat demand, an energy management optimal control problem is developed. An efficient RL framework based on a double-deep Q-learning (DDQL) algorithm is designed to solve the problem. The method does not rely on precise models and future traffic information. Dynamic programming (DP) and model predictive control (MPC) methods are also employed and compared with the proposed method under the training driving cycle and driving cycles randomly generated by the Markov-driver model. Experimental results show that the proposed method has efficient performance in maintaining battery state of charge (SOC) stability, real-time performance, fuel economy, and adaptability. The fuel economy can reach the 97.5% level of the DP-based strategy, which is approximately 5.9% higher than that of the MPC-based strategy.

Place, publisher, year, edition, pages
IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC , 2025. Vol. 11, no 1, p. 3062-3077
Keywords [en]
Heat engines; Energy management; Resistance heating; Batteries; Waste heat; Heat pumps; Fuel economy; Cabin heat demand; deep reinforcement learning (RL); energy management system; hybrid electric vehicles (HEVs); low-temperature environment; real time
National Category
Energy Engineering
Identifiers
URN: urn:nbn:se:liu:diva-211715DOI: 10.1109/TTE.2024.3434521ISI: 001416183200020Scopus ID: 2-s2.0-85200246853OAI: oai:DiVA.org:liu-211715DiVA, id: diva2:1938410
Note

Funding Agencies|Major Science and Technology Project of Jilin Province [20220301010GX]; International Scientific and Technological Cooperation [20240402071GH]

Available from: 2025-02-18 Created: 2025-02-18 Last updated: 2025-02-18

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Search in DiVA

By author/editor
Eriksson, Lars
By organisation
Vehicular SystemsFaculty of Science & Engineering
Energy Engineering

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 42 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf