TY - GEN
T1 - Beamspace and Frequency Domain ISAC Resource Allocation Using Reinforcement Learning
AU - Pulkkinen, Petteri
AU - Esfandiari, Majdoddin
AU - Koivunen, Visa
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2025
Y1 - 2025
N2 - In the emerging 6G systems, joint sensing and communications require efficient spatial and frequency domain waveform design and resource allocation to meet system performance objectives. Traditional structured optimization algorithms often face practical challenges such as non-convexity, high computational demand, lack of adaptability, and significant performance degradation when the assumed model is invalid or dynamic. Reinforcement learning (RL) offers a data-driven alternative that leverages observed data to overcome these deficits. This paper combines RL principles with acquired model awareness of radio environment dynamics, thereby enhancing the learning data efficiency and making the algorithm more interpretable than traditional model-free RL methods. We introduce an RL algorithm based on Thompson sampling for allocating sub carriers and beams in the beamspace domain to achieve desired performance levels in wireless communications and radar sensing tasks. The proposed RL method effectively learns from its experiences and balances sensing and communications functionalities, achieving superior target detection performance and competitive communication rates compared to traditional methods.
AB - In the emerging 6G systems, joint sensing and communications require efficient spatial and frequency domain waveform design and resource allocation to meet system performance objectives. Traditional structured optimization algorithms often face practical challenges such as non-convexity, high computational demand, lack of adaptability, and significant performance degradation when the assumed model is invalid or dynamic. Reinforcement learning (RL) offers a data-driven alternative that leverages observed data to overcome these deficits. This paper combines RL principles with acquired model awareness of radio environment dynamics, thereby enhancing the learning data efficiency and making the algorithm more interpretable than traditional model-free RL methods. We introduce an RL algorithm based on Thompson sampling for allocating sub carriers and beams in the beamspace domain to achieve desired performance levels in wireless communications and radar sensing tasks. The proposed RL method effectively learns from its experiences and balances sensing and communications functionalities, achieving superior target detection performance and competitive communication rates compared to traditional methods.
KW - frequency resource allocation
KW - Integrated sensing and communications (ISAC)
KW - reinforcement learning (RL)
KW - spatial resource optimization
KW - Thompson sampling
UR - http://www.scopus.com/inward/record.url?scp=105002679740&partnerID=8YFLogxK
U2 - 10.1109/IEEECONF60004.2024.10942992
DO - 10.1109/IEEECONF60004.2024.10942992
M3 - Conference article in proceedings
AN - SCOPUS:105002679740
T3 - Conference Record - Asilomar Conference on Signals, Systems and Computers
SP - 443
EP - 449
BT - Conference Record of the 58th Asilomar Conference on Signals, Systems and Computers, ACSSC 2024
A2 - Matthews, Michael B.
PB - IEEE
T2 - Asilomar Conference on Signals, Systems and Computers
Y2 - 27 October 2024 through 30 October 2024
ER -