Learning-based cell selection method for femtocell networks

Chaima Dhahri, Tomoaki Ohtsuki

研究成果: Conference contribution

24 被引用数 (Scopus)


In open-access non-stationary femtocell networks, cellular users (also known as macro users or MU) may join, through a handover procedure, one of the neighboring femtocells so as to enhance their communications/increase their respective channel capacities. To avoid frequent communication disruptions owing to effects such as the ping-pong effect, it is necessary to ensure the effectiveness of the cell selection method. Traditionally, such selection method is usually a measured channel/cell quality metric such as the channel capacity, the load of the candidate cell, the received signal strength (RSS), etc. However, one problem with such approaches is that present measured performance does not necessarily reflect the future performance, thus the need for novel cell selection that can predict the \textit{horizon}. Subsequently, we present in this paper a reinforcement learning (RL), i.e, Q- learning algorithm, as a generic solution for the cell selection problem in a non-stationary femtocell network. After comparing our solution for cell selection with different methods in the literature (least loaded (LL), random and capacity-based), simulation results demonstrate the benefits of using learning in terms of the gained capacity and the number of handovers.

ホスト出版物のタイトルIEEE 75th Vehicular Technology Conference, VTC Spring 2012 - Proceedings
出版ステータスPublished - 2012 8月 20
イベントIEEE 75th Vehicular Technology Conference, VTC Spring 2012 - Yokohama, Japan
継続期間: 2012 5月 62012 6月 9


名前IEEE Vehicular Technology Conference


OtherIEEE 75th Vehicular Technology Conference, VTC Spring 2012

ASJC Scopus subject areas

  • コンピュータ サイエンスの応用
  • 電子工学および電気工学
  • 応用数学


「Learning-based cell selection method for femtocell networks」の研究トピックを掘り下げます。これらがまとまってユニークなフィンガープリントを構成します。