
Machine learning for DCO-OFDM based LiFi
Author(s) -
Krishna Saha Purnita,
M. Rubaiyat Hossain Mondal
Publication year - 2021
Publication title -
plos one
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.99
H-Index - 332
ISSN - 1932-6203
DOI - 10.1371/journal.pone.0259955
Subject(s) - orthogonal frequency division multiplexing , dc bias , computer science , algorithm , visible light communication , electronic engineering , physics , telecommunications , voltage , optics , channel (broadcasting) , engineering , light emitting diode , quantum mechanics
Light fidelity (LiFi) uses different forms of orthogonal frequency division multiplexing (OFDM), including DC biased optical OFDM (DCO-OFDM). In DCO-OFDM, the use of a large DC bias causes optical power inefficiency, while a small bias leads to higher clipping noise. Hence, finding an appropriate DC bias level for DCO-OFDM is important. This paper applies machine learning (ML) algorithms to find optimum DC-bias value for DCO-OFDM based LiFi systems. For this, a dataset is generated for DCO-OFDM using MATLAB tool. Next, ML algorithms are applied using Python programming language. ML is used to find the important attributes of DCO-OFDM that influence the optimum DC bias. It is shown here that the optimum DC bias is a function of several factors including, the minimum, the standard deviation, and the maximum value of the bipolar OFDM signal, and the constellation size. Next, linear and polynomial regression algorithms are successfully applied to predict the optimum DC bias value. Results show that polynomial regression of order 2 can predict the optimum DC bias value with a coefficient of determination of 96.77% which confirms the effectiveness of the prediction.