Published in IEEE Transactions on Circuits and Systems, Volume 51, Issue 6, June 1, 2004, pages 1062-1074.
Copyright © 2004 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE. The definitive version is available at http://dx.doi.org/10.1109/TCSI.2004.829233.
NOTE: At the time of publication, the author Vladimir Prodanov was not yet affiliated with Cal Poly.
The I/Q imbalance is one of the performance bottlenecks in transceivers with stringent requirements imposed by applications such as 802.11a. The mismatch between the frequency responses of two analog low-pass filters, used, e.g., for channel selection in zero-IF receivers, makes this I/Q imbalance frequency dependent. Usually, frequency-dependent I/Q mismatch is estimated and corrected by adaptive techniques, which are complex to implement and may converge slowly due to noise. In this work, a simple, delay-based I/Q compensation scheme is proposed based on an extensive statistical analysis. Its digital implementation uses only two coefficients, which are tuned by a one-step two-tone error estimation. Simulations show that this hardware-efficient scheme significantly reduces the I/Q imbalance.
Electrical and Computer Engineering