Radar sensitivity and resolution in presence of range sidelobe reducing networks designed using linear programming

Binary phase-coded waveforms are widely used in modern radar systems. In linear receivers the sidelobes present at the output of the compression filter cause a reduction of the useful dynamic range. Many procedures have been developed to reduce the sidelobes based on a compensation of their Fourier spectra or on a design of an optimum single mismatched filter. The present paper uses linear programming technique to define an optimum sidelobe reducing network which introduces small mismatching losses when cascaded to a matched filter. The influence of quantization in a digital implementation is evaluated.

[1]  Charles E. Cook,et al.  Radar Signals: An Introduction to Theory and Application , 1967 .

[2]  Martin H. Ackroyd,et al.  Optimum Mismatched Filters for Sidelobe Suppression , 1973, IEEE Transactions on Aerospace and Electronic Systems.

[3]  August W. Rihaczek,et al.  Range Sidelobe Suppression for Barker Codes , 1971, IEEE Transactions on Aerospace and Electronic Systems.

[4]  Pravin Varaiya,et al.  Notes on Optimization , 1972 .