
Yury Polyanskiy
Curriculum Vitae 
Research:  Information theory, coding theory & related fields  
Education:  2010  Princeton University, Ph.D. (Advisers: H. V. Poor and S. Verdú)  
2005  MIPT, M. S. (honors), Dept. General and Applied Physics  
Address:  32D668, MIT (also: E17457, IDSS, MIT)  
77 Massachusetts Avenue  
Cambridge, MA 02139  
Phone:  (617) 3240047  
Email: 
About
Yury Polyanskiy is an Associate Professor of Electrical Engineering and Computer Science in the Dept. of EECS at MIT and a member of LIDS, IDSS and Center of Statistics.
Yury received the M.S. degree in applied mathematics and physics from the
Moscow Institute of Physics and Technology,
Moscow, Russia in 2005 and the Ph.D. degree in electrical engineering from
Princeton
University, Princeton, NJ in 2010. In 20002005 he lead the development of
the embedded software in the
Department of Surface Oilfield Equipment, Borets Company LLC (Moscow).
Currently, his research focuses on basic questions in information theory,
errorcorrecting codes, wireless communication and faulttolerant and
defecttolerant circuits.
Dr. Polyanskiy won the 2013 NSF CAREER award and 2011 IEEE Information
Theory Society Paper Award.
Back when he still had spare time, he enjoyed playing with mathematics
(especially, algebraic geometry and algebraic topology) and hacking Linux
kernel.


My current research interests spin around finding exact and approximate answers to nonasymptotic questions in communication theory. This direction has been explored in my thesis from an informationtheoretic (fundamental limits) angle. For example, the capacity of an additive white Gaussian noise (AWGN) channel with SNR=0 dB is given by Shannon's formula This classical result means that one can communicate at rates arbitrarily close to 0.5 bits per channel use with an arbitrary small probability of error ε in the limit of infinite number of channel uses (and no such communication is possible for any higher rate). This fundamental observation, however, means little for the practical engineer who is always limited by delay requirements. In such circumstances he would rather ask the question: Assume I agree to step back from the capacity by a factor η=0.9 and I also agree to tolerate a probability of error ε=10^{3}. What is the minimum number n of channel uses do I need? 2550 ≤ n ≤ 2850 n≈ 2750 n≈ 4120 For a much more detailed introduction to this line of research please check Chapter 1 of my thesis (requires no information theory background!). Bounds on maximal achievable rate for the AWGN(0 dB) 