The scientist’s investigation covers issues in Artificial intelligence, Reinforcement learning, Mathematical optimization, Machine learning and Upper and lower bounds. His study focuses on the intersection of Artificial intelligence and fields such as Active learning with connections in the field of Computational learning theory. His research in Reinforcement learning intersects with topics in End-to-end principle, Posterior probability, Human–computer interaction and Bellman equation.
His study in Mathematical optimization is interdisciplinary in nature, drawing from both Value, Markov decision process and Applied mathematics. The Machine learning study combines topics in areas such as Sampling, Bayesian probability and Doubly robust. His Thompson sampling research incorporates elements of Learning theory and Heuristic.
His primary areas of study are Artificial intelligence, Reinforcement learning, Machine learning, Mathematical optimization and Markov decision process. His study on Artificial neural network and Thompson sampling is often connected to Action as part of broader study in Artificial intelligence. His Reinforcement learning research includes elements of Bellman equation, Stationary distribution, Function approximation and Benchmark.
His study looks at the intersection of Machine learning and topics like Search engine with Ranking. His studies examine the connections between Mathematical optimization and genetics, as well as such issues in Estimator, with regards to Importance sampling. His Markov decision process research integrates issues from Algorithm, State and Probably approximately correct learning.
Lihong Li focuses on Reinforcement learning, Stationary distribution, Mathematical optimization, Applied mathematics and Benchmark. His Reinforcement learning research is multidisciplinary, incorporating perspectives in Markov decision process and Importance sampling. The concepts of his Mathematical optimization study are interwoven with issues in Estimator, Function space, Generalized estimating equation and Confidence interval.
His research investigates the connection with Applied mathematics and areas like Markov chain which intersect with concerns in Monte Carlo method, Constraint, Divergence and Queueing theory. His studies in Benchmark integrate themes in fields like Artificial neural network, Posterior probability and Thompson sampling. Lihong Li is researching Thompson sampling as part of the investigation of Regret and Artificial intelligence.
His primary scientific interests are in Reinforcement learning, Applied mathematics, Stationary distribution, Estimator and Mathematical optimization. His Reinforcement learning study frequently draws parallels with other fields, such as Importance sampling. His Importance sampling study combines topics in areas such as Bellman equation and Doubly robust.
His Applied mathematics study incorporates themes from Divergence, Constraint, Markov chain, Benchmark and Monte Carlo method. His Estimator study integrates concerns from other disciplines, such as Linear programming and Stability. Mathematical optimization is frequently linked to Black box in his study.
This overview was generated by a machine learning system which analysed the scientist’s body of work. If you have any feedback, you can contact us here.
A contextual-bandit approach to personalized news article recommendation
Lihong Li;Wei Chu;John Langford;Robert E. Schapire.
the web conference (2010)
Parallelized Stochastic Gradient Descent
Martin Zinkevich;Markus Weimer;Lihong Li;Alex J. Smola.
neural information processing systems (2010)
An Empirical Evaluation of Thompson Sampling
Olivier Chapelle;Lihong Li.
neural information processing systems (2011)
Contextual bandits with linear Payoff functions
Wei Chu;Lihong Li;Lev Reyzin;Robert E. Schapire.
international conference on artificial intelligence and statistics (2011)
Unbiased offline evaluation of contextual-bandit-based news article recommendation algorithms
Lihong Li;Wei Chu;John Langford;Xuanhui Wang.
web search and data mining (2011)
Sparse Online Learning via Truncated Gradient
John Langford;Lihong Li;Tong Zhang.
Journal of Machine Learning Research (2009)
PAC model-free reinforcement learning
Alexander L. Strehl;Lihong Li;Eric Wiewiora;John Langford.
international conference on machine learning (2006)
Doubly Robust Policy Evaluation and Learning
John Langford;Lihong Li;Miroslav Dud k.
international conference on machine learning (2011)
Doubly Robust Policy Evaluation and Learning
Miroslav Dudik;John Langford;Lihong Li.
arXiv: Learning (2011)
Towards a Unified Theory of State Abstraction for MDPs.
Lihong Li;Thomas J. Walsh;Michael L. Littman.
ISAIM (2006)
If you think any of the details on this page are incorrect, let us know.
We appreciate your kind effort to assist us to improve this page, it would be helpful providing us with as much detail as possible in the text box below:
Microsoft (United States)
Microsoft (United States)
Brown University
Citadel
Sichuan University
University of Alberta
Google (United States)
Stanford University
University of Alberta
Jingdong (China)