D-Index & Metrics Best Publications

D-Index & Metrics D-index (Discipline H-index) only includes papers and citation values for an examined discipline in contrast to General H-index which accounts for publications across all disciplines.

Discipline name D-index D-index (Discipline H-index) only includes papers and citation values for an examined discipline in contrast to General H-index which accounts for publications across all disciplines. Citations Publications World Ranking National Ranking
Computer Science D-index 34 Citations 12,703 80 World Ranking 7820 National Ranking 3655

Overview

What is he best known for?

The fields of study he is best known for:

  • Artificial intelligence
  • Statistics
  • Machine learning

The scientist’s investigation covers issues in Artificial intelligence, Reinforcement learning, Mathematical optimization, Machine learning and Upper and lower bounds. His study focuses on the intersection of Artificial intelligence and fields such as Active learning with connections in the field of Computational learning theory. His research in Reinforcement learning intersects with topics in End-to-end principle, Posterior probability, Human–computer interaction and Bellman equation.

His study in Mathematical optimization is interdisciplinary in nature, drawing from both Value, Markov decision process and Applied mathematics. The Machine learning study combines topics in areas such as Sampling, Bayesian probability and Doubly robust. His Thompson sampling research incorporates elements of Learning theory and Heuristic.

His most cited work include:

  • A contextual-bandit approach to personalized news article recommendation (1439 citations)
  • Parallelized Stochastic Gradient Descent (877 citations)
  • An Empirical Evaluation of Thompson Sampling (767 citations)

What are the main themes of his work throughout his whole career to date?

His primary areas of study are Artificial intelligence, Reinforcement learning, Machine learning, Mathematical optimization and Markov decision process. His study on Artificial neural network and Thompson sampling is often connected to Action as part of broader study in Artificial intelligence. His Reinforcement learning research includes elements of Bellman equation, Stationary distribution, Function approximation and Benchmark.

His study looks at the intersection of Machine learning and topics like Search engine with Ranking. His studies examine the connections between Mathematical optimization and genetics, as well as such issues in Estimator, with regards to Importance sampling. His Markov decision process research integrates issues from Algorithm, State and Probably approximately correct learning.

He most often published in these fields:

  • Artificial intelligence (45.59%)
  • Reinforcement learning (41.67%)
  • Machine learning (25.00%)

What were the highlights of his more recent work (between 2019-2021)?

  • Reinforcement learning (41.67%)
  • Stationary distribution (6.37%)
  • Mathematical optimization (21.57%)

In recent papers he was focusing on the following fields of study:

Lihong Li focuses on Reinforcement learning, Stationary distribution, Mathematical optimization, Applied mathematics and Benchmark. His Reinforcement learning research is multidisciplinary, incorporating perspectives in Markov decision process and Importance sampling. The concepts of his Mathematical optimization study are interwoven with issues in Estimator, Function space, Generalized estimating equation and Confidence interval.

His research investigates the connection with Applied mathematics and areas like Markov chain which intersect with concerns in Monte Carlo method, Constraint, Divergence and Queueing theory. His studies in Benchmark integrate themes in fields like Artificial neural network, Posterior probability and Thompson sampling. Lihong Li is researching Thompson sampling as part of the investigation of Regret and Artificial intelligence.

Between 2019 and 2021, his most popular works were:

  • GenDICE: Generalized Offline Estimation of Stationary Values (32 citations)
  • Doubly Robust Bias Reduction in Infinite Horizon Off-Policy Estimation (16 citations)
  • Off-Policy Evaluation via the Regularized Lagrangian. (12 citations)

In his most recent research, the most cited papers focused on:

  • Artificial intelligence
  • Statistics
  • Machine learning

His primary scientific interests are in Reinforcement learning, Applied mathematics, Stationary distribution, Estimator and Mathematical optimization. His Reinforcement learning study frequently draws parallels with other fields, such as Importance sampling. His Importance sampling study combines topics in areas such as Bellman equation and Doubly robust.

His Applied mathematics study incorporates themes from Divergence, Constraint, Markov chain, Benchmark and Monte Carlo method. His Estimator study integrates concerns from other disciplines, such as Linear programming and Stability. Mathematical optimization is frequently linked to Black box in his study.

This overview was generated by a machine learning system which analysed the scientist’s body of work. If you have any feedback, you can contact us here.

Best Publications

A contextual-bandit approach to personalized news article recommendation

Lihong Li;Wei Chu;John Langford;Robert E. Schapire.
the web conference (2010)

2267 Citations

Parallelized Stochastic Gradient Descent

Martin Zinkevich;Markus Weimer;Lihong Li;Alex J. Smola.
neural information processing systems (2010)

1334 Citations

An Empirical Evaluation of Thompson Sampling

Olivier Chapelle;Lihong Li.
neural information processing systems (2011)

1245 Citations

Contextual bandits with linear Payoff functions

Wei Chu;Lihong Li;Lev Reyzin;Robert E. Schapire.
international conference on artificial intelligence and statistics (2011)

741 Citations

Unbiased offline evaluation of contextual-bandit-based news article recommendation algorithms

Lihong Li;Wei Chu;John Langford;Xuanhui Wang.
web search and data mining (2011)

549 Citations

Sparse Online Learning via Truncated Gradient

John Langford;Lihong Li;Tong Zhang.
Journal of Machine Learning Research (2009)

527 Citations

PAC model-free reinforcement learning

Alexander L. Strehl;Lihong Li;Eric Wiewiora;John Langford.
international conference on machine learning (2006)

495 Citations

Doubly Robust Policy Evaluation and Learning

John Langford;Lihong Li;Miroslav Dud k.
international conference on machine learning (2011)

475 Citations

Doubly Robust Policy Evaluation and Learning

Miroslav Dudik;John Langford;Lihong Li.
arXiv: Learning (2011)

410 Citations

Towards a Unified Theory of State Abstraction for MDPs.

Lihong Li;Thomas J. Walsh;Michael L. Littman.
ISAIM (2006)

382 Citations

If you think any of the details on this page are incorrect, let us know.

Contact us

Best Scientists Citing Lihong Li

Jianfeng Gao

Jianfeng Gao

Microsoft (United States)

Publications: 94

Csaba Szepesvári

Csaba Szepesvári

University of Alberta

Publications: 63

Emma Brunskill

Emma Brunskill

Stanford University

Publications: 57

Maarten de Rijke

Maarten de Rijke

University of Amsterdam

Publications: 56

Shie Mannor

Shie Mannor

Technion – Israel Institute of Technology

Publications: 50

Rémi Munos

Rémi Munos

DeepMind (United Kingdom)

Publications: 43

Mihaela van der Schaar

Mihaela van der Schaar

University of Cambridge

Publications: 42

Mohammad Ghavamzadeh

Mohammad Ghavamzadeh

Google (United States)

Publications: 38

Shimon Whiteson

Shimon Whiteson

University of Oxford

Publications: 38

Sergey Levine

Sergey Levine

University of California, Berkeley

Publications: 37

John Langford

John Langford

Microsoft (United States)

Publications: 37

Michael L. Littman

Michael L. Littman

Brown University

Publications: 35

Olivier Pietquin

Olivier Pietquin

Google (United States)

Publications: 35

Alekh Agarwal

Alekh Agarwal

Google (United States)

Publications: 32

Joelle Pineau

Joelle Pineau

McGill University

Publications: 32

Thorsten Joachims

Thorsten Joachims

Cornell University

Publications: 32

Something went wrong. Please try again later.