
Preprints
Journal papers
 G. Neu and G. Bartók: Importance weighting without importance weights: An efficient algorithm for combinatorial semibandits. In Journal on Machine Learning Research (JMLR), vol. 17(154), pp. 121, 2016.
 L. Devroye, G. Lugosi and G. Neu: RandomWalk
Perturbations for Online Combinatorial Optimization. In IEEE Transactions on Information Theory,
vol. 61, pp. 40994106, 2015.
 G. Neu, A.
György, Cs. Szepesvári and A. Antos: Online Markov
Decision
Processes under Bandit Feedback. In IEEE Transactions on Automatic
Control, vol. 59., pp. 676691, 2014.
 A.
György and G.
Neu: NearOptimal
Rates for LimitedDelay Universal Lossy Source
Coding. In IEEE Transactions on Information Theory, vol. 60, pp.
28232834, 2014.
 G. Neu and
Cs.
Szepesvári: Training
Parsers by Inverse Reinforcement Learning. In Machine
Learning, vol. 77(2), pp. 303337, 2009.
Refereed conference papers
 G. Neu, N. Okolo: Efficient Global Planning in Large MDPs via Stochastic PrimalDual Optimization. In Proceedings of the 34th International Conference
on
Algorithmic Learning Theory (ALT), pp. 11011123, 2023.
 G. Gabbianelli, G. Neu, M. Papini: Online learning with offpolicy feedback. In Proceedings of the 34th International Conference
on
Algorithmic Learning Theory (ALT), pp. 620641, 2023.
 L. Viano, A. Kamoutsi, G. Neu, I. Krawczuk, V. Cevher: Proximal Point Imitation Learning. In Advances in Neural Information
Processing Systems 35
(NeurIPS), 2022.
 G. Neu, J. Olkhovskaya, M. Papini, L. Schwartz: Lifting the information ratio: An informationtheoretic analysis of Thompson sampling for Contextual bandits. In Advances in Neural Information
Processing Systems 35
(NeurIPS), 2022.
 G. Lugosi and G. Neu: Generalization bounds via convex analysis. In Proceedings of the 34th Annual Conference on Learning Theory (COLT), pp. 35243546, 2022. [slides]
 G. Neu and J. Olkhovskaya: Online learning in MDPs with linear function approximation and bandit feedback. In Advances in Neural Information
Processing Systems 34
(NeurIPS), pp. 1040710417, 2021.
 G. Neu, G. K. Dziugaite, M. Haghifam, D. M. Roy: InformationTheoretic Generalization Bounds for Stochastic Gradient Descent. In Proceedings of the 33nd Annual Conference on Learning Theory (COLT), pp. 35263545, 2021.
 J. BasSerrano, S. Curi, A. Krause and G. Neu: Logistic QLearning. In Proceedings of
the Twentyfourth International Conference on Artificial Intelligence and
Statistics (AISTATS), pp. 36103618, 2021. [slides]
 G. Neu and C. PikeBurke: A Unifying View of Optimism in Episodic Reinforcement Learning. In Advances in Neural Information
Processing Systems 33
(NeurIPS), pp. 13921403, 2020. [slides]
 G. Neu and J. Olkhovskaya: Efficient and robust algorithms for adversarial linear contextual bandits. In Proceedings of the 32nd Annual Conference on Learning Theory (COLT), pp. 30493068, 2020.
 G. Neu and N. Zhivotovskiy: Fast rates for online prediction with abstention. In Proceedings of the 32nd Annual Conference on Learning Theory (COLT), pp. 3030–3048, 2020.
 J. BasSerrano and G. Neu: Faster saddlepoint optimization for solving largescale Markov decision processes. In Conference on Learning for Dynamics and Control (L4DC), pp. 413–423, 2020.
 N. Mücke, G. Neu and L. Rosasco: Beating SGD saturation with tailaveraging and minibatching. In Advances in Neural Information
Processing Systems 32
(NeurIPS), pp. 1256812577, 2019.
 C. Riquelme, H. Penedones, D. Vincent, H. Maennel, S. Gelly, T. Mann, A. Barreto and G. Neu: Adaptive TemporalDifference Learning for Policy Evaluation with PerState Uncertainty Estimates. In Advances in Neural Information
Processing Systems 32
(NeurIPS), pp. 1187211882, 2019.
 W. Kotłowski and G. Neu: Bandit Principal Component Analysis. In Proceedings of the 32nd Annual Conference on Learning Theory (COLT), pp. 19942024, 2019. [slides]
 G. Lugosi, G. Neu and J. Olkhovskaya : Online influence maximization with local observations. In Proceedings of the 30th International Conference
on
Algorithmic Learning Theory (ALT), pp. 557580, 2019.
 G. Neu and L. Rosasco: Iterate averaging as regularization for stochastic gradient descent. In Proceedings of the 31st Annual Conference on Learning Theory (COLT), pp. 32223242, 2018.
 N. CesaBianchi, C. Gentile, G. Lugosi and G. Neu: Boltzmann exploration done right. In Advances in Neural Information
Processing Systems 30
(NeurIPS), pp. 62846293, 2017. [poster]
 G. Neu and V. Gómez: Fast rates for online learning in Linearly Solvable Markov Decision Processes. In Proceedings of the 30th Annual Conference on Learning Theory (COLT), pp. 15671588, 2017. [slides]
 T. Liu, G. Lugosi, G. Neu and D. Tao: Algorithmic stability and hypothesis complexity. In Proceedings of the 34th International Conference on Machine Learning (ICML), pp. 21592167, 2017.
 T. Kocák, G. Neu and M. Valko: Online learning with ErdősRényi sideobservation graphs. In Proceedings of
the 32nd Conference on
Uncertainty in Artificial Intelligence (UAI), pp. 339346, 2016.
 T. Kocák, G. Neu and M. Valko: Online learning with noisy side observations. In Proceedings of
the Nineteenth International Conference on Artificial Intelligence and
Statistics (AISTATS), pp. 11861194, 2016.
 G. Neu: Explore no more: Improved highprobability regret bounds for nonstochastic bandits. In Advances in Neural Information
Processing Systems
28
(NeurIPS), pp. 31503158, 2015. [poster] [slides]
 G. Neu: Firstorder regret bounds for combinatorial semibandits. In Proceedings of the 28th Annual Conference on Learning Theory (COLT),
pp. 13601375, 2015.[poster] [slides]
 G. Neu and M. Valko: Online Combinatorial
Optimization with Stochastic Decision Sets and Adversarial Losses.
In Advances in Neural Information Processing Systems
27
(NeurIPS), pp. 27802788, 2014. [poster] [slides]
 T. Kocák, G.
Neu, M. Valko and R. Munos: Efficient Learning
by Implicit Exploration
in Bandit Problems with Side Observations. In Advances in Neural
Information Processing Systems
27
(NeurIPS), pp. 613621, 2014. [poster] [slides]
 A. Sani, G. Neu and A. Lazaric: Exploiting Easy Data
in Online Optimization. In Advances in Neural Information
Processing Systems
27
(NeurIPS), pp. 810818, 2014. [poster] [spotlight] [talk]
 A. Zimin and G.
Neu: Online
Learning in Episodic Markov Decision Processes by Relative Entropy
Policy Search. In Advances in Neural Information Processing Systems
26
(NeurIPS), pp. 15831591, 2013. [poster] [slides]
 G. Neu and G.
Bartók: An
Efficient Algorithm for Learning with SemiBandit
Feedback. In Proceedings of the 24th International Conference
on
Algorithmic Learning Theory (ALT), pp. 234248, 2013. [poster] [slides] Full version in JMLR '16.
 L. Devroye, G.
Lugosi and G. Neu: Prediction
by RandomWalk Perturbation. In
Proceedings of the 26th Annual Conference on Learning Theory (COLT),
pp. 460473, 2013. [slides] Full version in IEEE TIT '15.
 G. Neu, A.
György, and Cs. Szepesvári: The Adversarial
Stochastic Shortest Path
Problem with Unknown Transition Probabilities. In Proceedings of
the
Fifteenth International Conference on Artificial Intelligence and
Statistics (AISTATS), pp. 805813, 2012. [supplement] [poster]
 A. György and G.
Neu: NearOptimal
Rates
for LimitedDelay Universal Lossy Source Coding. In 2011 IEEE
International Symposium
on
Information Theory, pp. 22182222, 2011.Full version in IEEE TIT '14.
 G. Neu, A.
György, Cs. Szepesvári and A. Antos: Online Markov
Decision
Processes under Bandit Feedback. In Advances in Neural Information
Processing Systems 23 (NeurIPS), pp. 18041812, 2010. [poster] [spotlight] Full version in IEEE TAC '14.
 G. Neu, A.
György, and Cs. Szepesvári: The Online Loopfree
Stochastic
ShortestPath Problem. In Proceedings of The 23rd Conference on
Learning Theory (COLT), pp. 231243, 2010.
 G. Neu and Cs.
Szepesvári: Apprenticeship
Learning using Inverse Reinforcement
Learning and Gradient Methods.
In Proceedings of the 23rd
Conference on
Uncertainty in Artificial Intelligence (UAI), pp. 295302, 2007.
Other

