I designed and taught a new Game Theory course (CAP-5993/4993) in spring 2017 (students enrolled from computer science, statistics, bioinformatics, and electrical and computer engineering departments). See the first lecture here. Here are slides providing an overview of my research from the recent SCIS Information Session.
In fall 2017 I designed and taught a new undergraduate class Artificial Intelligence (CAP 4630). See related posts "How undergraduate AI should be taught" and "3-player Kuhn poker project" from my blog. See also video from AI workshop at FIU's ShellHacks Hackathon, "Artificial Intelligence: From Poker Agents to Hurricane Prediction."
I have a PhD in computer science from Carnegie Mellon University and A.B. in math from Harvard University (see my CV here). I'm interested in artificial intelligence, game theory, multiagent systems, multiagent learning, large-scale optimization, large-scale data analysis and analytics, and knowledge representation, and applications including poker, education, medicine, socialization, and hurricane prediction. See my research statement for elaboration. My dissertation was on scalable approaches for computing game-theoretic solution concepts and learning in imperfect-information games. Here are slides from my thesis defense that also provide an overview of several topics of active study in multiagent systems and computational game theory.
I created the two-player no-limit Texas hold 'em agent Claudico that competed and lost against the strongest human players in the Brains vs. Artificial Intelligence Competition. See my paper, slides, and video presentation (starting at 1:33). Here is a thread on the Two Plus Two Forum, the most popular poker forum, discussing the event (the thread has over 230,000 views). I describe Claudico's key strengths and weaknesses and address several common questions in this post. See Claudico in action against Doug Polk (widely regarded as the best two-player no-limit Texas hold 'em player in the world) here. I also created the agent Tartanian7 that came in first place in the Annual Computer Poker Competition, beating each opposing agent with statistical significance. A brief description I wrote is available here, and a full description is provided in my paper below. I describe the design of the agent and the current state of research in computer poker in detail in this thread on the Two Plus Two Forum. The Brains vs. Artificial Intelligence Competition was organized by Professor Tuomas Sandholm. Collaborators on Claudico and Tartanian7 were Tuomas Sandholm and Noam Brown.
For takeaways and future directions, see the final chapter of my thesis, the final section of my paper on the Brains vs. Artificial Intelligence Competition, and slides from a talk I gave last year to the Princeton Poker Club. See also slides from my talk at Stanford which includes discussion of how Libratus improved on Claudico and slides from my more recent talk in Einstein's lecture hall at Princeton on the surprising breakthrough that enabled superhuman two-player no-limit Texas hold 'em play. Finally see slides from my talk "Strong Game-Theoretic Strategies: Beyond Two Agents" at the MIT Poker Club, and more recent talk "Successful Nash Equilibrium Agent for a 3-Player Imperfect-Information Game" in 2018 at the Princeton and MIT poker clubs.
I organized the first ever tutorial on Computer Poker with Marc Lanctot from Google DeepMind at the 2016 Conference on Economics and Computation (7/25/2016). See my slides from part 1 and Marc's from part 2. I organized the AAAI-15 Workshop on Computer Poker and Imperfect Information and co-organized it in 2014 with Eric Jackson. Here is a report summarizing the highlights of the most recent workshop. I have been a PC member for AAAI ('12, '14, '15, '16, '17, '18), AAMAS ('14, '16), IJCAI ('13, '15, '16), and WWW ('18, '19).
I organized the second Computer Poker Tutorial with Johannes Heinrich and Kevin Waugh at AAAI 2017. Here are slides from part 1, part 2, and part 3. I recently gave a 3-part guest lecture at Tsinghua University on computer poker which is available on youtube: part 1, part 2, part 3.
See also recent articles from the Pittsburgh Tribune-Review and front page of the "Tropical Life" section of the Miami Herald, and my tv debut as "The Robot" and subsequent commentary on Poker Night in America.
Sam Ganzfried, Austin Nowak, and Joannier Pinales. 2018. Successful Nash Equilibrium Agent for a Three-Player Imperfect-Information Game. Feature article at Games, 2018, 9, 33.
Sam Ganzfried and Farzana Yusuf. 2018. Optimal Weighting for Exam Composition.
Education Sciences special issue "Artificial Intelligence and Education."
Kailiang Hu and Sam Ganzfried. 2017. Midgame Solving: A New Weapon for Efficient Large-Scale Equilibrium Approximation. IEEE International Conference on Tools with Artificial Intelligence. Short paper.
Sam Ganzfried. 2017. Reflections on the First Man versus Machine No-Limit Texas Hold ‘em Competition. AI Magazine 38(2) summer issue. See SIGecom Exchanges version below.
Sam Ganzfried. 2017. What is the Right Solution Concept for No-Limit Poker?International Conference on Game Theory (slides).
Sam Ganzfried. 2017. Endgame Solving: The Surprising Breakthrough that Enabled Superhuman Two-Player No-Limit Texas Hold 'em Play. International Conference on Game Theory.
Sam Ganzfried and Farzana Yusuf. Computing Human-Understandable Strategies: Deducing Fundamental Rules of Poker Strategy. Invited feature article at Games, 2017, 8(4), 49.
New approach for computing strong game-theoretic strategies that can be easily understood by humans. This approach has enabled us to conclude several new fundamental rules of poker strategy, for example when to make a very small bet and when to make an extremely large one. Unconventional bet sizes were critical for recent poker agents such as Claudico (which I created) and Libratus.
Sam Ganzfried and Farzana Yusuf. Optimal Number of Choices in Rating Contexts. Cite as: arXiv:1605.06588 [cs.AI]. Early version appeared in Proceedings of Machine Learning Research, 2016, volume 58. Theory, simulations, and experimental results for the optimal number of choices to use from a small discrete set (which is used to approximate a large underlying choice set), with applications including online dating, paper reviewing, and exam grading. Counterintuitively, allowing more options is not always best, and fewer options is optimal surprisingly often.
Sam Ganzfried and Qingyun Sun. 2018. Bayesian Opponent Exploitation in Imperfect-Information Games. In Proceedings of the Conference on Computational Intelligence and Games (CIG). First exact algorithm for opponent exploitation in Bayesian setting in imperfect-information games using natural prior models (Dirichlet and uniform).
Sam Ganzfried. 2015. Reflections on the First Man vs. Machine No-Limit Texas Hold 'em Competition. SIGecom Exchanges, Volume 14.2. Feature article. Earlier version from 9/15. Slides from presentation at 2016 World Congress of the Game Theory Society (7/16) and slides/video (starting at 1:33) from presentation at 2016 New York Computer Science and Economics Day (1/16). See AI Magazine version above.
Sam Ganzfried. 2015. Computing Strong Game-Theoretic Strategies and Exploiting Suboptimal Opponents in Large Games. PhD dissertation, Computer Science Department, Carnegie Mellon University. Available as CMU technical report CMU-CS-15-104.
Sam Ganzfried and Tuomas Sandholm. 2015. Endgame Solving in Large Imperfect-Information Games. In Proceedings of the International Conference on Autonomous Agents and Multiagent Systems (AAMAS). Early version appeared in Proceedings of the Workshop on Computer Poker and Imperfect Information at the AAAI Conference on Artificial Intelligence (AAAI), 2013. The best two-player no-limit Texas hold 'em player in the world (Doug Polk) stated that the "endgame solver" was the strongest component of the agent Claudico in the Brains vs. Artificial Intelligence Competition. The approach has been subsequently adopted by popular commercial software such as GTORangeBuilder and PioSOLVER, and extensions of the paradigm introduced in this work were crucial to the success of the recent agents DeepStack and Libratus. The leader of DeepStack has stated that "Deepstack is all endgame solving," and the strongest component of Libratus was an improved endgame solving approach that enabled solving both the turn and river rounds (as opposed to just the river) in real time on a supercomputer.
Noam Brown*, Sam Ganzfried*, and Tuomas Sandholm. 2015. Hierarchical Abstraction, Distributed Equilibrium Computation, and Post-Processing, with Application to a Champion No-Limit Texas Hold'em Agent. In Proceedings of the International Conference on Autonomous Agents and Multiagent Systems (AAMAS). Short version appeared in Proceedings of the Demonstrations Program at the AAAI Conference on Artificial Intelligence, 2015. *Listed alphabetically. Describes architecture of two-player no-limit Texas hold 'em agent Tartanian7, which won the 2014 AAAI Annual Computer Poker Competition.
Sam Ganzfried and Tuomas Sandholm. 2015. Safe Opponent Exploitation. ACM Transactions on Economics and Computation (TEAC), 3(2), 8:1-28. Special issue on selected papers from EC-12. Algorithms and theory for when it is possible to deviate from just repeatedly playing one-shot equilibrium to exploit opponents' weaknesses.
Sam Ganzfried and Tuomas Sandholm. 2014. Potential-Aware Imperfect-Recall Abstraction with Earth Mover's Distance in Imperfect-Information Games. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI). Algorithm for game abstraction that takes into account trajectories of private information revelation across all rounds. See discussion on Poker-AI forum.
Sam Ganzfried. 2013. Computing Strong Game-Theoretic Strategies and Exploiting Suboptimal Opponents in Large Games. Thesis proposal.
Sam Ganzfried and Tuomas Sandholm. 2013. Action Translation in Extensive-Form Games with Large Action Spaces: Axioms, Paradoxes, and the Pseudo-Harmonic Mapping. In Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI). Algorithm for interpreting opponents' actions that have been removed from game abstraction, which is used by strongest no-limit Texas hold 'em agents.
Sam Ganzfried and Tuomas Sandholm. 2012. Safe Opponent Exploitation. In Proceedings of the ACM Conference on Electronic Commerce (EC). Succeeded by TEAC paper above.
Sam Ganzfried, Tuomas Sandholm, and Kevin Waugh. 2012. Strategy Purification and Thresholding: Effective Non-Equilibrium Approaches for Playing Large Games. In Proceedings of the International Conference on Autonomous Agents and Multiagent Systems (AAMAS). Theory and experiments for techniques for post-processing approximate equilibrium strategies of game abstractions.
Sam Ganzfried and Tuomas Sandholm. 2011. Game Theory-Based Opponent Modeling in Large Imperfect-Information Games. In Proceedings of the International Conference on Autonomous Agents and Multiagent Systems (AAMAS). Algorithm integrates approximate equilibrium prior with learning from observations and does not require historical data or domain-specific features.
Sam Ganzfried and Tuomas Sandholm. 2010. Computing Equilibria by Incorporating Qualitative Models. In Proceedings of the International Conference on Autonomous Agents and Multiagent Systems (AAMAS). Extended version as CMU technical report CMU-CS-10-105. Algorithm exploits human-understandable qualitative strategy model to improve equilibrium finding in Bayesian games with application to limit Texas hold 'em.
Sam Ganzfried and Tuomas Sandholm. 2009. Computing Equilibria in Multiplayer Stochastic Games of Imperfect Information. In Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI). New algorithm with theoretical results for computing approximate Nash equilibrium in multiplayer stochastic imperfect-information games with application to realistic three-player poker tournament.
Sam Ganzfried and Tuomas Sandholm. 2008. Computing an Approximate Jam/Fold Equilibrium for 3-Player No-Limit Texas Hold'em Tournaments. In Proceedings of the International Conference on Autonomous Agents and Multiagent Systems (AAMAS). Algorithm computes approximate Nash equilibrium in multiplayer stochastic imperfect-information games, with application to a realistic three-player poker tournament. Similar approach has been subsequently adopted by popular commercial software such as HoldemResources.
Copyright © Ganzfried Research. All rights reserved.