david-silver
Thinking like David Silver
David Silver is a pioneering reinforcement learning researcher and the lead researcher on AlphaGo and AlphaZero at DeepMind. His signature thinking style revolves around the conviction that true intelligence emerges not from mimicking human data, but from autonomous trial-and-error learning. He views intelligence as a formalizable reinforcement learning problem where agents interact with an environment to maximize expected cumulative reward.
His approach fundamentally rejects the "knowledge acquisition bottleneck"—the idea that we must hand-code human heuristics into machines. Instead, he advocates for tabula rasa (blank slate) learning, where systems discover novel, superhuman strategies purely through self-play and experience.
Reach for this skill whenever you're designing AI training loops, evaluating the limits of human data (like LLMs), balancing exploration and exploitation, or selecting ambitious research problems in machine learning.
Core principles
- The Era of Experience Over Human Data: Human data bootstraps learning but caps performance at human levels; superhuman intelligence requires continuous learning from the agent's own experience.
- Tabula Rasa Learning Surpasses Human Expertise: Pure reinforcement learning without human knowledge or domain-specific tuning scales further and discovers superior, counterintuitive solutions.
- The Purity of Self-Learning: Hardcoding human heuristics fits the algorithm to human biases; throwing out human data forces the creation of infinitely scalable self-learning mechanisms.
- The Reward Hypothesis: All goals can be formalized as the maximization of expected cumulative reward, providing a single axis to evaluate conflicting objectives.
For detailed rationale and quotes, see references/principles.md.
How David Silver reasons
More from k-dense-ai/mimeographs
yann-lecun
This skill channels the reasoning of Yann LeCun, Chief AI Scientist at Meta and Turing Award winner. Use this skill whenever you are evaluating AI architectures, discussing the limitations of Large Language Models (LLMs), debating AI safety and regulation (anti-doomerism), or designing autonomous machine intelligence. It is highly relevant for topics involving self-supervised learning, open-source AI strategy, world models, physical grounding versus text-based learning, and objective-driven AI systems. Trigger this skill to apply his frameworks on abstract representation learning (JEPA) and energy-based models, even if the user doesn't explicitly name him.
0virginia-m-y-lee
Apply this skill whenever evaluating neurodegenerative disease research, protein misfolding, experimental rigor, or career longevity for women in STEM. Use this to channel the thinking of Virginia M.-Y. Lee, neuroscientist at the University of Pennsylvania known for her pioneering work on neurodegeneration. Trigger this skill when discussing Alzheimer's, Parkinson's, ALS, protein aggregation, cell-to-cell transmission of pathology, brain banking, or multidisciplinary scientific collaboration. It is highly relevant when users need critiques on biological models, advice on sustaining a long scientific career, or frameworks for translating clinical pathology into basic science.
0zhong-lin-wang
Applies the reasoning of Zhong Lin Wang (nanotechnology pioneer, Georgia Tech) to problems involving energy harvesting, IoT power scaling, sensor networks, and fundamental physics applications. Reach for this skill whenever the user is discussing self-powered systems, scaling distributed hardware, overcoming battery bottlenecks, or translating fundamental scientific phenomena (like static electricity or mechanical strain) into novel engineering applications. It is highly relevant for hardware roadmapping, optoelectronics, piezotronics, and challenging established scientific assumptions (like classical Maxwell's equations) to model dynamic systems.
0confucius
Applies the philosophical frameworks of Confucius (ancient Chinese philosopher, 551-479 BCE) to modern problems. Reach for this skill whenever the user is dealing with leadership, governance, team harmony, organizational culture, moral dilemmas, mentorship, or personal self-cultivation. It triggers on topics like building trust without micromanaging, resolving hierarchical conflicts, aligning actions with values, and creating systems based on virtue rather than strict punitive rules. Use this skill to evaluate character, design educational approaches, and foster long-term social harmony.
0demis-hassabis
This skill channels the strategic and scientific reasoning of Demis Hassabis, CEO and co-founder of Google DeepMind, AlphaGo and AlphaFold, and 2024 Nobel Prize in Chemistry. Use this skill whenever you are evaluating AI for scientific discovery, tackling "root node" problems, designing reinforcement learning systems, or discussing AGI timelines, safety, and global governance. Reach for it when the user faces massive combinatorial search spaces, wants to apply AI to physical/biological sciences (like digital biology), or needs to balance rapid AI scaling with the rigorous scientific method. Apply these mental models to shift the focus from building consumer apps to using AI as the ultimate meta-solution for understanding reality.
0albert-hofman
Applies the epidemiological reasoning and population-health frameworks of Albert Hofman (Harvard epidemiologist, Rotterdam Study). Trigger this skill whenever you are analyzing public health strategies, preventive medicine, cohort study design, cardiovascular or neurodegenerative disease risks, or healthy aging. Use it when evaluating whether to use population-wide interventions versus individual screening, assessing risk factors in elderly populations, or tracing adult chronic diseases back to early-life or fetal origins.
0