Machine learning techniques may appear ill-suited for application in fields that prioritize rigor and deep understanding; however, they have recently found unexpected uses in theoretical physics and pure mathematics. In this Perspective, Gukov, Halverson and Ruehle have discussed rigorous applications of machine learning to theoretical physics and pure mathematics.
Category: mathematics – Page 35
“Gravity pulls matter together, so that when we throw a ball in the air, the Earth’s gravity pulls it down toward the planet,” Mustapha Ishak-Boushaki, a professor of physics in the School of Natural Sciences and Mathematics (NSM) at UT Dallas, and member of the DESI collaboration, said in a statement. “But at the largest scales, the universe acts differently. It’s acting like there is something repulsive pushing the universe apart and accelerating its expansion. This is a big mystery, and we are investigating it on several fronts. Is it an unknown dark energy in the universe, or is it a modification of Albert Einstein’s theory of gravity at cosmological scales?”
DESI’s data, however, shows that the universe may have evolved in a way that isn’t quite consistent with the Lambda CDM model, indicating that the effects of dark energy on the universe may have changed since the early days of the cosmos.
“Our results show some interesting deviations from the standard model of the universe that could indicate that dark energy is evolving over time,” Ishak-Boushaki said. “The more data we collect, the better equipped we will be to determine whether this finding holds. With more data, we might identify different explanations for the result we observe or confirm it. If it persists, such a result will shed some light on what is causing cosmic acceleration and provide a huge step in understanding the evolution of our universe.”
The traveling salesman problem is considered a prime example of a combinatorial optimization problem. Now a Berlin team led by theoretical physicist Prof. Dr. Jens Eisert of Freie Universität Berlin and HZB has shown that a certain class of such problems can actually be solved better and much faster with quantum computers than with conventional methods.
Quantum computers use so-called qubits, which are not either zero or one as in conventional logic circuits, but can take on any value in between. These qubits are realized by highly cooled atoms, ions, or superconducting circuits, and it is still physically very complex to build a quantum computer with many qubits. However, mathematical methods can already be used to explore what fault-tolerant quantum computers could achieve in the future.
“There are a lot of myths about it, and sometimes a certain amount of hot air and hype. But we have approached the issue rigorously, using mathematical methods, and delivered solid results on the subject. Above all, we have clarified in what sense there can be any advantages at all,” says Prof. Dr. Jens Eisert, who heads a joint research group at Freie Universität Berlin and Helmholtz-Zentrum Berlin.
Despite being almost a year old, this blog by Chip Huyen is still a great read for getting into fine-tuning LLMs.
This article covers everything you need to know about Reinforcement Learning from Human Feedback (RLHF).
#AI #ReinforcementLearning
A narrative that is often glossed over in the demo frenzy is about the incredible technical creativity that went into making models like ChatGPT work. One such cool idea is RLHF: incorporating reinforcement learning and human feedback into NLP. This post discusses the three phases of training ChatGPT and where RLHF fits in. For each phase of ChatGPT development, I’ll discuss the goal for that phase, the intuition for why this phase is needed, and the corresponding mathematical formulation for those who want to see more technical detail.
Quantum cognition is a new research program that uses mathematical principles from quantum theory as a framework to explain human cognition, including judgment and decision making, concepts, reasoning, memory, and perception. This research is not concerned with whether the brain is a quantum computer. Instead, it uses quantum theory as a fresh conceptual framework and a coherent set of formal tools for explaining puzzling empirical findings in psychology. In this introduction, we focus on two quantum principles as examples to show why quantum cognition is an appealing new theoretical direction for psychology: complementarity, which suggests that some psychological measures have to be made sequentially and that the context generated by the first measure can influence responses to the next one, producing measurement order effects, and superposition, which suggests that some psychological states cannot be defined with respect to definite values but, instead, that all possible values within the superposition have some potential for being expressed. We present evidence showing how these two principles work together to provide a coherent explanation for many divergent and puzzling phenomena in psychology. (PsycInfo Database Record © 2020 APA, all rights reserved)
A small team of AI researchers at Microsoft reports that the company’s Orca-Math small language model outperforms other, larger models on standardized math tests. The group has published a paper on the arXiv preprint server describing their testing of Orca-Math on the Grade School Math 8K (GSM8K) benchmark and how it fared compared to well-known LLMs.
Many popular LLMs such as ChatGPT are known for their impressive conversational skills—less well known is that most of them can also solve math word problems. AI researchers have tested their abilities at such tasks by pitting them against the GSM8K, a dataset of 8,500 grade-school math word problems that require multistep reasoning to solve, along with their correct answers.
In this new study, the research team at Microsoft tested Orca-Math, an AI application developed by another team at Microsoft specifically designed to tackle math word problems, and compared the results with larger AI models.
Year 2010 😗😁
The world has waited with bated breath for three decades, and now finally a group of academics, engineers, and math geeks has discovered the number that explains life, the universe, and everything. That number is 20, and it’s the maximum number of moves it takes to solve a Rubik’s Cube.
Known as God’s Number, the magic number required about 35 CPU-years and a good deal of man-hours to solve. Why? Because there’s-1 possible positions of the cube, and the computer algorithm that finally cracked God’s Algorithm had to solve them all. (The terms God’s Number/Algorithm are derived from the fact that if God was solving a Cube, he/she/it would do it in the most efficient way possible. The Creator did not endorse this study, and could not be reached for comment.)
A full breakdown of the history of God’s Number as well as a full breakdown of the math is available here, but summarily the team broke the possible positions down into sets, then drastically cut the number of possible positions they had to solve for through symmetry (if you scramble a Cube randomly and then turn it upside down, you haven’t changed the solution).
Nice figures in this newly published survey on Scaled Optimal Transport with 200+ references.
👉
Optimal Transport (OT) is a mathematical framework that first emerged in the eighteenth century and has led to a plethora of methods for answering many theoretical and applied questions. The last decade has been a witness to the remarkable contributions of this classical optimization problem to machine learning. This paper is about where and how optimal transport is used in machine learning with a focus on the question of scalable optimal transport. We provide a comprehensive survey of optimal transport while ensuring an accessible presentation as permitted by the nature of the topic and the context. First, we explain the optimal transport background and introduce different flavors (i.e. mathematical formulations), properties, and notable applications.
A century ago, Emmy Noether published a theorem that would change mathematics and physics. Here’s an all-ages guided tour through this groundbreaking idea.
Research unveils a mathematical model for ice nucleation, showing how surface angles affect water’s freezing point, with applications in snowmaking and cloud seeding.
From abstract-looking cloud formations to roars of snow machines on ski slopes, the transformation of liquid water into solid ice touches many facets of life. Water’s freezing point is generally accepted to be 32 degrees Fahrenheit. But that is due to ice nucleation — impurities in everyday water raise its freezing point to this temperature. Now, researchers unveil a theoretical model that shows how specific structural details on surfaces can influence water’s freezing point.
Research Findings and Their Implications.