Physics Colloquium
Quantitative reasoning tasks which can involve mathematics, science, and programming are often challenging for machine learning models in general and for language models in particular. We show that transformer-based language models obtain significantly better performance on math and science questions when trained in an unsupervised way on a large, math-focused dataset. Performance can be further improved using prompting and sampling techniques including chain-of-thought and majority voting. Minerva, a model that combines these techniques, achieves SOTA on several math and science benchmarks. I will describe the model, its capabilities and limitations.
Join via Zoom:
https://caltech.zoom.us/j/85811994621
Meeting ID: 858 1199 4621
The colloquium is held in Feynman Lecture Hall, 201 E. Bridge.
In person is open to those with a valid Caltech ID.