Deep learning methods in natural language processing generally become more effective with larger datasets and bigger networks. But it is not evident whether the same is true for more specialized domains such as cheminformatics. Frey and colleagues provide empirical explorations of chemistry models and find that neural-scaling laws hold true even for the largest tested models and datasets.
- Nathan C. Frey
- Ryan Soklaski
- Vijay Gadepally