Role overview
Direct message the job poster from JAM
Dan Kirkpatrick
Dan Kirkpatrick
Helping employers find hidden Data Science and Machine Learning talent
A fast-growing AI lab with compute (including B200s) on par with top big-tech research groups and that is pioneering the next generation of LLMs and developing architectures that go beyond today’s transformer frameworks. You'll explore new attention mechanisms, scaling strategies, and hybrid architectures to unlock more efficient and expressive intelligence. The role involves leading large-scale pre-training runs, designing empirical studies, and collaborating with an elite research team focused on efficiency and scaling.
If you have a very strong background in modern sequence modeling, distributed training, and transformer-based architectures alongside an excellent academic record plus publishing record this could be the role you're looking for. This is an opportunity to shape how foundational models evolve where you'll be working on problems that define the future of large-scale AI research.
Show more
Show less
Seniority level
Mid-Senior level
Employment type
Full-time
Job function
Research
Industries
Technology, Information and Media