Astera Logo

Astera

Senior Research Scientist - Simplex

Posted 2 Days Ago
Be an Early Applicant
In-Office
Emeryville, CA, USA
180K-400K Annually
Senior level
In-Office
Emeryville, CA, USA
180K-400K Annually
Senior level
The Senior Research Scientist will lead research on intelligence theory, design experiments, and direct junior researchers while collaborating closely with the team.
The summary above was generated by AI
What We're Building at Simplex

At Simplex, we're building a science of intelligence. Our aim is to develop and apply a rigorous theory of latent internal structure in neural networks, and how that structure relates to computation and behavior.

We believe that when dealing with intelligence, understanding is safety. Without genuine understanding, we can't reliably monitor, control, or even reason clearly about what these systems are doing. But these same systems also present us with a new opportunity. For the first time, we have AI complex enough to serve as testbeds for theories of intelligence, including biological. We aim to build a theory applicable to intelligence, both artificial and biological.

We have the beginnings of such a theory, grounded in the physics of information and experimentally verified in transformers. Now, we are scaling our team. Some of the near-term goals we have are building unsupervised methods that recover belief geometries in real LLMs, extending the theory to more complex cognitive tasks, and pushing toward tools reliable enough to matter for safety.

The Role

This is a senior position on a small team. We're looking for someone who will shape the direction of our research, not just execute within it. That means identifying the most important open questions, designing the experiments and theory to address them, and driving projects from conception through to results that move the field.

You'll work closely with the rest of the team. We do our best work when thinking together, and the best ideas here tend to emerge from conversation. But you'll also be expected to carry significant research threads independently, direct more junior researchers, and take a leadership role in figuring out what Simplex should be working on next.

We are also hiring for Research Scientist positions.

Who We're Looking For

We're looking for people who can do rigorous mathematics and get their hands dirty with real models and data; ideally someone who moves naturally between theory and experiment, and feels deeply driven to understand intelligence.

You learn across fields. Our work draws on many fields, like dynamical systems, probability, deep learning, physics, information theory, and neuroscience. You don't need to know all of it coming in, but you're the kind of person who picks things up, and follows your curiosity—and surprising experimental results—wherever they lead.

You have taste. You know the difference between a problem that matters and a problem that's merely publishable. You have strong opinions about research directions, not just techniques. You've developed these opinions through experience. You’ve seen enough research programs succeed and fail to have real judgement about what’s worth pursuing, and how.

You set directions. You identify which problems are worth solving. You've led research efforts before, and you've developed the instinct for when to push harder on a path and when to change course.

You communicate. You can explain your ideas clearly to collaborators, in writing, and on a whiteboard. You can also explain them to people outside of your subfield. Science is a team activity for us, and at a senior level that means helping others think more clearly, not just thinking clearly yourself.

You build. You're at home in front of a whiteboard and in a terminal. We are building new theory, new code, and new experiments. You think big, but you're serious about it, and you actually try to make things happen rather than just ideating. You use AI tools, you tinker, you're excited about what's becoming possible.

You have a body of work that demonstrates depth and originality in a quantitative field such as physics, mathematics, neuroscience, machine learning, etc. A PhD is typical but not required if you've found another way to show this type of track record.

Current Projects

Belief discovery at scale
Finding belief-state geometry in large language models without supervision. Can we automatically identify the internal structures that encode what a model knows about the world?

Building a theory of intelligence
We have the beginning steps of a theory, but it needs to be extended and refined in a number of ways, in order to, e.g., capture internal world models of different types and apply to other neural systems (e.g., RL and biological brains).

Generalization
Why and how do neural networks generalize? Our framework suggests ways in which internal structures support out-of-distribution behavior.

Red Teaming
We have an entire team dedicated to stress-testing our own framework. Finding the boundaries, the edge cases, the places where the theory breaks down in service of figuring out what's actually true.

Biological Intelligence
The same mathematics that reveals structure in transformers might apply to biological neural networks. We plan on testing this on real brain data because ultimately we're interested in intelligence wherever it appears.

Learn More About Our Work

Our foundational result (manuscript, blog post) showed that transformers trained on next-token prediction spontaneously organize their activations into geometries predicted by Bayesian belief updating over hidden states of a world model. Even when trained on simple token sequences from hidden Markov models, complex fractals emerge in the residual stream, structures far removed from the surface statistics of the training data. We think of this work as providing the first steps into an understanding of what fundamentally we are training AI systems to do, and what representations we are implicitly training them to have.

Since then, we've pushed in several directions. In Constrained Belief Updating Explains Transformer Representations, we asked how attention implements belief updating when Bayesian inference is fundamentally recurrent. We found that attention parallelizes recurrence by decomposing belief updates spectrally across heads, and we were able to make verified predictions about embeddings, OV vectors, attention patterns, and residual stream geometry at different layers.

We've also developed a theory of in-context learning grounded in training data structure. When training data mixes multiple sources, models must infer not just what hidden state the generator is in, but which source is active. This hierarchical belief updating necessarily produces power-law loss scaling with context length and explains why induction heads emerge.

We've been asking what the most general computational framework for understanding neural network representations might be. Our initial work implied activations should lie in simplices, but we've now shown that networks discover quantum and post-quantum belief geometries when these are the minimal way to model their training data. This offers a new foundation for thinking about features, superposition, and what representations neural networks use on their own terms.

Most recently, we’ve shown that transformers naturally decompose their world model into interpretable parts. These factored belief representations provide an exponential-dimensional advantage, and suggest that we can understand and surgically intervene upon low-dimensional subspaces of large models.

For a comprehensive overview of where we are and where we're headed, see our July 2025 progress report on the Alignment Forum. You can also watch Paul and Adam discuss the research program at the FAR Seminar or read this recent interview from August 2025.

Preferred Qualities
  • PhD or equivalent in Physics, CS, neuroscience, math, or comparable

  • Extensive ML experience

  • Experience in interpretability

We're especially interested in people who might be overlooked by traditional hiring, unconventional backgrounds, unusual paths, the kind of candidate who doesn't fit neatly into a box but has something real. If you're not sure whether you're qualified, we'd still like to hear from you.

About Neuro-AI and Astera Institute

Astera Institute is an independent research organization with a $3B+ endowment, operating outside the constraints of markets and academia. We run more like a startup than a foundation — small team, minimal bureaucracy, high risk tolerance.

Neuro-AI is Astera's program for charting a path in a world with thinking machines. It combines three complementary efforts — Neuro (decoding how biological neural activity becomes conscious experience), AGI (neuroscience-informed approaches to engineering intelligence), and Safety (building a scientific understanding of intelligence itself). The Simplex team leads the Safety effort within Neuro-AI, and additionally has a team in London.

This role is employment with Astera Institute. Learn more at astera.org.

Top Skills

Ai Tools
Bayesian Inference
Deep Learning

Similar Jobs

18 Minutes Ago
In-Office
57K-95K Annually
Mid level
57K-95K Annually
Mid level
Information Technology • Internet of Things • Mobile • On-Demand • Software
The Vertical Account Executive, Healthcare will build partnerships with healthcare providers, drive sales of fiber services, and collaborate with various teams while meeting revenue goals.
Top Skills: ExcelMicrosoft PowerpointMicrosoft WordSalesforce
4 Hours Ago
Remote or Hybrid
United States
97K-145K Annually
Senior level
97K-145K Annually
Senior level
Cloud • Fintech • Software • Business Intelligence • Consulting • Financial Services
As an Audit Manager, you'll oversee audit engagements, guide teams, ensure compliance with accounting standards, and build client relationships.
Top Skills: Accounting SoftwareFasb Accounting StandardsGaap
4 Hours Ago
Remote or Hybrid
United States
97K-145K Annually
Senior level
97K-145K Annually
Senior level
Cloud • Fintech • Software • Business Intelligence • Consulting • Financial Services
Manage audit engagements, provide performance feedback, oversee operations, assist clients with GAAP compliance, and lead training programs while engaging in business development efforts.
Top Skills: Capitalized SoftwareConvertible DebtFasb Accounting StandardsGaapRevenue RecognitionSafeStock-Based Compensation

What you need to know about the San Francisco Tech Scene

San Francisco and the surrounding Bay Area attracts more startup funding than any other region in the world. Home to Stanford University and UC Berkeley, leading VC firms and several of the world’s most valuable companies, the Bay Area is the place to go for anyone looking to make it big in the tech industry. That said, San Francisco has a lot to offer beyond technology thanks to a thriving art and music scene, excellent food and a short drive to several of the country’s most beautiful recreational areas.

Key Facts About San Francisco Tech

  • Number of Tech Workers: 365,500; 13.9% of overall workforce (2024 CompTIA survey)
  • Major Tech Employers: Google, Apple, Salesforce, Meta
  • Key Industries: Artificial intelligence, cloud computing, fintech, consumer technology, software
  • Funding Landscape: $50.5 billion in venture capital funding in 2024 (Pitchbook)
  • Notable Investors: Sequoia Capital, Andreessen Horowitz, Bessemer Venture Partners, Greylock Partners, Khosla Ventures, Kleiner Perkins
  • Research Centers and Universities: Stanford University; University of California, Berkeley; University of San Francisco; Santa Clara University; Ames Research Center; Center for AI Safety; California Institute for Regenerative Medicine

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account