The Astera Institute is excited to launch a major new neuroscience research effort led by Dr. Doris Tsao, who will be joining as Chief Scientist for Astera Neuro. We seek to understand one of the deepest mysteries of science: how the brain produces conscious experience, cognition, and intelligent behavior. Astera will support this effort with $600M+ over the next decade.
Doris has spent her career developing one of the most detailed accounts of how neural activity gives rise to perception through work on the neural code and circuitry underlying face and object recognition. This work shows how a complex visual percept, object identity, is represented by a principled geometric code. Her recent work explores a new computational framework for how symbols first arise in the brain through specialized circuits for object tracking.
What are we doing?
Across every moment of our lives, the brain transforms raw sensory input into a coherent world filled with objects, relationships, meanings, and a sense of self. Yet we still do not understand the fundamental computational principles the brain uses to construct this internal world. Uncovering these principles would transform both neuroscience and technology–revealing the mechanism responsible for generating conscious experience, and at the same time, providing a new framework for AGI.
At the heart of our new effort is the conviction that true understanding of the brain’s internal model means being able to manipulate it in a controlled way. Towards this goal, we are betting that the brain’s representational architecture is compositional, built from elemental units and a neural syntax for combining them. By identifying these fundamental units and the rules that create and link them, we can uncover the brain’s infinitely generative internal code. This, in turn, would provide a principled way to construct or modify internal representations, much as knowing the words and grammar of a language allows the creation of an unlimited range of sentences and meanings. Such capability would mark a profound advance in understanding.
The compositional framework remains a hypothesis, but pursuing it opens a path for fundamentally new kinds of experiments. The first step will be to measure neural activity through large-scale recordings across a rich variety of stimuli and behaviors, allowing us to characterize the underlying neural code. We will then attempt to write in hypothesized neural codes and thereby construct or alter internal representations according to proposed compositional rules. In this way, we can move neuroscience beyond passive observation and towards active, engineering-style tests of a model. Whether or not our hypothesis proves fully correct, this approach will accelerate our understanding of how the brain’s internal model is built.
A field ready for a paradigm shift
The ability to precisely map and modify the brain’s internal model may sound like a lofty goal and indeed, for decades, progress in neuroscience was limited by technology. But that barrier has largely fallen, and we believe now is the right time for our moonshot. We now have the tools to interrogate the brain at unprecedented resolution and scale.
What is needed next is a coordinated engineering effort to fully harness these tools. Advances in large-scale neural recording, targeted stimulation, chronic high-density interfaces, and computational modeling have created a unique moment where a focused, non-clinical, scientifically driven program can push far beyond what academic labs or clinically oriented companies alone can achieve. We intend to fill this essential gap between traditional basic research and clinically driven neurotechnology.
Progress towards our goals opens major branches of independent inquiry:
- Inspiring new approaches to building and steering AI systems: Understanding the brain’s computational strategies—the architectural principles and representations—could reveal fundamentally different approaches to building AI systems that are orders of magnitude more efficient and naturally aligned with human cognition. Industry pursues only a narrow slice of what’s possible. We believe reverse-engineering the only generalized intelligence in existence could open up new pathways to general artificial intelligence.
- Deepening our fundamental understanding of biological intelligence and conscious experience: The brain is one of the universe’s wonders. What is the structure of neural activity required for a specific experience? What are the primitives of perception and thought? How does the brain represent itself? How do disruptions in the brain manifest as psychiatric and neurological conditions? We seek to develop a theory of conscious experience that successfully predicts the experiences that emerge when we write specific patterns to the brain.
- Opening pathways to revolutionary neural interventions: Today’s brain-machine interfaces work at the periphery, translating motor commands or delivering basic sensory inputs. But understanding deeper computational structures could enable interfaces that engage with the brain’s core representational system. This could have major therapeutic applications, for example, a visual prosthesis for the blind that restores vivid, naturalistic visual experience, not just pixelated sight.
Why Astera is pursuing this work
Since the founding of the Astera Institute in 2020, Obelisk, Astera’s AGI research program, has pursued the hypothesis that a better understanding of how intelligence arises in natural systems could reveal computational principles missing from current AI paradigms. The brain achieves flexible, general intelligence with roughly 20 watts of power. It constructs everything we experience—every object we see, every thought, every feeling—from patterns of electrical activity across ~100 billion neurons. It learns continuously from sparse data. It plans, imagines, and constructs a coherent model of the world. We don’t yet understand how.
Astera Neuro brings deep experimental neuroscience into direct dialogue with this work. We hope to create a tight iterative loop across teams where experimental findings shape AI architecture research, and computational questions drive new lines of neuroscientific inquiry.
We believe Doris has developed what may be the most detailed empirical account of how neural activity produces perception so far. The potential of her work requires long-term investment. We are excited to work with Doris to test her model and systematically explore how the brain constructs reality in direct collaboration with Obelisk engineers and researchers exploring alternative approaches to AGI. The iteration between these basic and applied research efforts will surface things neither could find separately.
Research will be shared exclusively outside traditional journals as a forcing function for developing faster, more open, and more useful outputs that represent the full scientific process. As we’ve seen with other efforts, we believe such an approach will enable greater alignment of scientific goals and values across the team. We will also be iterating on ways to make these outputs more compatible with AI-driven discovery.
Building the team
We are excited for the opportunity to build this moonshot. We have a chance to experiment with how science can be done by designing our team and approaches in a purposeful way. This work requires capabilities that don’t typically collaborate as part of a cohesive iterative circuit at an institutional scale: neuroscientists who can design experiments on complex natural behaviors, ML engineers who can build models from massive neural datasets, optical engineers working on holographic optogenetics and advanced imaging, systems builders who can create scalable experimental infrastructure, and metascience innovators dedicated to accelerating all aspects of this work.
Doris brings decades of foundational work on neural coding. For her next chapter with Astera, she is joined by an exceptional founding team (soon to be announced) whose contributions span large-scale reading and writing to neural circuits, clarifying the neural basis for cognition, and understanding brain function during naturalistic behavior.
We are now looking for a Chief Operating Officer who will work in direct partnership with Jed and Doris to transform their scientific vision into operational reality. They will be orchestrating collaboration across disciplines, building systems that support both rigor and speed, and helping create an organization capable of tackling problems at this scale.
What do standardized, low-cost space telescopes, ultra-high-performance bio-inspired materials, and fusion energy that costs under 1¢/kWh all have in common? For one, each of these domains holds incredible potential to further human flourishing. And secondly, each represents a new idea that will be pursued over the next year at our Emeryville, CA campus.
We’re delighted to introduce three new residents to our Residency Program — a program in which residents receive a salary, a budget of up to $2M for team and expenses, compute access, lab space, and an exceptional community of talented like-minded peers, mentors, and investors.
Read on to learn more about our three new ambitious entrepreneurs, along with a brief overview of their work. In the coming months, we’ll be sharing more detailed profiles of the residents and their projects.
If you’re interested in applying to be a future resident, you can reach us at residency@astera.org, or subscribe here to receive our next call for applications, coming in early 2026.
Aaron Tohuvavohu – Cosmic Frontier Labs
Dr. Aaron Tohuvavohu is a physicist, astronomer, and explorer designing the next generation of space telescopes. He has designed missions and experiments across the electromagnetic and multi-messenger spectrum, with expertise spanning black holes, relativistic explosions, UV and X-ray instrumentation, and space systems engineering. Most recently, he led an 11-month sprint from clean sheet to launch of the highest-performance UV detector in orbit, and drove major upgrades to NASA’s Swift Observatory, significantly expanding its scientific reach, impact, and efficiency.
Project description
Cosmic Frontier Labs is building a new class of scientific tools to accelerate discovery and exploration of the Universe. We are expanding humanity’s cosmic horizons by scaling up the number and capability of orbital observatories, bringing Hubble-quality to fleets of telescopes rather than single flagships. By redesigning precision instruments for manufacturability and iteration, the team is moving space astronomy from an era of scarcity to one of abundance, continuous innovation, and exponential discovery.
These telescopes will form a platform for science that evolves as quickly as the questions we ask. We will build the platform iteratively, to continuously integrate new detectors, optics, and algorithms on successive units. In this near future, exploring the cosmos won’t depend on waiting decades for the next great observatory, but on a living, growing constellation of instruments; each a window into the expanding frontier of human understanding.
Open roles: Contact info@cosmicfrontier.org if you’re interested in the mission and want to explore ways to contribute!
Damien Scott – 1cFE
Damien Scott is a technologist and founder. Homeschooled in Botswana and shaped by science fiction, his north star is to build energy systems that move humanity up the Kardashev scale toward post-scarcity. His first entrepreneurial venture was founding Marain, an electric and autonomous-vehicle simulation and optimization company that was acquired by General Motors in 2022. His career has spanned energy and mobility systems across startups and large companies, including the extreme engineering environment of Formula 1 at Williams F1. Beyond racing, he worked on a wide variety of initiatives, from adapting uranium-enrichment centrifuge concepts, to electromechanical flywheel energy storage, to hybrid hypercars and automated mining systems. He has a BSc in physics from the University of Sydney and an MS and MBA from Stanford University.
Project description
Everything humanity values depends on abundant, inexpensive energy. Most usable energy across the universe is fusion…with extra steps. The last decade has brought major public and private progress towards cutting out those steps, to directly generate electricity from fusion, and bring us closer to abundant, low-cost energy. The 1cFE initiative builds on this progress to set our ambitions higher: could the cost of fusion reach below-1¢/kWh LCOE within the next ten years? We map cost-first corridors to sub-cent power, integrating physics, engineering, and manufacturing. We will also publish open analyses, and test how emerging AI capabilities can radically improve and compress cycles across science, first-of-a-kind engineering, and deployment. Our outputs are intended to steer R&D, capital allocation, and policy toward the fastest corridors to sub-cent fusion energy, thereby pushing humanity up the Kardashev scale and upgrading our civilization.
Open roles: Theoretical Physicist and Systems Engineer
Tim McGee – Impossible Fibers
Tim McGee is a biologist and materials innovator developing new ways for proteins and composites to self-assemble into high-performance materials. Trained in Biomolecular Science and Engineering at UCSB, his mission is to translate biology into design and manufacturing. As an early pioneer of bio-inspired design at Biomimicry 3.8, IDEO, and later his own firm, LikoLab, he has worked with global teams on challenges ranging from advanced coatings for food, to novel textile manufacturing, to the biophilic design of urban environments. Most recently, McGee founded Impossible Fibers at Speculative Technologies, leading a DARPA-funded collaboration to predict fiber properties directly from amino acid sequences. His work integrates biology, design, and engineering to create new manufacturing capabilities where materials are assembled from the nanoscale to the macroscale.
Project description
The Impossible Fibers Lab is building a new manufacturing environment that enables proteins to self-assemble into exceptional materials; fibers and composites with electrical, optical, and mechanical properties beyond what’s achievable today. Existing fiber production systems were designed a century ago, and were made for cellulose and plastics, not for the complexity of proteins. McGee’s team combines microfluidics engineering, encapsulation chemistry, automated liquid handling and robotics, and novel spinning techniques to explore how protein composites form, align, and transform during fiber fabrication. The resulting structured dataset will map the relationships between molecular sequence, process conditions, and material outcomes, creating the foundation for predictive, bio-inspired materials design.
In the long term, Impossible Fibers seeks to make matter programmable, from quantum interactions to custom product-scale performance, laying the foundations for a new era of materials manufacturing.
—
Extending a warm welcome to our new residents, and stay tuned for a deeper dive into their work!
We’re moving into an age in which agents are our partners across all aspects of science. Machines will systematically process data, spot patterns, propose experiments, and even generate hypotheses across more and more of our work. I’m excited for how this will accelerate science.
A technological shift of this magnitude requires a similarly big shift in what we study and how we go about it. For one, there should be significantly more emphasis on what data is important to generate and how we build, share, and scale datasets. For another, there should be much more emphasis on funding system architecture that best enables systematic AI-driven discovery, as opposed to primarily funding individual labs to expand datasets.
Today, I’m happy to announce a new $5M funding initiative in structural biology that will experiment with how we make this shift.
The role of scientists in science
Humans remain essential in science. This is a critical moment to stake out where we can most uniquely contribute. What can’t AI solve yet? And for the things it can, how can we leverage its capabilities to do even more creative, generative exploration beyond that radius?
Machines excel at synthesizing large amounts of information through systematic data processing, pattern recognition, and probability calculations. We should replace ourselves in those types of analyses where possible. It’s an uncomfortable, but necessary transition.
There’s still plenty of upstream and downstream work that only we can do:
- Downstream: ML predictions are hypotheses, not conclusions. We must test them through research and reuse, closing the loop to validate and improve predictions.
- Upstream: Machines work with the data we give them; the nature and design of this data sets the ceiling for what’s possible to predict. We decide what questions to ask and how to architect the systems to answer them.
This piece focuses on the upstream. Too many research proposals focus on generating more data without asking how those expensive exercises will get us somewhere worthwhile. Many brute-force scaling efforts will hit diminishing returns unless we first rethink the kinds of data and data systems we need. This is where human ingenuity will matter most.
The next PDB is the PDB
We need to work smarter, not just harder. As a structural biologist, I’ll give an example close to my heart. And one I’m funding next.
A common trope in funding circles is “What’s the next Protein Data Bank (PDB)?” The question is asked as if structural biology were “done” post-AlphaFold. However, one next big challenge is the PDB; it’s in moving from static protein structures to predicting how they move.
Since form informs function, we use protein structures as clues for what proteins might do in cells. But proteins inherently work through motion: traveling, binding, catalyzing, breathing, and changing conformations. Unlocking protein dynamics would be a giant leap toward uncovering more protein function. This helps us better understand, engineer, and manipulate biology. Biotech companies are already attempting to leverage local data or simulations about protein dynamics for drug modeling.
So one of the “next PDBs” is the PDB itself. It’s not just a remake; it’s PDB 2.0, wherein we can transition from static snapshots to movies. PDB 2.0 can bring protein structures to life.
To get there, we need to think carefully about what conditions are required for such a breakthrough. PDB 1.0’s success depended on being large and standardized, with open data norms. But it also hinged on two other major elements:
- Getting the right slice of information The PDB didn’t contain all protein structures. But it had enough variation in sequences and folds to reveal key design principles. Data was structured for computation (e.g., multiple sequence alignments) to enable breakthroughs. Today, we know more about ML needs, so we can be intentional about focusing our efforts on the data that is most information-rich.
- Scaling through embedded practice The PDB scaled because crystallography naturally produced standardized data as a byproduct of routine work across methods, hardware, file formats, and infrastructure. This reduced marginal costs as the practice spread. It’s very different from a brute-force approach to data generation, where costs keep rising and the volume of data increases.
Finding the right information and embedding it in practice go hand in hand: the right “plumbing” motivates and enables the scientific community to capture valuable data.
We should challenge ourselves to be more cost-effective with PDB 2.0, now that we know more about what we are working towards. We also have machines to help us run quick calculations on what data is most information-dense and therefore most valuable. Instead of taking fifty years and costing tens of billions of dollars, could PDB 2.0 be faster and more efficient?
Principles in practice: introducing The Diffuse Project
With a $5M seed for The Diffuse Project from Astera, work will be coordinated across several universities (UCSF, Vanderbilt, Cornell), scientists associated with national labs and light sources (Los Alamos, Lawrence Berkeley, CHESS), and a team within Astera Institute. Our goal: redesign multiple components of the X-ray crystallography process to capture oft-discarded data on protein conformations.
Traditional crystallography focuses on Bragg scattering, the bright spots that represent the averaged conformation in a protein crystal. Other protein conformations in a crystal produce a more distributed signal — diffuse scattering — which is usually ignored due to its inherent complexity. But today, we have more tools to deal with the messy heterogeneity of diffuse scattering than ever before. Our technical aim is to transform diffuse scattering into a usable signal for hierarchical models of conformational ensembles. I’m excited by how current technologies allow us to better embrace biological complexity in this way.
In close collaboration with experts across the crystallography workflow, we’ve devised a strategy to test whether diffuse scattering could become mainstream practice in structural biology. Our parallel mission is to make the workflow useful, easy, and affordable for anyone studying protein function.
Our team is committed to radical open science by releasing data quickly and sharing progress entirely outside traditional journals — and there have already been many benefits. It has made our brainstorming and collaboration processes much more ambitious and fun. It’s freed us up to focus on outputs that are most useful and representative of our process.
An exciting aspect of the team’s open approach is the chance to experiment in real-time with what scientific publishing could look like in the future. Our open science team is collaborating with our researchers to figure out how they quickly share outputs that are most useful and most representative of our process. As AI becomes more central, our publishing will become more data-centric as well. This is an invaluable opportunity to iterate on this alongside active science to solve real problems on the ground.
We’ve only been in operation for about a month, and we already have data and results to share! You can follow all of this work on The Diffuse Project website.
Utility is the north star
Without journal constraints, we can think more clearly about what will make our science as impactful, accelerated, and rigorous as possible. We keep coming back to the mantra: utility is the north star.
What might high-utility success look like? Our answers span different time horizons, but we strive to be concrete. Here are some examples (which will likely evolve as we go):
- Adoption of diffuse scattering by experts outside the initiative
- Adoption of diffuse scattering by non-experts
- Integration of other modalities with a growing X-ray diffuse scattering dataset
- ML models for protein motion trained on diffuse scattering data
- Biotech start-ups founded from these models
- New therapeutics on the market developed from these models
The diffuse scattering work could take up 7–10 years to reach a scaling inflection point. That’s longer than most projects, but shorter than the time required for the original PDB to scale to a size that enabled AlphaFold. I’m optimistic that we could speed this up if The Diffuse Project truly succeeds in enabling the broader community. It’s a challenge that our team is excited to take on.
A note for funders
Funders, this is the time to flip the script on what science we support. We must move from primarily funding individuals to funding coordinated efforts that “lift all boats” in agent-led discovery systems.
Effective data systems will be a bottleneck for transformational ML. Systems have interdependent parts; if you don’t redesign multiple components at once, you risk getting stuck at a local maximum, due to dependencies between components. This is a well known principle in evolution and engineering. But we’re not great at applying this scientific framework to our own science. We need funding mechanisms that enable sustained, coordinated design sprints across interconnected entities with different areas of expertise.
This kind of systems-level innovation is risky for individual researchers. Not only does it require grappling with the unknown, it also requires grappling with many unknown pieces at once. Any proposal for a project like this would be disingenuous if it provided a super specific roadmap with a high degree of certainty.
Instead, funders need to redefine what success means so that researchers can embrace risk. Namely, teams should prioritize learning and sharing, even about their failures, which can often be informative. The goal can’t be to “win.” It needs to be to learn. The learning process is dynamic (just like proteins!) and we should fund and publish research with that in mind.
Iterating on data systems is also operationally risky. It requires proper infrastructure, engineering, compute access, and ML collaborations. We’re experimenting with unique ways to get this done without creating more bureaucracy or coordination overhead. Astera is providing compute access through Voltage Park and hiring a team for much of the computational and modeling work. This is another avenue by which funders can create an outsized impact through their support.
And at a minimum, we must insist on open science and open data practices. Otherwise, we risk limiting reuse, rigor, and impact. It’s the greatest insurance funders have for a return on their public good investment. Based on my own experiences, I’m very optimistic that we can hold the line on this without sacrificing talent. With the open science requirement, I’ve been able to quickly filter through to some of the smartest scientists, with the most abundant mindsets, with whom to go after big problems. The Diffuse Project’s team is awesome. I could not be more thrilled about the stellar set of scientists leading the research.
A first step towards the future
The Diffuse Project is not an outlier. It’s one example of the broader shift we need to make towards a data-centric future of discovery.
This shift will allow us scientists to work smarter. As machine learning systems become more prominent in hypothesis generation, simulation, and interpretation, the role of scientists will not disappear — it will evolve. Our ability to be creative about what data we generate, how we go about it, and what it enables, will become an increasingly central part of how we contribute to discovery.
This is not a loss. It’s an opportunity. It’s a chance — for both scientists and funders — to reclaim the architectural layer of science as a thoughtful, creative, and essential realm of experimentation and discovery. And it’s a place where human judgment still matters — perhaps more than ever.
If you’re working on something that could benefit from this kind of innovation, I’d love to see it published openly so I can follow and learn through open discourse. In the spirit of open science, I generally prefer not to respond to private emails or DMs. I would love to see open science publishing in effect at the ideation stage too.
Interpretability is one of the biggest open questions in artificial intelligence. In other words, what’s going on inside these models?
Adam Shai and Paul Riechers are the co-founders of Simplex, a research organization and Astera grantee, focused on unpacking this question. They’re working to understand how AI systems represent the world — a critical component of AI safety with major societal implications.
We sat down with them for a conversation on internal mental structures in AI systems, belief-state geometries, what we can learn about intelligence at large, and why this all urgently matters.
How did you both decide to leave academic tracks to start Simplex?
Adam: My background was in experimental neuroscience. I spent over a decade in academia — during my postdoc, I was training rats to perform tasks, recording their neural activity, and trying to reverse-engineer how their brain neural activity leads to intelligent behavior. I thought I had some handle on how intelligence works.
During that time, ChatGPT came out. I was stunned at its abilities. I started digging into the transformer architecture, the underlying technology that powers ChatGPT, and I was even more shocked. The transformer architecture didn’t intuitively tell a story about its behavior.
I realized two things at that time. The first was unsettling. I’d thought of myself as a neuroscientist with a pretty good intuition for the mechanistic underpinnings of intelligence. But that intuition seemed to be completely incorrect here.
The second was even scarier. What are the societal implications for having this kind of intelligence at scale? The social need to understand these systems, and to understand intelligence more broadly, is more important than it’s ever been before. At the same time, there’s also this new opportunity to understand intelligence, not just for artificial systems, but maybe even for humans — to really learn more about our own cognition.
Paul: I came from a background in theoretical physics, thinking about the fundamental role of information in quantum mechanics and thermodynamics, with a recurring theme of predicting the future from the past — how to do that as well as possible. That’s traditionally a core part of physics — your equations of motion help you predict how the stars move, for example. But it also gets into chaos theory, where you have very few details available to predict a chaotic, complex world. And you quickly get to the ultimate limits of predictability and inference.
At the time, this had nothing to do with neural nets and I hadn’t paid much attention to them — they work on predictions of tokens, or chunks of words — but they seemed unprincipled compared to the beautiful theory work I was doing in physics. But then they started working — and not just working, but working really well. To the point where I started feeling uneasy about the future societal implications.
And given my background in prediction, I then started wondering what it might look like to apply the components of a principled mathematical framework to what these neural networks are doing. What are their internal representations? What emergent behaviors can we anticipate? I linked up with Adam and we tried the smallest possible thing we could. And it worked better than we could have hoped for.
That’s what grew into Simplex — we realized there wasn’t really a space for this yet, a program to solve interpretability all the way, and there’s a lot to build on. Our mission at Simplex is to develop a principled science of intelligence, to help make advanced AI systems compatible with humanity. It’s been a really fun collaboration and it’s leading us to a better understanding of how AI models are working internally, which then gives insight into what the heck we’re building as AI systems scale. We’re optimistic these insights can help create more options and guide the development of AGI towards better outcomes for humanity.
Thanks for reading Human Readable! Subscribe to receive news and updates from Astera Institute.
You’ve described AI models not as engineered systems, but as ‘grown’ ones…what do you mean by that?
Adam: Yeah, I think this is an under-appreciated aspect of this new technology. Most people assume they’re engineered programs. But they’re not like other software programs — they’re more like grown systems. Engineers set up the context and rules for growing, and then press play on the growing process. What they grow into and the path of their development isn’t engineered or controlled.
And so we’ve ended up with these systems that are incredibly powerful. They are writing, deploying, and debugging code, solving complex mathematical problems at near-human expert levels, and really peering over the edge of what humans can do. This isn’t a future scenario, that’s where we are right now. But we have very little idea of how these systems work internally. And that’s a problem — a lot of safety issues come from this unknown relationship between the behaviors of these systems and their internal processing.
Paul: For example, AI systems are increasingly showing signs of deception, especially when under safety evaluation. Some actively evade safeguards, while others generate plausible explanations that don’t match their true reasoning. These mismatches between external behavior and internal thought process of the AI highlight the need to understand the geometry and evolution of internal activations — without it, we’re largely blind to the system’s intentions.
So how do you go about studying that internal structure?
Paul: We’ve come up with a framework that started from a part of physics called computational mechanics — basically exploring the fundamental limits of learning and prediction by asking what features of the past are actually useful in predicting the future. And then we leveraged this structural part, which looks at the geometric relationship between these different features of the past.
Adam: We discovered that AI models organize information in consistent patterns, like a mental map consisting of specific types of shapes and structures. We call these patterns belief-state geometries, because fundamentally they are the AI model’s internal representation of what’s going on in the world. These patterns often have repeated structure at different scales, leading to intricate fractals. For instance, when we studied a transformer learning simple patterns, we found it creates a fractal structure that looks like a Sierpinski triangle — where each point represents a different belief about the probabilities of all possible futures. As the AI reads more text, it moves through this geometric space, updating its beliefs in mathematically precise ways that our theory predicted. Now that we have this framing, we can start to anticipate those structures. We can start to predict where in the network to look for them, and test whether our theories hold and we can refine them. It’s a huge shift from just poking around and hoping something interpretable falls out.
How does this differ from what others in the field are doing?
Adam: There’s a field of research called interpretability, which is trying to understand how the internals of AI systems carry out their behavior. It is very similar to the type of neuroscience I used to do, but applied to AI systems instead of to biological brains. There’s been an enormous amount of progress over the last few years, and a lot of interest because of the growth of LLMs. In many ways interpretability has even been able to surpass the progress in neuroscience.
However, the approach is often highly empirical. Despite all the progress, the field is often left wondering to what extent the findings apply to other situations that weren’t tested, or if their methods are really robust, or how to make sense of their findings. In a lot of ways, we still don’t really know what the thing we are looking for is, exactly. What does it look like to understand a system of interacting parts (whether they be biological or artificial neurons) whose collective activity brings about interesting cognitive behavior? What’s the type signature of that understanding?
Paul: What’s missing is a theoretical framework to guide empirical approaches. Our unique advantage is that, by taking a step back and really thinking about the nature of computation in these trained systems, we can anticipate the relationship between the internal structure and behaviors of AI systems. One of the most important points of this framework is that it is both rigorous and also amenable to actual experiments. We aren’t doing theory for theory’s sake. The point is to understand the reality of these systems. So it’s this feedback loop between theory and experiment that allows us to make foundational progress that we can trust and build on in a way that’s different from most other players in the field.
How might this type of geometric relationship between features of the past differ across sensory modalities — language, images, sounds, etc.? Is there a difference between how it works in humans and AI systems?
Paul: This is an area we’re really interested in exploring more. There’s some evidence that suggests that neural networks trained on different modalities converge on similar geometric representations of the world. It seems to be that no matter which modalities an intelligent system uses to interact with the world, it’s trying to reconstruct the whole thing.
It raises some fascinating questions — to what extent are different modalities and even different intelligences converging on a shared sense of understanding? Is there a unique answer to what it’s like to be intelligent? And if so, maybe that’s useful for increasing the bandwidth of communication among different intelligent beings? And if not — if each intelligent thing understands the world in a valid but incompatible way — that’s maybe not great for hopes of us being able to come to a shared understanding and aligned goals? So that’s one thing we’re very interested in as we learn how models represent the world.
Adam: I think there’s also a really interesting opportunity to understand ourselves better. It opens up this entire new field of access to understand intelligence. If you’re a neuroscientist, you no longer have to decide between studying a human brain and having very low access, or studying a rat and having more access at the expense of cognitive behavior. With neural networks, you can look at everything.
We can even potentially engineer neural networks to whatever level of complexity, whatever kind of data, and whatever kind of behavior we want to study, even at a level that exceeds human performance. And it opens up this new, really fast feedback loop between theory and experiment — it’s an unprecedented opportunity to understand intelligence in a very general sense.
Given your work is really about understanding intelligence more broadly, beyond just AI systems, where do you hope it will lead?
Adam: Previously, there’s been no framework that gives us any kind of foothold to talk about this relationship between internal structure and outward behavior. We’re trying to build this kind of principled framework for how to think about these questions. It could be applied to LLMs in order to understand them and make them safer. But the general framework also has the promise of being applicable to other systems where we’re trying to understand the relationship between the internal structure and outward behavior. And those other systems could be biological brains.
Paul: Part of the value we’re providing is also a shared understanding, or ground truth for how these systems work. Today, people have different opinions about what these systems are — some maintain that the current AI paradigm will fall short of AGI or superhuman intelligence, while a growing contingent finds it obvious that you can bootstrap a minimal amount of intelligence to become superhuman across the board. More concerning is that informed technologists disagree about whether ASI (artificial superintelligence) will most likely lead to human extinction or flourishing. Even among experts, people really talk past each other. AI safety may or may not be solvable. Part of our work is to establish a scientific foundation for coming to a consensus on that, and identifying paths forward.
I’m hopeful that we can elevate the conversation by creating a shared understanding of what the science says, so it’s no longer doomers and optimists, but rather people working together to figure out the implications of what we’re building, and how we steer towards the kind of future we want. It’s a lofty goal but we think it’s possible. As we continue to build our understanding of structures in our own networks, we’ll hopefully be able to leverage that for a societal conversation for what it is we want to be building towards.
Learn more about Simplex’s insights and follow their technical progress here.
What does it take to make — and keep — a planet habitable?
For over 50 years, humans have explored space, seeking new homes for life. Life on Mars is the stuff of great science fiction. And the work of actually creating sustainable habitats and ecosystems beyond Earth has, by extension, been a far-flung future. Now, that may be changing.
Edwin Kite is a planetary scientist and current resident at Astera who, together with his team and collaborators, is working on defining a contemporary Mars terraforming research agenda. We spoke with him about what it would take to warm Mars up enough for life to thrive, how open source tools and datasets help research communities build towards the future, and what drives a scientist to investigate how people might create ecosystems beyond Earth.
You’ve been working on Mars science for years. Why this planet?
The biggest unanswered question in Earth science is how and why our planet stayed habitable for life.
For example, for nine-tenths of Earth history, our planet has been uninhabitable for humans: we don’t know why oxygen levels rose.
These are especially interesting questions when you consider that Mars was once habitable, but lost its ability to sustain life. Mars holds a record of that environmental catastrophe, and may hold traces of life that established itself there before that natural disaster. We are in a golden age of Mars science today, with two plutonium-powered rovers on the surface and an international fleet of spacecraft in orbit. We can deeply explore the planet for signs of this record and seek answers to our questions about what happened. It’s a great time to be doing Mars science.
The biggest unanswered question in Earth science is how and why our planet stayed habitable for life.
Understanding what made and what ended Mars’s early habitability can also help us better understand Earth’s history of life and explore the possibility of re-making Mars habitable. Mars has plenty of water and carbon, and its surface receives about as much sunlight as does all of Earth’s land. Sunlight powers almost all of our biosphere, so it’s tantalizing to think of what kind of biosphere sunlight might support in the future on Mars. It’s by coevolution with photosynthetic life that people built cities.
Our ancestors were, in Darwin’s words, hairy forest-dwellers. They moved outwards and built tools like spacesuits and sealskin coats to allow human life in the face of once-unimaginable hazards, like hard vacuum and winter snow. But this approach can only take us so far. Throughout time, space has been forbidding to life, with radiation, micrometeoroids, and cold. If we are going to have an adventure that’s endless, we’ll need to adapt the environment to ourselves. I don’t know how we’ll do this, but the bigger rocky and icy worlds of our solar system seem like a logical place to start. Many are rich in life-essential-volatile elements, all can offer radiation protection, and all have enough gravity to hold onto a stable atmosphere.
What’s the scope of your current project at Astera, focused on terraforming?
It’s been understood for over 50 years that there would be two steps to making Mars more Earth-like: first warm the planet up to allow photosynthesis, which is relatively quick and easy, then build up the oxygen level using photosynthesis. We’re looking at the first step, surface warming.
Mars is too cold for stable liquid water — the average temperature is around 210 K (about -60º C), and atmospheric pressure is only ~6 millibars. Warming the planet by 30 – 50°C could melt near-surface ice, enabling surface habitability and photosynthesis. There are lots of ways to warm Mars, including greenhouse gases and orbiting mirrors. Our team at Astera is investigating a warming approach based on engineered aerosols — specifically, nanoparticles that can forward-scatter sunlight and block thermal infrared. Compared to greenhouse gases, they’re four orders of magnitude more mass-efficient. That kind of efficiency matters — you want to get the biggest radiative payoff. If we want to make progress in this century, we need to use the materials that are already readily available on Mars, rather than shipping them in.
Along with the simulation work and delivery prototyping that we’re doing at Astera, we are working with collaborators at Northwestern University to batch-manufacture and test the most promising particles. This work is part of an extended collaboration involving scientists from Aeolis Research, JPL, the University of Central Florida, MIT Haystack Observatory, and the University of Chicago Climate Systems Engineering Initiative, among other institutions.
What properties of aerosols lead to warming the planet, rather than cooling it?
On a clear-sky night when we can see the stars, it’s typically cooler than on a cloudy night. So clouds (a form of aerosol) act as a warm blanket. Clouds also bounce sunlight back to space (cooling effect). For any aerosol, the net effect (warming minus cooling) depends on the size, shape, and composition of the aerosol. To warm Mars, we need to choose/design a combination of size, shape, and composition that gives a strong warming effect.
We also need to pay attention to particle mass. In our recent Science Advances paper, we showed that certain nanoparticle designs can achieve the same warming effect as fluorocarbon gases (a particularly potent greenhouse gas), but with ~50,000 times less mass. That matters, because even with the improved launch economics we’re seeing, getting mass to Mars is still expensive, so we need to keep the particle factory as lightweight as possible.
How would these nanoparticles be made?
The current concept involves dispersing nanoparticles into Mars’ atmosphere, where they remain aloft for long periods. These particles — which could include graphene disks or metal ribbons or even natural salts — selectively scatter shortwave solar radiation while blocking outgoing infrared. This alters the radiative balance, raising surface temperatures.
We’re exploring multiple production pathways. For example, it may be possible to fabricate aerosols using materials from Mars regolith, or using Mars’ CO2-rich air as the feedstock for making graphene disks as a byproduct of oxygen production. For a solar-powered graphene production, the basic ingredients for warming Mars could be Mars’ air and sunlight.
Particles also have another advantage: they start working within months and stop working when removed. That makes the system controllable and potentially reversible. So we could switch off or adjust a well-designed intervention as warming proceeds.
We hear your team has already made advances on researching particles 👀
Yes! One of our researchers, Alex Kling, developed an open-source screening tool to assess the warming efficiency of different particle types. It’s motivated in part by a conversation at the Tenth Mars conference, where Mars climate researchers agreed we needed something like this for natural aerosols as well. And I hope it will also be used by exoplanet researchers. Natural aerosols can be really important in extending the habitable zone, in both warming and cooling directions (for example, the type of high altitude organic haze found on Saturn’s moon Titan can be really effective in cooling a world’s surface).
Our tool is already available on GitHub, and it’s forming the basis for our next set of experiments. You can find it here:
We’re also going to batch-manufacture and test particle interactions and production protocols — essentially, answer questions like, how hard is it to make these materials from scratch? What kinds of impurities or degradation modes do we need to model? And we’ll continue sharing our results openly through preprints, on Zenodo, and on Substack.
Why is it important to do this work as open science?
Making all our findings public is necessary for informing a spacefarer’s consensus about what to do with those findings. The Outer Space Treaty says, “The exploration and use of outer space […] shall be the province of all humankind,” and every country has signed up to that.
Because we’re starting a new field, we have to work in a way that others can plug into. We have to seed open standards, protocols, and cultural momentum now, to create infrastructure that enables many organizations to work on terraforming in parallel. We’re building tools, datasets, and protocols that others can test, benchmark against, and improve. That includes publishing aerosol designs, experimental methods, and model inter-comparison frameworks — both the successes and the failures. We need to create a foundation that the whole research community can build on. And culturally, this aligns with NASA’s approach — raw rover images are public as soon as they arrive — and it’s a model that’s worked well for decades.
How do you think about the ethical questions that arise from a proposal to terraform Mars?
Perhaps Mars should remain untouched, as a planet-sized wilderness park that people bypass on our way to the stars? Or we could see Mars as an environmental restoration problem: undoing the collapse of a once-habitable planet?
Personally, I think a night sky full of living worlds is better than one that’s dead. Mars is Galactic gardening for beginners. If Mars is lifeless, then there’s no ecosystem to protect, and creating a new one might be the most meaningful thing we could do with it. But more work is needed to search for life on Mars, and decisions about Mars’ future should be made within a democratic framework. Any serious terraforming effort would take decades, if not centuries. On those long timescales, our politics and our worldviews will change. What won’t change are the basic scientific constraints, and it’s the science that we propose to study.
How can people learn more?
You can read the essay with Robin Wordsworth or check out our latest technical study. And keep an eye out for our forthcoming technical Substack!
Throughout a meandering career in government, academia, policy think tanks, startups, and investing, I have spent a lot of time thinking about what stalls progress and what to do about it to catalyze a more abundant future. I’m thrilled to have joined Astera to put this into practice by leading our strategic investments in the visionary science and transformative technologies that can yield an abundant future.
What excited me most about coming to Astera was the opportunity to develop a theory of where to deploy our philanthropic resources. Astera’s values and broad approach provide a framework for how we are going to deploy capital — efficiently, catalytically, creatively, boldly. But what we invest in could take many forms. We want our philanthropy to increase human flourishing, innovation, abundance, transformative science, and other abstract ideals. But there is no “human flourishing store” we can walk down to to make a purchase. We have to deploy resources into specific projects. Which ones? Why?
Any theory we develop to answer these questions will be iterative, subject to continual feedback and refinement. We expect to have some failures and to talk openly about the lessons we learn along the way. And yet I do think there is a good meta-framework for thinking about these issues. Last year I had the privilege of speaking at the Progress Conference 2024. My talk focused on explaining the causes of the Great Stagnation. Some of my charts were pretty grim — I called it the Dark Steven Pinker portion of the talk — but I ended on a more optimistic note: a way forward for the progress movement.
It’s a three-pronged approach involving technoeconomics, sociopolitics, and entrepreneurship.
Technoeconomics refers to the study and analysis of how technological capabilities and constraints interact with economic factors — like costs, productivity, and market dynamics — to shape the feasibility, performance, and adoption of technologies to grow the economy. What technologies are technically feasible and economically viable? How badly do customers want them? How do changes in input costs, learning curves, or performance metrics affect competitiveness? What is the optimal configuration of a system given technological and economic trade-offs?
These questions are critical if we want to have a concrete vision of the future we are trying to build. It’s one thing to say that a good future should include abundant energy. But if we are going to actually build a future with abundant energy, we need to start developing some definite opinions about which sources of energy can contribute in what ways. What are our projections for solar deployments? Can nuclear fission ever be cheap again? Why is it so expensive in the first place, given that it was so cheap in the 1970s? What about geothermal and fusion? What variants of each are most promising? How do solar and batteries fit into this?
Only after we engage with these kinds of questions will we have a basis for steering resources into energy technologies. We need to start with a concrete vision, and technoeconomics is the way to get one.
Thanks for reading Human Readable! Subscribe to receive news and updates from Astera Institute.
Sociopolitics refers to the intersection and interaction of social and political factors in shaping institutions, decisions, and outcomes. Technoeconomics tells us what’s fundamentally possible and what customers want, but sociopolitics is the source of cultural and institutional barriers to what is possible—what the rest of society wants. What kinds of innovations will be culturally rejected or refused by regulators or other institutions? What kinds of innovators and innovations are rewarded with money or status?
The ways that sociopolitical issues impact innovation start far upstream in the scientific process. Where does the funding for scientific and technological innovation come from? What are the incentives that these funding and reward mechanisms create for the kinds of questions that are asked in the first place? With our open science policy, we made the decision to defund journal publication, in part because journals create an incentive to develop polished narratives instead of discovering useful truths. As a science and technology funder, we are acutely aware of the shortcomings of other funding sources and also of how hard it can be for us to skillfully deploy funding. What are we tacitly incentivizing? How do we account for the fact that people tell us what we want to hear? We won’t always get everything right, but we are committed to grappling with these questions and working to foster institutions that are aligned with our mission of supporting visionary science and transformative technology.
Many people think of entrepreneurship as the activity of setting up a business. I like a different definition I got from Tom Kalil. Tom says that an entrepreneur is someone who is not limited by resources under direct control. An entrepreneur finds a way to get the job done while outperforming what they “should” be able to do with their starting resources. You can be a business entrepreneur, but you can also be a philanthropy entrepreneur or a science entrepreneur.
Resources are not just money. They are whatever is required to overcome the problems identified in the technoeconomics and sociopolitics part of the analysis. Projects require buy-in from all kinds of stakeholders, not just funders — what does it take to achieve that buy-in? For Astera to be unreasonably successful, we need to persuade the world’s most innovative scientists and technologists to work with us in sometimes unconventional ways. And we also need to catalyze change in other philanthropic organizations in the process. True success requires us to be entrepreneurial.
I believe that applying these three lenses will serve us well as we decide what kinds of projects to fund to drive transformative scientific and technological change. But — surprise! — it’s also what I think we should look for in the teams we support. We are looking for teams with ambitious and concrete ideas for what a compelling future could look like (technoeconomics), who are clear-eyed about the nontechnical challenges their projects may face (sociopolitics), and who have a relentless determination to surmount those challenges (entrepreneurship).
These attitudes are not the default in science and technology. What’s more common is an indefinite, nonspecific sense that the future will be better or worse. When there is a concrete vision, it is often not accompanied by an interest in engaging with the work required to instantiate it. And with the first two elements lacking, entrepreneurship in science and technology often takes on a performative, going-through-the-motions form.
If you are a scientist or technologist with a concrete vision for a better future, an intense interest in the obstacles to actualize it, and an unflagging drive to make it happen, please reach out. I’d love to hear from you.
Investors sign hundreds of term sheets over their careers. Founders only do so a handful of times. When you’re raising money to bring your idea to life, there are a myriad of decisions you need to make — what sources of funding to pursue, how to structure your company, what timeline you should work towards. Often, these are decisions you’re making for the first time. And when you search for fundraising advice, most of what you find is directed at tech startups and told from an investor’s point of view.
We’re flipping that script. The Founder’s Guide to Funding Health and Science Organizations is a 50+ page handbook by founders Andrea Coravos and Rachel Katz, that captures their own experiences of building and successfully exiting companies, along with dozens of behind-the-scenes conversations with other health and science founders.
It covers what traditional guides leave out — how to build momentum, how to negotiate when you’re not sure you have leverage, and strategies to make funding decisions that actually align with your mission. Whether you’re exploring venture, grants, philanthropic capital, or revenue-first models, this guide is meant to help you navigate complexity and find sources of funding that let you build your company on your own terms.
At Astera, we are eager to support others who can transform their hard-won insights into public goods that enable others who are building new technologies. You can find links to this guide and other resources for startups and nonprofits on our website. Please reach out if you have insights and tools to share for guiding science and technology toward greater impact!
In Abundance, Ezra Klein and Derek Thompson make the case that the biggest barriers to progress today are institutional. They’re not because of physical limitations or intellectual scarcity. They’re the product of legacy systems — systems that were built with one logic in mind, but now operate under another. And until we go back and address them at the root, we won’t get the future we say we want.
I’m a scientist. Over the past five years, I’ve experimented with science outside traditional institutes. From this vantage point, one truth has become inescapable. The journal publishing system — the core of how science is currently shared, evaluated, and rewarded — is fundamentally broken. And I believe it’s one of the legacy systems that prevents science from meeting its true potential for society.
It’s an unpopular moment to critique the scientific enterprise given all the volatility around its funding. But we do have a public trust problem. The best way to increase trust and protect science’s future is for scientists to have the hard conversations about what needs improvement. And to do this transparently. In all my discussions with scientists across every sector, exactly zero think the journal system works well. Yet we all feel trapped in a system that is, by definition, us.
I no longer believe that incremental fixes are enough. Science publishing must be built anew. I help oversee billions of dollars in funding across several science and technology organizations. We are expanding our requirement that all scientific work we fund will not go towards traditional journal publications. Instead, research we support should be released and reviewed more openly, comprehensively, and frequently than the status quo.
This policy is already in effect at Arcadia Science and Astera Institute, and we’re actively funding efforts to build journal alternatives through both Astera and The Navigation Fund. We hope others cross this line with us, and below I explain why every scientist and science funder should strongly consider it.
Journals are the Problem
First, let me explain why this is such a big deal to those who are new to this issue. It might seem like publishing is a detail. Something that happens at the end of the process, after the real work of science is done. But in truth, publishing defines science.
The currency of value in science has become journal articles. It’s how scientists share and evaluate their work. Funding and career advancement depend on it. This has added to science growing less rigorous, innovative, and impactful over time. This is not a side effect, a conspiracy, or a sudden crisis. It’s an insidious structural feature.
For non-scientists, here’s how journal-based publishing works:
After years of research, scientists submit a narrative of their results to a journal, chosen based on field relevance and prestige. Journals are ranked by “impact factor,” and publishing in high-impact journals can significantly boost careers, visibility, and funding prospects.
Journal submission timing is often dictated by when results yield a “publishable unit” — a well-known term for what meets a journal’s threshold for significance and coherence. Linear, progressive narratives are favored, even if that means reordering the actual chronology or omitting results that don’t fit. This isn’t fraud; it’s selective storytelling aimed at readability and clarity.
Once submitted, an editor either rejects the paper or sends it to a few anonymous peer reviewers — two or three scientists tasked with judging novelty, technical soundness, and importance. Not all reviews are high quality, and not all concerns are addressed before editorial acceptance. Reviews are usually kept private. Scientific disagreements — essential to progress — rarely play out in public view.
If rejected, the paper is re-submitted elsewhere. This loop generally takes 6–12 months or more. Journal submissions and associated data can circulate in private for over a year without contributing to public discussion. When articles are finally accepted for release, journals require an article processing fee that’s often even more expensive if the article is open access. These fees are typically paid for by taxpayer-funded grants or universities.
Several structural features make the system hard to reform:
- Illusion of truth and finality: Publication is treated as a stamp of approval. Mistakes are rarely corrected. Retractions are stigmatized.
- Artificial scarcity: Journals want to be first to publish, fueling secrecy and fear of being “scooped.” Also, author credit is distributed through rigid ordering, incentivizing competition over collaboration. In sum, prestige is then prioritized.
- Insufficient review that doesn’t scale: Three editorially-selected reviewers (who may have conflicts-of-interest) constrain what can be evaluated, which is a growing problem as science becomes increasingly interdisciplinary and cutting edge. The review process is also too slow and manual to keep up with today’s volume of outputs.
- Narrow formats: Journals often seek splashy, linear stories with novel mechanistic insights. A lot of useful stuff doesn’t make it into public view, e.g. null findings, methods, raw data, untested ideas, true underlying rationale.
- Incomplete information: Key components of publications, such as data or code, often aren’t shared to allow full review, reuse, and replication. Journals don’t enforce this, even for publications from companies. Their role has become more akin to marketing.
- Limited feedback loops: Articles and reviews don’t adapt as new data emerges. Reuse and real-world validation aren’t part of the evaluation loop. A single, shaky published result can derail an entire field for decades, as was the case for the Alzheimer’s scandal.
Stack all this together, and the outcome is predictable: a system that delays and warps the scientific process. It was built about a century ago for a different era. As is often the case with legacy systems, each improvement only further entrenches a principally flawed framework. It’s time to walk away and think about what makes sense now.
What We’ve Learned So Far
We’re in a bit of a catch-22 as a scientific community in that we don’t have a solution to jump to, but we also can’t develop one well if we continue with journals. Prohibiting journals is our deliberate forcing function as we support such development at Astera and The Navigation Fund. By removing journals as an option, our scientists have to get more thoughtful about how, when, and why they publish. We’ve started to see some shapes of the future.
We began this as an experiment at Arcadia a few years ago. At the time, I expected some eventual efficiency gains. What I didn’t expect was how profoundly it would reshape all of our science. Our researchers began designing experiments differently from the start. They became more creative and collaborative. The goal shifted from telling polished stories to uncovering useful truths. All results had value, such as failed attempts, abandoned inquiries, or untested ideas, which we frequently release through Arcadia’s Icebox. The bar for utility went up, as proxies like impact factors disappeared.
Peer review has also become better and faster for us at Arcadia. It’s a real mechanism for improving our rigor, not a secret editorial gate. We often get public feedback, and we use it to openly improve our work in real time. Another recent example of accelerating public peer review was a study about room temperature superconductivity that was released in 2023, got bombarded on twitter, and was then countered by several independent validation studies in less than a month. The controversial work happened outside of journals, and it wasn’t just peer reviewed, it was peer tested. Evidence-based community consensus happened at lightning speed.
It’s important to note that you don’t have to opt-out of academia to try something new. Astera recently funded a major structural biology project involving multiple academic groups, and the scientists enthusiastically agreed to forge a path without journals. It has been a delightful experience to think more clearly with them about the true impact of their work. The potential outcomes have to be so valuable for what they are — in this case, scalable X-ray crystallography methods that advance our understanding of how proteins move — that they transcend journal proxies. Expansive, iterative reuse of their methods is a more worthwhile goal than shiny comments from three anonymous reviewers. Those are the kinds of ambitious projects we like to fund.
Pre-prints are also a great way for anyone to participate now. And we need scientists to experiment with more radical formats. Pre-prints still maintain many journal features and are typically released close-in-time to journal articles, for which they are ultimately designed. In contrast, digital notebooks designed for computational work, such as Jupyter, allow for entirely different paradigms of publishing. Arcadia and others are now playing around with ways to automate conversion of such notebooks to publishable, dynamic outputs that can self-update as linked data evolves (see here, here, and here).
These experiences have converted me completely. I can’t unsee this new world. I look forward to making that true for more people by helping them take the first step.
What Could Happen Next
So how do we start? It’s important to define core publishing requirements before trying stuff. In 2016, a group of scientists, publishers, and funding agency representatives put forth the FAIR (Findable, Accessible, Interoperable, and Reusable) Principles, and they can be summarized as follows:
- Findable means we can discover stuff across digital space and time. Published items need to be linked to information about their creators and a long-lasting unique identifier, like a DOI (or Digital Object Identifier).
- Accessible means that you can easily find and search for publications using normal things like Google scholar or even ChatGPT. There shouldn’t be extra barriers, like paywalls, to finding published work.
- Interoperable means that you should be able to connect information across different formats and venues, which will only get more important as we leverage more AI tools over time.
- Reusable means that it’s possible for others to build on published work, which requires information and permissive licenses.
Work also needs to be permanently archived so that it’s accessible in the long term, which is not a new problem and remains largely unsolved. We especially need to figure this out for large datasets and repositories that the community relies on.
Scientists should probably be putting out shorter narratives, datasets, code, and models at a faster rate, with more visibility into their thinking, mistakes, and methods. In this age of the internet, almost anything could technically be a “publishable unit.” It doesn’t even have to sound nice or match the human attention span anymore, given our increasing reliance on AI agents.
In more general terms, we need publishing to be a reliable record that approximates the true scientific process as closely as possible, both in content and in time. The way we publish now is so far from this goal that we’re even preventing our own ability to develop useful large language models (LLMs) that accelerate science. Automated AI agents can’t learn scientific reasoning based on a journal record that presents only the glossy parts of the intellectual process, and not actual human reasoning. We are shooting ourselves in the foot.
What Could Be Possible
What are some practical non-journal options for publishing available now? Many FAIR options exist outside of journals today that are ready to go, and we wrote up our recommendations for Astera scientists recently.
However, we are far from achieving what’s possible. In addition to the many issues inherent to journal processes, it’s a system that was never designed to scale with our needs today. When journals first emerged in the 17th-18th century, they were responsible for handling outputs from hundreds of scientists. In 2021, a UNESCO report estimated that there’s approximately 9 million full-time researchers around the world that publish millions of articles across about 40,000 journals.
It’s also not possible to architect a viable alternative system in a vacuum without users to iterate with. We urgently need more scientists to try strategies that stretch our imagination for what the future could hold. Below are just a few ideas that could be fun to explore:
- Tools to automate the finding and collating of peer review, wherever it lives, e.g. published reviews, comments, social media, meeting notes, conferences, reuse of the science in the real world, etc.
- Ways to directly share lab notebooks and rely on LLMs to dynamically organize and synthesize information, over time. These outputs can be customized based on the readers’ interests.
- LLMs or knowledge graphs that help anyone quickly landscape scientific areas, flag conflicting data, or quantitatively score the novelty of new studies.
- Autonomous agents that can analyze actual data to generate new hypotheses propose alternative interpretations of published conclusions.
- The ability to connect reviews with real world outcomes at a later point-in-time through betting markets or interpretability work on different autonomous agents.
I don’t know if these are the right ideas, which is why we need to try them. But all of these ideas ladder up to giving both authors and readers more agency and optionality to meet their own needs. You drive your own content. You can curate and assess science without an editor. And everyone doesn’t necessarily need to agree on a single, centralized solution. Today’s technologies allow for a multitude of strategies that can be layered on top of the internet and augmented over time. We can have choices.
Towards a Better Return-on-Investment
You might think that scientific publishing would be too costly to revamp, even if we had clear solutions. But journal-based publishing already costs the global scientific community $10-25 billion per year for subscriptions and article processing fees, most of which are paid for using taxpayer-funded grants. In addition, a conservative estimate of millions of scientist hours are spent on journal publications annually. Currently, a significant portion of the science community outside the U.S. can’t even afford to participate via journals.
This is an expensive price to pay in exchange for an unreliable record, immeasurable delays, opportunity costs, and degradation of public trust. I cannot think of a worse return-on-investment for scientists, science funders, and society than continuing to enable journal publishing.
We don’t need to wait for permission to fix this. The future of science is not going to be rescued by journals or legacy institutions. We need to reclaim science’s role in serving society. I often hear scientists say they can’t abandon journals because they will lose their funding. As a funder, I’m letting you know that we’re not just comfortable with new publishing strategies, we require it.
If our approach sounds exciting to you, send us your ideas, apply for funding, and participate in our experiments. If you’re a fellow science funder, I’d love for you to join us in holding the line for change.
Seemay Chou is Astera’s Co-Founder and Interim Chief Scientist.
When Dakota Gruener first began her career in global health policy, she was driven by a simple but powerful question: what are the biggest levers we can pull to improve human wellbeing? In pursuing that question, she’s traveled across continents and disciplines, from vaccine access in Cameroon to digital identity systems in Bangladesh. After a stark revelation about climate displacement, she’s now taking on one of the most complex and controversial frontiers in climate science: sunlight reflection.
Today, as a Resident at the Astera Institute, Dakota leads Reflective, a research organization focused on evaluating the science, risks, and governance of methods that could reflect sunlight away from Earth and potentially cool the planet. We sat down with her to explore her approach, why sunlight reflection demands urgent research, and what it means to pursue climate solutions that could prove essential, even when they come with a great deal of unknowns.
Read the full interview on our Substack: https://asterainstitute.substack.com/p/reflecting-on-and-for-our-future