Category theory for legible software and corrigible world modeling
About the Topos Institute
Mission: to shape technology for public benefit by advancing sciences of connection and integration.
Three pillars of our work, from theory to practice to social impact:
- Collaborative modeling in science and engineering
- Collective intelligence, including theories of systems and interaction
- Research ethics
. . .
- Physics simulations (PDEs) with Decapodes.jl
- Reaction networks with AlgebraicPetri.jl
- Epidemiological modeling with StockFlow.jl
- Agent-based modeling with AlgebraicRewriting.jl
- Interactive GUIs with Semagrams
- Vision: topos.institute
- Research: topos.site
- Blog: topos.site/blog
Hi, I’d like to start by introducing my organization, the Topos Institute. You might know our founders, Brendan Fong and David Spivak, or some researchers such as Sophie Libkind or Owen Lynch. We are a nonprofit research institute aimed at social impact via socially-conscious technology. Of our three pillars of work, I’ll just focus on the first one. (…)
We have a lot of projects that exemplify the principles I’m going to showcase today, but I want to be conscious about what can actually be communicated in the space of a talk, so I hope you can forgive that my actual examples are going to be very small toy programs that can fit on a single slide, since I want you to deeply understand the difference between a traditional approach and our approach . Feel free to interrupt if something isn’t clear, since I’d rather get through one example deeply than breeze through all three examples I have planned.
Why is this relevant to AI safety?
- We can use verified, interpretable models (AI or otherwise) without losing too much economic value
- Reduce incentive to hand authority over to uninterpretable black boxes.
In this slide I’m going to draw some connections between our collaborative modeling and some basic ideas in AI safety. I’m just going to make a lot of claims, and then we’ll return here at the end to see whether or not I was able to show why, in principle, that our category theoretic approach to modeling and software engineering can be helpful in these ways.
So to start with #1: this image from Davidad’s ARIA’s Program. It’s about “shifting the incentives from the race to the bottom, by derisking credible paths to AI systems where we have strong reasons to expect confidence in the safety of systems”.
Our compositional approach to modeling is uniquely poised to allow both large, complex models (which are needed for the economic value) while retaining the verifiability and interpretability.
Why is this relevant to AI safety?
- We can use verified, interpretable models (AI or otherwise) without losing too much economic value
- Reduce incentive to hand authority over to uninterpretable black boxes.
- AI can target domains with a well-specified syntax and semantics allows for legibility and explanation
- Davidad’s ARIA proposal: world-models as autonomous AI output
- Anna Leshinaskaya’s talk on moral decisionmaking w/ combinatorial grammar
- Yoshua Bengio’s emphasis of separating model from implementation
As a second point, I’ve listed some examples of people who understand we don’t want raw decisions to be the output of our model, rather we want to target languages which we can interpret and interrogate. This shows up in the ARIA plan, where the AI is only interfacing the decisionmaking via providing world models which we can prove things about. Likewise, in the last two FAR seminars, we saw how moral decisionmaking can be made safer by stipulating a graph-like grammar with actors and outcomes, which can then be used to compute a decision but also explain why that decision came about. Likewise, this is the essence of Yoshua’s point about separating the model from the implementation. Making this easier is precisely one of the strengths of Topos’ approach to modeling and software.
Why is this relevant to AI safety?
- We can use verified, interpretable models (AI or otherwise) without losing too much economic value
- Reduce incentive to hand authority over to uninterpretable black boxes.
- AI can target domains with a well-specified syntax and semantics allows for legibility and explanation
- Davidad’s ARIA proposal: world-models as autonomous AI output
- Anna Leshinaskaya’s talk on moral decisionmaking w/ combinatorial grammar
- Yoshua Bengio’s emphasis of separating model from implementation
- We facilitate collaboration between diverse stakeholders, better alignment of values
- Opens opportunities for more equitable representation of values from all stakeholders (rather than just those who’ve mastered dependent type theory)
Lastly, on the topic of aligning AI with our values, it’s widely understood that any formalization of values should be the result of a kind of dialogue with relevant stakeholders, but programmers, therapists, ecologists, etc. are all coming from different perspectives and different dialects for talking about their knowledge of the world and understanding of what harm or value is.
Topos offers a sophisticated account of domain specific languages for domain experts that can be formal and interoperable, which are critical for this kind of task.
Outline
I. Future of software engineering
II. Mini examples of the future paradigm
III. Intuition for how structural mathematics works
Ok with those to introductory notes out of the way, let me get to the main portion of the talk.
Here is the structure.
Note: it’s not feasible to cover category theory at a technical level in such a short talk, rather I want to show what you can do with it. So I’ll be avoiding jargon whenever possible.
Stages of development for software engineering
- We solve problems one at a time, as they come.
. . .
- Abstractions act as a solution to an entire class of problems.
Common abstractions are things like functions / datatypes / scripts in our language.
. . .
- We migrate abstractions to allow them to address changes in problem specification.
Understand structure of the abstraction + implementation + the relationship between them.
This slide summarizes everything I want to talk about today. (Stage 1) You can imagine a story told about the history of programming where, once upon a time, we solved problems one at a time. New problem, write some new code.
(…) (Stage 2) Then something magical happened, we learned to use abstractions. These are parameterized and for each set of input parameters we solve another problem in some problem class. This seems great, but to the extent these entire problem classes feel related, even this feels repetitive. If you unfocus your eyes a bit and see that you could think of this entire problem class as a kind of ‘big problem’ and we haven’t made any progress at all, in fact!
(…) (Stage 3) I want to propose our vision for a future paradigm, where we have the right language for talking about the structure of the new problem class as it relates to the old one, we ought to be able to automatically derive what the new abstraction is that solves that class.
There will be lots of examples illustrating how we can move from Stage 2 to 3 when we use abstractions that we understand well. I’ll try to clearly show what’s going on for toy examples.
Three examples of abstractions
In this section of the talk: I’ll make reference to three specific types of abstractions which I’ll introduce one at a time. These are databases (here I’m just showing a particular toy database schema), Petri nets (also called bipartite graphs), and wiring diagrams
Abstraction: databases (SQL)
CREATE SCHEMA COLORGRAPH;
CREATE TABLE VERT (ID PRIMARY KEY, COLOR VARCHAR);
CREATE TABLE EDGE (ID PRIMARY KEY,
FOREIGN KEY TGT REFERENCES VERT(ID),
FOREIGN KEY SRC REFERENCES VERT(ID));
INSERT INTO VERT VALUES ('B')
INSERT INTO VERT VALUES ('R')
INSERT INTO VERT VALUES ('Y')
INSERT INTO EDGE VALUES (1,2)
INSERT INTO EDGE VALUES (2,3)
INSERT INTO EDGE VALUES (3,1)
Here is an example specific database schema. Schemas are abstractions which codify various logical constraints, such as “every edge has a source and target vertex” and “every vertex has a color”.
Below are two representations of a particular instance of this database. To the right are some SQL statements which would create such a triangle.
Abstraction: databases (Python)
class ColorGraph():
def __init__(
self,
int, int}},
edges:Set{Tuple{int},
vertices:Set{int, str}):
colors:Dict{ ...
= Schema(
SchemaColorGraph =["V", "E"],
Ob=[("src", "E", "V"), (tgt, "E", "V")],
Hom=["Color"],
AttrType=[("color", "V", "Color")]
Attr
)= acset_type(SchemaColorGraph, [str]) ColorGraph
def triangle(graph=ColorGraph()):
= [add_vertex(graph, color)
a, b, c for color in "BRY"]
for
[add_edge(graph, s, t) in [[a,b],[b,c],[c,a]]]
(s, t) return graph
. . .
In addition to the SQL way of framing this, we could also do this with a class or struct, though admittedly that would not capture the foreign key requirements. On the right, I show how we declare this data structure in AlgebraicJulia, which is declaring the database schema. (…)
For our first concrete example of abstraction reuse: consider this big triangle. We could always just code this up again from scratch, but what if we recognize that the big triangle is just three different triangles that have been glued together? It’s a mark of a good abstraction for our triangle
function that it permits us to do this.
Gluing: side-by-side comparison
"""Modify to allow it to overlap existing vertex IDs"""
def triangle(graph, v1=nothing, v2=nothing, v3=nothing):
= [add_vertex(graph, color) if isnothing(v) else v
vs for (v, color) in [v1=>"B", v2=>"R", v3=>"Y"]]
for i,j in [[1,2],[2,3],[3,1]]:
add_edge(graph, vs[i], vs[j])return vs
"""Now expressible as an abstraction of `triangle!`"""
def big_tri(graph):
= triangle(graph)
blue, red, _ = triangle(graph, blue)
_, _, yellow triangle(graph, nothing, red, yellow)
def replace(graph, fk:str, find:int, repl:int):
# TODO
def big_tri(graph):
triangle(graph)
triangle(graph)
triangle(graph)for fk in [:src, :tgt]:
5, 2)
replace(graph, fk, 7, 1)
replace(graph, fk, 9, 5)
replace(graph, fk,
for v in [5,7,9]] [delete_vertex(graph, v)
Here are two different ways we could leverage our triangle!
function to make a big colored triangle, thought of as three colored triangles which a related by a kind of gluing relationship.
The first one requires us to modify triangle!
. No longer does it just add a triangle to some preexisting database instance, but it allows one to pass in parameters so that the triangle has the option of reusing existing vertices. Not only did we have to substantial refactor of the function, it was also convenient to change the behavior of the old function a bit (by having it return vertices rather than the graph itself), so there may be other places in our codebase where we have to fix broken code.
The second one doesn’t require us to modify the function: it just creates three disjoint triangles, and then ‘manually’ modifies the result to manipulate the foriegn keys and finally delete the isolated vertices. Although we did not need to modify the function, we certainly needed a deep understanding of its implementation in order to do this, plus this code is very clunky and not at all re-usable for other problems (whereas the first solution has potential to be used for other compositions of triangles).
Gluing: side-by-side comparison
"""Modify to allow it to overlap existing vertex IDs"""
def triangle(graph, v1=nothing, v2=nothing, v3=nothing):
= [add_vertex(graph, color) if isnothing(v) else v
vs for (v, color) in [v1=>"B", v2=>"R", v3=>"Y"]]
for i,j in [[1,2],[2,3],[3,1]]:
add_edge(graph, vs[i], vs[j])return vs
"""Now expressible as an abstraction of `triangle!`"""
def big_tri(graph):
= triangle(graph)
blue, red, _ = triangle(graph, blue)
_, _, yellow triangle(graph, nothing, red, yellow)
def replace(graph, fk:str, find:int, repl:int):
# TODO
def big_tri(graph):
triangle(graph)
triangle(graph)
triangle(graph)for fk in [:src, :tgt]:
5, 2)
replace(graph, fk, 7, 1)
replace(graph, fk, 9, 5)
replace(graph, fk,
for v in [5,7,9]] [delete_vertex(graph, v)
AlgebraicJulia
# Building blocks
= [triangle() for _ in 1:3]
T1,T2,T3 = [add_vertex(ColorGraph(), c) for c in "BRY"]
b,r,y
# Relation of building blocks to each other
= Relation(
pattern =["t1", "t2"], r=["t2", "t3"]; y=["t3", "t1"]
b
)
# Composition
= glue(pattern, T1=T1, T2=T2, T3=T3,
big_tri =b, r=r, y=y) b
On the bottom here I have some code built on top of AlgebraicJulia. We construct the big triangle in three steps. We declare the building blocks, we declare how those building blocks are related to each other, and then we compose the building blocks according to the relation. On the right I show what we call an undirected wiring diagram - this is what is created in line 6. We are going to stick triangles into T1, T2, T3, and we will stick an isolated vertex of the appropriate color in b
, r
, and y
. The lines in the diagram say that b
must be related to T1
and T3
in some way. It turns out there is only one way this can be done, by mapping the blue vertex to the blue vertex of the T1
(and of T3
). So there is no further information needed to disambiguate what composing these building blocks means.
Gluing: side-by-side comparison
"""Modify to allow it to overlap existing vertex IDs"""
def triangle(graph, v1=nothing, v2=nothing, v3=nothing):
= [add_vertex(graph, color) if isnothing(v) else v
vs for (v, color) in [v1=>"B", v2=>"R", v3=>"Y"]]
for i,j in [[1,2],[2,3],[3,1]]:
add_edge(graph, vs[i], vs[j])return vs
"""Now expressible as an abstraction of `triangle!`"""
def big_tri(graph):
= triangle(graph)
blue, red, _ = triangle(graph, blue)
_, _, yellow triangle(graph, nothing, red, yellow)
def replace(graph, fk:str, find:int, repl:int):
# TODO
def big_tri(graph):
triangle(graph)
triangle(graph)
triangle(graph)for fk in [:src, :tgt]:
5, 2)
replace(graph, fk, 7, 1)
replace(graph, fk, 9, 5)
replace(graph, fk,
for v in [5,7,9]] [delete_vertex(graph, v)
AlgebraicJulia
# Building blocks
= [triangle() for _ in 1:3]
T1,T2,T3 = [add_vertex(ColorGraph(), c) for c in "BRY"]
b,r,y
# Relation of building blocks to each other
= Relation(
pattern =["t1", "t2"], r=["t2", "t3"]; y=["t3", "t1"]
b
)
# Composition
= glue(pattern, T1=T1, T2=T2, T3=T3,
big_tri =b, r=r, y=y) b
This is what the diagram represents with the building blocks substituted in. Although there’s a technical term for this, called a colimit, we informally call this gluing building blocks along shared boundaries (which are, in this case, those isolated vertices).
Structural mathematics allows us to have a good understanding of what this gluing process is even without knowing what is inside the bubbles and the junctions. We didn’t have to know much about the internal structure of a triangle or how it was constructed to do this - we simply stated how these existing things were related to each other, something external to the graph building blocks. We just declared what we wanted as a piece of data, rather than writing another piece of hacky code that will become someone else’s problem to debug down the line.
Abstraction: Petri nets
class PetriNet():
def __init__(self,
str],
S:Set[str,
T:Dict[str],
Tuple[List[str]]]
List[
): ...
= Schema(
SchemaPetriNet =["S", "T", "I", "O"],
Ob=[("is", "I", "S"), ("os", "O", "S"),
Hom"it", "I", "T"), ("ot", "O", "T")]
(
)= acset_type(SchemaPetriNet) PetriNet
= PetriNet({"S", "I"},
sis "inf": (["S","I"], ["I","I"]),
{"rec": (["I"], ["S"])})
def ODE(p:PetriNet, rates, init, time): ...
def stochastic(p:PetriNet, rates, init, time): ...
. . .
Next, I want to move onto Petri nets, which are a way of looking at bipartite graphs, where we interpret one of the vertex types, call them transitions, as hyperedges which have multiple inputs and multiple outputs. This is especially relevant to modeling epidemiology or supply chains or chemical reaction networks. It might be of interest that these can also be thought of as instances of a certain database schema, which means if we understand the basic structure of databases then we do not have to write much Petri-net-specific code at all.
Let’s start with the example of a world where there are susceptible and infected people, and it takes a susceptible + an infected person to make two infected people, and the infected people can recover spontaneously too. This is just a specification of what sorts of transitions are possible without saying anything about their rates. There are two particular ways of taking a positive real number weight for each transition and obtaining a dynamical system: one of which is to create an ODE using mass-action kinetics (like in chemistry) and another is to do a stochastic simulation where you keep track of individual tokens living inside all of the blue circles and randomly trigger the transitions to fire. (…)
I’m going to talk about a new problem, but should first mention that the gluing problem from before is still relevant to petri nets. We can do the exact same AlgebraicJulia solution from before (flip back to previous slide) if we wanted without having to write any new code, whereas our two code-based solutions from before would have to be rewritten from scratch, something that feels very similar but still is completely different because these data structures look pretty different. But to keep things interesting let me talk about a different kind of abstraction reuse problem in the Petri net scenario.
Now consider that we may realize that, after running our simulation, that we should have made a distinction: what we thought was one homogenous group of susceptible people is better modeled by two US and a European populations, and likewise for infected people. It’s possible for someone to travel by plane and move from one of these populations to the other, too. We could express this itself as a Petri net with two states and two transitions, and only might we realize that our new model is really the multiplication of our starting Petri net with another one, in a sense that can be made precise. It would be nice to be able to declare that and be done with it, but first I’ll walk you through how we’d to do this in the old paradigm.
Multiplying: side-by-side comparison
def prepend(pre): return lambda s: pre + "_" + s
=
US, EU, E, W map(prepend, ["US", "EU", "east", "west"])
def mult_EU(petri:PetriNet):
= PetriNet()
res for state in petri.S:
= add_state(res, US(state))
us = add_state(res, EU(state))
eu =>[eu])
add_transition(res, E(state), [us]=>[us])
add_transition(res, W(state), [eu]
for (T, (I, O)) in petri.T:
=>US.(O))
add_transition(res, US(T), US.(I)=>EU.(O))
add_transition(res, EU(T), EU.(I)
return res
def multiply(r1:PetriNet, r2:PetriNet):
= PetriNet()
res for (s1, s2) in itertools.product(r1.S, r2.S):
f"{s1}_{s2}")
add_state(res,
for rx1 in r1.T:
for s2 in r2.S:
add_transition(res, rename_rxn(rx1, s2))
for rx2 in r2.T:
for s1 in r1.S:
add_transition(res, rename_rxn(rx2, s1))
return RxnNet(rs)
def rename_rxn(rxn, species:str): # TODO
Our first approach is going to be a function which takes an arbitrary Petri net and ‘multiplies’ it with the EU petri net, i.e. for every state in the original Petri net, it creates a US and an EU version of it (and a corresponding travel transition in each direction). We then need a copy of the original transitions living in both the US and EU parts of the Petri net.
The second approach is more general but requires more work. It is going to multiply two arbitrary Petri nets, by first creating a species for every pair of species in the original, and then a transition for every pair of species and transition in the other Petri net.
Both of these require a lot of string name hacking, and this code completely breaks if we were to try to do something similar on a different data structure.
Multiplying: back to basics
I want to explain why we are in some sense “multiplying” the two Petri nets. Let’s start off with our normal notion of multiplying numbers: take \(2 \times 3 = 6\). The essence of this is that we have two sets of things and multiplying is asking how many pairs there are where we take something from one set and something from the other. So here we have the male dogs Alex and Bob and the females Amy, Beth, and Cathy. And the fact that there are six possible pairings is what is expressed by \(2 \times 3 = 6\).
Multiplying: back to basics
Now suppose our sets can have two kinds of things, say cats and dogs. When we take pairs of these, we’re only going to allow the ones which are of the same kind, so we won’t pair female cats with male dogs, for example.
Multiplying: back to basics
Now this slide presents the same information as the previous one, but rather than depicting the things we’re multiplying directly as two different kinds of things, we instead have at the bottom a common language of types of things (in this case, just cats and dogs) and our two sets have functions which assign to each thing what type of thing it is. This is the general form of multiplication that we’ll need in order to multiply our Petri nets.
Our generalized notion of gluing things together, did not rely on the internals of what we were gluing but just how they were externally related to each other (their relation to the boundary we glued them along). Here I hope you see the analogous thing happening for multiplicaiton: rather than multiplying these two sets based on what type of thing each element is in itself, we multiply the sets based on the external realtion of the two sets to some common set of types of things.
Multiplying: side-by-side comparison
def prepend(pre): return lambda s: pre + "_" + s
=
US, EU, E, W map(prepend, ["US", "EU", "east", "west"])
def mult_EU(petri:PetriNet):
= PetriNet()
res for state in petri.S:
= add_state(res, US(state))
us = add_state(res, EU(state))
eu =>[eu])
add_transition(res, E(state), [us]=>[us])
add_transition(res, W(state), [eu]
for (T, (I, O)) in petri.T:
=>US.(O))
add_transition(res, US(T), US.(I)=>EU.(O))
add_transition(res, EU(T), EU.(I)
return res
def multiply(r1:PetriNet, r2:PetriNet):
= PetriNet()
res for (s1, s2) in itertools.product(r1.S, r2.S):
f"{s1}_{s2}")
add_state(res,
for rx1 in r1.T:
for s2 in r2.S:
add_transition(res, rename_rxn(rx1, s2))
for rx2 in r2.T:
for s1 in r1.S:
add_transition(res, rename_rxn(rx2, s1))
return RxnNet(rs)
def rename_rxn(rxn, species:str): # TODO
AlgebraicJulia
= PetriNet(
P_base =1, T=2, I=2, O=2, is=1, os=1, it=[1,2], ot=[1,2]
S
)
# Label P1 and P2 transitions as green or orange
= homomorphism(P1, P_base, init=...)
h1 = homomorphism(P2, P_base, init=...)
h2
# Composition of old problems
= multiply(h1, h2) US_EU_SIS
So instead of multiplying sets, we are multiplying Petri nets. We actually have a Petri net of colors. Here we see there is only one type of species and two types of transitions. The coloring of the circles and boxes of the SIS and US_EU models illustrates the data of these maps into the base Petri net of colors.
Multiplying: side-by-side comparison
def prepend(pre): return lambda s: pre + "_" + s
=
US, EU, E, W map(prepend, ["US", "EU", "east", "west"])
def mult_EU(petri:PetriNet):
= PetriNet()
res for state in petri.S:
= add_state(res, US(state))
us = add_state(res, EU(state))
eu =>[eu])
add_transition(res, E(state), [us]=>[us])
add_transition(res, W(state), [eu]
for (T, (I, O)) in petri.T:
=>US.(O))
add_transition(res, US(T), US.(I)=>EU.(O))
add_transition(res, EU(T), EU.(I)
return res
def multiply(r1:PetriNet, r2:PetriNet):
= PetriNet()
res for (s1, s2) in itertools.product(r1.S, r2.S):
f"{s1}_{s2}")
add_state(res,
for rx1 in r1.T:
for s2 in r2.S:
add_transition(res, rename_rxn(rx1, s2))
for rx2 in r2.T:
for s1 in r1.S:
add_transition(res, rename_rxn(rx2, s1))
return RxnNet(rs)
def rename_rxn(rxn, species:str): # TODO
AlgebraicJulia
= PetriNet(
P_base =1, T=2, I=2, O=2, is=1, os=1, it=[1,2], ot=[1,2]
S
)
# Label P1 and P2 transitions as green or orange
= homomorphism(P1, P_base, init=...)
h1 = homomorphism(P2, P_base, init=...)
h2
# Composition of old problems
= multiply(h1, h2) US_EU_SIS
We might also think that it makes sense for susceptible people to be able to travel internationally but not sick people. A slight modification of this diagram is sufficient to allow us to still just declare how the two models are related and have the multiplication do the right thing.
Abstraction: Wiring diagrams
def goods_flow(steps:int, initial:List[float],
input:Callable):
= []
output = initial
trade₁, trade₂ for step in 1:steps:
= [
trade₁, (trade₂, out) input(step), trade₂),
trader₁(
trader₂(trade₁)
]
output.append(out)
return output
"""
A + B = C
±√(C) = (B, D)
Find all integer solutions (A, D) up to `n`
"""
def solve_constraint(n::Int):
def cond(AD):
= AD
A, D = -D
B = A + B
C return C == D^2
return filter(cond, filter(Iterators.product(
range(n), range(n))))
Here are two ways we might be doing something conceptually specified by the flowchart.
Abstraction: Wiring diagrams
A wiring diagram substituion expresses that a process fits into another process as a subprocess. The subtitution is possible because this subprocess’ outer box matches the interface of this inner box.
Substituting: side-by-side comparison
def goods_flow(steps:int, initial:List[float],
input:Callable):
= []
output = initial
trade₁, trade₂ for step in 1:steps:
= [
trade₁, (trade₂, out) input(step), trade₂),
trader₁(
trader₂(trade₁)
]
output.append(out)
return output
\(\Rightarrow\)
def goods_flow(steps:int, initial:List[float],
input::Callable):
= []
output = initial
trade2, trade3, trade4 for step in 1:steps:
= [
trade2, trade3, (trade4, out)
trader2(trade3),input(step), trade2),
trader3(
trader4(trade3, trade2),
]
output.append(out)
return output
This is our process for updating the trading simulation with respect to this high-level change.
If we’d had a piece of code which solves the 3 4 system, we couldn’t re-use it.
Substituting: side-by-side comparison
"""
A + B = C
±√(C) = (B, D)
Find all integer solutions (A, D) up to `n`
"""
def solve_constraint(n::Int):
def cond(AD):
= AD
A, D = -D
B = A + B
C return C == D^2
return filter(cond, filter(Iterators.product(
range(n), range(n))))
\(\Rightarrow\)
"""
A + B = C
C * B = D
±√(D) = (B, E)
Find all integer solutions (A, E) up to `n`
"""
def solve_constraint(n::Int)
def constr(AE):
= AE
A, E = -E
B = A + B
C = C * B
D return D == E^2
return filter(constr, itertools.product(
range(n), range(n)))
This is our process for updating the equation solver with respect to this high-level change.
If we’d had a piece of code which solves the 3 4 system, we couldn’t re-use it.
Substituting: side-by-side comparison
= Trader(2, 1, lambda x,y: ...)
sys1 = Trader(1, 2, lambda x: ...)
sys2 = Trader(2, 1, lambda x,y: ...)
sys3 = Trader(2, 1, lambda x,y: ...) sys4
= Relation(2, 1, lambda xs,ys: xs[1] + xs[2] == ys[1])
sys1 = Relation(1, 2, lambda xs,ys: sqrt(xs[1]) == ys[1]
sys2 and ys[1]== -ys[2])
= sys1
sys3 = Relation(2, 1, lambda xs,ys: xs[1] * xs[2] == ys[1]) sys4
AlgebraicJulia
= WiringDiagram(["A"], ["A"])
wd12 = add_box(wd12, Box("op1", ["A","D"], ["C"]))
b1 = add_box(wd12, Box("op2", ["C"], ["D","A"]))
b2
add_wires(wd12, [1), (b1, 2),
(input_id(wd12), 1), (b2, 1),
(b1, 1), (b1, 1),
(b2, 2), (output_id(wd12), 1)
(b2, ])
= oapply(wd12, [sys1,sys2])
model = ocompose(wd12, 1, wd34)
wd234 = oapply(wd234,[sys1,sys3,sys4]) new_mod
With AlgebraicJulia, rather than just having the flowchart be something that inspires us from the whiteboard to write some code, we reason about it as a piece of data. The code we would’ve written is constructed automatically by saying what goes in each box and having a mathematical understanding of what is in general required to go from saying what is in each box to saying what the outer box does.
Changing from one abstraction to another
In addition to the previous examples which show re-use of abstractions within each of the three data types, this example shows how we can preserve an abstraction even when we change the data type: one might write a chemical reaction simulator that accepts Petri nets, but for other parts of one’s analysis one might not be working with general Petri nets. There are many algorithms and concepts that make sense for ordinary directed graphs that don’t make sense for Petri nets, so one ought work for as long as possible in the restricted setting with the knowledge that, at the end, one can identify graphs with a special kind of Petri net (i.e. those which have transitions with exactly one input and one output). What may be surprising is that this transformation process is completely automated, just given a simple relationship between the two database schemas at the top. Structural mathematics tells us how to specify the relationships between specifications such that implementations can be migrated from one to the other in a meaning-preserving way.
Changing from one abstraction to another
This high level data to relate the two schemas can be also used to move data in the other direction. So maybe what we need at the end of the day is a petri net, but we want to do our fancy graph algorithms with certain graphs and then convert them to Petri nets at the end.
Changing abstractions: side-by-side
"""
Species and transitions are vertices.
Inputs and outputs are edges.
"""
def petri_to_graph(p:PetriNet):
= Graph()
grph = {s: add_vertex(grph) for s in p.S}
vs for (t, (i, o)) in pairs(p.T):
= add_vertex(grph)
t for e in i:
add_edge(grph, vs[e], t)for e in o:
add_edge(grph, t, vs[e])
return grph
"""
Vertices are species.
Edges are transitions with one input and
one output.
"""
def graph_to_petri(g:Graph):
= PetriNet()
P for i in G.vertices:
f"v{i}")
add_species(P,
for (e, (s, t)) in enumerate(G.edges):
f"e{e}",
add_transition(
([P.S[s]], [P.S[t]]))
return P
. . .
AlgebraicJulia
= FinFunctor(SchPetri, SchGraph,
F = "V", T = "V", I => "E", O => "E")
S
= SigmaMigration(F, my_petri) my_graph
= FinFunctor(SchPetri, SchGraph,
F => "V", T => "E", I => "E", O => "E")
S
= DeltaMigration(F, my_graph) my_petri
Here we simply declare the relationship between the two schemas and say how we want to interpret it (the name of action which pushes data forward along the map is called sigma
and the one which moves data the opposite direction is called delta
).
Replacing code with data
Common theme: writing code vs declaring relationships between abstractions.
Problem | Python solution | AlgebraicJulia solution |
---|---|---|
Different pieces of a model need to be glued together. | Write a script which does the gluing or modifies how pieces are constructed. | Declare how overlap relates to the building blocks. (colimits) |
Different aspects of a model need to be combined / a distinction is needed. | Write a script which creates copies of one aspect for every part of the other aspect. | Declare how the different aspects interact with each other. (limits) |
We want to integrate systems at different levels of granularity. | Refactor the original code to incorporate the more detailed subsystem. | Separate syntax/semantics. Declare how the part relates to the whole at syntax level. (operads) |
We make a new assumption and want to migrate old knowledge into our new understanding. | Write a script to convert old data into updated data. | Declare how the new way of seeing the world (i.e. schema) is related to the old way. (data migration) |
So to summarize these past toy examples, there are some very general kinds of ways that we recognize a problem has shifted and we wish to reuse our old abstractions towards the new problem. In any specific scenario, one can always write an ad hoc function or script to handle this, but once you do this time after time, these one-off solutions become unmaintainable spaghetti. This is in contrast to the eat-your-cake-and-have-it-too solution on the right, where we have general solutions which do not need to be rewritten for every new scenario that pops up. And the way these compose together is transparent.
Of course, it’s not magic, and this strategy isn’t immediately available whenever we dream up of a new domain. Structural mathematics shows us what structure to look for in a new domain that allows us to do these things, though it’s still a case-by-case process to take domain-expert’s knowledge and find this structure. In the next section I’m going to give a high level picture of how this works.
Why is this relevant to AI safety?
- We can use verified, interpretable models (AI or otherwise) without losing too much economic value
- Reduce incentive to hand authority over to uninterpretable black boxes.
In this slide I’m going to draw some connections between our collaborative modeling and some basic ideas in AI safety. I’m just going to make a lot of claims, and then we’ll return here at the end to see whether or not I was able to show why, in principle, that our category theoretic approach to modeling and software engineering can be helpful in these ways.
So to start with #1: this image from Davidad’s ARIA’s Program. It’s about “shifting the incentives from the race to the bottom, by derisking credible paths to AI systems where we have strong reasons to expect confidence in the safety of systems”.
Our compositional approach to modeling is uniquely poised to allow both large, complex models (which are needed for the economic value) while retaining the verifiability and interpretability.
Why is this relevant to AI safety?
- We can use verified, interpretable models (AI or otherwise) without losing too much economic value
- Reduce incentive to hand authority over to uninterpretable black boxes.
- AI can target domains with a well-specified syntax and semantics allows for legibility and explanation
- Davidad’s ARIA proposal: world-models as autonomous AI output
- Anna Leshinaskaya’s talk on moral decisionmaking w/ combinatorial grammar
- Yoshua Bengio’s emphasis of separating model from implementation
As a second point, I’ve listed some examples of people who understand we don’t want raw decisions to be the output of our model, rather we want to target languages which we can interpret and interrogate. This shows up in the ARIA plan, where the AI is only interfacing the decisionmaking via providing world models which we can prove things about. Likewise, in the last two FAR seminars, we saw how moral decisionmaking can be made safer by stipulating a graph-like grammar with actors and outcomes, which can then be used to compute a decision but also explain why that decision came about. Likewise, this is the essence of Yoshua’s point about separating the model from the implementation. Making this easier is precisely one of the strengths of Topos’ approach to modeling and software.
Why is this relevant to AI safety?
- We can use verified, interpretable models (AI or otherwise) without losing too much economic value
- Reduce incentive to hand authority over to uninterpretable black boxes.
- AI can target domains with a well-specified syntax and semantics allows for legibility and explanation
- Davidad’s ARIA proposal: world-models as autonomous AI output
- Anna Leshinaskaya’s talk on moral decisionmaking w/ combinatorial grammar
- Yoshua Bengio’s emphasis of separating model from implementation
- We facilitate collaboration between diverse stakeholders, better alignment of values
- Opens opportunities for more equitable representation of values from all stakeholders (rather than just those who’ve mastered dependent type theory)
Lastly, on the topic of aligning AI with our values, it’s widely understood that any formalization of values should be the result of a kind of dialogue with relevant stakeholders, but programmers, therapists, ecologists, etc. are all coming from different perspectives and different dialects for talking about their knowledge of the world and understanding of what harm or value is.
Topos offers a sophisticated account of domain specific languages for domain experts that can be formal and interoperable, which are critical for this kind of task.
Why does it work?
Informal definition: a category is a bunch of things that are related to each other
. . .
Key intuition: category theory is concerned with universal properties.
- These change something that we once thought of as a property of an object into a kind of relationship that object has towards related objects.
Goal of talk is not to communicate how category theory works at a technical level. This is why this is the only definition which will be featured in the talk. (…)
But it’s important to get an intuition for why it works to know it’s not magic. I think the philosophy of category theory, which encourages working with universal properties, is a good explanation. I’ll try to describe one example of the simplest universal property just to give you some intuition for how it works.
Example universal property: emptiness
Consider mathematical sets which are related to each other via functions.
The empty set is the unique set which has no elements in it.
But if we we look at how the empty set relates to all the other sets, we’ll eventually notice something about these relations.
. . .
The empty set is the unique set which has exactly one function into every other set.
So if you were to tell a high schooler what an empty set is, this is a perfectly good definition. However, it doesn’t automatically generalize to telling us what an empty matrix is, or what an empty dynamical system or graph is. (…)
Example universal property: emptiness
Consider colored graphs related to each other via vertex mappings which preserve color and edges.
The empty graph uniquely has no vertices nor edges in it.
But if we we look at how it relates to all the other graphs, we’ll eventually notice something characteristic.
The empty graph is the unique graph which has exactly one graph mapping into every other graph.
We can see that the yellow vertex lives within the triangle in one way but in the big triangle in two ways, etc. If I wanted to say why this is the empty graph, of course my first instinct would be to say “Just look at it! There are no vertices, no edges!”. But look that our same definition from earlier which says to look for the object which is related to every other object exactly once also applies here.
Universal properties and generalizable abstractions
Category theory enforces good conceptual hygeine - one isn’t allowed to depend on “implementation details” of the things which feature in its definitions.
This underlies the ability of models built in AlgebraicJulia to be extended and generalized without requiring messy code refactor.
When you structure your code around these definitions, you inherit this kind of generalizability.
Takeaways
CT is useful for the same reason interfaces are generally useful.
In particular, CT provides generalized1 notions of:
- multiplication / multidimensionality
- adding things side-by-side
- gluing things along a common boundary
- looking for a pattern
- find-and-replace a pattern
- parallel vs sequential processes
- Mad Libs style filling in of wildcards
- Zero and One
- “Open” systems
- Subsystems
- Enforcing equations
- Symmetry
. . .
These abstractions all fit very nicely with each other:
- conceptually built out of basic ideas of limits, colimits, and morphisms.
We can use them to replace a large amount of our code with high level, conceptual data.
I hope to have shown that just a few basic concepts in category theory, in particular morphisms, limits, and colimits, are a very versatile vocabulary of good abstractions. Various combinations of these are sufficient to capture all of the diverse kinds of activities you see up here in a very generic way that can be specialized to your domain once you view it categorically.
Decapodes.jl: multiphysics modeling
"""Define the multiphysics"""
= @decapode DiffusionQuantities begin
Diffusion ::Form0{X}
C::Form1{X}
ϕ== k(d₀{X}(C)) # Fick's first law
ϕ end
= @decapode DiffusionQuantities begin
Advection ::Form0{X}
C::Form1{X}
(V, ϕ)== ∧₀₁{X}(C,V)
ϕ end
= @decapode DiffusionQuantities begin
Superposition ::Form0{X}
(C, Ċ)::Form1{X}
(ϕ, ϕ₁, ϕ₂)== ϕ₁ + ϕ₂
ϕ == ⋆₀⁻¹{X}(dual_d₁{X}(⋆₁{X}(ϕ)))
Ċ ∂ₜ{Form0{X}}(C) == Ċ
end
= @relation (C, V) begin
compose_diff_adv diffusion(C, ϕ₁)
advection(C, ϕ₂, V)
superposition(ϕ₁, ϕ₂, ϕ, C)
end
"""Geometry"""
= loadmesh(Torus_30x10()) mesh
"""Assign semantics to operators"""
= sym2func(mesh)
funcs :k] = Dict(:operator => 0.05 * I(ne(mesh)),
funcs[:type => MatrixFunc())
:⋆₁] = Dict(:operator => ⋆(Val{1}, mesh,
funcs[=DiagonalHodge()), :type => MatrixFunc());
hodge:∧₀₁] = Dict(:operator => (r, c, v) -> r .=
funcs[∧(Tuple{0,1}, mesh, c, v), :type => InPlaceFunc())
Extending what I’ve said so far to the design of simulators amounts to the following:
- simulations should be compiled to code rather than written in code
- This is because we can do high level conceptual programming (e.g. limits and colimits) when we are working with data, whereas representing our simulation via code will forever condemn us to having every conceptual update require a painstaking manual code update that could be very challenging.
When we work at the low level of code, introducing a simple mathematical idea (for example adding a gravitation force) forces us to remind ourself of all the implementation details - there could be a cascade of changes required.
I can only briefly gesture at the work done by some of my colleagues in a compositional language for multiphysics. The idea is that there is a graphical language for specifying equations which rigorously corresponds to things like Fick’s law of diffusion and conservation of mass. This is just like how our Petri Nets were a nice graphical language which could rigorously be interpreted as Chemical Reaction Networks and how spans of models can be rigorously interpreted as models glued along an overlap.
There is some interesting math behind how this works which I won’t be able to get into, involving something called the “discrete exterior calculus”, but from a user perspective you can see us declaring each of these small diagrams and then composing them together along shared variables into a multi-physics.
We then need to give a computational semantics by associating linear operators with some of the primitive building blocks, such as “multiplication by k” being 0.05 times the identity matrix.
We decoupled the high level physics from the geometry and the implementation.
Decapodes.jl: simulation
Resources
- Papers and talks: algebraicjulia.org
- Blog posts: algebraicjulia.org/blog and topos.site/blog
- Code: github.com/AlgebraicJulia
If you want to follow up on any the ideas in this talk, we have lots of papers, a couple of which I point out here to be potentially of interest to scientists. We also write blog posts intended to be more accessible and conversational, and I highlight a few here. Our code is always open source and we’re delighted to see it getting used.
Footnotes
Defined by universal properties.↩︎