This makes little sense to me. Ontologies and all that have been tried and have always been found to be too brittle. Take the examples from the front page (which I expect to be among the best in their set): human_activity => climate_change. Those are such a broad concepts that it's practically useless. Or disease => death. There's no nuance at all. There isn't even a definition of what "disease" is, let alone a way to express that myxomatosis is lethal for only European rabbits, not humans, nor gold fish.
Given we've tried to develop such ontologies constantly for thousands of years now. What do you think the cause for such hopeless optimism might be? If only we had a database of causal relationships to consult...
Democritus (b 460BCE) said, “I would rather discover one cause than gain the kingdom of Persia,” which suggests that finding true causes is rather difficult.
"According to the Greek historian Herodotus, Xerxes's first attempt to bridge the Hellespont ended in failure when a storm destroyed the flax and papyrus cables of the bridges. In retaliation, Xerxes ordered the Hellespont (the strait itself) whipped three hundred times, and had fetters thrown into the water."
Not so sure one should take stories about who said something in ancient times at face value ;)
[1] https://en.wikipedia.org/wiki/Xerxes_I
If you think you can use logic to determine human behavior in the past, well, it doesn't even work for modern behavior lol You'd be surprised what kind of beliefs about the world led to what kind of actions in history
Felix, qui potuit rerum cognoscere causas. [0]
Virgil.
[0] https://en.m.wikipedia.org/wiki/Felix,_qui_potuit_rerum_cogn...
Virgili, hoc postremo dico: mihi nomen non est Felix.
Or is less of a hassle.
I totally agreed that in the past years of hammering out an ontology for a particular area just results in a common understanding between those who wrote the ontology and a large gulf between them and the people they want to use it ( everyone else ).
What's perhaps different is that the machine, via LLM's, can also have an 'opinion' on meaning or correctness.
Going fully circle I wonder what would happen if you got LLM's to define the ontology....
>what would happen if you got LLM's to define the ontology.
https://deepsense.ai/resource/ontology-driven-knowledge-grap...
>hammering out an ontology for a particular area just results in a common understanding between those who wrote the ontology and a large gulf between them and the people they want to use it
This is the other side of the bitter lesson, which is just the empirical observation of a phenomenon that was to be expected from first principles (algorithmic information theory): a program of minimal length must get longer if the reality it models becomes more complex.
For ontologists, the complexity of the task increases as the generality is maintained while model precision is increased (top down approach), or conversely, when precision is maintained the "glue" one must add to build up a bigger and bigger whole while keeping it coherent becomes more and more complex (bottom up approach).
Koller and Friedman write in "Probabilistic Graphical Models" about the "clarity test", so that state variables should be clear for an all seeing observer.
States like "human_activity" are not objectively measurable.
Fairly PGMs and causal models are not the same, but this way of thinking about state variables is an incredible good filter.
> States like "human_activity" are not objectively measurable.
Well, or at least they would need a heavy dose of operationalisation.
But “disease => death” + AI => surely at least few billion in VC funding.
The best thing about this statement is that it can be read as 'the fact that disease causes death, plus the application of AI, will surely lead to billions VC funding' but it can also be read as 'disease is to death as AI is to a few billion in VC funding'. :D
Exactly. In some cases disease causes death. In others it causes immunity which in turn causes “good health” and postpones death.
Contradictory cause-effect examples, each backed up with data, are a reliable indicator of a class of situations that need a higher chain-effect resolution.
Which is directly usable knowledge if you are building out a causal graph.
In the meantime, a cause and effect representation isn't limited to only listing one possible effect. A list of alternate disjoint effects, linked to a cause, is also directly usable.
Just as an effect may be linked to different causes. Which if you only know the effect, in a given situation, and are trying to identify cause, is the same problem in reverse time.
It is my opinion that if we examine any factor closely, it will have multiple disjoint effects. As in nothing is absolutely unilateral in its effects. Some of those effects will depend on certain conditions. If it is possible to specify condition, annotations, and other nuances such as levels of confidence or source of the opinion, such a database might be pretty useful.
It's pretty easy to outline a high level ontology and let LLMs annotate/link it into something pretty useful, you can even have a benchmark suite using that ontology via LLM as a judge to progressively optimize it.
Ontology, not ontologies, have been tried.
We have quite a good understanding that a system cannot be both sound a complete, regardless people went straight in to make a single model of the world.
> a system cannot be both sound a complete
Huh, what do you mean by this? There are many sound and complete systems – propositional logic, first-order logic, Presburger arithmetic, the list goes on. These are the basic properties you want from a logical or typing system. (Though, of course, you may compromise if you have other priorities.)
My take is that the GP was implicitly referring to Gödel’s Incompleteness Theorems with the implication being that a system that reasons completely about all the human topics and itself is not possible. Therefore, you’d need multiple such systems (plural) working in concert.
That doesn't make much sense.
If you take multiple systems and make them work in concert, you just get a bigger system.
> If you take multiple systems and make them work in concert, you just get a bigger system.
The conclusion may be wrong, but a "bigger system" can be larger than the sum of its constituents. So a system can have functions, give rise to complexity, neither of its subsystems feature. An example would be the thinking brain, which is made out of neurons/cells incapable of thought, which are made out of molecules incapable of reproduction, which are made from atoms incapable of catalyzing certain chemical reactions and so on.
This is just emergence, though? How is emergence related to completeness?
This happens over and over with the relatively new popularization of a theory: the theory is proposed to be the solution to every missing thing in the same rough conceptual vector.
It takes a lot more than just pointing in the general direction of complexity to propose the creation of a complete system, something which with present systems of understanding appears to be impossible.
> How is emergence related to completeness?
I didn't make that argument. I think, the original conclusion above isn't reasonable. However, "a concert" isn't "just" a bigger system either, which is my point.
It just depends on your definition of system, doesn’t it?
Sort of, the guardrail here IMO is you have an ontology processor that basically routes to a submodule, and if there isn't a submodule present it errors out. It is one large system, but it's bounded by an understanding of its own knowledge.
Concerts - again plural. And naturally you only bring in appropriate instruments.
Turtles all the way down?
A collection of systems is itself a system. The theorem would not recognize the distinction.
I believe, neither the expansion of Gödel's theorems to "everything", non-formalized systems, nor the conclusion of a resolution by harnessing multiple systems in concert, are sound reasoning. I think, it's a fallacious reductionism.
What is a non-formalized system?
I am very curious on this. In particular, if you are able to split systems into formalized and non formalized, then I thinks there are quite some praise and a central spot in all future history books for you!
I am not a native speaker, so please don't get hung up on particular expressions.
I meant, the colloquial philosophies and general ontology are not subject of Gödel's work. I think, the forgone expansion is similar to finding evidence for telepathy in the pop-sci descriptions of quantum entanglement. Gödel's theorems cover axiomatic, formal systems in mathematics. To apply it to whatever, you first have to formalize whatever. Otherwise, it's an intuition/speculation, not sound reasoning. At least, that's my understanding.
Further reading: https://en.wikipedia.org/wiki/G%C3%B6del's_incompleteness_th...
Yep - when you use a multiplum of systems, then some systems can be regarded complete while other systems are sound.
This is in contrast to just one system that attempts to be sound and complete.
Could you define sound and complete in this context ? IIRC Rust's borrow checker is sound (will not mark something dysfunctional as functional) but not complete: some programs would take too long to verify, the checker times out, and compilation fails even though the program is potentially correct.
The meaning of the word person is ~sound (ie. Well defined) when two lawyers speak.
But when a doctor tells the lawyer that they operated a person, the lawyer can reasonably say "huh" - the concept of a person has shifted with the context.
What is an ontology exactly? I see Palantir talking about it all the time and it just sounds like vague marketing.
It comes from "the knowledge of being," and has been used to describe real-world knowledge representation, in particular hierarchical(-ish) semantic networks in AI since its early days.
When I see Palantir talk about it in a press release is that something real or just fluffy marketing?
These days anyone can spin up a developer account and check it out. Near as I could tell, you can create abstract 'objects' and link them to datasets/columns in the environment. And then you can link objects together. It's basically just an ER modeling tool, but they have great sales and seemed to have convinced people that they are constructing ontologies.
Agreed. About the strongest we can hope for are causal mechanisms, and most of those will be at most hypotheses and/or partial explanations that only apply under certain conditions.
Honestly, I don’t know understand how these so-ontologies have persisted. Who is investing in this space, and why?
Even more importantly, it's not even a simple probability of death, or a fraction of a cause, or any simple one-dimensional aspect. Even if you can simplify things down to an "arrow", the label isn't a scalar number. At a bare minimum, it's a vector, just like embeddings in LLMs are!
Even more importantly, the endpoints of each such causative arrow are also complex, fuzzy things, and are best represented as vectors. I.e.: diseases aren't just simple labels like "Influenza". There's thousands of ever-changing variants of just the Flu out there!
A proper representation of a "disease" would be a vector also, which would likely have interesting correlations with the specific genome of the causative agent. [1]
Next thing is that you want to consider the "vector product" between the disease and the thing it infected to cater for susceptibility, previous immunity, etc...
A hop, skip, and a small step and you have... Transformers, as seen in large language models. This is why they work so well, because they encode the complex nuances of reality in a high-dimensional probabilistic causal framework that they can use to process information, answer questions, etc...
Trying to manually encode a modern LLM's embeddings and weights (about a terabyte!) is futile beyond belief. But that's what it would take to make a useful "classical logic" model that could have practical applications.
Notably, expert systems, which use this kind of approach were worked on for decades and were almost total failures in the wider market because they were mostly useless.
[1] Not all diseases are caused by biological agents! That's a whole other rabbit hole to go down.
You're losing interpretability and scrutability, but gaining detail and expressiveness. You have no way to estimate the vectors in a causal framework, all known methods are correlational. You have no clean way to map the vectors to human concepts. Vectors are themselves extremely compressed representations, there is no clear threshold beyond which a representation becomes "proper".
That was very well said.
One quibble, and really mean only one:
> a high-dimensional probabilistic causal framework
Deep learning models aka neural network type models, are not probabilistic frameworks. While we can measure on the outside a probability of correct answers across the whole training set, or any data set, there is no probabilistic model.
Like a Pachinko game, you can measure statistics about it, but the game itself is topological. As you point out very clearly, these models perform topological transforms, not probabilistic estimations.
This becomes clear when you test them with different subsets of data. It quickly becomes apparent that the probabilities of the training set are only that. Probabilities of the exact training set only. There is no probabilistic carry over to any subset, or for generalization to any new values.
They are estimators, approximators, function/relationship fitters, etc. In contrast to symbolic, hard numerical or logical models. But they are not probabilistic models.
Even when trained to minimize a probabilistic performance function, their internal need to represent things topologically creates a profoundly "opinionated" form of solution, as apposed to being unbiased with respect to the probability measure. The measure never gets internalized.
What’s the relationship between what you’re saying and the concepts of “temperature” and “stochasticity”? The model won’t give me the same answer every time.
The probability distribution that the model outputs is deterministic. The decoding method that uses that distribution to decide what next token to emit may or may not be deterministic. If we decide to define the decoding method as part of "the model", then I guess the model is probabilistic.
It's also worth noting that the parameters (weights and biases) of the model are random variables, technically speaking, and this can be considered probabilistic in nature. The parameter estimates themselves are not random variables, to state the obvious. The estimates are simply numbers.
As I understand it, this is a dataset of claimed causation. It should contain vaccines->autism, not because it's true, but because someone, in public, claimed that it was.
So, by design, it's pretty useless for finding new, true causes. But maybe it's useful for something else, such as teaching a model what a causal claim is in a deeper sense? Or mapping out causal claims which are related somehow? Or conflicting? Either way, it's about humans, not about ontological truth.
Also, it seems to mistake some definitions as causes.
A coronavirus isn't "claimed" to cause SARS. Rather, SARS is a name given to the disease cause by a certain coronavirus. Or alternatively, the name SARS-nCov-1 is the name given to the virus which causes SARS. Whichever way you want to see it.
For a more obvious example, saying "influenza virus causes influenza" is a tautology, not a causal relationship. If influenza virus doesn't cause influenza disease, then there is no such thing as an influenza virus.
Yes, I agree there are a lot of definitions or descriptions masquerading as explanations, especially in medicine and psychology. I think maybe insurance has a lot to do that. If you just describe a lot of symptoms, insurance won't know whether to cover it or not. But if you authoritatively name that symptom set as "BWZK syndrome" or something, and suddenly switch to assuming "BWZK syndrome" is a thing, the unknown cause to the symptoms, then insurance has something it can deal with.
But this description->explanation thing, whatever the reason, is just another error people make. It's not that different from errors like "vaccines cause autism". Any dataset collecting causal claims people make is going to contain a lot of nonsense.