I can't pretend that I'm able to keep up with technology developments associated with AI but almost every week there is something that gets my attention. Today it's research from Meta into LLMs based on concepts rather than individual tokens.
Large Concept Models: Language Modeling in a Sentence Representation Space, December 15, 2024.
Link: https://arxiv.org/abs/2412.08821
In the past year, I adjusted my mapping method to match what the technology could handle in terms of knowledge graphs, moving from multiple words and sometimes sentences to individual words, and paying much more attention to the connecting words. Now it is quite possible that the technology will evolve to mimic how we think and I can return to mapping in sentences rather than individual words.
"In practice, a concept would often correspond to a sentence in a text document, or an equivalent speech utterance. We posit that a sentence is an appropriate unit to achieve language independence, in opposition to single words. This is in sharp contrast to current LLMs techniques which are heavily English centric and token based."
We don't think in individual words. We think in combinations of words that express concepts. This assumes that concepts aren't individual words but rather snippets of sentences. This is going back to the way I mapped conversations at NASA, using snippets rather than individual concepts. This actually makes much more sense to me and always has.
This research looks a language modeling in a sentence representation space. I jumped the gun and immediately connected that to knowledge graphs and the challenges I've encountered trying to connect the dots around knowledge graphs and concept mapping. I'll have to dissect this and figure out if I'm totally off base. There may be subtle and not to subtle differences between concept mapping and concept or knowledge modeling that I am not grasping.
The "conceptual" approach described in the paper has potential implications for knowledge graphs:
- It could enhance semantic understanding: By focusing on concepts rather than individual tokens, large concept models can better capture and represent the relationships and hierarchies in knowledge graphs.
- It could advance reasoning capabilities: By operating at the level of concepts, large concept models could potentially perform more sophisticated reasoning based on the interconnected data in knowledge graphs, enhancing decision-making processes and inferencing.
- It could also improve traceability and our understanding of how the models operate. Conceptual representation would be easier to follow just like a concept map made of sentences is much easier to follow than a concept map made up of words and connecting sentences.