Abstract
There is a common misconception across the lntelligence Community (IC) to the effect that information trapped within multiple heterogeneous data silos can be semantically integrated by the sorts of meaning-blind statistical methods employed in much of artificial intelligence (Al) and natural language processlng (NLP). This leads to the misconception that incoming data can be analysed coherently by relying exclusively on the use of statistical algorithms and thus without any shared framework for classifying what the data are about. Unfortunately, such approaches do not yield sustainable results where we are dealing with widely distributed, highly heterogeneous and often changing bodies of data. We argue here that the needed integration requires the use of what we call an lntegrating Semantic Framework (ISF), which provldes a consistent set of categories and relationships that can be reused over and over again to tag successive bodies of data in ways which foster more coherent representation and reasoning.