Reading, Analysing and Activating Complex Text


The Orion Semantic AI system has a vocabulary of about 45,000 words, and about 10,500 phrases. The first step in analysing a new piece of text is for the machine to read the text, looking for words or phrases it doesn’t know, and for any words it doesn’t know in the dictionary definitions of the unknown words or phrases. The resulting word files are checked by hand if time allows (a dictionary definition can be very narrow, or missing an obvious definition, including literal use), and then the word and its meaning structure is created in the network.

A word can have several parts of speech (up to 5), if a verb then several types of verb, then possibly many senses. When reading the text, direct connections are immediately made so that the word is identified within the grammatical structure . Often, that is not possible – the POS or senses are reduced, but not reduced to one – the word is left to be revisited when more information is available – sometimes the necessary context is a hundred pages away, or in a linked document that hasn’t been opened yet. As an example, “bar” can be a noun, a verb, a preposition. As a noun, it can be a simple physical object (an iron bar or a sand bar), an abstract object (a bar on forever chemicals), a figurative object (the bar on a high-jump frame – he raised the bar), a collective noun - the Bar for lawyers collectively). It is obviously important that we don’t get these things mixed up – our Unconscious Mind does all the work.

In the machine version, a hierarchical operator is used. If the word has one or more parts of speech and several senses for the particular POS, a MULTIPARENT2 operator is used. If the POS is a verb, and the verb can take several forms (Intransitive, Transitive, TransInfinitive, etc. – about a hundred) a MULTIPARENT3 is used. The operator looks expensive, but only costs one link for each word in the text.

Figure 1 MULTIPARENT2 Operator

What happens when the parser cannot decide between senses, or even whether the word is being used as a noun or a verb. An operator is used which shows a decision is pending, and an attempt to clear the backlog is made at the end of a paragraph, a section, a chapter. A similar thing happens when a wordgroup may have been encountered – is it a wordgroup, in which case which meaning, or a literal use of the words.

Some examples:

Leave someone to do something

He left his wife to pursue his own interests.

He left his wife to do the washing up while he watched the cricket.

He left his sons to finish his work when he was gone.

He raised the bar

He raised the bar on semiconductor track widths, using far ultraviolet.

He raised the bar on forever chemicals in the drinking water at the next meeting, saying the benefits don’t justify the costs.

“Bar” can mean the figurative use of the bar on a high jump frame, where raising it makes it harder to jump over, or one can raise (“mention”) a prohibition (a “bar”) at a meeting.

Figure 2 Senses of Bar as a noun

The point of the examples is to show that unless you know the meanings of words, there are many mistakes to be made, especially on a large piece of text describing something complex.

LLMs were developed for Search Engines, and didn’t need to know the meanings of words. Someone found that reasoning was possible on small simple pieces of text, and used the magical thinking of “In the next release!” to convince people this minor problem could be solved. Of course it can only be solved by a system that has a sophisticated knowledge of what words (tens of thousands of different words and phrases) mean in context, with many of the words not in the text being read, but “understood”. An LLM is a very crude simulacrum of a knowledgeable human reader (and a human reader has limits – reading thousand page pieces of text were not part of the design brief a million years ago, whereas for Semantic AI they are).

Isn’t English too difficult for a computer to grasp? Won’t you be back to relying on statistics to decide which sense of a word is meant?

An undirected network of objects and operators grasps what English is about, and controls the computer by sending states and values through it, or building new structure or altering existing structure. As to who does what, it is horses for courses. If a chatbot can reliably tell a customer how to take out a home loan from a bank, it is not our turf. If the text runs to hundreds or thousands of pages or reading requires some knowledge on the part of the reader (the intricacy of a credit foncier home loan with variable rates can be confusing), it is our turf. Legislation has domains of criminal law, civil law, military law, where words have different meanings, some words not in the text become active and some not (civil law uses “a balance of probabilities” for its standard of proof, criminal law uses “beyond a reasonable doubt”) and all of it looks different to specifications for boats or planes or what have you - yes, there will be statistics, but over a controlled domain. We are already controlling the context of some senses – (“of a person”, “of a scheduled event”).

The intention is to handle things which are too complex for a human to grasp at the big picture level and in fine detail, and yet it is essential to get both right. The human component is an essential part of solving complex problems, so this gets a heavy emphasis in the Semantic AI machine’s design – the human’s emotional states and sometimes irrationality.

Comments

Popular Posts