Tag Archives: semantics

The Semiotic Triangle

Words and their Objects

When we speak, we use words to describe the objects in our environment. With words, however, we do not possess the objects, but only describe them, and as we all know, words are not identical to the objects they describe. It is obvious that there is no identity.

Some funny examples of the not always logical use of words can be found in the following text (in German), which explains why the quiet plays loudly and the loud plays quietly.

Fig 1: The piano (the quiet one)

Fig. 2: The lute (the wood)

But how does the relationship between words and objects look if it is not an identity? It cannot be a 1:1 relationship, because we use the same word to describe different objects. Conversely, we can use several words for the same object. The relationship is not fixed either, because the same word can mean something different depending on the context. Words change relentlessly over time, they change their sound and their meaning.

The relationship between words and designated objects is clearly illuminated by Ogden and Richards’1 famous depiction of the semiotic triangle from 1923.

The Semiotic Triangle

 

Fig 3: The semiotic triangle according to Ogden and Richards1

 

Fig 4: The semiotic triangle exemplified by the word ‘table’

The idea of the triangle has many predecessors, including Gottlob Frege, Charles Peirce, Ferdinand de Saussure and Aristotle.

Ogden and Richards use the semiotic triangle to emphasise that we should not confuse words, objects and concepts. The three points of the triangle point to three areas that are completely different in nature.

The tricky thing is that we are not just tempted but entirely justified in treating the three points as if they were identical. We want the word to describe an object precisely. We want our concepts to correspond exactly with the words we use for them. Nevertheless, the words are not the objects, nor are they concepts.

Ogden and Richards have this to say: «Between the symbol [word] and the referent [object] there is no other relevant relationship than the indirect one, which consists in the fact that the symbol is used by someone [subject] to represent a referent.»

The relationship between the word (symbol, sign) and the object (referent) is always indirect and runs via someone’s thought, i.e. the word activates a mental concept in ‘someone’, i.e. a human subject, speaker or listener. This inside image is the concept.

Fig. 5: This is how Ogden and Richards see the indirect relationship between symbol and indicated object (referent). The thoughts contain the concepts.

The diminished baseline in Fig. 5 can also be found in the original of 1923. Symbol (word) and referent (reference object) are only indirectly connected via the thoughts in the interpreting subjects. This is where the mental concepts are found. The concepts are the basic elements of our thoughts.

When we deal with semantics, it is essential to take a look at the triangle. Only the concepts in our head connect the words with the objects. Any direct connection is an illusion.

This is a post on the topic of semantics.

Translation: Juan Utzinger


1 Ogden C.K. und Richards I.A. 1989 (1923): The Meaning of Meaning. Orlando: Harcourt.

Semantics and Linguistics

What is semantics?

A simple and easily understandable answer is that semantics is the meaning of signals. The signals can exist in any form: as text, as an image, etc. The most frequently studied semantics is that of words.

This is a good reason to examine the relationship of linguistics and semantics. Can semantics be regarded as a subdiscipline of linguistics?

Linguistics and semantics

Linguistics, the science of language and languages, has always examined the structure (grammar, syntax) of languages. Once the syntax of a sentence has been understood, linguists see two further tasks, i.e. secondly to examine the semantics of the sentence and thirdly to examine its pragmatics. “Semantics” is about the meaning of sentences, “pragmatics” about the “why” of a statement.

The linguists’ three steps

In the linguists’ eyes, there are thus three steps in understanding language: syntax -> semantics -> pragmatics. These three fields are weighted very differently by linguists: a conventional textbook predominantly deals with syntax, whereas semantics and pragmatics play a marginal role – and always on the basis of the previously conducted syntactic analysis. The linguists’ syntactic analysis thus already sets the course for what is based on it, namely semantics and pragmatics.

This is not really ideal for semantics. When you deal with semantics in more detail, it becomes clear that the grammar and other properties of individual languages constitute externals which may circumscribe the core of the statements – their meaning – in an occasionally very elegant manner, but they merely circumscribe them and do not represent them completely, let alone directly. A direct formal representation of what is meant by a text, however, would be the actual objective of a scientific semantics.

Can this objective be attained? First, we will have to clarify the relationship between words and concepts – words and concepts are not the same. Concepts are the basic elements of semantics and have a special, but not entirely simple relationship with the words of a language.

Word does not equal concept

One could flippantly assume that there is a one-to-one relationship between words and concepts, i.e. that behind every word, there is a concept which summarises the meaning of the word. But this is precisely what is wrong. Words and concepts cannot unequivocally be mapped on each other. The fact that this is the case can be recognised by everybody who observes himself while reading, talking and thinking.

It is obvious that a word can have several meanings depending on the context in which it is uttered. Occasionally, a word may even have no meaning at all, for instance if it is a technical term and I don’t know the specialist field. In such a case, I may be able to utter the word, but it remains devoid of meaning for me. Yet somebody who understands the specialist field will understand it.

Meaning has much to do with the addressee

Even perfectly normal words which we all know, not always have an unequivocal meaning but can evoke slightly different ideas (meanings) depending on the listener or the context. This does not only concern abstract words or words to which various values are attached, such as happiness, democracy, truth, etc.: absolutely concrete terms like leg, house and dog are interpreted differently by different people, too. The reception of the words as meaningful concepts has much to do with the addressee, his situation and expectations. There is definitely no 1:1 relation between words and concepts.

Meanings vary

Even in ourselves, there are quite different ideas for the same word; depending on the situation, we associate different ideas with the same word, depending on the situation and the everchanging state of our momentary knowledge of words and topics.

A dynamic process

The transition from one language to another shows how the link between words and concepts is a dynamic process in time and changes the meaning of the words. The English word ‘brave’ is the same word as the word ‘bravo’ in Italian, which we use if a musical performance inspires us. But the same word also exists in German, where today it means prissy or well-behaved – certainly not exactly the same as brave, though it is the same word and once meant the same in German as in English.

Semantics examines the play of meanings

We have to accept that a word and a concept cannot be mapped on each other just like that. Although in individual cases it may seem that there is precisely one concept (one semantics) behind every word, this idea is completely inappropriate in reality. And it is this idea which prevents the play of meanings from being understood correctly. Yet it is precisely this play of meanings which, in my view, constitutes semantics as a field of knowledge. In this field, it is possible to represent concepts formally in their own proper structure – which is completely independent from the formal representation of words.


Translation: Tony Häfliger and Vivien Blandford

What Can I Know?


This website is powered by the question of how thinking works.


Information and Interpretation

How is data assigned a meaning? What does information consist of? The answer seems clear, as the bit is generally regarded as its building block.

Entropy is the quantity by which information appears in physics – thanks to C. E. Shannon, the inventor of the bit. Bits measure entropy and are regarded as the measure of information. But what is entropy and what does it really have to do with information?


Artificial Intelligence (AI)

Today there is a lot of talk about AI. I have been creating such systems for forty years – but without labelling them with this publicity term.

  • The big difference: corpus-based and rule-based AI
  • How real is the probable?
  • Which requires more intelligence: jassen (a popular Swiss card game) or chess?
  • What distinguishes biological intelligence from machine intelligence?

What today is called AI are always neural networks. What is behind this technology? Neural networks are extremely successful – but are they intelligent?

-> Can machines be intelligent? 


Logic

Mathematical logic, to many, appears to be the ultimate in rationality and logic. I share the respect for the extraordinary achievements of the giants on whose shoulders we stand. However, we can also think beyond this:

  • Are statements always either true or false?
  • Can classical logic with its monotonicity really be used in practice?
  • How can time be incorporated into logic?
  • Can we approach logical contradictions in a logically correct way?

Aristotle’s classical syllogisms still influence our view of the world today. This is because they gave rise to the ‘first order logic’ of mathematics, which is generally regarded as THE classical logic. Is there a formal way out of this restrictive and static logic, which has a lot to do with our static view of the world?

-> Logic: From statics to dynamics


Semantics and NLP (Natural Language Processing)

Our natural language is simply ingenious and helps us to communicate abstract ideas. Without language, humanity’s success on our planet would not have been possible.

  • No wonder, then, that the science that seeks to explain this key to human success is considered particularly worthwhile. In the past, researchers believed that by analysing language and its grammar they could formally grasp the thoughts conveyed by it, which is still taught in some linguistics departments today. In practice, however, the technology ‘Large Language Model’ (LLM) of Google’s has shattered this claim.

As a third option, I argue in favour of a genuinely semantic approach that avoids the gaps in both the grammar and the LLMs. We will deal with the following:

  • Word and meaning
  • Semantic architectures
  • Concept molecules

-> Semantics and Natural Language Processing (NLP)


Scales: Music and Maths

A completely different topic, which also has to do with information and the order in nature, is the theory of harmony. Rock and hits are based on a simple theory of harmony, jazz and classical music on complex ones. But why do these information systems work? Not only can these questions be answered today, the answers also provide clues to the interplay between the forces of nature.

  • Why do all scales span an octave?
  • The overtone series is not a scale!
  • Standing waves and resonance
  • Prime numbers and scales

-> How musical scales evolved


The author

My name is Hans Rudolf Straub. Information about my person can be found here.


Books

On the topics of computational linguistics, philosophy of information, NLP and concept molecules:

The Interpretive System, H.R. Straub, ZIM-Verlag, 2020 (English version)
More about the book

Das interpretierende System, H.R. Straub, ZIM-Verlag, 2001 (German version)
More about the book

On the subject of artificial intelligence:

Wie die künstliche Intelligenz zur Intelligenz kommt, H.R. Straub, ZIM-Verlag, 2021 (Only available in German)
More about the book
Ordering the book from the publisher

You can order a newsletter here.


Thank You

Many people have helped me to develop these topics. Wolfram Fischer introduced me to the secrets of Unix, C++ and SQL and gave me the opportunity to build my first semantic interpretation programme. Norbert Frei and his team of computer scientists actively helped to realise the concept molecules. Without Hugo Mosimann and Maurus Duelli, Semfinder would neither have been founded nor would it have been successful. The same applies to Christine Kolodzig and Matthias Kirste, who promoted and supported Semfinder in Germany. Csaba Perger and Annette Ulrich were Semfinder’s first employees, full of commitment and clever ideas and – as knowledge engineers – provided the core for the emerging knowledge base.

Wolfram Fischer actively helped me with the programming of this website. Most of the translations into English were done by Vivien Blandford and Tony Häfliger, as well as Juan Utzinger.

Thank you sincerely!

The three innovations of rule-based AI

Have the neural networks outpaced the rule-based systems?

It cannot be ignored: corpus-based AI has overtaken rule-based AI by far. Neural networks are making the running wherever we look. Is the competition dozing? Or are rule-based systems simply incapable of yielding equivalent results to those of neural networks?

My answer is that both methods are predisposed for performing very different functions as a matter of principle. A look at their respective modes of action makes clear what the two methods can usefully be employed for. Depending on the problem to be tackled, one or the other has an advantage.

Yet the impression remains: the rule-based variant seems to be on the losing side. Why is that?

In what dead end has rule-based AI got stuck?

In my view, rule-based AI is lagging behind because it is unwilling to cast off its inherited liabilities – although doing so would be so easy. It is a matter of

  1. acknowledging semantics as an autonomous field of knowledge,
  2. using complex concept architectures,
  3. integrating an open and flexible logic (NMR).

We have been doing this successfully for more than 20 years. What do the three points mean in detail?

Point 1: acknowledging semantics as an autonomous field of knowledge

Usually, semantics is considered to be part of linguistics. In principle, there would not be any objection to this, but linguistics harbours a trap for semantics which is hardly ever noticed: linguistics deals with words and sentences. The error consists in perceiving meaning, i.e. semantics, through the filter of language, and assuming that its elements have to be arranged in the same way as language does with words. Yet language is subject to one crucial limitation: it is linear, i.e. sequential – one letter follows another, one word comes after another. It is impossible to place words in parallel next to each other. When we are thinking, however, we are able to do so. And when we investigate the semantics of something, we have to do so in the way we think and not in the way we speak.

Thus we have to find such formalisms for the concepts as occur in thought. The limitation imposed by the linear sequence of the elements and the resulting necessity to reproduce compounds and complex relational structures with grammatical tricks in a makeshift way, and differently in every language – this structural limitation does not apply to thinking, and this results in structures on the side of semantics that are completely different from those on the side of language.

Word ≠ concept

What certainly fails to work is a simple “semantic” annotation of words. A word can have many and very different meanings. One meaning (= a concept) can be expressed with different words. If we want to analyse a text, we must not look at the individual words but always at the general context. Let’s take the word “head”. We may speak of the head of a letter or the head of a company. We cannot integrate the context into our concept by associating the concept of <head< with other concepts. Thus there is a <body part<head< and a <function<head<. The concept on the left (<body part<) then states the type of the concept on the right (<head<). We are thus engaged in typification. We look for the semantic type of a concept and place it in front of the subconcept.

Consistantly composite data elements

The use of typified concepts is nothing new. However, we go further and create extensive structured graphs, which then constitute the basis for our work. This is completely different from working with words. The concept molecules that we use are such graphs possess a very special structure to ensure that they can be read easily and quickly by both people and machines. This composite representation has many advantages, among them the fact that combinatorial explosion is countered very simply and that the number of atomic concepts and rules can thus be drastically cut. Thanks to typification and the use of attributes, similar concepts can be refined at will, which means that by using molecules we are able to speak with a high degree of precision. In addition, the precision and transparency of the representation have very much to do with the fact that the special structure of the graphs (molecules) has been directly derived from the multifocal concept architecture (cf. Point 2).

Point 2: using complex concept architectures

Concepts are linked by means of relations in the graphs (molecules). The above-mentioned typification is such a relation: when the <head< is perceived as a <body part<, then it is of the <body part< type, and there is a very specific relation between <head< and <body part<, namely a so-called hierarchical oris-a’ relation – the latter because in the case of hierarchical relations, we can always say ‘is a”, i.e. in our case: the <head< is a <body part<.

Typification is one of the two fundamental relations in semantics. We allocate a number of concepts to a superordinate concept, i.e. their type. Of course this type is again a concept and can therefore be typified again in turn. This results in hierarchical chains of ‘is-a’ relations with increasing specification, such as <object<furniture<table<kitchen table<. When we combine all the chains of concepts subordinate to a type, the result is a tree. This tree is the simplest of the four types of architecture used for an arrangement of concepts.

This tree structure is our starting point. However, we must acknowledge that a mere tree architecture has crucial disadvantages which preclude the establishment of semantics which are really precise. Those who are interested in the improved and more complex types of architecture and their advantages and disadvantages, will find a short description of the four types of architecture on the website of meditext.ch.

In the case of the concept molecules, we have geared the entire formalism, i.e. the intrinsic structure of the rules and molecules themselves, to the complex architectures. This has many advantages, for the concept molecules now have precisely the same structure as the axes of the multifocal concept architecture. The complex folds of the multifocal architecture can be conceived of as a terrain, with the dimensions or semantic degrees of freedom as complexly interlaced axes. The concept molecules now follow these axes with their own intrinsic structure. This is what makes computing with molecules so easy. It would not work like this with simple hierarchical trees or multidimensional systems. Nor would it work without consistently composite data elements whose intrinsic structure follows the ramifications of the complex architecture almost as a matter of course.

Point 3: integrating an open and flexible logic (NMR)

For theoretically biased scientists, this point is likely to be the toughest, for classic logic appears indispensable to most of them, and many bright minds are proud of their proficiency in it. Classic logic is indeed indispensable – but it has to be used in the right place. My experience shows me that we need another logic in NLP (Natural Language Processing), namely one that is not monotonic. Such non-monotonic reasoning  (NMR) enables us to attain the same result with far fewer rules in the knowledge basis. At the same time, maintenance is made easier. Also, it is possible for the system to be constantly developed further because it remains logically open. A logically open system may disquiet a mathematician, but experience shows that an NMR system works substantially better for the rule-based comprehension of the meaning of freely formulated text than a monotonic one.

Conclusion

Today, the rule-based systems appear to be lagging behind the corpus-based ones. This impression is deceptive, however, and derives from the fact that most rule-based systems have not yet succeeded in jumping ahead of themselves and becoming more modern. This is why they are either

  • only applicable for ckear tasks in a small and well defined domain , or
  • very rigid and therefore hardly employable, or
  • they require an unrealistic use of resources and become unmaintainable.

If, however, we use consistently composite data elements and a higher degree of concept architectures, and if we deliberately refrain from monotonic conclusions, a rule-based system will enable us to get further than a corpus-based one – for the appropriate tasks.

Rule-based and corpus-based systems differ a great deal from each other, and depending on the task in hand, one or the other has the edge. I will deal with this in a later post.

The next post will deal with the current distribution of the two AI methods.

This is a post about artificial intelligence.


Translation: Tony Häfliger and Vivien Blandford