2. Information and Its Structural Manifestation
The idea of the meaning of information as a relationship between the two informational entities of a sign and a denotation can be applied to any conceptualization of information. However, in such a vague formulation, it does not have much value for work on specific methods of informational semantics. For this reason, it will be necessary to make a decision on the choice of the definition of information and to make reference to the theory of information built upon this definition [
3,
4].
The present author’s definition of information is based on only one categorial (non-definable) concept of the one-many opposition. This minimalistic conceptual framework for information has its advantage in the development of its formal theory consisting in a direct connection to mathematical concepts. Information is defined as an identification of a variety understood as that which makes one out of many. This can be achieved by a selection of one component out of the many (selective manifestation of information) or by equipping the many with a structure that unites it (structural manifestation of information). These are two coexisting manifestations of information, not two types of information as one always requires the presence of the other, although possibly for a different variety. While the reference to selection is quite straightforward and does not require elaborate explanation, the reference to a structure can generate a demand for explanation. Of course, the formalism of a theoretical model of information has to address both manifestations.
Thus, the concept of information requires a variety (many), which can be understood as an arbitrary set S (called a carrier of information). The information system is this set S equipped with the family of subsets M satisfying two conditions: (1) the entire set S belongs to the family M and (2) together with every subfamily of M, its intersection belongs to M, i.e., M is a
Moore family of subsets. Of course, this means that we have a
closure operator defined on S (i.e., a function
f on the power set 2
S of a set S) such that [
5,
6]:
- (1)
For every subset A of S, A f(A);
- (2)
For all subsets A, B of S, A B f(A) f(B);
- (3)
For every subset A of S, f(f(A)) = f(A)).
The set S with a closure operator f defined on it is usually called a closure space and is represented by the symbol <S, f>. Alternatively, this closure space can be defined as a distinction of the Moore family M of subsets of S.
The Moore family M of subsets is simply the family
f-Cl of all closed subsets, i.e., subsets A of S, such that A =
f(A). The family of closed subsets M =
f-Cl is equipped with the structure of a complete lattice
Lf by the set theoretical inclusion of sets.
Lf can play a role in the generalization of logic for (not necessarily linguistic) information systems, although it does not have to be a Boolean algebra. In many cases, it maintains all the fundamental characteristics of a logical system [
6].
Information itself is a distinction of a subfamily ℑ of M, such that it is closed with respect to (pair-wise) intersection and is dually hereditary, i.e., with each subset belonging to ℑ, all subsets of S including it belong to ℑ (i.e., ℑ is a filter in the lattice Lf).
The Moore family M can represent a variety of structures of a particular type (e.g., geometric, topological, algebraic, logical, etc.) defined on the subsets of S. This corresponds to the structural manifestation of information. Filter ℑ in turn, in many mathematical theories associated with localization, can be used as a tool for identification, i.e., selection of an element within the family M, and under some conditions within the set S. For instance, in the context of Shannon’s selective information based on a probability distribution of the choice of an element in S, ℑ consists of elements in S with the probability measure 1, while M is simply the set of all subsets of S.
Now, when we have the basic mathematical formalism for information, we can proceed to the formalization of the theory of morphisms, functions that preserve informational structure, or in the mathematical language of lattice theory, homorphisms of closure spaces. This type of mapping is of crucial importance for our understanding of symbolic representation as it is defined as a mapping of the information of the sign to the information of the denotation.
If we have two closure spaces <S, f> and <T, g>, then a function φ: S → T is called a homomorphism of closure spaces if it satisfies the condition: ∀A ⊆ S: φ(f(A)) ⊆ g(φ(A)).
It can be easily recognized that this is exactly the same condition that defines continuous functions in the case of topological spaces (topological information), and as in topology, for general transitive closure spaces it is equivalent to the requirement that the inverse image of every g-closed subset is f-closed. It is important to notice that homomorphisms of closure spaces can be defined between closure spaces of very different types in diverse disciplines of mathematics.
Now, when we add the condition that the function φ is bijective and satisfies the stronger condition ∀A ⊆ S: φ(f(A)) = g(φ(A)), we get an isomorphism of closure spaces. From the point of view of mathematical theory, isomorphic closure spaces can be considered identical. Finally, isomorphisms from <S, f> on itself (i.e., when S = T and f = g) are called automorphisms or transformations of closure spaces. It can be easily shown that the class of all automorphisms on a closure space <S, f> forms a group with respect to the composition of functions.
The original choice of the closure space formalism made by the author to develop a general theory of information was guided by purely pragmatic considerations. Neither the definition of information formulated by the author as an identification of a variety, nor the large variety of continuing attempts to formulate a theory of structural information from the famous 1972 book by René Thom “
Structural Stability and Morphogenesis” [
7] to recent contributions [
8], compels us to its use, other than the fact that the formalism is very general and allows us to consider practically all existing formally formulated theories of information as special cases. Closure space morphisms give us transitions between very different types of information. For instance, we can consider a symbolic relationship between linguistic systems governed by classical logic described in terms of the consequence closure operation and its denotation described in terms of geometry, topology, or some other form of morphology.
Certainly, this is a common practice of mathematical science to look for a formalism in which we have representation of all theoretical terms and that describes the subject of the study adequately. Nothing more is expected in mathematical practice. Yet, the choice of formalism calls for a more careful philosophical reflection going beyond pragmatic justification.
Another issue motivating this study is the question about the meaning of the concept of a structure frequently used but rarely defined outside of particular contexts. We have well-defined structures in many mathematical theories (e.g., relational structures, algebraic structures, topological structures, etc.), but under more careful inspection, the question “What does structure mean in general?” is rarely asked and, if asked, is far from being answered. The problem is that we often refer to “equivalent” definitions of structures without any formal justification or explanation for this equivalence outside of the particular cases. Our own two definitions of a closure space presented above, one involving the concept of a closure operator and the other involving the concept of a Moore family of subsets, are considered equivalent. We can switch between the two definitions freely, but how do we describe this equivalence outside of this specific context?
In standard situations, the equivalence of structures is defined by isomorphisms. Two isomorphic structures are identical for everyday mathematical practice. Therefore, the question is how to formalize the relationship of equivalence of structures when due to differences in used concepts we cannot consider isomorphisms. The name “cryptomorphism” appeared in this context and became a standard expression. Garrett Birkhoff in 1967 introduced the concept of cryptomorphism (actually he used two terms apparently understood the same way, the other was “crypto-isomorphism”) in the context of abstract algebras defined in alternative ways, excluding the use of isomorphism as a criterion for their equivalence [
6].
Birkhoff illustrated the need to go beyond the usual association of algebraic structures through isomorphisms by the example of the concept, omnipresent in mathematics, of a group. A group can be considered as an abstract algebra with four different signatures (types of operations according to the number of operands entering algebraic operations): (1, 2), (0, 1, 2), (2, 2), and (2). Algebras cannot be isomorphic if they have different signatures, and the idea of “polymorphism” presented by Bourbaki in their early attempts to formalize the general concept of a structure does not work here. Birkhoff proposed a solution for the association of algebraic structures, but it is heavily dependent on the context of abstract algebras and can be applied only to so-called varieties (classes of algebras defined by polynomial equations). His approach cannot be extended to other types of structures, for instance topological or geometric. Thus far, the concept of cryptomorphism has not acquired any formal general definition, and the term is used as a generic description of an ad hoc translation between different conceptual frameworks for the study of particular mathematical objects.
The present author started his attempt to develop a general theory of structures from the formulation of the concept of symmetry based on closure space theory [
9,
10]. Following the general idea of Felix Klein’s Erlangen Program, structures have been associated with the invariants of symmetries. This study is intended as a link between the symmetry-based general concept of structure and the more traditional methodology of the study of abstract concepts based on equivalence relations.
The limited space and scope of this paper does not allow for more than a very brief and general description of the triangular relationship between the three concepts of an equivalence relation, a group action on a set, and the lattice of substructures. The latter two concepts appear in the formulation of the theory of general symmetry, but without any association with the concept of equivalence relations fundamental for the process of abstraction [
9,
10]. The mutual cryptomorphic interdependence of all three concepts sets the foundation for structural abstraction, i.e., for considering classes of structures as one structural object.
The cryptomorphic character of the relationship does not lead to an error of circularity, because we consider here only three specific structures and their cryptomorphisms are defined for a very specific context. We do not attempt to engage arbitrary structures as tools, but rather to engage the three specific structures as tools for the inquiry of arbitrary structures. Therefore, the three particular instances of cryptomorphisms are used for the purpose of raising the level of abstraction.
It is appropriate to explain a little bit more about why these three specific structures are so important for our purposes. Equivalence relations are fundamental not only for mathematics, but for any form of abstract thinking. Abstraction is a process of transition from the lower level of individual objects to the higher one of abstract concepts, reducing complexity by the elimination of individual differences irrelevant for the purpose of our consideration. Instead of dealing with the huge variety of individual properties of objects that we study in reference to only a few properties or relations of interest, we consider entire classes of mutually equivalent individuals sharing relevant properties as individuals of a higher level of abstraction.
Group actions on sets are fundamental tools of natural sciences, in particular in physics. Group actions are used for the conceptualization of symmetry or local invariance of selected (symmetric) collectives under transformation that globally may change all individuals. An example of such a symmetry could be a symmetric configuration of points forming some structure, but also a human being or any biological organism is an invariant structure whose material components are constantly being replaced by new ones.
The least obvious is the third structure, the lattice of substructures. Probably the most disappointing aspect of Birkhoff’s extensive study of lattices was the early discovered fact that the lattice of substructures does not identify the structure uniquely. Otherwise, lattice theory would have already given the ultimate answer to the general question “What is a structure?”. For instance, non-isomorphic groups can have isomorphic lattices of subgroups [
6]. This however does not disqualify lattices as very convenient tools. For our purposes, it turns out that instead of any specific lattice, such as the lattice of substructures, we can study several lattices of substructures invariantly with respect to the action of subgroups of structural automorphisms.
Thus, the tool for the general concept of a structure is not one of the three structures described above, but their mutual interdependence. An early description of the method briefly outlined here can be found in another paper by the author [
11], while a more elaborate description is in preparation.