PART 4:
INFORMATION, INFORMATION THEORY AND INFORMATION TRANSFER

(01Mar97)

If we think in terms of PERCEPTION, then we are most likely to think in terms of THINGS -- because things are what we perceive and have mental-image pictures of stored in our memory library. The incoming signals through the eye are processed as signals through a number of systems before they end up as thing-images.


It is relatively certain that our "understanding" processes undergo something quite similar, if not identical.

When we think in terms of THINGS, then we think in terms of objects, their shapes, sizes, colors, their meaning as an IT. We also think in terms of the distances between objects, their placement with regard to each other.


If we think of subjects or topics, we do so by first converting them into an IT-THING: for example, consider biology. IT is a science, as most know whether they know anyTHING more about IT.

The most fundamental basis of most consensus realities consists of IT-THINGS, and the most essential nomenclature utilized is set up to identify it-things. And this is the case even regarding philosophical abstractions, which, too, are it-things -- e.g., IT is an abstraction whatever IT is.

The general purpose of the first organized psychical research organizations set up during the 1880s was to witness, inspect, identify, separate and categorize what later came to be called "paranormal" phenomena.


But in order to proceed, the phenomena first had to be given identifiers, and which turned the phenomena into IT-THINGS. "IT is clairvoyance," for example. "IT is levitation," "IT is mediumship," "IT is thought-transference" (a term-concept later replaced by "IT is telepathy"), and finally "IT is psychic" whatever it is.

Phenomena are not just phenomena, but different kinds of them, and which need to be differentiated, distinguished and identified one from another. But sometimes this differentiation doesn't work very well if one doesn't really understand what IT is in the first place.


For example, in spite of about 100 years to do so, exceedingly great confusions continue to persist in making differentiation between clairvoyance and telepathy.

But generally speaking, differentiation is achieved by making an IT out of different kinds of phenomena and then assigning a nomenclature bit (or byte) in order to talk or write about any of them. When this is accomplished, we can thenceforth "know" what is being referred to because it has been rendered into an IT-IS kind of THING.

The first essential goal of organized parapsychology (circa the 1930s) was not only to inspect ESP phenomena, but to do so only within the parameters of recognized and approved scientific methods.


Extra-sensory perception (ESP) was an it-identifier of "perceptions" that could not be attributed to any of the five physical senses, and so it could be said those perceptions were external to or outside the physical senses.

To test for the presence of ESP in given individuals or subjects, "targets" were utilized, and there came into existence standardized forms of targets (among them the famous Zener cards) which mostly consisted of pictures of geometric shapes or colors. A "target" is always an IT.


The goal of the testing was to determine if the subjects could perceive the "targets" via senses other than the physical five.
The targets, of course, were IT-THINGS - expressed as "It is a circle," "It is a square," or "IT (the target) is the wavy lines."

Now, in the "universe" of IT-THINK, there is only one basic way to judge "success" - whether one perceives-sees IT or doesn't see IT.


Thus, the parapsychology ESP subjects either "got the target" or didn't get it." Or, "hit" the target, or "missed" it.

As we shall see in later essays, the "hit-miss" paradigm that arose in parapsychology led to some rather dreadful situations regarding comprehension, morale and defeatism.


But nonetheless it was a perfectly logical approach within the contexts of IT-THING-THINK, and which contexts are universal everywhere and in all cultures.

The concepts of PERCEPTION are intimately and permanently linked to IT-THINGS, because if you examine any of them very carefully one can only perceive an IT. And even then, as has been reviewed in Part 3, the IT-PERCEPTION is a mental-image reconstruction, the sum of which is of the perceiver, and not exactly of the IT itself.

It is worth the time to review a few of the numerous definitions of THING:

  • a separate and distinct individual quality, fact, idea, concept or entity;

  • a material or substance of a given kind;

  • a piece of information or news;

  • an event, deed, act or circumstance;

  • a state of affairs in general, or within a specific or implied sphere.

The five definitions of THING given above can and do account for almost, but not quite, everything - and which is why we refer to everything AS every-thing. And so our perceptions are geared to perceive, identify, and discriminate among THINGS - and which then emerge in conscious awareness as reconstructed images.

There is absolutely nothing wrong with basic IT-THINK, and indeed it permits survival on about a 90 per cent basis - except when there are holes or gaps in it.

But IT-IS gaps can be somewhat corrected within the contexts of consensus realities in that IT-IS perception that is consistent with consensus reality is considered proper or successful perception, while perception that is not is considered improper or aberrant-undesirable - or at least non-conforming.

In general, however, any gap-difficulties along these lines are sort of smoothed over in that the nomenclature of a given consensus reality is the concepto-nomenclature everyone within it speaks and writes with - and tends to think with, too.

Just outside the enormous, collective IT-THINK syndromes of our species is a slightly different THINK format.
This "level" of thinking has to do with RELATIONSHIPS between and among IT-THINGS.


Identifying it-things, and identifying them as it-things, only goes so far, although that process is entirely serviceable to a certain degree.

One can identify it-things, endlessly so, but only because they become perceptually concrete in some form - even an idea takes on a sort of concrete-ness if it becomes shared and approved of.

Relationships among it-things, however, are usually of a far different matter because, in the first instance, they have to be deduced. For example, the relationship between hydrogen atoms and hydrogen bombs is not readily apparent, and thus had to be deduced before it became identifiable.

This is to point up that although the arrangement of IT-THINK to IT-THINGS is usually on a one-to-one basis, the arrangement of IT-THINK to relationships among and between IT-THINGS is not on any kind of one-to-one basis - excepting the most gross and familiar samples of it.


The reason for this difficulty is that relationships between it-things can be many and varied and include anything from the imaginable to the unimaginable, from the boring to the fantastic.

Another difficulty arises because once IT-THINK becomes properly installed it tends to run on automatic with the mind-boggling speed encountered in Part 3 regarding the basic ten-step processes of perception.

DEDUCTIVE-THINK regarding relationships, however, usually never runs on automatic unless the deductions have themselves been pre-reduced to common understanding, at which time those particular deductions have taken on the clothing of IT-THINK.

Relationships of it-things to one another can be explicit or implicit, with the explicit ones being easier to identify, this type of thing usually being referred to as logic.

Implicit relationships, however, are identified as such because there is very little in the way of objective or explicit cues involved.


Thus, the deducing (detecting) of implicit relationships can escape the deductive processes of almost everyone - with the exception of those who somehow chance to "notice" them.


And those who DO notice them are quite likely to be attributed as intuitives. And, indeed, if it were up to me, I'd itemize the deduction of implicits as the basic and most broadly-shared type of intuition's many other types. And here is a basic clue regarding "enhancing" one's intuition - by first enhancing one's deductive processes regarding implicit relationships.
As it is in our present consensus reality, we reinforce the processes regarding explicit relationships, but pay very little attention to strengthening the much more wide parameters of implicit relationships.

One of the more recent definitions of "genius" is that a genius is one "who sees what others cannot." Although this clearly involves a lot of factors, the deducing of implicit relationships probably is fundamental here - since most rely on explicit rather than on implicit deducing.

Now to move speedily on.


The relationship, for example, between ESP and perception seems explicit enough, and therefore seems logical -- especially when a long line of "psychics" say "I perceive" thus and so.


They are correct in saying that they do perceive. But what they perceive is in fact whatever has been processed through their perception-making systems, the sum of these processes being the perception.


And as we have seen these end products are not at all one-to-one images. And so what they report "seeing" may or may not correspond with the actual facts or conditions of what they have "seen" as perceptions.

This is a situation that has not gone unrecognized in parapsychology.


In testing for ESP, researchers encounter many more "misses" than "hits" and the frequency of the misses has condensed into the theory of "Psi-missing." It is thought that Psi-missing is somehow related to "avoidance" of the "target," and as such constitutes some kind of unidentified psychological factors.

You see, "paraPSYCHOLOGY" is, after all, majorly conceived of as a branch of psychology -- not as a branch of perception study. And when it was understood by the rest of science that "perceptions" mostly consisted of "cognitive" versus physiological factors, perception, too, began to be thought of as predominantly having a psychological basis.

In any event, ESP and perception of IT targets are thought to go hand-in-hand, and all explicit and implicit considerations along these lines are shared not only in parapsychology, but throughout science, philosophy, and in our present general consensus realities as well.

Furthermore, the web of Psi-Perceptions is linked throughout by the IT-making nomenclature commonly utilized.


If, then, one refers to Psi or ESP, it is automatically understood everywhere that you are referring to special formats of perception that have been assigned IT nomenclature: psychic, clairvoyance, telepathy, intuition, and etc.


It is even commonly understood that "special" refers NOT to perception per se, but to the unusual other-than-sensory ways it is achieved -- if and when it is achieved.

Well, this "prevailing paradigm," as it should properly be termed, has actually prevailed for about 100 years, and has been unsuccessfully approached and tested in the light of every angle conceivable.


The only thing that has been achieved is to document beyond any shadow of doubt that ESP processes do exist, but whose presence by parapsychological methodologies are found at only very low statistical levels (which will be discussed in a later essay).

So, "psychic" perceptions have been tested for from every angle possible -- which is to say, very angle consistent with the prevailing consensus reality hypotheses that ESP and Perception are interrelated both explicitly and implicitly, so much so that you can't have the one without the other.

But what if this consensus reality concept isn't complete enough? In other words, what if it has a "gaping hole" in its interconnecting line-up of conceptualizing -- one of those invisible gaping holes that are not at all obvious because the apparent picture seems complete and logical enough?

And what if what is needed to fill this hole has been around for about fifty or more years, but has been excluded because the prevailing concepts are considered sufficient unto themselves? And because if the needed factors were to be included, the entire consensus making nomenclature appropriate to Psi-Perceptions would either explode or be useless and vacated.

This would mean that everyone has cloned the wrong stuff, so to speak, and what they have cloned in this regard has been acting as mental information processing viruses all along.


Ye gads! This would imply a radical reality shift - one which, in its first instance, would big-time EMBARRASS those possessed of the cloned viruses - not only in parapsychology, but in science and philosophy as well, to say nothing of the consensus realities involved.


Information
The essential definitions of the verb TO INFORM, and the noun INFORMATION, never have been ambiguous, but quite precise and clear.

INFORM is said to have been derived from the Latin verb INFORMARE from IN + FORMA.


However, the Latin FORMA was a noun, and even though the preposition of IN is added to it, it still remains a noun. And nouns, of course, refer to and are meant to identify it-things, not activities which verbs indicate.

FORMA referred to the shape and structure of something as distinguished from its material or constituent parts.


The preposition IN refers to inclusion of some kind, most usually a spatial inclusion, but also inclusion in something that does not have spatial-material form such as belief, faith, opinion or assumption (i.e., in the faith, only in belief, in his or her opinion or assumption, etc., and of course, IN his or her conception or misconception.)

The key concept of FORMA refers to shape and structure, and so INFORM refers to what has structural shape, has taken on structural shape, or been put into structural shape.


So, technically speaking INFORM remains a noun with regard to whatever form a form is in, becoming a verb only when referring to an activity which puts something into shape-structure.

In English, however, IN + FORM as referring to structural shape has been used only rarely, this meaning having early been replaced with the concept of MESSAGES - meaning that messages convey information, and that information is used to convey messages.


If the above seems mildly confusing, it's because it is. So don't worry too much at this point.


You see, on the receiver's part, the actual message is what one deduces from the words (or "signals") which the sender believed represented the message he or she was trying to send. This "process" takes a good deal of "encoding" on the sender's part and a good deal of "decoding" on the receiver's part. But I digress.

Additionally, when we think of something formed we tend to think in terms of FORM only, not that something has PUT whatever it is INTO form or format.

I now caution each who chances to read the above to slow down, focus a little, and notice two important factors:

  • that there is a vast and very incompatible raw difference between messages and the structure and shape of something; and

  • when we think of form as form, we tend to think of it as an IT object or subject, not as something which has been brought into or put into form by various shape-making, structure-making processes of some kind.

In other words, something which is formed or has achieved form is the RESULT of whatever has caused it to take on shape-structure.

In English, then, the concept of "into form" has been dropped or vacated, and so we tend not to think in terms of how and why something has come into whatever form it has.


But this is somewhat typical of English nomenclature, which tends to IT-identify end products as things in themselves, not as the result of processes - which is to say, formative processes that have to be structural in order to arrive at any given in-formed state.


This is best perceived not via words, but by a diagram. I'll provide one in the context of a more refined essay further on. But anyone can make one for themselves by diagramming how an IT does take on form.

To help in enhancing clarity here, when we think of those superpower faculties that result in some kind of clairvoyance, we tend to think the images the clairvoyant "sees" ARE the clairvoyance.


I.e., he or she "sees" things that others don't, and by means other's don't have active - hence the clairvoyant angle. We mistake WHAT the clairvoyant sees as the clairvoyance, and fail to notice that the informative processes which permit the seeing are the real clairvoyance.

In other words, into-form-making PROCESSES always precede the resulting images.


Thus, if clairvoyance is possible, the IN + FORM clairvoyance-making processes pre-exist what they yield - for what they yield is what the clairvoyant sees. If the processes are not active, then the clairvoyant will not see anything.


If we compare this to perception-making processes, we know that the perceptions are the sum result of whatever they have been processed through. The superpower faculties apparently "work" in the same exact way.

It is interesting, and important, to trace the ENGLISH etymologies of INFORM and INFORMATION. The OXFORD DICTIONARY OF THE ENGLISH LANGUAGE summarizes when and under what conditions English nomenclature can be noted as first in use.

With regard to INFORM, the OXFORD identifies the primary ancient Latin sense of INFORMARE (to give form to, shape or fashion), but notes: "The primary sense had undergone various developments in ancient and medieval Latin, and in French, before the word appeared in English."


This is a clever way of saying that when INFORM came into English usage it did not mean putting into a form.)

This appearance in English seems to have taken place during the 1300s, but seems more than anything else to have referred to "formative principle, or formative character."

Used in this sense, the first ENGLISH uses of INFORM were probably drawn from French rather than directly from Latin.
It is certain that the word INFORMATION is drawn from French, not directly from Latin. Its first usage's in English, again during the 1300s, are exactly those of the French:

"The action of informing [specifically as] forming or molding of the mind or character, training, teaching or instructing; communicating of instructive knowledge."

In this sense, then, from French into English, INFORMATION referred to mind-shaping, out of which would emerge "character" - such having been a particularly French preoccupation ever since.

After this shift in usage-meaning, in English INFORMATION then appears to have separated into two components, both utilizing the same nomenclature term, INFORMATION.

The first component remained the same, almost up until the 1930s when it began to be identified as "mind-programming."

The second component had to do with providing evidence, either for or against someone, and usually the latter regarding criminal court cases, heresy examinations and trials.


It would appear that "evidence" found acceptable or logical in the light of certain consensus realities was accepted as "information" - while "evidence" found unacceptable was rejected as something else.

INFORMATION was still being thought of in exactly this way among the world's intelligence agencies and systems when I chanced to fall into the government-sponsored "Psi-spy" research project at Stanford Research Institute in 1972.


Also, during that same epoch, the then hopeful and exceedingly well-funded realm of "scientific" futurology (now generally defunct) also had adapted to this same concept of information, and was being tortured by it - which is to say, adapted to the concept that information consists only of whatever is found acceptable, or logical within a given consensus reality.

"Consensus reality," however, was considered by futurologists to consist of the majority opinion of "informed specialists" and/or their vote. Since majority opinions can be wrong at least as often as right, one does wonder how futurology every got off the ground. However, one doesn't need to wonder why it "failed."

During the 1600s, and specifically as the result of certain Renaissance activities, a new concept-context regarding INFORMATION was added into this or that drift of meanings.


The earliest noted uses of this meaning occurred about 1649, and we find the gist of this meaning more or less unchanged in WEBSTER'S of 1828, the original edition of the first American dictionary of the English language.

In that dictionary this meaning is given as the FIRST meaning of INFORM. And I quote:

"INFORM, verb transitive: - Properly, to give form or shape to, but in this sense NOT USED. [Emphasis added.]
"1. To animate; to give life to; to activate by vital powers.
"2. To instruct; to tell to; to acquaint; to communicate knowledge to; to make known to by word or writing."

"INFORM, verb intransitive: - To give intelligence, as in: `He might either teach in the same manner, or inform how he had been taught.' And: "To inform against, to communicate facts by way of accusation."

"INFORMATION:
"1. Intelligence via notice, news or advice communicated by word or writing.
"2. Knowledge derived from reading or instruction.
"3. Knowledge derived from the senses or from the operation of the intellectual faculties.
"4. Communication of facts for the purpose of accusation."

As of 1828, then, long gone is the concept of IN + FORMA, as is indicated by WEBSTER'S 1828 itself - and not reactivated until the advent of Information Theory, as will be discussed ahead (save to mention here that information theory cannot survive without that concept.)

In WEBSTER'S 1828, the first definition of INFORM - to animate; to give life to; actuate [i.e., activate] by vital powers - reflects the central hypothesis of VITALISM, which we have already encountered.


However, the term VITAL-ISM apparently had not evolved as of 1828, since it is not given in that same dictionary. (The concept of an ism itself seems to have surfaced only in about the 1780s.)

However, a brief review of this topic is important - because there are significant links between essential vitalism, information, and activation of the superpower faculties. (An individual essay regarding vitalism will be provided within this series of essays.)

You see, IF information (intelligence) is accurate enough, it is broadly accepted that it can activate or vitalize activity, and which would be akin to animating or reanimating them.


On the other hand, if information (intelligence) is cluttered with information viruses, one would not normally expect activation. Rather, one would anticipate de-activation, or devitalization - and which, if it could happen, would result in all sorts of de-evolutionary stuff.

VITALISM was crushed and beat into non-existence about 1920, at which time the consensus realities of philosophical materialism acquired the contexts of science proper and thenceforth prevailed. And any science based in philosophical materialism simply has to be an IT-MAKING science.

Prior to that, philosophical vitalism (technically in existence roughly since about 1533 during the Renaissance) and philosophical materialism (technically in existence since about 1845) had been seen as sister sciences.


The advocates of the two philosophical orientations were soon antagonistic to each other. An enormous conflict, now quite forgotten, ensued, lasted for about eighty years, with the materialists being the ultimate victors. Vitalism was snuffed in academia, and references to it were deleted from consensus reality sources which then prevailed as logical and rational.

In spite of all the philosophical imbroglios that are brought forth to explain the victory, the actual reason is quite simple.
By 1920, the material sciences had demonstrated they could produce products of enormous, even fabulous economic value. The vitalism sciences did not produce much of economic meaning. Funding therefore went to the material sciences. End of that story.

There were two essential definitions regarding vitalistic principles, to which a number of other concepts were derived. Be sure that I am not digressing or drifting here.

1. That the functions of a living organism are due to a vital principle distinct from physical-chemical forces;

2. That the processes of life are not explicable by the laws of physics and chemistry alone - and that life is in some part self-determining and self-informing.

Please read self-informing as IN + FORM, meaning self-making into form.

For conceptual clarity, any use of the term VITAL within vitalism's contexts should immediately be replaced with ANIMATING - at least to discriminate between animate and inanimate conditions.

In the end, all of the nomenclature that might be associable to vitalism and/or its two essential concepts was stringently, and with something akin to a vengeance, expunged from modernist consensus reality-making literature. Any even glancing reference to those terms was enough to occasion loss of professional standing, potential funding, and etc.


Thus, cutting-edge scientists have to walk gingerly, and talk around such concepts if and when they chance to encounter any possibility of their real existence.

In any event, this brief review of the etymological history of INFORM and INFORMATION indicates that only one concept of them prevails, the concept that information is what one reads and learns from.


We can note, too, that two important concepts have more or less fallen into disuse and oblivion: IN + FORMA, and INFORM as it relates to animating principles.


And it is in this consensus reality condition that information theory arose.


Information Theory
So, what IS information theory?


And why might it be of fundamental importance with regard to activating (vitalizing) the superpower faculties?

Most sources dealing with information theory are somewhat or completely inaccessible (unintelligible) to those who haven't developed the mental information processing grids or nomenclature to deal with it.


However, THE NEW COLUMBIA ENCYCLOPEDIA (1975) has a rather neat rendering, at least as regards the early developmental hypotheses.

The theory is indicated as a mathematical one, principally formulated as of 1948 by the American scientist Claude E. Shannon, to explain aspects and problem of information and communication ("communication" later being thought of as information-transfer, especially in the psychoenergetic research of the former USSR.)

The entry in the encyclopedia is worth quoting in its entirety, and I'll do this first.


I caution you not to get confused if you don't understand parts or all of it.


After quoting it, I'll lift out the signal, easy to conceptualize, part and clarify it with respect to opening new cognitive channels toward activating the superpowers.

I never recommend anything, but sometimes I "suggest." If you have any desire at all to approach an activation of any of the superpowers, I suggest you pay serious attention to the quoted materials below, even to the point of memorizing them (i.e., installing them quite firmly in your memory library.)

One preliminary note, though. Shannon et. al. seized upon the term ENTROPY and included it in the discursive part of the theory. This is a term properly belonging to thermodynamics, and has otherwise since been defined in a number of different ways. In information theory it means "noise," and so I'll replace "entropy" with noise, indicating that I did so.


Synopsis of the 1948 Information Theory

"In this theory, the term INFORMATION is used in a special sense; it a measure of the freedom of choice with which a message is selected from the set of all possible messages.


"Information is thus distinct from meaning, since it is entirely possible for a string of nonsense words and a meaningful sentence to be equivalent with respect to information content.


"Numerically, information is measured [via the theory] in BITS (short for binary digit; see Binary System.)


"One bit is equivalent to the choice between two equally likely choices. For example, if we know that a coin is to be tossed but are unable to see it as it falls, a message telling whether the coin came up heads or tails gives us one bit of information.
 

"When there are several equally likely choices, the number of bits is equal to the logarithm of the number of choices taken to the base two. For example, if a message specifies one of sixteen equally likely choices, it is said to contain four bits of information.


"When the various choices are not equally possible, the situation is more complex.


"Interestingly, the mathematical expression for information content closely resembles the expression for ENTROPY in thermodynamics. The greater the information in a message, the lower its randomness, or `noisiness,' and hence the smaller its entropy [i.e., the smaller its noise content.]


"Often, because of constraints such as grammar [language, and the way it is expressed], a source does not use its full range of choice. A source that uses just 70% of its freedom of choice would be said to have a relative noise ratio [entropy] of 0.7. The redundancy of such a source is defined as 100% minus the relative entropy, or, in this case, 30% [meaning 30% message-signal adulterated by 70% noise].


"The redundancy of English is about 50%; i.e., about half of the elements used in writing or speaking are freely chosen, and the rest are required by the structure of the language.


"A message proceeds along some channel from the source to the receiver. Information theory defines for any given channel a limiting capacity or rate at which it can carry information, expressed in bits per second.


"In general, it is necessary to process, or encode, information from a source before transmitting it through a given channel.


"For example, a human voice must be encoded before it can be transmitted by radio.


"An important theorem of information theory states that if a source with a given entropy feeds information to a channel with a given capacity, and if the noise in the source is less than the channel capacity, a code exists for which the frequency of errors may be reduced as low as desired.


"If the channel capacity is less than the noise source, no such code exists.


"The theory further shows that noise, or random disturbance of the channel, creates uncertainty as to the correspondence between the received signal and the signal transmitted.


"The average uncertainty in the message when the signal is known is called the equivocation.


"It is shown that the net effect of noise is to reduce the information capacity of the channel. However, redundancy in a message, as distinguished from redundancy in a source, makes it more likely that the message can be reconstituted at the receiver without error.


"For example, if something is already known as a certainty, then all messages about it give no information and are 100% redundant, and the information is thus immune to any disturbances of the channel.


"Using various mathematical means, Shannon was able to define channel capacity for continuous signals, such a music and speech.


"While the theory is not specific in all respects, it proves the existence of optimum coding schemes without showing how to find them. For example, it succeeds remarkably in outlining the engineering requirements of communication systems and the limitations of such systems."

SEE C. E. Shannon and Warren Weaver, THE MATHEMATICAL THEORY OF COMMUNICATION (1949).


Formats of (or regarding) Information
When we begin to think of what information IS, most of us probably will think it is what we hear or read in some kind of printed or visual format. We think this because this concept "dwells" in consensus realities as such, and we have cloned it quite nicely. And from any number of aspects that concept is serviceable - as far as it goes.


But. By the time "information" reaches a spoken, printed or visual format, it is an end-product of the processes which have organized and produced it in those formats.


Nonetheless, this end-product can act as a "source" of information and we can more or less duplicate it in our own heads.

"Duplicate," of course, means reproduce or copy it into our own heads, the ostensible goal being to understand it. In this sense, then, the information we in-put into our heads has been CONVEYED by the spoken, printed or visual format.


After the in-put, however, the "conveyance" of the information continues getting into our heads by being filtered through the mental information processing grids of the recipient. The grids are extensions of the memory library earlier referred to.

In THIS processes, the "information" will ultimately reach steps 8 and 9 of the perceptual processes. Meaning that the "information" that finally comes out as understanding will be the sum of the in-put plus whatever the in-put gets filtered through in the case of each individual.


If matches to the in-put "information content" are found in the memory library, THEN a kind of duplication can take place. The duplication is called "understanding."

But if matches are not found, then the information content probably will be routed through the nearest similarity in the memory library. In this case, we are now one-step or more removed from duplication (and removed from "complete understanding.")


If no matches are found, then the recipient of the in-put information content will "draw a blank" - for example, regarding twelve types of snow, camels, telepathy or clairvoyance.

In other words, INFORMATION is what we understand, even if only in a partial way. If the in-put does not result in "understanding," then it is NOT information.


Information Transfer
The whole of the above, and its obvious problem areas, is what some information theorists refer to as the information transfer process.


One of the central concepts of information theory is that all information is available all of the time.


Some of the theorists mitigate this all-inclusive concept by saying that information sources are everywhere.


Others opine that information can be drawn from everything and anything.

In the sense of all of the above, the EXISTENCE of information is not in question. What is problematical, in big-time ways, is the TRANSFER of it into a system wherein it can be duplicated, misduplicated, or blanked out.


In the sense of the human, the prevailing consensus reality concepts usually hold that the "system" being referred to is "the mind" and its mental information processes.

"The mind," however, when spoken of this way is applicable as a generality to every human specimen, and which is good enough for a theory.


In matters of actual PERFORMANCE, though, the "individual mind" should be substituted for the all-inclusive generality - because even if information does exist everywhere, it is the individual mind that produces duplication, misduplication, or the blanking out, and which in turn result in understanding, misunderstanding, or nothing at all.

Please note that the term PERFORMANCE has been emphasized above because it is entirely relevant toward activating the superpowers, "activating" having to do with performance. And here I foreshadow a topic that will require at least two essays among those several more to come.


Information Signals
Information transfers via speech, print or in visual formats, actually contain two MODES or MODULATIONS of information content.


But to get at this, it must FIRST be comprehended that the words of speech or writing/print the images, charts, etc., of the visual formats are NOT the information content itself, but merely symbols and signs for it.


In this sense, the symbols and signs are the OBJECTIVE "carriers" of the information content - which is to say that they are SIGNALS that will stimulate duplication of the content simply because the receivers associate MEANING to the signals - IF the meanings of the signals are shared in common.


If the meanings are not shared among the recipients, then the signals will be "inaccessible" to all those who do not.

And here is one of the most apparent bases for language and its concepto-nomenclature - to establish a shared and sharable basis for the sending and receiving of information content.


This is to say that pre-set meanings are encoded into nomenclature and images, and the consensus reality learning networks transfer the encoded meanings into the memory storage of their citizens so that there can be a mutual basis of information transfer and exchange. An intrasocial collective or group is thus formatted regarding transfer of information within it.

The best pre-set words or images to effect this information transfer unity are those that have precise meanings encoded into them, since the meaning-information-content can be "recognized" most easily.

Any increasing permutations of meanings regarding a given information transfer signal tend to decrease the cohesion of the unity within the collective, and tend to permit distortions of meaning within individuals.

One would therefore think that precise and exact meanings for signals would be stringently established by social consensus necessity. And indeed this IS the case where an absolute need to do so is apparent, the "need" being intimately related to performance, and especially where it is found to be dangerous not to be precise.


For example, no one becomes an electrician based only on the general consensus reality that electricity lights up bulbs and turns the toaster on.


A suitable and precise nomenclature has to be evolved and become shared among potential electricians - or else they can get fried all too easily. Airline pilots can not become one simply because airplanes fly. Arctic people cannot deal with snow simply if it is snow, and Arab Bedouins will be out-maneuvered in the economics of the camel market if they think a camel is a camel.

However, within any given social unity where there is no perceived absolute need to INCREASE nomenclature, that kind of effort is not usually undertaken - because the average citizen within the unity, and with regard to average performance within it, can function quite well via a lesser rather than an increase in signal-carrying nomenclature.


And, to begin with, the so-called average citizen probably won't ever "acquire" a nomenclature in terms of quantity that extends beyond his or her recognized need to do so, or beyond what it takes to fit into the consensus reality they desire to fit into (or, sometimes, are trapped within.)

So the average citizen within any given consensus reality had no explicit or necessary need to add more specific nomenclature; but there is also a need not to have too little, either.


The way this is apparently resolved is to establish a number of IT-IDENTIFIERS that do not require much further break-apart into it-TYPES, into increasing refinement of comprehensions of types of something, and which would require the increase of nomenclature.


In this way, then, people who do not need to use different types of snow for survival can be content with snow as something that falls in winter and needs to be shoveled when it interferes with traffic or might crush the roof in. So, among such people, SNOW is snow. It is a perfectly good information signal, and the need for any increasingly refined differentiation beyond that probably has to do only with amounts of it.

So, among such people "SNOW" is a "clean" and "clear" signal regarding information transfer, whereas among the Arctic peoples barely fifty years ago it would have been as "noisy" as Times Square at New Year's Eve.


In much the same way, people who don't realize that different types of clairvoyance exist will not have any need to identify them - meaning that the single use of this one nomenclature signal is perceived by them to be sufficient.


But not to anyone who wants to learn how to be clairvoyant. The best instructors of clairvoyance I am familiar with have to begin, as they do, by breaking the single concept apart, at least into "aspects" of clairvoyance.

So, here we now approach the concept of "clear" and "noisy" signals, this concept revolving around whether or not the carrier (word or image) of a signal is a precise, thus a clean one, or whether it induces noise into the signal load.

And it is at this point that the essential problems of information transfer integrate with the basic information theory offered up by Shannon in 1948, the basic problem regarding information transfer OF ANY KIND having to do with the ratio between "signal" and "noise."


Please note that in preparation for this series of essays, an earlier essay dealing exclusively with the SIGNAL-TO-NOISE RATIO has been available in this database for several months. That essay can now be appended to this series' essays as Part 4A.


Information Noise
As stipulated within information theory by Shannon, a message (information content) proceeds along some channel from the source to the receiver.


In line with our interests, information is in-put via some kind of "channel" to the receiver, who then out-put it in terms of information encoded into concept-nomenclature for further information transfer.


But the in-put itself is an information transfer from "a source" wherever or whatever it might consist of.

We are thus dealing with TWO information transfers:

(a) the in-put transfer, and
(b) the out-put transfer.

Between (a) and (b), however, is "a channel," and after (b) is concluded another "channel" is necessary to further accomplish an information transfer.

So we can think in terms of the in-put channel and the out-put channel, the in-put channel having to do with reception of the information, the out-put one having to do with what we call "communication."

In the human sense of all of this, the out-put transfer (the "communicating") must first be encoded into concepto-nomenclature that can be transferred to others simply because their mental information processing equipment is already encoded to receive and duplicate it.


All of this seems clear enough, doesn't it.

However, there is one serious glitch. You see, the in-put transfer ALSO has to be processed INTO the same mental information processing equipment in order that it CAN be "received."

If that mental information processing equipment (which now has to do DOUBLE duty regarding in-put AND out-put) is not pre-formatted with some exactness regarding both quantity and quality of the in-put, then the "channel capacity" will be LESS than it needs to transfer the full information load into the receiver system.


If this is the case, then the out-put transfer will be only a partial one, or perhaps hardly anything at all. If it would be the case that the in-put and out-put channel cannot MATCH any of the signal, then the signal will disappear into the blanked out thing.

In basic information theory, anything that hampers, distorts, confuses, obliterates the signal is referred to as "noise."
In this sense, if the noise "in" the channel is less than the signal, then a code exists (or can be established) for which the frequency of errors (noise) may be reduced as low as desired.


If the "noise" in the channel is greater than the signal, then the signal may not be identified; it can still exist in the channel, although so embedded in the noise that it cannot register, be picked up, or identified.

In the sense we are interested, the human sense, it turns out that human mental information processes ending up in "perception" can produce not only signal-laden but noise-laden conceptualizations and mental image pictures with hardly any way to discriminate which is which.


Where Does Information Processing Noise Come From?
In answer to this question, the daring among us will assume that the noise originates in our own heads - and which is usually the case.


But a deeper inspection of noise sources reveals that what's in our heads and which contributes to the noise may not be innately present to begin with.

A better part of the noise sources in our mental information processes is ACQUIRED - usually by the enculturization processes that make us fit in our given consensus realities.


This understanding is rather broadly accepted in some echelons of human inquiry, especially if the consensus reality social processes drift into mind-programming rather than overall efficient education.

But there is another far more powerful, but far more LESS obvious, noise source, and it is one we all adapt to in order to learn to communicate.


Language itself.

As Shannon pointed up in his information theory (and much to the shock of many at the time) that one is "constrained" to utilize language - and with language comes the concepto-nomenclature that becomes lodged, by necessity, into our memory library.


I'll paraphrase how Shannon put it.

Regarding English, some fifty per cent of the concept-nomenclature we lean upon is required by the structure [and familiar usage] of the language. The other 50 per cent is open to free choice of concepts and nomenclature.


Shannon's implication was that if the language-determined part was inhabited with noise-making redundancies, then any adaptation to the language would induce these into mental information processes of ALL those who utilized it.
So, you see, we are not at each individual level "guilty" of faulty information processing - at least 50 per cent of the time.

But whatever their source, even the 50 per cent presence of noise-making viruses can easily decrease or prevent performance ever activating.

As it turns out, although noise-making redundancies can be identified in every area of human endeavor, some are more prone to a larger percentage than others, especially those that have become adapted to ambiguity. Dare I mention politics and over-bloated administrations? Or the present conditions of the "fine" arts? Or the parameters of "love," "hate," "sex?" Of course, I'll not mention the realms of "psychic phenomena" - since everyone knows what they are.

In any event, it might be said that where over-simplification and ambiguity prevail, so too do noise-making redundancies - all of which bury the signal within the noise, no matter how fashionable is the noise.


It's somewhat worth mentioning, generally speaking anyway, an area of human endeavor thickly populated with noise-making redundancies tends to be "volcanic" in nature. Such areas can exist peacefully within their own parameters, stabilized by their own consensus realities. But if intruded upon, or if THEY intrude upon, things begin to heat up.

The topics of information and information transfers will be picked up again in additional essays.


It is now desirable to devote Part 5 to a correlation of what has been discussed in Parts 1 - 4.


In Part 6, we'll discuss not only the noise-making redundancies embedded and perpetuated within ambiguities, but their utterly destructive viral effect on clean, clear "signals." Ambiguous concepts induce structure-lessness, hence they wreck any signal-awareness of STRUCTURE, and without knowledge of the structure of anything very little else can ever be known about it. As we shall see in subsequent essays, STRUCTURE is the IN + FORM, or the format, of something - and as such is what needs to be worked with or within, not against.

In any event, any real attempt to activate any of the superpowers must encompass the reality that signal-to-noise ratios are intimately involved. Thus, the presence in any system of disinformation or misinformation can act as if it is infected with viruses.
 

Back to Contents