Digital Journal

S2Schat shows a radical difference from GPT algorithms

0

Singapore, September 16, 2024

Arllecta Group approached the initial implementation of the sense-to-sense (S2S) algorithm based on its own mathematical theory Sense Theory [1] specifically designed for the creation of self-identifying AI. 

S2SChat

Now we will briefly describe the technological approach used in our solution.

The uniqueness of this algorithm in comparison with the current most advanced GPT models lies in at least four of its characteristic differences. 

The first difference is that this algorithm does not require gigabytes of data and billions of normalized parameters to train the algorithm. The architecture of the sense-algorithm has, as a special case, an implementation of a limited Boltzmann machine. However, at the core lies the implementation of the zero-object paradigm as a vector quantity of non-constant length. This approach allows us to move from numerical (statistical) analysis to sense (semantic) analysis, which is extremely important when creating self-learning and adaptive AI. Without the ability to create “new” knowledge, no AI will succeed in self-identification (self-consciousness). 

The second difference is that in the implementation of such extremely important AI tasks for the result as clustering and data typing, we use two innovative tools – the Neuro-Amorphic Function (NAF) and Sense Diagrams.  NAF allows us to extrapolate many physical phenomena in nature and very well describes the distribution of senses (meanings) in the text received as input. A sense diagram, created by analogy with Bohr’s atomic theory, very accurately describes large groups of sense (semantic) sets made up of input data, as well as which is extremely important, semantic connections between each element of these sets. It is worth noting that the computational time required to construct and analyze such sets is polynomial.

The third difference is that the sense-algorithm allows us to meaningfully connect objects of different nature from a large volume of data. For example, if there are 1,000,000 sets classified according to any criterion and having different origins of their elements, the sense-algorithm allows one to find a semantic connection between any two objects of these sets.

The fourth and perhaps one of the most important differences is that we do not use the transformer architecture at all, since we believe that this architecture has several critical defects that greatly distort the meaning of the processed input text. Two obvious defects are that, firstly, the combinational law for the scalar product of a query vector and a key vector cannot be satisfied in the attention mechanism, secondly, normalizing the values ​​of the output vector of the attention mechanism introduces distortions in the decoder module, since any type of context vector, when normalized, loses the main thing – the semantic connection between its elements.

 

Now we will briefly describe the AI focus in our solution.

According to the author of the mathematical theory Sense Theory and the creator of 25 main software modules that make up the software core of S2Schat [2], Egger Mielberg, the current task of our AI is to obtain only one sentence as the result of an answer when analyzing not only a separate small passage from a book or a specialized article, but also a full-fledged one books of 500 sheets or more.

The main cognitive task of our solution is to implement 2 important directions. The first direction is to search and linguistically describe the zero object of the first level as a vector axis that determines the main meaning of the processed text. 

The second direction is to search and describe the depth of the semantic connection between zero objects of the second and other levels. For this implementation, we use Sense Derivatives [3] and the Sense entropy [4] value. 

The uniqueness of the meaning of sense entropy lies in the possibility of describing the degree of semantic connections between objects of different natures. This feature is missing in traditional mathematics.

 

And now we’ll tell you a little about the numbers obtained when comparing our S2Schat solution (13 modules out of 25 implemented) with the most advanced GPT solution now in the market, in our opinion, ChatGPT 4o from OpenAI. The reason we chose ChatGPT is that the vast majority of other current GPT-solutions in the global AI market now are based on ChatGPT.

ChatGPT S2Schat 
The sense amplification coefficient (%) [4] 16 97 the more tokens the algorithm identifies in the text and uses for analysis, the lower the value
The sense efficiency coefficient (n zero objects – SEC) [4] 5 1 the less semantic connection between tokens, the higher the value
The Mielberg sense cycle [4] SEC not const SEC const if SEC remains constant when the text sample changes, the result of finding the main meaning is maximum
The Sense Entropy (n) [4] 0.3 0.9 =1 – one meaning (sense) of the analyzed text, 

<1 – several meanings (senses) of the analyzed text

 

For our text analysis we used the American Ways book (XX, On Understanding excerpt).

Below are the exact specifications of our solution and approximate ChatGPT’s.

ChatGPT S2Schat
Number of servers >>100 2 S2S: 4CPU, 32 RAM, 1Tb SSD
Number of parameters >100 mln 25 S2S: 25 core modules determine 1 parameter each
Architecture transformer & others sense-to-sense model S2S: sense derivatives, sense limit & sense sets
Training database different text open sources user downloaded text S2S: the algorithm works on processing text in A4 format
Training mode pretrained real-time S2S: the algorithm uses user-uploaded text
Team number >50 4 S2S: each engineer covers two or three positions
Calculation methods of  traditional mathematics Sense Theory  S2S: a number of computational tools of this theory are used

 

In our algorithmic calculations, we are considering “sense energy”, which has a vector nature and therefore, it will be more accurate to define the Law of Conservation of Sense [4]: 

The total sense energy (SE) of any open sense space (OSS) is constant if the conditions of the Mielberg cycle

are satisfied in this sense space.

In other words, to create AI with self-identification, it is extremely important for the implemented algorithm that the meaning (sense) of the analyzed book, article, abstract and other source remains the same. 

The law of conservation of sense, like the law of conservation of energy, shows the constancy of the existence of the object of study only in different semantic (energy) forms.

The law of conservation of sense, in contrast to the Turing test, qualitatively determines the degree of “humanity” of digital AI.

Resources:

[1] Sense Theory. Part 1.

https://vixra.org/pdf/1905.0105v1.pdf

[2] S2Schat.

https://www.s2schat.com

[3] Sense Derivative.

https://www.researchgate.net/publication/344876659_Sense_Derivative

[4] Sense Entropy. The Law of Conservation of Sense.

https://www.researchgate.net/publication/369295558_Sense_Entropy_The_Law_of_Conservation_of_Sense



Information contained on this page is provided by an independent third-party content provider. Binary News Network and this Site make no warranties or representations in connection therewith. If you are affiliated with this page and would like it removed please contact [email protected]

Paying All-Cash vs. Getting a Home Improvement Loan: Which is best?

Previous article

Ainur Ryskulova’s “Contemporary Classics” Illuminates White Box Gallery

Next article

You may also like

Comments

Comments are closed.