From Memetics


What is it, and how do we relate to it.

there is a problem with the shannon definition on the atomic and axiomatic level. it refers to subjective manipulation of information our definition of information is related to both von neumann entropy and shannon information entropy

through this can be measured in the brain, through pathways. this means memetic innoculation can be inbuilt and trained

use of bayesian and time updating will lead to algorithmic information relates to information and probability space. of informational events. this is a good primer on the post bayes algo theory this too states more on algorithmic information theory itself: states that it has notions of randomness, which were not found in shannons work. and stated above, fisher information : - still need to disseminate this, would need to be considered as paradigms of prior and posterior. the other wiki post on this has more of a atomic definition which is less like shannons :

the brain as constant predicting states of kalman filters. [1]

kalman interfaces for brain interfaces see :

also the use of bayesian networks [2]for neural nets also shows some form of information abstraction as a method of informational use.

good primer on GANs :

still need to cover overlap of computation and abstraction : see reference on fodor and dennett

hence why we use the notion of entropy, rather than information itself, as information has connotations, that have issues on the definitional aspect of memetics, information as a subset of entropy. information as also objective, but not the abstraction of said information is objective.