Guest guest Posted November 23, 2006 Report Share Posted November 23, 2006 Namaskarams to everyone. Just to give the discussion another dimension, I have posted this reply. There are few opinions or formulation I want to discuss here. 1. The assumption that Vedas are teaching of Ishwara to give a lift for the human thoughts is very naïve. As Shri.Aurobindo says in his book "The secret of the Vedas", Vedas are not merely teaching or specific, singular information that was conceived by human consciousness. Instead it is a form like music or Shruti that already exists and will exist ever, whether Homo-sapiens understands it or not. As regards how it came to human readable form for him to be able to express, it is the inspiration that led to it rather than an evolution of human brain. There are still some neuron or part in human brain (or method of process in human brain) which is as old or as antiquated as that of first life on earth. Inspiration is running of drops from the peak of mountain as a result of the order of the gravity of this planet. This is what Vedas has been through it is flow of the peak of drops of fluid from the peak of clarity, due to gravity in an human understandable form. Vedas are correlated to Goddess Saraswati, but Saraswati was also a river. There is a connection there. It is very raw on our part to come to conclusion that because the Saraswati river valley civilization led to the congregation of Vedas and Vedic rishis the correlation between Sarawathi and the Vedas. The point here for us to note is that the river saraswathi just flows from mountain going from the mountain valley to the plains and then to the desert. In the similar way, the Vedic scriptures are that of inspiration from the deeper (higher) consciousness of the human brain. It is a deeper ear to hear the already existing, flowing music shruti the Vedas. Vedas are considered to be shabdha Brahman. The cosmic music, that flows, which we call time, the yug, constricting its bandwidth only to the earth planet leaving others (space) devoid of it, is what is called the Vedas. The universe is made of that. To express more of what I am trying to communicate, I have of given the quote of T.S Eliot the great poet. Quoting T.S. Eliot: "For myself, I can only say that a knowledge of the springs which released a poem is not necessarily a help toward understanding the poem: too much information about the origins of a poem might even break my contact with it … I am even prepared to suggest that there is, in all great poetry, something which must remain unaccountable however complete might be our knowledge of the poet, and that is what matters most. When the poem has been made, something new has happened, something that cannot be wholly explained by anything that went before." James Shannon the U.S. scientist who was the first visionary of the Information Theory says about entropy for information. I have explained below the content of the entropy theory postulated by James Shannon. An intuitive understanding of information entropy relates to the amount of uncertainty about an event associated with a given probability distribution. As an example, consider a box containing many coloured balls. If the balls are all of different colours and no colour predominates, then our uncertainty about the colour of a randomly drawn ball is maximal. On the other hand, if the box contains more red balls than any other colour, then there is slightly less uncertainty about the result: the ball drawn from the box has more chances of being red (if we were forced to place a bet, we would bet on a red ball). Telling someone the colour of every new drawn ball provides them with more information in the first case than it does in the second case, because there is more uncertainty about what might happen in the first case than there is in the second. Intuitively, if there were no uncertainty as to the outcome, then we would learn nothing by drawing the next ball, and so the information content would be zero. As a result, the entropy of the "signal" (the sequence of balls drawn, as calculated from the probability distribution) is higher in the first case than in the second. Shannon, in fact, defined entropy as a measure of the average information content associated with a random outcome. It is important to remember that entropy is a quantity defined in the context of a probabilistic model for a data source. Independent fair coin flips have entropy of 1 bit per flip. A source that always generates a long string of A's has entropy of 0, since the next character will always be an 'A'. On how this is related to the Vedas and origin of human evolution is as below. The Vedas are not from the single source of human brain. It is an eclectic collection of inspirations from various brain structures. So in all probability the measure of average information content will always be new and entropy greater that zero with this associated random outcome of information. When the entropy is zero, the information continues to exist but in a rather unreadable form. No one can deny this fact if anyone accepts the sound or music could have been there without human being actually arriving at the planet for them to be conceived and understood. So it is not that it existed only after human arriving at the planet, it is there as light, sound, or science existed before and will exist after homo-sapiens. Quote Link to comment Share on other sites More sharing options...
Recommended Posts
Join the conversation
You are posting as a guest. If you have an account, sign in now to post with your account.
Note: Your post will require moderator approval before it will be visible.