Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Collapse
Code Project
  1. Home
  2. The Lounge
  3. I live the chaos, but I don't understand it :(

I live the chaos, but I don't understand it :(

Scheduled Pinned Locked Moved The Lounge
commcphelpquestion
33 Posts 10 Posters 0 Views 1 Watching
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • 0 0x01AA

    I have to use up the unused vacation for 2023. I usually use this as an opportunity to do a little further training. At the end of this year 2023, I have decided to finally understand the term entropy in information technology to the end. But I fail once again:( All I can confirm so far is the statement of 'John von Neumann' (John von Neumann - Wikipedia[^]) to Claude Shannon - Wikipedia[^] : “You should call it entropy. […] Nobody knows what entropy really is, so in a debate “You will always have the advantage.” :sigh: :laugh: [Edit] Btw. any idea which forum here is ok to ask questions about that theme? [Edit1] Now I think it becomes out of 'lounge'disussion, sorry for that. Still I'm looking for a place where one can disuss that. I'm pretty shure that @KornfeldEliyahuPeter can help me of this ;)

    Mike HankeyM Offline
    Mike HankeyM Offline
    Mike Hankey
    wrote on last edited by
    #2

    0x01AA wrote:

    entropy

    An Entro to Py

    As the aircraft designer said, "Simplicate and add lightness". PartsBin an Electronics Part Organizer - Release Version 1.3.0 JaxCoder.com Latest Article: SimpleWizardUpdate

    1 Reply Last reply
    0
    • 0 0x01AA

      I have to use up the unused vacation for 2023. I usually use this as an opportunity to do a little further training. At the end of this year 2023, I have decided to finally understand the term entropy in information technology to the end. But I fail once again:( All I can confirm so far is the statement of 'John von Neumann' (John von Neumann - Wikipedia[^]) to Claude Shannon - Wikipedia[^] : “You should call it entropy. […] Nobody knows what entropy really is, so in a debate “You will always have the advantage.” :sigh: :laugh: [Edit] Btw. any idea which forum here is ok to ask questions about that theme? [Edit1] Now I think it becomes out of 'lounge'disussion, sorry for that. Still I'm looking for a place where one can disuss that. I'm pretty shure that @KornfeldEliyahuPeter can help me of this ;)

      G Offline
      G Offline
      Gary Stachelski 2021
      wrote on last edited by
      #3

      Here is a link to a youtube video on Shannon Information Entropy. See if this makes more sense. Information entropy | Journey into information theory | Computer Science | Khan Academy - YouTube[^]

      0 1 Reply Last reply
      0
      • 0 0x01AA

        I have to use up the unused vacation for 2023. I usually use this as an opportunity to do a little further training. At the end of this year 2023, I have decided to finally understand the term entropy in information technology to the end. But I fail once again:( All I can confirm so far is the statement of 'John von Neumann' (John von Neumann - Wikipedia[^]) to Claude Shannon - Wikipedia[^] : “You should call it entropy. […] Nobody knows what entropy really is, so in a debate “You will always have the advantage.” :sigh: :laugh: [Edit] Btw. any idea which forum here is ok to ask questions about that theme? [Edit1] Now I think it becomes out of 'lounge'disussion, sorry for that. Still I'm looking for a place where one can disuss that. I'm pretty shure that @KornfeldEliyahuPeter can help me of this ;)

        J Offline
        J Offline
        Jo_vb net
        wrote on last edited by
        #4

        Learning this could be a first step https://machinelearningmastery.com/what-is-information-entropy/[^]

        0 1 Reply Last reply
        0
        • G Gary Stachelski 2021

          Here is a link to a youtube video on Shannon Information Entropy. See if this makes more sense. Information entropy | Journey into information theory | Computer Science | Khan Academy - YouTube[^]

          0 Offline
          0 Offline
          0x01AA
          wrote on last edited by
          #5

          Thanks for this. My usual problem is they state 'the probability is this and that' but no explanation why it is ;)

          G 1 Reply Last reply
          0
          • J Jo_vb net

            Learning this could be a first step https://machinelearningmastery.com/what-is-information-entropy/[^]

            0 Offline
            0 Offline
            0x01AA
            wrote on last edited by
            #6

            Thank you for this. I read a lot, but I'm stumbling again and again. E.g: from here https://www.physik.uni-wuerzburg.de/fileadmin/11030300/_imported/fileadmin/tp3/ThermoEDynamik/Entropie.pdf[^]

            Quote:

            Da man mit n Bits bekanntlich 2^n verschiedene Bitmuster bilden kann, ist sofort klar, dass ein System mit 2^n Zuständen mit einer n-Bit-Datei vollständig beschrieben werden kann, so dass in diesem Fall H=n is

            For me nothing is clear... :( Especally that one can describe e.g. the 2^8 conditions in an 8 bit file. But most probably this is a lack of my understanding

            J 1 Reply Last reply
            0
            • 0 0x01AA

              I have to use up the unused vacation for 2023. I usually use this as an opportunity to do a little further training. At the end of this year 2023, I have decided to finally understand the term entropy in information technology to the end. But I fail once again:( All I can confirm so far is the statement of 'John von Neumann' (John von Neumann - Wikipedia[^]) to Claude Shannon - Wikipedia[^] : “You should call it entropy. […] Nobody knows what entropy really is, so in a debate “You will always have the advantage.” :sigh: :laugh: [Edit] Btw. any idea which forum here is ok to ask questions about that theme? [Edit1] Now I think it becomes out of 'lounge'disussion, sorry for that. Still I'm looking for a place where one can disuss that. I'm pretty shure that @KornfeldEliyahuPeter can help me of this ;)

              A Offline
              A Offline
              Amarnath S
              wrote on last edited by
              #7

              0x01AA wrote:

              Nobody knows what entropy really is

              As a Mechanical Engineer by education, I can only say that Second Law of Thermodynamics dictates that the entropy of the Universe is continuously increasing, dS > 0.

              0 1 Reply Last reply
              0
              • A Amarnath S

                0x01AA wrote:

                Nobody knows what entropy really is

                As a Mechanical Engineer by education, I can only say that Second Law of Thermodynamics dictates that the entropy of the Universe is continuously increasing, dS > 0.

                0 Offline
                0 Offline
                0x01AA
                wrote on last edited by
                #8

                But there is a limit to this? Something like 2.58###

                1 Reply Last reply
                0
                • 0 0x01AA

                  I have to use up the unused vacation for 2023. I usually use this as an opportunity to do a little further training. At the end of this year 2023, I have decided to finally understand the term entropy in information technology to the end. But I fail once again:( All I can confirm so far is the statement of 'John von Neumann' (John von Neumann - Wikipedia[^]) to Claude Shannon - Wikipedia[^] : “You should call it entropy. […] Nobody knows what entropy really is, so in a debate “You will always have the advantage.” :sigh: :laugh: [Edit] Btw. any idea which forum here is ok to ask questions about that theme? [Edit1] Now I think it becomes out of 'lounge'disussion, sorry for that. Still I'm looking for a place where one can disuss that. I'm pretty shure that @KornfeldEliyahuPeter can help me of this ;)

                  J Offline
                  J Offline
                  jschell
                  wrote on last edited by
                  #9

                  As a suggestion attempting to learn one part of a system of study well, probably is always going to require that one learns more about the system that contains it first. So study information theory first.

                  0 1 Reply Last reply
                  0
                  • J jschell

                    As a suggestion attempting to learn one part of a system of study well, probably is always going to require that one learns more about the system that contains it first. So study information theory first.

                    0 Offline
                    0 Offline
                    0x01AA
                    wrote on last edited by
                    #10

                    Quote:

                    So study information theory first.

                    I would say 'Shannon's' therorie about Entropy is exactly that very basic theory ;)

                    1 Reply Last reply
                    0
                    • 0 0x01AA

                      Thank you for this. I read a lot, but I'm stumbling again and again. E.g: from here https://www.physik.uni-wuerzburg.de/fileadmin/11030300/_imported/fileadmin/tp3/ThermoEDynamik/Entropie.pdf[^]

                      Quote:

                      Da man mit n Bits bekanntlich 2^n verschiedene Bitmuster bilden kann, ist sofort klar, dass ein System mit 2^n Zuständen mit einer n-Bit-Datei vollständig beschrieben werden kann, so dass in diesem Fall H=n is

                      For me nothing is clear... :( Especally that one can describe e.g. the 2^8 conditions in an 8 bit file. But most probably this is a lack of my understanding

                      J Offline
                      J Offline
                      Jo_vb net
                      wrote on last edited by
                      #11

                      I'm sure you understand it (but you do not know that)! Wortbreite 8 Bit Bereiche mit Vorzeichen -128 ~ 127 Bereich ohne Vorzeichen 0 ~ 255 2^8 = 256 and 1 to 256 you can define with 8 bit as shown above with 0 to 255. Now you should know that you already knew that :-D

                      0 1 Reply Last reply
                      0
                      • J Jo_vb net

                        I'm sure you understand it (but you do not know that)! Wortbreite 8 Bit Bereiche mit Vorzeichen -128 ~ 127 Bereich ohne Vorzeichen 0 ~ 255 2^8 = 256 and 1 to 256 you can define with 8 bit as shown above with 0 to 255. Now you should know that you already knew that :-D

                        0 Offline
                        0 Offline
                        0x01AA
                        wrote on last edited by
                        #12

                        I think , I'm very save with binary numbers. But again translated by google, I don't get the point on that:

                        Quote:

                        Since it is well known that 2^n different bit patterns can be formed with n bits, it is immediately clear that a system with 2^n states can be completely described with an n-bit file, so that in this case H=n

                        How one can desribe 256 condistions in a file with one byte? I'm pretty sure I have a problem understanding the article, but I'm also very happy if somebody can explain what I'm missinterpreting ;)

                        J B 2 Replies Last reply
                        0
                        • 0 0x01AA

                          I have to use up the unused vacation for 2023. I usually use this as an opportunity to do a little further training. At the end of this year 2023, I have decided to finally understand the term entropy in information technology to the end. But I fail once again:( All I can confirm so far is the statement of 'John von Neumann' (John von Neumann - Wikipedia[^]) to Claude Shannon - Wikipedia[^] : “You should call it entropy. […] Nobody knows what entropy really is, so in a debate “You will always have the advantage.” :sigh: :laugh: [Edit] Btw. any idea which forum here is ok to ask questions about that theme? [Edit1] Now I think it becomes out of 'lounge'disussion, sorry for that. Still I'm looking for a place where one can disuss that. I'm pretty shure that @KornfeldEliyahuPeter can help me of this ;)

                          B Offline
                          B Offline
                          BobbyStrain
                          wrote on last edited by
                          #13

                          There is a simple equation that defines entropy. Chemical Engineers make use of the term to describe the behavior of substances. We use it, for example, to evaluate the performance of a steam turbine. It has been misused by zealous promoters to obfuscate information.

                          0 J 2 Replies Last reply
                          0
                          • B BobbyStrain

                            There is a simple equation that defines entropy. Chemical Engineers make use of the term to describe the behavior of substances. We use it, for example, to evaluate the performance of a steam turbine. It has been misused by zealous promoters to obfuscate information.

                            0 Offline
                            0 Offline
                            0x01AA
                            wrote on last edited by
                            #14

                            And that simple equation is?

                            B 1 Reply Last reply
                            0
                            • 0 0x01AA

                              I think , I'm very save with binary numbers. But again translated by google, I don't get the point on that:

                              Quote:

                              Since it is well known that 2^n different bit patterns can be formed with n bits, it is immediately clear that a system with 2^n states can be completely described with an n-bit file, so that in this case H=n

                              How one can desribe 256 condistions in a file with one byte? I'm pretty sure I have a problem understanding the article, but I'm also very happy if somebody can explain what I'm missinterpreting ;)

                              J Offline
                              J Offline
                              Jo_vb net
                              wrote on last edited by
                              #15

                              A file with one byte can describe how many states a system has (state value 1 or 2 or ... 256). If you want describe 256 states you need 256 * 8 bit.

                              1 Reply Last reply
                              0
                              • 0 0x01AA

                                I have to use up the unused vacation for 2023. I usually use this as an opportunity to do a little further training. At the end of this year 2023, I have decided to finally understand the term entropy in information technology to the end. But I fail once again:( All I can confirm so far is the statement of 'John von Neumann' (John von Neumann - Wikipedia[^]) to Claude Shannon - Wikipedia[^] : “You should call it entropy. […] Nobody knows what entropy really is, so in a debate “You will always have the advantage.” :sigh: :laugh: [Edit] Btw. any idea which forum here is ok to ask questions about that theme? [Edit1] Now I think it becomes out of 'lounge'disussion, sorry for that. Still I'm looking for a place where one can disuss that. I'm pretty shure that @KornfeldEliyahuPeter can help me of this ;)

                                0 Offline
                                0 Offline
                                0x01AA
                                wrote on last edited by
                                #16

                                Now I think it becomes out of 'lounge'disussion, sorry for that. Still I'm looking for a place where one can disuss that. I'm pretty shure that @KornfeldEliyahuPeter can help me of this ;)

                                1 Reply Last reply
                                0
                                • 0 0x01AA

                                  Thanks for this. My usual problem is they state 'the probability is this and that' but no explanation why it is ;)

                                  G Offline
                                  G Offline
                                  Gary Stachelski 2021
                                  wrote on last edited by
                                  #17

                                  In the case of the video, the probabilities were made up to create two different machines. They then went on to illustrate how you could calculate the information entropy for each machine to compare them. The machine with the lower calculated entropy was more predictable Or better organized. The higher the calculated entropy the more disorganized a system is, so it is harder to predict what the output of the next cycle will be.

                                  J 1 Reply Last reply
                                  0
                                  • B BobbyStrain

                                    There is a simple equation that defines entropy. Chemical Engineers make use of the term to describe the behavior of substances. We use it, for example, to evaluate the performance of a steam turbine. It has been misused by zealous promoters to obfuscate information.

                                    J Offline
                                    J Offline
                                    jschell
                                    wrote on last edited by
                                    #18

                                    Been a while but I don't think that is the same thing as entropy in information theory. But perhaps are related.

                                    1 Reply Last reply
                                    0
                                    • 0 0x01AA

                                      And that simple equation is?

                                      B Offline
                                      B Offline
                                      BobbyStrain
                                      wrote on last edited by
                                      #19

                                      [4.6 Entropy - University Physics Volume 2 | OpenStax](https://openstax.org/books/university-physics-volume-2/pages/4-6-entropy)

                                      0 1 Reply Last reply
                                      0
                                      • 0 0x01AA

                                        I think , I'm very save with binary numbers. But again translated by google, I don't get the point on that:

                                        Quote:

                                        Since it is well known that 2^n different bit patterns can be formed with n bits, it is immediately clear that a system with 2^n states can be completely described with an n-bit file, so that in this case H=n

                                        How one can desribe 256 condistions in a file with one byte? I'm pretty sure I have a problem understanding the article, but I'm also very happy if somebody can explain what I'm missinterpreting ;)

                                        B Offline
                                        B Offline
                                        BernardIE5317
                                        wrote on last edited by
                                        #20

                                        my understanding is that a system at any particular moment is in a state . a single state . of course a system can not be in more than one state at any moment unless of course we are discussing Quantum Mechanics which of course we are not i presume . if it is known the system can be in any state of 256 possible states then at any moment only 8 bits are required to specify that state . QED

                                        0 J 2 Replies Last reply
                                        0
                                        • 0 0x01AA

                                          I have to use up the unused vacation for 2023. I usually use this as an opportunity to do a little further training. At the end of this year 2023, I have decided to finally understand the term entropy in information technology to the end. But I fail once again:( All I can confirm so far is the statement of 'John von Neumann' (John von Neumann - Wikipedia[^]) to Claude Shannon - Wikipedia[^] : “You should call it entropy. […] Nobody knows what entropy really is, so in a debate “You will always have the advantage.” :sigh: :laugh: [Edit] Btw. any idea which forum here is ok to ask questions about that theme? [Edit1] Now I think it becomes out of 'lounge'disussion, sorry for that. Still I'm looking for a place where one can disuss that. I'm pretty shure that @KornfeldEliyahuPeter can help me of this ;)

                                          B Offline
                                          B Offline
                                          BernardIE5317
                                          wrote on last edited by
                                          #21

                                          video below gives a nice explanation of entropy in information theory . Information entropy | Journey into information theory | Computer Science | Khan Academy - YouTube[^]

                                          0 1 Reply Last reply
                                          0
                                          Reply
                                          • Reply as topic
                                          Log in to reply
                                          • Oldest to Newest
                                          • Newest to Oldest
                                          • Most Votes


                                          • Login

                                          • Don't have an account? Register

                                          • Login or register to search.
                                          • First post
                                            Last post
                                          0
                                          • Categories
                                          • Recent
                                          • Tags
                                          • Popular
                                          • World
                                          • Users
                                          • Groups