h a l f b a k e r y
The halfway house for at-risk ideas

meta:

account: browse anonymously, or get an account and write.

 user: pass:
register,

# Information Rocket

Think of the information content as an object in orbit, which is function of mutation rate ("entropy field"), and replication rates ("centrifugal force field"), and be inspired to create an information rocket ;)
 (+2, -1) [vote for, against]

The law of gravity implies that a body of mass M within a field of gravity, is attracted by force F, dependent on the strength and direction of the gravitational field. The force F manifests as change of velocity delta-v, and change of coordinate (delta-d). If, however, there is tangential motion of the velocity s, it may compensate for the orbital fall delta-d, and so satellites don't fall from the sky.

Suppose that the mass of "information object" (e.g., bit string of length N) is defined as Kolmogorov complexity ("shortest computer program (in a predetermined programming language) that produces the object as output").

Suppose that "information object" is in the field of entropy (e.g., "random noise/temperature"), that is defined as "information loss", and its integral (just like the integral of deceleration) with respect to time, results in loss of information (in number of bits erased), in a similar way that object of matter in the field of gravity results in object loss of speed, and eventually altitude (number of meters).

Suppose that "information object" is in the field of random replication (e.g., field that engenders random copies of its substrings), that is defined as "information gain", and its integral (just like integral of acceleration) with respect to time, results in gain of information (in number of bits created), in a way similar to how object of matter with an engine of propulsion results in gain of speed, and eventually altitude (number of meters).

Suppose the source of entropy is a ball in a direction perpendicular to the arrow of time (just like often sources of gravity are balls in the directions perpendicular to orbital satellite motions). E.g., assume that we are traveling in a temporal curvature, i.e., if the 5 billion years ago looks temporarily behind, perhaps 10 billion years ago is actually not directly behind, but at an angle like satellites traveling in a circular or elliptic orbits are, because a planet is a ball- like structure.

Suppose the source of entropy is of a fixed density (i.e., just like fixed mass of a massive object, this would be fixed entropy, i.e., temperature or thing that mutates bits) at a distance D of imaginary time (i.e., time perpendicular to the tangent of our time flow), where entropy is inversely proportional to the square distance to its source. (Note -- think of imaginary time as expressed in the imaginary part of real numbers, i.e., the one perpendicular to our timeline, e.g., if the universe has existed for 13.7 billion years, this represents the distance covered in Real part of it, while time perpendicular to normal years is unknown, it may be different, depending on what happened to the complexity of the universe, -- i.e., if it became more random, then it approached the entropy source, if it became more ordered, it may have increased the distance from it.).

With these assumptions, begin plotting an information rocket, to capture the bits in our minds, and let them escape that entropy.

 — Mindey, Feb 12 2019

Centrifetal Centripugal_20Force
[MaxwellBuchanan, Feb 12 2019]

Infinity -- "Beating the Entropy" https://inf.li/#/en/@/topic/69/
The problem associated with the idea. [Mindey, Feb 15 2019]

Everything-List -- "On information Rockets..." https://groups.goog...ng-list/KRQXyndJv58
Thoughts on Everything-List in the search for rationale for the above. [Mindey, Feb 15 2019]

To illustrate anno by [Ian Tindale]. [Mindey, Feb 15 2019]

Yes, but have you accounted for Cole's Law?
 — MaxwellBuchanan, Feb 12 2019

 Umm, it shouldn't be information as mass.... Indeed, because information actually does not attract (or warm things up). Or.. maybe it does, if it interferes with other information...

To some extent, the distance between events in time-space, is the computational complexity between them.
 — Mindey, Feb 12 2019

 [Max], no, but I fixed the typo. :) Should it be "centripetal" force, rather than "centrifugal"?

So, will we have a prize for launching something like events (call them "processes") into timespace, and and having them fly like ... almost forever?
 — Mindey, Feb 12 2019

//Should it be "centripetal" force, rather than "centrifugal"// There's a link for that.
 — MaxwellBuchanan, Feb 12 2019

Murphy will have something to say about all of this.
 — RayfordSteele, Feb 13 2019

Suppose a duck is actually a walrus. [-]
 — Voice, Feb 13 2019

 [Voice] is not giving anything like Levenshtein distance a chance... Hmm.

 It's just like in that case, where you can actually define work in physics as an integral of force over time, not as an integral of force over distance: most people will have hard time to get it.

I'm putting this out as a half-baked idea. Baked one would be actually completely worked-out.
 — Mindey, Feb 13 2019

Information cannot be defined in terms of the physics definition of work. The other definition of work also cannot be defined in terms of the physics definition of work. You're saying "information is like energy and therefore work is like rockets." and again I answer: bollocks.
 — Voice, Feb 13 2019

 //Centrifetal

A novel use of centrifuges by midwives?
 — not_morrison_rm, Feb 13 2019

 I read it, and parts of it made me think of things.

It is a fun way to look at things, but I am confused. What if you make two of the entropy balls at a distance to the other making a gravityless laGrange point? That would seem to be a place to store data absent entropy. The thing is, that locating something between two entropy sources does not seem like it would preserve data. One possibility is that if the balls were constructed one bit out of synchronization with each other and then when a bit was pulled from 1 to 0 by one of the balls, the other ball would push it from 0 to 1. The thing is though then they wouldn't be entropy balls and it might only work on things that were 1 bit big, or however many bits fit on a surface between the two balls.
 — beanangel, Feb 13 2019

 //Suppose a duck is actually a walrus.//

... then we'll strive to get all our walruses in a row, which would be admirable. [+]
 — pertinax, Feb 14 2019

 [beanangel], indeed, as your prediction tells, we would have to have a laGrange point, where information would be safe, but this is not the case! Very very interesting.

 Naturally, for heat sources, an increased mutation rate is expected. However, if we think of distance, not as physical distance, but as complexity distance from one object to another (i.e., how much mutation do we need for an event of time-space to collide with another event (become one process), and how much mutation do we need for an event to disappear (i.e., a process to die, become randomness)), then we could imagine that the distance to the entropy source is the number of mutations needed for the object to become one with (i.e., in-sync with) entropy source.

 I guess, it depends on whether or not the entropy field has wave-like properties. Falling towards entropy balls actually does not mean becoming more random, but becoming more similar to (more in-sync with) these entropy sources.

Having entropy balls of opposite phase would make laGrange points exist, that is, if entropy, like gravity, is a wave-like phenomenon, and the waveform is dependent on the direction (between the events, i.e., time- space objects, which the entropy balls are too).
 — Mindey, Feb 14 2019

Regarding attraction between events... Do events attract? Some say they do. Do processes tend to make other processes more similar to them? Sometimes they do. For example, some events, like earthquakes, and criminal gangs, were found to cluster (as self-exciting point processes - Hawkes processes).
 — Mindey, Feb 14 2019

Are we imagining entropy balls as observable and, if so, how would you know if you were looking at one?
 — pertinax, Feb 14 2019

[pertinax], if we had eyes that can see events (processes) in time-space like we see objects in space, I guess we would see them as objects of high complexity, that makes other objects alike them (or in-sync with them).
 — Mindey, Feb 14 2019

So you mean we're nearby a neighboring computational universe with different laws of physics, and the aliens in it are attracting us to their kind of randomness?
 — Inyuki, Feb 14 2019

Yeah, e.g., why the future is more random than the past? We may be on a collision course with an alien universe, and we need a technological singularity to amplify our kind of randomness to escape their pull.
 — Mindey, Feb 14 2019

Like some sort of giant IBM Selectric golf ball.
 — Ian Tindale, Feb 14 2019

I think you should copy this to your https://inf.li site.
 — beanangel, Feb 15 2019

[beanangel] well, it's definitely matching with the problem of "Beating Entropy" (see link), but the https://inf.li doesn't work well enough, I want to rewrite it, cause now some features are confusing.
 — Mindey, Feb 15 2019

//we need a technological singularity to amplify our kind of randomness//Aha! Of course, the Infinite Improbability Drive is well worked out over the trilogy in five parts, and the many iterations of radio and film. It could work for this application. Or not. Too bad Douglas is dead; he'd like these crazy times in which we live.
 — Sgt Teacup, Feb 15 2019

 //If we had eyes that can see events (processes) in time-space//

 We *do* have eyes that can see events in time-space. The theory of relativity, IIRC, is founded on observable events in time- space. Unless "time-space" is somehow different from "space- time", in which case please explain the difference.

Processes, on the other hand, are things we infer from observed events. So, does "see events (processes)" mean "see events and infer processes from them" or something else? And, if something else, then what?
 — pertinax, Feb 15 2019

My simple two cents is, there is confusion between information (virtual definitions) and the media (the paper and ink, mass and energy movements that holds the information. It's the media that is subject to physical interactions.
 — wjt, Feb 16 2019

 [pertinax], I guess, there is a different possible ways to define the word "event":

 (A) - time slice of minimal interval, say of Plank time-interval (B) - an extrusion of a 3D object over time dimension

 When I am talking about events above, I mean the (B). For example, you can think of a piece of rock, that decays over time, becomes smaller, and then disappears. You can imagine that as a sweep extrusion of a 3D object over time. That's what I mean by an "event" (in B-sense). In the B-sense, every person is an event of an object (with some internal complexity), that spans over years, and then decays (e.g., dies and gets dismantled). The event in B-sense also coincides with what we think of as "process". A piece of rock actually is a process. A human being is a process. That's why I say "events (processes)". These events cannot be "seen" in the literal sense, - they can be experienced.

 > Processes, on the other hand, are things we infer from observed events. So, does "see events (processes)" mean "see events and infer processes from them" or something else? And, if something else, then what?

 Since you here use the event in the (A)-sense (i.e., atomic events), yes, it means seeing events and inferring processes of high complexity, -- for example, hot things, randomly jumpy things, like the "IBM Selectric golf ball" (see link), pointed out by [Ian Tindale].

 For hot balls, you can measure ("observe") their temperature, which may seem like instantaneous measurement, but we shouldn't forget that the measurement of temperature is a statistic of multiple data inputs by the particles, so, it's not instantaneous, and you'd see no temperature in an interval of Plank time. To observe the temperature we watch the changes in one measurement with respect to another, with the assumption that they are related. It is impossible with just a single measurement of a Plank-time event.

To observe the entropy ball, one would have to create two identical probes as different events, e.g., -- on different times (e.g., two identical rocks that lived on same place on different centuries). E.g., one probe 20 years ago, and another probe 10 years from now. Then, measure the discrepancy of their decay rates. To point its direction, create more of these probes on different locations, triangulate that to the general time-space direction. The closest I can think of what's been done, is the Kilogram experiment, however, the samples were manufactured at the same time. The identical probes would have to be manufactured at different times. The timescales and the kind of object sensitivity (is it a rock, or a clock, a cpu, or a time crystal) for entropy types would have to be chosen depending on the kind (e.g., "wavelength") of entropy we want to measure (e.g., very-short term, brownian motion, or a longer term random flushing of sea-shores, or some inherent quantum-entropy of space).
 — Mindey, Feb 16 2019

Sorry, can I change my annotation? What I meant to say is, the information in those scattered chicken bones is looking clearer and clearer.
 — wjt, Feb 16 2019

 So... if I've got this straight, (no bets on that mind you), given the many worlds interpretation every 'instant' of every possibility is a slide in an infinite slide show and so it should be possible to... encapsulate information in entropy-less bubbles to escape being lost to the flow of time?

Is that right?
 — 2 fries shy of a happy meal, Feb 16 2019

[2 fries shy of a happy meal], I suppose so, and these entropy-less bubbles might be the laGrange points between the universes in orbit around each others.
 — Mindey, Feb 16 2019

 Well it would make a decent premise for a sci fi novel.

Thing is though, if true, there can be no loss of information if only consciousness is trapped by the flow of time. The overall multiverse slide show remains static and is itself an entropy-less bubble. Maybe one of many.
 — 2 fries shy of a happy meal, Feb 16 2019

 // I want to rewrite it, cause now some features are confusing. //

And because, currently, if you try to load a page without having already allowed all of the scripts it needs, it destroys its own URL, so you have to go back to the page where you clicked the link to it and click the link again after unblocking everything.
 — notexactly, Feb 16 2019

 So, whatever information source actually can be thought of entropy source, following a certain probability distribution, and it's attraction may be thought of a distributional drag (any information source within another information source may be affect to drag to acquire the pattern of that other source).

The Lagrange points may exist, when there is pseudo-random pattern interferences, or a stable pattern-engendering probability-distributional drags.
 — Mindey, Mar 01 2019

 [annotate]

back: main index