Toggle light / dark theme

To achieve interstellar travel, the Kline Directive instructs us to be bold, to explore what others have not, to seek what others will not, to change what others dare not. To extend the boundaries of our knowledge, to advocate new methods, techniques and research, to sponsor change not status quo, on 5 fronts, Legal Standing, Safety Awareness, Economic Viability, Theoretical-Empirical Relationships, and Technological Feasibility.

In this post I will explore Technological Feasibility. At the end of the day that is the only thing that matters. If a hypothesis is not able to vindicate itself with empirical evidence it will not become technologically feasible. If it is not technologically feasible then it stands no chance of becoming commercially viable.

If we examine historical land, air and space speed records, we can construct and estimate of velocities that future technologies can achieve, aka technology forecasting. See table below for some of the speed records.

Year Fastest Velocity Craft Velocity (km/h) Velocity (m/s)
2006 Escape Earth New Horizons 57,600 16,000
1976 Capt. Eldon W. Joersz and Maj. George T. Morgan Lockheed SR-71 Blackbird 3,530 980
1927 Car land speed record (not jet engine) Mystry 328 91
1920 Joseph Sadi-Lecointe Nieuport-Delage NiD 29 275 76
1913 Maurice Prévost Deperdussin Monocoque 180 50
1903 Wilbur Wright at Kitty Hawk Wright Aircraft 11 3

A quick and dirty model derived from the data shows that we could achieve velocity of light c by 2151 or the late 2150s. See table below.

Year Velocity (m/s) % of c
2200 8,419,759,324 2808.5%
2152 314,296,410 104.8%
2150 274,057,112 91.4%
2125 49,443,793 16.5%
2118 30,610,299 10.2%
2111 18,950,618 6.3%
2100 8,920,362 3.0%
2075 1,609,360 0.5%
2050 290,351 0.1%
2025 52,384 0.0%

The extrapolation suggests that on our current rate of technological innovation we won’t achieve light speed until the late 2150s. The real problem is that we won’t achieve 0.1c until 2118! This is more than 100-years from today.

In my opinion this rate of innovation is too slow. Dr. David Neyland, of DARPA’s TTO was the driving force behind DARPA’s contribution to the 100-year Starship Study. When I met up with Dr. David Neyland during the first 100YSS conference, Sept. 30 to Oct 2, 2011, I told him “I’ll drop a zero”. That is I expect interstellar travel to be achievable in decades not centuries. And to ramp up our rate of technological innovation we need new theories and new methods of sifting through theories.

Previous post in the Kline Directive series.

—————————————————————————————————

Benjamin T Solomon is the author & principal investigator of the 12-year study into the theoretical & technological feasibility of gravitation modification, titled An Introduction to Gravity Modification, to achieve interstellar travel in our lifetimes. For more information visit iSETI LLC, Interstellar Space Exploration Technology Initiative.

Solomon is inviting all serious participants to his LinkedIn Group Interstellar Travel & Gravity Modification.

To achieve interstellar travel, the Kline Directive instructs us to be bold, to explore what others have not, to seek what others will not, to change what others dare not. To extend the boundaries of our knowledge, to advocate new methods, techniques and research, to sponsor change not status quo, on 5 fronts, Legal Standing, Safety Awareness, Economic Viability, Theoretical-Empirical Relationships, and Technological Feasibility.

In this post I discuss the third and final part, Concepts and Logical Flow, of how to read or write a journal paper, that is not taught in colleges.

A paper consists of a series of evolving concepts expressed as paragraphs. If a concept is too complex to be detailed in a single paragraph, then break it down into several sub-concept paragraphs. Make sure there is logical evolution of thought across these sub-concepts, and across the paper.

As a general rule your sentences should be short(er). Try very hard not to exceed two lines of Letter or A4 size paper at font size 11. Use commas judicially. Commas are not meant to extend sentences or divide the sentence into several points!!! They are used to break up a sentence into sub-sentences to indicate a pause when reading aloud. How you use commas can alter the meaning of a sentence. Here is an example.

And this I know with confidence, I remain and continue …

Changing the position of the commas, changes the meaning to

And this I know, with confidence I remain and continue …

We see how ‘confidence’ changes from the speaker’s assessment of his state of knowledge, to the speaker’s reason for being. So take care.

When including mathematical formulae, always wrap. Wrap them with an opening paragraph and a closing paragraph. Why? This enhances the clarity of the paper. The opening paragraph introduce the salient features of the equation(s), i.e. what the reader needs to be aware of in the equation(s), or an explanation of the symbols, or why the equation is being introduced.

The closing paragraph explains what the author found by stating the equations, and what the reader should expect to look for in subsequent discussions, or even why the equation(s) is or is not relevant to subsequent discussions.

Many of these concept-paragraphs are logically combined into sections, and each section has a purpose for its inclusion. Though this purpose may not always be stated in the section, it is important to identify what it is and why it fits in with the overall schema of the paper.

The basic schema of a paper consists of an introduction, body and conclusion. Of course there are variations to this basic schema, and you need to ask the question, why does the author include other types of sections.

In the introduction section(s) you summarize your case, what is your paper about, and what others have reported. In the body sections you present your work. In the conclusion section you summarize your findings and future direction of the research. Why? Because a busy researcher can read your introduction and conclusion and then decide whether your paper is relevant to his or her work. Remember we are working within a community of researchers in an asynchronous manner, an asynchronous team, if you would. As more and more papers are published every year, we don’t have the time to read all of them, completely. So we need a method of eliminating papers we are not going to read.

An abstract is usually a summary of the body of the paper. It is difficult to do well and should only be written after you have completed your paper. That means you are planning ahead and have your paper written and abstracts completed when you receive the call for papers.

An abstract tells us if the paper could be relevant, to include in our list of papers to be considered for the shortlist of papers to be read. The introduction and conclusion tells if the paper should be removed from our short list. If the conclusion fits in with what we want to achieve, then don’t remove the paper from the short list.

I follow a rule when writing the introduction section. If I am writing to add to the body of consensus, I state my case and then write a review of what others have reported. If I am negating the body of consensus, then I write a review of what others have reported, and then only do I state my case of why not.

As a general rule, you write several iterations of the body first, then introduction and finally the conclusion. You’d be surprised by how your thinking changes if you do it this way, This is because you have left yourself open to other inferences that had not crossed your mind from the time you completed your work, to the time you started writing your paper.

If someone else has theoretical or experimental results that apparently contradicts your thesis, then discuss why and why not, and you might end up changing your mind. It is not a ‘sin’ to include contradictory results, but make sure you discuss this intelligently and impartially.

Your work is the sowing and growing period. Writing the paper is the harvesting period. What are you harvesting? Wheat, weeds or both? Clearly the more wheat you harvest the better your paper. The first test for this is the logical flow of your paper. If it does not flow very well, something is amiss! You the author, and you the reader beware! There is no substitute but to rethink your paper.

The second test is, if you have tangential discussions in your paper that seem interesting but are not directly relevant. Prune, prune & prune. If necessary split into multiple concise papers. A concise & sharp paper that everyone remembers is more valuable than a long one that you have to plough through.

Go forth, read well and write more.

Previous post in the Kline Directive series.

Next post in the Kline Directive series.

—————————————————————————————————

Benjamin T Solomon is the author & principal investigator of the 12-year study into the theoretical & technological feasibility of gravitation modification, titled An Introduction to Gravity Modification, to achieve interstellar travel in our lifetimes. For more information visit iSETI LLC, Interstellar Space Exploration Technology Initiative.

Solomon is inviting all serious participants to his LinkedIn Group Interstellar Travel & Gravity Modification.

To achieve interstellar travel, the Kline Directive instructs us to be bold, to explore what others have not, to seek what others will not, to change what others dare not. To extend the boundaries of our knowledge, to advocate new methods, techniques and research, to sponsor change not status quo, on 5 fronts, Legal Standing, Safety Awareness, Economic Viability, Theoretical-Empirical Relationships, and Technological Feasibility.

In this post I discuss part 2 of 3, Mathematical Construction versus Mathematical Conjecture, of how to read or write a journal paper that is not taught in colleges.

I did my Master of Arts in Operations Research (OR) at the best OR school in the United Kingdom, University of Lancaster, in the 1980s. We were always reminded that models have limits to their use. There is an operating range within which a model will provide good and reliable results. But outside that operating range, a model will provide unreliable, incorrect and even strange results.

Doesn’t that sound a lot like what the late Prof. Morris Kline was saying? We can extrapolate this further, and ask our community of theoretical physicists the question, what is the operating range of your theoretical model? We can turn the question around and require our community of theoretical physicists to inform us or suggest boundaries of where their models fail “ … to provide reasonability in guidance and correctness in answers to our questions in the sciences …”

A theoretical physics model is a mathematical construction that is not necessarily connected to the real world until it is empirically verified or falsified, until then these mathematical constructions are in limbo. Search the term ‘retrocausality’ for example. The Wikipedia article Retrocausality says a lot about how and why of the origins of theoretical physics models that are not within the range of our informed common sense. Let me quote,

“The Wheeler–Feynman absorber theory, proposed by John Archibald Wheeler and Richard Feynman, uses retrocausality and a temporal form of destructive interference to explain the absence of a type of converging concentric wave suggested by certain solutions to Maxwell’s equations. These advanced waves don’t have anything to do with cause and effect, they are just a different mathematical way to describe normal waves. The reason they were proposed is so that a charged particle would not have to act on itself, which, in normal classical electromagnetism leads to an infinite self-force.”

John Archibald Wheeler and Richard Feynman are giants in the physics community, and these esteemed physicists used retrocausality to solve a mathematical construction problem. Could they not have asked the different questions? What is the operating range of this model? How do we rethink this model so as not to require retrocausality?

This unfortunate leadership in retrocausality has led to a whole body of ‘knowledge’ by the name of ‘retrocausality’ that is in a state of empirical limbo and thus, the term mathematical conjecture applies.

Now, do you get an idea of how mathematical construction leads to mathematical conjecture? Someone wants to solve a problem, which is a legitimate quest because that is how science progresses, but the solution causes more problems (not questions) than previously, which leads to more physicists trying to answer those new problems, and so forth .… and so forth .… and so forth .…

In Hong Kong, the Cantonese have an expression “chasing the dragon”.

Disclaimer: I am originally from that part of the world, and enjoyed tremendously watching how the Indian and Chinese cultures collided, merged, and separated, repeatedly. Sometimes like water and oil, and sometimes like water and alcohol. These two nations share a common heritage, the Buddhist monks, and if they could put aside their nationalistic and cultural pride, who knows what could happen?

Chasing the dragon in the Chinese cultural context “refers to inhaling the vapor from heated morphine, heroin, oxycodone or opium that has been placed on a piece of foil. The ‘chasing’ occurs as the user gingerly keeps the liquid moving in order to keep it from coalescing into a single, unmanageable mass. Another more metaphorical use of the term ‘chasing the dragon’ refers to the elusive pursuit of the ultimate high in the usage of some particular drug.”

Solving a mathematical equation always gives a high, and discovering a new equation gives a greater high. So when we write a paper, we have to ask ourselves, are we chasing the dragon of mathematical conjecture or chasing the dragon of mathematical construction? I hope it is the latter.

Previous post in the Kline Directive series.

—————————————————————————————————

Benjamin T Solomon is the author & principal investigator of the 12-year study into the theoretical & technological feasibility of gravitation modification, titled An Introduction to Gravity Modification, to achieve interstellar travel in our lifetimes. For more information visit iSETI LLC, Interstellar Space Exploration Technology Initiative.

Solomon is inviting all serious participants to his LinkedIn Group Interstellar Travel & Gravity Modification.

To achieve interstellar travel, the Kline Directive instructs us to be bold, to explore what others have not, to seek what others will not, to change what others dare not. To extend the boundaries of our knowledge, to advocate new methods, techniques and research, to sponsor change not status quo, on 5 fronts, Legal Standing, Safety Awareness, Economic Viability, Theoretical-Empirical Relationships, and Technological Feasibility.

I was not intending to write Part 5, but judging from the responses I thought it was necessary to explain how to read a journal paper – and a good read cannot be done without a pen and paper. If you are writing a paper, when you have completed it, I would suggest you set it aside for at least a week. Don’t think about your paper or the topic during this shmita period. Then come back to your paper with a pen & paper and read it afresh. You’d be surprised by the number of changes you make, which means you have to start well before your deadline.

Note, you can find articles on how to review or write papers and here is one, by IOP (Institute of Physics, UK) titled Introduction to refereeing, and is a good guide to read before reading or writing a paper. This is especially true for physics but applies to all the sciences and engineering disciplines.

Note, for those who have been following the comments on my blog posts, IOP explicitly states “Do not just say ‘This result is wrong’ but say why it is wrong…” and “be professional and polite in your report”. So I hope, we as commentators, will be more professional in both our comments and the focus of our comments. Thanks.

In this post I will address what is not taught in colleges. There are three things to look out for when reading or writing a paper, Explicit and Implicit Axioms, Mathematical Construction versus Mathematical Conjecture, and finally, Concepts and Logical Flow. In this first part I discuss Explicit and Implicit Axioms.

This may sound silly but 1+1 = 2 is not an axiom. Alfred North Whitehead and Bertrand Russell proved that 1+1 adds to 2. Therefore, we see, that the immense success of the modern civilization compared to all other previous civilizations, is due to the encroachment of the imperceptible mathematical rigor in our daily lives by nameless, faceless scientist, engineers and technicians. Now that is something to ponder about. If we lose that rigor we lose our society. We can discuss economic and political theory but without this mathematical rigor, nothing else works.

Any theoretical work is based on axioms. For example in Pythagorean geometry, one assumes that surfaces are flat in such a manner the sum of the angles of a triangle adds to 180º. In Riemann geometry this is not the case. Explicit axioms are those stated in the paper.

Implicit axioms are axioms that are taken for granted to be true and therefore not stated, or considered too trivial to be mentioned. More often than not, the author is not aware he or she is using or stating an implicit axiom.

For example, mass causes a gravitational field is an implicit axiom, as we cannot with our current theoretical foundations nor with our current technologies prove either way that mass is or is not the source of a gravitational field. This axiom is also considered trivial because what else could?

But wait, didn’t Einstein … ? Yes correct, he did .…

Mass is a carryover from Newton. It shows how difficult it is to break from tradition even when we are breaking from tradition! Since Newton figured out that mass was an excellent means (i.e. “proxy” to be technically rigorous) to determining gravitational acceleration in mathematical form, therefore mass had to be the source. All tests pertaining to Einstein’s relativity test the field characteristics, not how the source creates the gravitational field.

But our understanding of the world has changed substantially since both Newton and Einstein. We know that quarks are at the center of matter and exist in the same ‘amount’ as mass. So how does one tell the difference between quark interaction and mass as the gravitational source?

The importance of implicit axioms in particular and axioms in general, is that when we recognize them we can change them and drive fundamental changes in theory and technologies. I asked the questions, what is gravity modification and how can we do it? These questions are at best vague, but they were as good a starting point as any? But life happens backwards. We get the answer and then only do we recognize the precise question we were attempting to ask!

When I started researching gravity modification in 1999, I just had this sense that gravity modification should be possible in our lifetimes, but I did not know what the question was. It was all vague and unclear at that time, but I was very strict about the scope of my investigation. I would only deal with velocity and acceleration.

I spent 8 years searching, examining, discarding, testing and theorizing anomalies, trying to get a handle on what gravity modification could be. Finally in 2007 I started building numerical models of how gravitational acceleration could work in spacetime. In February 2008 I discovered g=τc2 and at that moment I knew the question: Can gravitational acceleration be described mathematically without knowing the mass of the planet or star?

So the implicit axiom, mass is required for gravitational acceleration, is no longer valid, and because of that we now have propulsion physics.

If, in the spirit of the Kline Directive, you want to explore what others have not, and seek what others will not, my advice is that when you read a paper ask yourself, what are the implicit and explicit axioms in the paper?

Previous post in the Kline Directive series.

Next post in the Kline Directive series.

—————————————————————————————————

Benjamin T Solomon is the author & principal investigator of the 12-year study into the theoretical & technological feasibility of gravitation modification, titled An Introduction to Gravity Modification, to achieve interstellar travel in our lifetimes. For more information visit iSETI LLC, Interstellar Space Exploration Technology Initiative.

Solomon is inviting all serious participants to his LinkedIn Group Interstellar Travel & Gravity Modification.


…here’s Tom with the Weather.
That right there is comedian/philosopher Bill Hicks, sadly no longer with us. One imagines he would be pleased and completely unsurprised to learn that serious scientific minds are considering and actually finding support for the theory that our reality could be a kind of simulation. That means, for example, a string of daisy-chained IBM Super-Deep-Blue Gene Quantum Watson computers from 2042 could be running a History of the Universe program, and depending on your solipsistic preferences, either you are or we are the character(s).

It’s been in the news a lot of late, but — no way, right?

Because dude, I’m totally real
Despite being utterly unable to even begin thinking about how to consider what real even means, the everyday average rational person would probably assign this to the sovereign realm of unemployable philosophy majors or under the Whatever, Who Cares? or Oh, That’s Interesting I Gotta Go Now! categories. Okay fine, but on the other side of the intellectual coin, vis-à-vis recent technological advancement, of late it’s actually being seriously considered by serious people using big words they’ve learned at endless college whilst collecting letters after their names and doin’ research and writin’ and gettin’ association memberships and such.

So… why now?

Well, basically, it’s getting hard to ignore.
It’s not a new topic, it’s been hammered by philosophy and religion since like, thought happened. But now it’s getting some actual real science to stir things up. And it’s complicated, occasionally obtuse stuff — theories are spread out across various disciplines, and no one’s really keeping a decent flowchart.

So, what follows is an effort to encapsulate these ideas, and that’s daunting — it’s incredibly difficult to focus on writing when you’re wondering if you really have fingers or eyes. Along with links to some articles with links to some papers, what follows is Anthrobotic’s CliffsNotes on the intersection of physics, computer science, probability, and evidence for/against reality being real (and how that all brings us back to well, God).
You know, light fare.

First — Maybe we know how the universe works: Fantastically simplified, as our understanding deepens, it appears more and more the case that, in a manner of speaking, the universe sort of “computes” itself based on the principles of quantum mechanics. Right now, humanity’s fastest and sexiest supercomputers can simulate only extremely tiny fractions of the natural universe as we understand it (contrasted to the macro-scale inferential Bolshoi Simulation). But of course we all know the brute power of our computational technology is increasing dramatically like every few seconds, and even awesomer, we are learning how to build quantum computers, machines that calculate based on the underlying principles of existence in our universe — this could thrust the game into superdrive. So, given ever-accelerating computing power, and given than we can already simulate tiny fractions of the universe, you logically have to consider the possibility: If the universe works in a way we can exactly simulate, and we give it a shot, then relatively speaking what we make ceases to be a simulation, i.e., we’ve effectively created a new reality, a new universe (ummm… God?). So, the question is how do we know that we haven’t already done that? Or, otherwise stated: what if our eventual ability to create perfect reality simulations with computers is itself a simulation being created by a computer? Well, we can’t answer this — we can’t know. Unless…
[New Scientist’s Special Reality Issue]
[D-Wave’s Quantum Computer]
[Possible Large-scale Quantum Computing]

Second — Maybe we see it working: The universe seems to be metaphorically “pixelated.” This means that even though it’s a 50 billion trillion gajillion megapixel JPEG, if we juice the zooming-in and drill down farther and farther and farther, we’ll eventually see a bunch of discreet chunks of matter, or quantums, as the kids call them — these are the so-called pixels of the universe. Additionally, a team of lab coats at the University of Bonn think they might have a workable theory describing the underlying lattice, or existential re-bar in the foundation of observable reality (upon which the “pixels” would be arranged). All this implies, in a way, that the universe is both designed and finite (uh-oh, getting closer to the God issue). Even at ferociously complex levels, something finite can be measured and calculated and can, with sufficiently hardcore computers, be simulated very, very well. This guy Rich Terrile, a pretty serious NASA scientist, sites the pixelation thingy and poses a video game analogy: think of any first-person shooter — you cannot immerse your perspective into the entirety of the game, you can only interact with what is in your bubble of perception, and everywhere you go there is an underlying structure to the environment. Kinda sounds like, you know, life — right? So, what if the human brain is really just the greatest virtual reality engine ever conceived, and your character, your life, is merely a program wandering around a massively open game map, playing… well, you?
[Lattice Theory from the U of Bonn]
[NASA guy Rich Terrile at Vice]
[Kurzweil AI’s Technical Take on Terrile]

Thirdly — Turns out there’s a reasonable likelihood: While the above discussions on the physical properties of matter and our ability to one day copy & paste the universe are intriguing, it also turns out there’s a much simpler and straightforward issue to consider: there’s this annoyingly simplistic yet valid thought exercise posited by Swedish philosopher/economist/futurist Nick Bostrum, a dude way smarter that most humans. Basically he says we’ve got three options: 1. Civilizations destroy themselves before reaching a level of technological prowess necessary to simulate the universe; 2. Advanced civilizations couldn’t give two shits about simulating our primitive minds; or 3. Reality is a simulation. Sure, a decent probability, but sounds way oversimplified, right?
Well go read it. Doing so might ruin your day, JSYK.
[Summary of Bostrum’s Simulation Hypothesis]

Lastly — Data against is lacking: Any idea how much evidence or objective justification we have for the standard, accepted-without-question notion that reality is like, you know… real, or whatever? None. Zero. Of course the absence of evidence proves nothing, but given that we do have decent theories on how/why simulation theory is feasible, it follows that blithely accepting that reality is not a simulation is an intrinsically more radical position. Why would a thinking being think that? Just because they know it’s true? Believing 100% without question that you are a verifiably physical, corporeal, technology-wielding carbon-based organic primate is a massive leap of completely unjustified faith.
Oh, Jesus. So to speak.

If we really consider simulation theory, we must of course ask: who built the first one? And was it even an original? Is it really just turtles all the way down, Professor Hawking?

Okay, okay — that means it’s God time now
Now let’s see, what’s that other thing in human life that, based on a wild leap of faith, gets an equally monumental evidentiary pass? Well, proving or disproving the existence of god is effectively the same quandary posed by simulation theory, but with one caveat: we actually do have some decent scientific observations and theories and probabilities supporting simulation theory. That whole God phenomenon is pretty much hearsay, anecdotal at best. However, very interestingly, rather than negating it, simulation theory actually represents a kind of back-door validation of creationism. Here’s the simple logic:

If humans can simulate a universe, humans are it’s creator.
Accept the fact that linear time is a construct.
The process repeats infinitely.
We’ll build the next one.
The loop is closed.

God is us.

Heretical speculation on iteration
Even wonder why older polytheistic religions involved the gods just kinda setting guidelines for behavior, and they didn’t necessarily demand the love and complete & total devotion of humans? Maybe those universes were 1st-gen or beta products. You know, like it used to take a team of geeks to run the building-sized ENIAC, the first universe simulations required a whole host of creators who could make some general rules but just couldn’t manage every single little detail.

Now, the newer religions tend to be monotheistic, and god wants you to love him and only him and no one else and dedicate your life to him. But just make sure to follow his rules, and take comfort that your’re right and everyone else is completely hosed and going to hell. The modern versions of god, both omnipotent and omniscient, seem more like super-lonely cosmically powerful cat ladies who will delete your ass if you don’t behave yourself and love them in just the right way. So, the newer universes are probably run as a background app on the iPhone 26, and managed by… individuals. Perhaps individuals of questionable character.

The home game:
Latest title for the 2042 XBOX-Watson³ Quantum PlayStation Cube:*
Crappy 1993 graphic design simulation: 100% Effective!

*Manufacturer assumes no responsibility for inherently emergent anomalies, useless
inventions by game characters, or evolutionary cul de sacs including but not limited to:
The duck-billed platypus, hippies, meat in a can, reality TV, the TSA,
mayonaise, Sony VAIO products, natto, fundamentalist religious idiots,
people who don’t like homos, singers under 21, hangovers, coffee made
from cat shit, passionfruit iced tea, and the pacific garbage patch.

And hey, if true, it’s not exactly bad news
All these ideas are merely hypotheses, and for most humans the practical or theoretical proof or disproof would probably result in the same indifferent shrug. For those of us who like to rub a few brain cells together from time to time, attempting to both to understand the fundamental nature of our reality/simulation, and guess at whether or not we too might someday be capable of simulating ourselves, well — these are some goddamn profound ideas.

So, no need for hand wringing — let’s get on with our character arc and/or real lives. While simulation theory definitely causes reflexive revulsion, “just a simulation” isn’t necessarily pejorative. Sure, if we take a look at the current state of our own computer simulations and A.I. constructs, it is rather insulting. So if we truly are living in a simulation, you gotta give it up to the creator(s), because it’s a goddamn amazing piece of technological achievement.

Addendum: if this still isn’t sinking in, the brilliant
Dinosaur Comics might do a better job explaining:

(This post originally published I think like two days
ago at technosnark hub www.anthrobotic.com.
)

I cannot let the day pass without contributing a comment on the incredible ruling of multiple manslaughter on six top Italian geophysicists for not predicting an earthquake that left 309 people dead in 2009. When those who are entrusted with safeguarding humanity (be it on a local level in this case) are subjected to persecution when they fail to do so, despite acting in the best of their abilities in an inaccurate science, we have surely returned to the dark ages where those who practice science are demonized by the those who misunderstand it.

http://www.aljazeera.com/news/europe/2012/10/20121022151851442575.html

I hope I do not misrepresent other members of staff here at The Lifeboat Foundation, in speaking on behalf of the Foundation in wishing these scientists a successful appeal against a court ruling which has shocked the scientific community, and I stand behind the 5,000 members of the scientific community who sent an open letter to Italy’s President Giorgio Napolitano denouncing the trial. This court ruling was ape-mentality at its worst.

The Kline Directive: Theoretical-Empirical Relationship (Part 4)

Posted in business, cosmology, defense, economics, education, engineering, nuclear weapons, particle physics, philosophy, physics, policy, scientific freedom, spaceTagged , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , | 11 Comments on The Kline Directive: Theoretical-Empirical Relationship (Part 4)

To achieve interstellar travel, the Kline Directive instructs us to be bold, to explore what others have not, to seek what others will not, to change what others dare not. To extend the boundaries of our knowledge, to advocate new methods, techniques and research, to sponsor change not status quo, on 5 fronts, Legal Standing, Safety Awareness, Economic Viability, Theoretical-Empirical Relationship, & Technological Feasibility.

In this post I have updated the Interstellar Challenge Matrix (ICM) to guide us through the issues so that we can arrive at interstellar travel sooner, rather than later:

Interstellar Challenge Matrix (Partial Matrix)

Propulsion Mechanism Relatively Safe? Theoretical-Empirical Relationship?
Conventional Fuel Rockets: Yes, but susceptible to human error. Known. Theoretical foundations are based on Engineering Feasible Theories, and have been evolving since Robert Goddard invented the first liquid-fueled rocket in 1926.
Antimatter Propulsion: No. Extensive gamma ray production (Carl Sagan). Issue is how does one protect the Earth? Capable of an End of Humanity (EOH) event. Dependent on Millennium Theories. John Eades states in no uncertain terms that antimatter is impossible to handle and create.
Atomic Bomb Pulse Detonation: No, because (Project Orion) one needs to be able to manage between 300,000 and 30,000,000 atomic bombs per trip. Known and based on Engineering Feasible Theories.
Time Travel: Do Not Know. Depends on how safely exotic matter can be contained. Dependent on a Millennium Theory. Exotic matter hypotheses are untested. No experimental evidence to show that Nature allows for a breakdown in causality.
String / Quantum Foam Based Propulsion: Do Not Know. Depends on how safely exotic matter can be contained. Dependent on a Millennium Theory. String theories have not been experimentally verified. Exotic matter hypotheses are untested. Existence of Quantum Foam now suspect (Robert Nemiroff).
Small Black Hole Propulsion: No. Capable of an End Of Humanity (EOH) event Don’t know if small black holes really do exist in Nature. Their theoretical basis should be considered a Millennium Theory.

It is quite obvious that the major impediments to interstellar travel are the Millennium Theories. Let us review. Richard Feynman (Nobel Prize 1965) & Sheldon Lee Glashow (Nobel Prize 1979) have criticized string theory for not providing novel experimental predictions at accessible energy scales, but other theoretical physicists (Stephen Hawking, Edward Witten, Juan Maldacena and Leonard Susskind) believe that string theory is a step towards the correct fundamental description of nature. The Wikipedia article String Theory gives a good overview, and notes other critics and criticisms of string theories. In What is String Theory? Alberto Güijosa explains why string theories have come to dominate theoretical physics. It is about forces, and especially about unifying gravity with the other three forces.

Note, strings expand when their energy increases but the experimental evidence aka Lorentz-Fitzgerald transformations tell us that everything contracts with velocity i.e. as energy is increased.

In my opinion, the heady rush to a theory of everything is misguided, because there is at least one question that physics has not answered that is more fundamental than strings and particles. What is probability and how is it implemented in Nature?

Probabilities are more fundamental than particles as particles exhibit non-linear spatial probabilistic behavior. So how can one build a theory of everything on a complex structure (particles), if it cannot explain something substantially more fundamental (probabilities) than this complex structure? The logic defies me.

We can ask more fundamental questions. Is this probability really a Gaussian function? Experimental data suggests otherwise, a Var-Gamma distribution. Why is the force experienced by an electron moving in a magnetic field, orthogonal to both the electron velocity and the magnetic field? Contemporary electromagnetism just says it is vector cross product, i.e. it is just that way. The cross product is a variation of saying it has to be a Left Hand Rule or a Right Hand Rule. But why?

Is mass really the source of a gravitational field? Could it not be due to quark interaction? Can we device experiments that can distinguish between the two? Why do photons exhibit both wave and particle behavior? What is momentum, and why is it conserved? Why is mass and energy equivalent?

Can theoretical physicists construct theories without using the laws of conservation of mass-energy and momentum? That would be a real test for a theory of everything!

In my research into gravity modification I found that the massless formula for gravitational acceleration, g=τc2, works for gravity, electromagnetism and mechanical forces. Yes, a unification of gravity and electromagnetism. And this formula has been tested and verified with experimental data. Further that a force field is a Non Inertia (Ni) field, and is present where ever there is a spatial gradient in time dilations or velocities. This is very different from the Standard Model which requires that forces are transmitted by the exchange of virtual particles.

So if there is an alternative model that has united gravity and electromagnetism, what does that say for both string theories and the Standard Model? I raise these questions because they are opportunities to kick start research in a different direction. I answered two of these questions in my book. In the spirit of the Kline Directive can we use these questions to explore what others have not, to seek what others will not, to change what others dare not?

That is why I’m confident that we will have real working gravity modification technologies by 2020.

In concluding this section we need to figure out funding rules to ensure that Engineering Feasible and 100-Year Theories get first priority. That is the only way we are going to be able to refocus our physics community to achieve interstellar travel sooner rather than later.

Previous post in the Kline Directive series.

Next post in the Kline Directive series.

—————————————————————————————————

Benjamin T Solomon is the author & principal investigator of the 12-year study into the theoretical & technological feasibility of gravitation modification, titled An Introduction to Gravity Modification, to achieve interstellar travel in our lifetimes. For more information visit iSETI LLC, Interstellar Space Exploration Technology Initiative.

Solomon is inviting all serious participants to his LinkedIn Group Interstellar Travel & Gravity Modification.

The Kline Directive: Theoretical-Empirical Relationship (Part 3)

Posted in cosmology, defense, education, engineering, particle physics, philosophy, physics, policy, spaceTagged , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , | 17 Comments on The Kline Directive: Theoretical-Empirical Relationship (Part 3)

To achieve interstellar travel, the Kline Directive instructs us to be bold, to explore what others have not, to seek what others will not, to change what others dare not. To extend the boundaries of our knowledge, to advocate new methods, techniques and research, to sponsor change not status quo, on 5 fronts:

1. Legal Standing. 2. Safety Awareness. 3. Economic Viability. 4. Theoretical-Empirical Relationship. 5. Technological Feasibility.

In Part 1, we learned that Einstein was phenomenally successful because his work was deeply meshed with the experimental evidence of the day. In Part 2, we learned that to be successful at developing new useful theories and discovering new fundamental properties of Nature that will bring forth new interstellar travel technologies, we need to avoid hypotheses that are not grounded in experimental data, as these are purely mathematical conjectures.

In my book on gravity modification I classified physics hypotheses and theories into 3 categories, as follows:

A. Type 1: The Millennium Theories
These are theories that would require more than a 100 years and up to 1,000 years to prove or disprove. Mathematically correct but inscrutable with physical verifiable experiments, even in the distant future.

String and quantum gravity theories fall into this category. Why? If we cannot even figure out how to engineer-modify 4-dimensional spacetime, how are we going to engineer-modify a 5-, 6-, 9-, 11- or 23-dimensional universe?

How long would it take using string theories to modify gravity? Prof. Michio Kaku in his April 2008 Space Show interview had suggested several hundred years. Dr. Eric Davis in his G4TV interview had suggested more than 100 years maybe 200 years. So rightly, by their own admission these are Millennium Theories. It should be noted that Richard Feynman (Nobel Prize 1965) & Sheldon Lee Glashow (Nobel Prize 1979) were against string theory, but their opinions did not prevail.

Even hypotheses that conjecture time travel should be classified as Millennium Theories because they require ‘exotic’ matter. John Eades, a retired CERN senior scientist, in his article Antimatter Pseudoscience, states in no uncertain terms that antimatter is impossible to handle and create in real quantities. Then what about exotic matter?

For that matter any hypothesis that requires antimatter or exotic matter should be classified a Millennium Theory.

B. Type 2: The 100-Year Theories
These are theories that show promise of being verified with technologies that would require several decades to engineer, test and prove.

These types of theories do not lend themselves to an immediate engineering solution. The engineering solution is theoretically feasible but a working experiment or technology is some decades away, because the experimental or physical implementation is not fully understood.

Note there is this gap. We do not have 100-Year Theories in our repertoire of physical theories to keep the pipeline supplied with new and different ways to test the physical Universe.

C. Type 3: The Engineering Feasible Theories
These are theories that lend themselves to an engineering solution, today. They are falsifiable today, with our current engineering technologies. They can be tested and verified in the laboratory if one knows what to test for and how to test for these experimental observations.

Today Relativity falls into this category because we have the engineering sophistication to test Einstein’s theory, and it has been vindicated time and time again. But, there is a very big ‘but’. But Relativity cannot give us gravity modification or new propulsion theories, because it requires mass. We need to stand on Einstein’s shoulders to take the next step forward.

Therefore, if we are to become an interstellar civilization, in the spirit of the Kline Directive, we need to actively seek out and explore physics in such a manner as to bring forth Engineering Feasible and 100-Year Theories.

We need to ask ourselves, what can we do, to migrate the theoretical physics research away from Theory of Everything research to the new field of propulsion physics? Gravity modification is an example of propulsion physics. Here is the definition of gravity modification, from my book:

“Gravity modification is defined as the modification of the strength and/or direction of the gravitational acceleration without the use of mass as the primary source of this modification, in local space time. It consists of field modulation and field vectoring. Field modulation is the ability to attenuate or amplify a force field. Field vectoring is the ability to change the direction of this force field.”

Note by this definition requiring no mass, relativity, quantum mechanics and string theories cannot be used to theorize propulsion physics. Therefore, the urgent need to find genuinely new ways in physics, to achieve interstellar travel.

Can we get there? The new physics? To answer this question let me quote Dr. Andrew Beckwith, Astrophysicist, Ph.D.(Condensed Matter Theory) who wrote the Foreword to my book:

“I believe that Quantum Mechanics is an embedded artifact of a higher level deterministic theory, i.e. much in the same vein as G. t’Hooft, the Nobel prize winner. In this sense, what Benjamin has done is to give a first order approximation as to what Quantum Mechanics is actually a part of which may in its own way shed much needed understanding of the foundations of Quantum Mechanics well beyond the ‘Pilot model’ of DICE 2010 fame (this is a conference on the foundations of Quantum Mechanics and its extension given once every two years in Pisa , Italy, organized by Thomas Elze).”

Why does Dr. Andrew Beckwith reference quantum mechanics in a book on gravity modification?

Because my investigation into gravity modification led me to the conclusion that gravitation acceleration is independent of the internal structure of the particle. It does not matter if the particle consists of other particles, strings, pebbles or rocks. This led me to ask the question, so what is the internal structure of a photon? I found out that the photon probability is not Gaussian but a new distribution, Var-Gamma. Therefore I believe Robert Nemiroff’s three photon observation will be vindicated by other physicist-researchers sifting through NASA’s archives for gamma-ray burst.

Previous post in the Kline Directive series.

—————————————————————————————————

Benjamin T Solomon is the author & principal investigator of the 12-year study into the theoretical & technological feasibility of gravitation modification, titled An Introduction to Gravity Modification, to achieve interstellar travel in our lifetimes. For more information visit iSETI LLC, Interstellar Space Exploration Technology Initiative.

Solomon is inviting all serious participants to his LinkedIn Group Interstellar Travel & Gravity Modification.

To achieve interstellar travel, the Kline Directive instructs us to be bold, to explore what others have not, to seek what others will not, to change what others dare not. To extend the boundaries of our knowledge, to advocate new methods, techniques and research, to sponsor change not status quo, on 5 fronts:

1. Legal Standing. 2. Safety Awareness. 3. Economic Viability. 4. Theoretical-Empirical Relationship. 5. Technological Feasibility.

From Part 1 … “that mathematics has become so sophisticated and so very successful that it can now be used to prove anything and everything, and therefore, the loss of certainty that mathematics will provide reasonability in guidance and correctness in answers to our questions in the sciences”.

We need to note that there are several different relationships between the mathematics of physics and the physics of the real world.

The first relationship and most common type is that several different types of equations in physics describe the same physics of the world. Gravity is a great example. The three mathematical theories on gravity are relativity, quantum and string theories. All three model the same single physical phenomenon, gravitational fields. So if one is correct than the other two must be wrong. All three cannot be correct. So which is it?

Just for argument sake, there is another alternative — all three are wrong. But wait didn’t all those experiments and observations prove that General Relativity is correct? Remember for argument’s sake, that proving that something fits the experimental observation does not mean that is how Nature works. That is why theoretical physicists spend so much time, money and effort considering alternatives like quantum and string theories.

The second relationship is that different mathematical descriptions can be ascribed to different parts of a physical phenomenon. For example Einstein’s General Relativity describes spacetime as tensor calculus, a very complex mathematical model which he did not get right on his first attempt. General Relativity addresses the question of gravity’s source as an energy-momentum tensor. To put it simply, these equations are complex.

Whereas in my work I realized at some point during my investigation into gravity modification, that to develop technologies that could modify gravity we needed a mathematical equation (g=τc2) that would describe the phenomenon of gravitational acceleration without needing to include mass. I discovered this equation, g=τc2, after very extensive numerical modeling of gravitational accelerations in spacetime, where tau, τ is the change in time dilation divided by change in distance (for more look up my Physics Essays paper, “Gravitational Acceleration Without Mass And Noninertia Fields”). Consider how elegantly simple this equation is and without mass we can now replace the source with something more technology friendly.

And the third type of relationship is the mathematics of physics that cannot or cannot yet be verified with experimental evidence. String theories are great examples of this. From what I know, there is nothing in the string theories (which have not been borrowed for quantum mechanics) that is experimentally verifiable. And yet we go on. Why?

Consider this. The experimental evidence proves that nothing with mass can be accelerated past the velocity of light (aka Lorentz-Fitzgerald transformations), and yet Dr. Eric Davis agrees with Dr. Richard Obousy that using string quantum theory that the maximum velocity one can reach is 1032 x c (100,000,000,000,000,000,000,000,000,000,000 x velocity of light). Now what would you believe, experimental evidence or mathematical conjecture?

Now, do you agree that that mathematics has become so sophisticated and so very successful that it can be used to prove anything and everything, and therefore, the loss of certainty that mathematics can provide reasonability in guidance and correctness in answers to our questions in the sciences?

Don’t get me wrong. Mathematics is vital for the progress of the sciences, but it needs to be tempered with real world experimental evidence, otherwise it is just conjecture, and retards our search for interstellar travel technologies.

Previous post in the Kline Directive series.

Next post in the Kline Directive series.

—————————————————————————————————

Benjamin T Solomon is the author & principal investigator of the 12-year study into the theoretical & technological feasibility of gravitation modification, titled An Introduction to Gravity Modification, to achieve interstellar travel in our lifetimes. For more information visit iSETI LLC, Interstellar Space Exploration Technology Initiative.

Solomon is inviting all serious participants to his LinkedIn Group Interstellar Travel & Gravity Modification.

To achieve interstellar travel, the Kline Directive instructs us to be bold, to explore what others have not, to seek what others will not, to change what others dare not. To extend the boundaries of our knowledge, to advocate new methods, techniques and research, to sponsor change not status quo, on 5 fronts:

1. Legal Standing. 2. Safety Awareness. 3. Economic Viability. 4. Theoretical-Empirical Relationship. 5. Technological Feasibility.

In Part 1 of this post I will explore Theoretical-Empirical Relationship. Not theoretical relationships, not empirical relationships but theoretical-empirical relationships. To do this let us remind ourselves what the late Prof. Morris Kline was getting at in his book Mathematics: The Loss of Certainty, that mathematics has become so sophisticated and so very successful that it can now be used to prove anything and everything, and therefore, the loss of certainty that mathematics will provide reasonability in guidance and correctness in answers to our questions in the sciences.

History of science shows that all three giants of science of their times, Robert Boyle, Isaac Newton & Christiaan Huygens believed that light traveled in aether medium, but by the end of the 19th century there was enough experimental evidence to show aether could not be a valid concept. The primary experiment that changed our understanding of aether was the Michelson–Morley experiment of 1887, which once and for all proved that aether did not have the correct properties as the medium in which light travels.

Only after these experimental results were published did, a then unknown Albert Einstein, invent the Special Theory of Relativity (SRT) in 1905. The important fact to take note here is that Einstein did not invent SRT out of thin air, like many non-scientists and scientists, today believe. He invented SRT by examining the experimental data to put forward a hypothesis or concept described in mathematical form, why the velocity of light was constant in every direction independent of the direction of relative motion.

But he also had clues from others, namely George Francis FitzGerald (1889) and Hendrik Antoon Lorentz (1892) who postulated length contraction to explain negative outcome of the Michelson-Morley experiment and to rescue the ‘stationary aether’ hypothesis. Today their work is named the Lorentz-Fitzgerald transformation.

So Einstein did not invent the Special Theory of Relativity (SRT) out of thin air, there was a body of knowledge and hypotheses already in the literature. What Einstein did do was to pull all this together in a consistent and uniform manner that led to further correct predictions of how the physics of the Universe works.

(Note: I know my history of science in certain fields of endeavor, and therefore use Wikipedia a lot, not as a primary reference, but as a starting point for the reader to take off for his/her own research.)

Previous post in the Kline Directive series.

Next post in the Kline Directive series.

—————————————————————————————————

Benjamin T Solomon is the author & principal investigator of the 12-year study into the theoretical & technological feasibility of gravitation modification, titled An Introduction to Gravity Modification, to achieve interstellar travel in our lifetimes. For more information visit iSETI LLC, Interstellar Space Exploration Technology Initiative.

Solomon is inviting all serious participants to his LinkedIn Group Interstellar Travel & Gravity Modification.