Notes on a Static Universe: Incredibly, an expanding universe does imply an expanding earth on most cosmological models that follow Einstein and or Friedmann. As space expands, so does everything in it. This is why, if the redshift signified cosmological expansion even the very atoms making up the matter in the universe would also have to expand. There would be no sign of this in rock crystal lattices etc since everything was expanding uniformly as was the space between them. This expansion occurred at the Creation of the Cosmos as the verses you listed have shown.
It is commonly thought that the progressive redshift of light from distant galaxies is evidence that this universal expansion is still continuing. However, W. Q. Sumner in Astrophysical Journal 429:491-498, 10 July 1994, pointed out a problem. The maths indeed show that atoms partake in such expansion, and so does the wavelength of light in transit through space. This "stretching" of the wavelengths of light in transit will cause it to become redder. It is commonly assumed that this is the origin of the redshift of light from distant galaxies. But the effect on the atoms changes the wavelength of emitted light in the opposite direction. The overall result of the two effects is that an expanding cosmos will have light that is blue-shifted, not red-shifted as we see at present. The interim conclusion is that the cosmos cannot be expanding at the moment (it may be contracting).
Furthermore, as Arizona astronomer William Tifft and others have shown, the redshift of light from distant galaxies is quantised, or goes in "jumps". Now it is uniformly agreed that any universal expansion or contraction does not go in "jumps" but is smooth. Therefore expansion or contraction of the cosmos is not responsible for the quantisation effect: it may come from light-emitting atoms. If this is so, cosmological expansion or contraction will smear-out any redshift quantisation effects, as the emitted wavelengths get progressively "stretched" or "shrunk" in transit. The final conclusion is that the quantised redshift implies that the present cosmos must be static after initial expansion. [Narliker and Arp proved that a static matter-filled cosmos is stable against collapse in Astrophysical Journal 405:51-56 (1993)].
Therefore, as the heavens were expanded out to their maximum size, so were the earth, and the planets, and the stars. I assume that this happened before the close of Day 4, but I am guessing here. Following this expansion event, the cosmos remained static. (Barry Setterfield, September 25, 1998)
In 2002, he authored the article "Is the Universe Static or Expanding", in which he presents further research on this subject.
Setterfield: The evidence that an accelerating expansion is occurring comes because distant objects are in fact further away than anticipated given a non-linear and steeply climbing red-shift/distance curve.
Originally, the redshift/distance relation was accepted as linear until objects were discovered with a redshift greater than 1. On the linear relation, this meant that the objects were receding with speeds greater than light, a no-no in relativity. So a relativistic correction was applied that makes the relationship start curving up steeply at great distances. This has the effect of making large redshift changes over short distances. Now it is found that these objects are indeed farther away than this curve predicts, so they have to drag in accelerating expansion to overcome the hassle.
The basic error is to accept the redshift as due to an expansion velocity. If the redshift is NOT a velocity of expansion, then these very distant objects are NOT travelling faster than light, so the relativistic correction is not needed. Given that point, it becomes apparent that if a linear redshift relation is maintained throughout the cosmos, then we get distances for these objects that do not need to be corrected. That is just what my current redshift paper does. (January 12, 1999)
Setterfield: It has been proved recently by aberration experiments that distant starlight from remote galaxies arrives at earth with the same velocity as light does from local sources. This occurs because the speed of light depends on the properties of the vacuum. If we assume that the vacuum is homogeneous and isotropic (that is has the same properties uniformly everywhere at any given instant), then light-speed will have the same value right throughout the vacuum at any given instant. The following proposition will also hold. If the properties of the vacuum are smoothly changing with time, then light speed will also smoothly change with time right throughout the cosmos.
On the basis of experimental evidence from the 1920's when light speed was measured as varying, this proposition maintains that the wavelengths of emitted light do not change in transit when light-speed varies, but the frequency (the number of wave-crests passing per second) will. The frequency of light in a changing c scenario is proportional to c itself. Imagine light of a given earth laboratory wavelength emitted from a distant galaxy where c was 10 times the value it has now. The wavelength would be unchanged, but the emitted frequency would be 10 times greater as the wave-crests are passing 10 times faster. As light slowed in transit, the frequency also slowed, until when it reaches earth at c now, the frequency would be the same as our laboratory standard as well as the wavelength. Trust that this reply answers your question. (June 15, 1999)
Setterfield: Thank you for your questions; let me see what I can do to answer them. Firstly about missing mass. There are two reasons why astronomers have felt that mass is "missing". The first is due to the apparent motion of galaxies in clusters. That motion in the outer reaches of the clusters appears to be far too high to allow the clusters to hold together gravitationally unless there is extra mass somewhere. However, that apparent motion is all measured by the redshift of light from those galaxies. Tifft's work with the quantised redshift indicates that redshift is NOT a measure of galaxy motion at all. In fact, motion washes out the quantisation. On that basis, the whole foundation on which the missing mass in galaxy clusters is built is faulty, as there is very little motion of galaxies in clusters at all. The second area where astronomers have felt that mass is "missing" is due to the behaviour of rotation rates of galaxies as you go out from their centres. The behaviour of the outer regions of galaxies is such that there must be more mass somewhere there to keep the galaxies from flying apart as their rotation rate is so high. It seems that there might indeed be a large envelope of matter in transparent gaseous form around galaxies that could account for this discrepancy. Alternatively, some astronomers are checking for Jupiter sized solid objecdts in the halo of our galaxy that could also overcome for the problem. A third possibility is that the Doppler equations on which everything is based may be faulty at large distances due to changes in light speed. I am looking into this.
Setterfield: The entire discussion about dark energy and dark matter is based, at least partly, on a mis-interpretation of the redshift of light from distant galaxies. But before I elaborate on that, let me first specifically address a recent article in Science (“Breakthrough of the Year: Illuminating the Dark Universe” Charles Seife, Science 302, 2038-2039, 2003) espousing the proof for dark energy. I find this amazing in view of the recent data that has come in from the European Space Agency's (ESA) X-ray satellite, the XMM-Newton. According to the ESA News Release for December 12, 2003 the data reveal "puzzling differences between today's clusters of galaxies and those present in the Universe around seven thousand million years ago." The news release says that these differences "can be interpreted to mean that 'dark energy' which most astronomers now believe dominates the universe simply does not exist." In fact, Alain Blanchard of the Astrophysical Observatory in the Pyrenees says the data show "There were fewer galaxy clusters in the past". He goes on to say that "To account for these results you have to have a lot of matter in the Universe and that leaves little room for dark energy."
In other words, we have one set of data which can be interpreted to mean that dark energy exists, while another set of data suggests that it does not exist. In the face of this anomaly, it may have been wiser for Science to have remained more circumspect about the matter. Unfortunately, the scientific majority choose to run with an interpretation they find satisfying, and tend to marginalize all contrary data. I wonder if the European data may not have been published by the time that Science went to print on the issue. Thus there may be some embarrassment by these later results, and they may be marginalized as a consequence.
The interpretation being placed on the WMAP observations of the microwave background is that it is the "echo" the Big Bang event, and all other data is interpreted on this basis. But Takaaki Musha from Japan pointed out in an article in Journal of Theoretics (3:3, June/July 2001) that the microwave background may well be the result of the Zero Point Energy allowing the formation of virtual tachyons in the same way that it allows the formation of all other kinds of virtual particles. Musha demonstrated that all the characteristics of the microwave background can be reproduced by this approach. In that case, the usual interpretation of the WMAP data is in error and the conclusions drawn from it should be discarded and a different set of conclusions deduced.
However, the whole dark matter/dark energy discussion points to the fact that anomalies exist which current theory did not anticipate. Let me put both of these problems in context. Dark matter became a necessity because groups of galaxies seemed to have individuals within the group which appeared to be moving so fast that they should have escaped long ago if the cosmos was 14 billion years old. If the cosmos was NOT 14 billion years old but, say, only 1 million or 10,000 years old, the problem disappears. However, another pertinent answer also has been elaborated by Tifft and Arp and some other astronomers, but the establishment does not like the consequences and tends to ignore them as a result. The answer readily emerges when it is realized that the rate of movement of the galaxies within a cluster is measured by their redshift. The implicit assumption is that the redshift is a measure of galaxy motion. Tifft and Arp pointed out that the quantized redshift meant that the redshift was not intrinsically a measure of motion all but had another origin. They pointed out that, in the centre of the Virgo cluster of galaxies, where motion would be expected to be greatest under gravity, the ACTUAL motion of the galaxies smeared out the quantization. If actual motion does this, then the quantized redshift exhibited by all the other galaxies further out means that there is very little actual motion of those galaxies at all. This lack of motion destroys the whole basis of the missing matter argument and the necessity for dark matter then disappears. The whole missing matter or dark matter problem only arises because it is in essence a mis-interpretation of what the redshift is all about. In a similar fashion, the whole dark energy (or necessity for the cosmological constant) is also a redshift problem. It arises because there is a breakdown in the redshift/distance relationship at high redshifts. That formula is based on the redshift being due to the galaxies racing away from us with high velocities. Distant supernovae found up to 1999 proved to be even more distant than the redshift/distance formula indicated. This could only be accounted for on the prevailing paradigm if the controversial cosmological constant (dark energy) were included in the equations to speed up the expansion of the universe with time. The fact that these galaxies were further away than expected was taken as proof that the cosmological constant (dark energy) was acting. Then in October this year, Adam Riess announced that there were 10 even more distant supernovae whose distance was CLOSER than the redshift relation predicted. With a deft twist, this was taken as further proof for the existence of dark energy. The reasoning went that up to a redshift of about 1.5 the universal expansion was slowing under gravity. Then at that point, the dark energy repulsion became greater than the force of gravity, and the expansion rate progressively speeded up.
What in fact we are looking at is again the mis-interpretation of the redshift. Both dark matter and dark energy hinge on the redshift being due to cosmological expansion. If it has another interpretation, and my latest paper being submitted today [late December, 2003] shows it does, then the deviation of distant objects from the standard redshift/distance formula is explicable, and the necessity for dark matter also disappears. Furthermore, the form of the actual equation rests entirely on the origin of the Zero Point Energy. The currently accepted formula can be reproduced exactly as one possibility of several. But the deviation from that formula shown by the observational evidence is entirely explicable without the need for dark energy or dark matter or any other exotic mechanism.
I trust that this gives you a feel for the situation.
Question: I have just recently come across your work and feel like I have come home! For years I have squared the accounts of the bible and science (being from a science backgound myself) by passing Genesis off (in particular) as 'allegorical' and always telling christian friends that we must not ignore the scientific evidence. I was uncomfortable with all the scientific explanations, though, and suspected things were not quite right. I was aware that data was data and that the scientific explanations throughout history had been modified (in some cases drastically) in the light of new evidence. Hence, the Newtonian view becoming the Einsteinian view. So I was left with..'well this is what I currently believe until such time as new evidence indicates otherwise'.Setterfield: It is a delight that you have found us, and many thanks for the encouragement it gives to us as well! You have had quite a journey, but the Lord has brought you to some sort of haven of rest. We trust that you will be able to absorb what we have put on our website and share it with others as the Lord directs.
As for your specific questions, I would like to deal with the missing mass/dark matter problem. This really is a deep problem for gravitational astronomers. Currently these is no assured answer from that direction except to modify the laws of gravity and so on - all very messy procedures. However, what must be realized is that there are two aspects to the problem. The first that was noted was the missing mass in galaxy clusters. What was happening was that the redshifts of light from these galaxy clusters was being measured and interpreted as a Doppler shift, like the drop in pitch of a police siren passing you. The redshift on this basis was taken to mean that galaxies are racing away from us. When the average redshift of a cluster of galaxies was taken, and then the highest redshifts of galaxies in the cluster were compared with that, then, if the redshift was due to motion, these high redshift galaxies were moving so fast they would have escaped the cluster long ago and moved out on their own. As a result, since they are obviously still part of the cluster, there must be a lot of additional material in the cluster to give a gravitational field strong enough to hold the cluster together. The whole chain of reasoning depends on the interpretation of the redshift as galaxy motion.
In contra-distinction to this interpretation of the redshift, the Zero Point Energy approach indicates that the redshift is simply the result of a lower ZPE strength back in the past which affects atomic orbits so they emit redder light (probably in quantum steps as indicated by Tifft and others). On that basis, the redshift has nothing to do with velocity, so the galaxy clusters have no mass "missing" and hence no "dark matter." We can see the actual motion of galaxies in the center of clusters where the motion actually washes out the quantization, so we are on the right track there. On this approach, the clusters of galaxies are "quiet" without high velocities that pull them apart.
The second reason for the search for "missing mass/dark matter" is the actual rotation of individual galaxies themselves.As we look at our solar system, the further out the planet is from the sun, the slower it moves around the sun, because of the Laws of gravity. Astronomers expected the same result when they looked at the velocities of stars in the disks of galaxies as they moved around the center of those galaxies. That is, they expected the stars in the outer part of the disk would move at a slower pace than stars nearer the center. They dont! The stars in the outer part of the disk orbit with about the same velocity as those closer in. For this to happen on the gravitational model, there must be some massive amounts of material in and above/below the disk that we cannot see to keep the rotation rate constant. Since we have been unable to find it, we call it missing matter or dark matter. That is the conundrum. Another suggestion is that the laws of gravity need amending.
However, if all galaxies were formed by a plasma process, as indicated by Anthony Peratt's simulations, then the problem is solved. In the formation of galaxies by plasma physics, Peratt has shown that the miniature galaxies produced in the laboratory spin in exactly the same way as galaxies do in outer space. No missing mass is needed; no modification to the laws of gravity. All that is needed is to apply the laws of plasma physics to the formation of galaxies, not the laws of gravitation. These laws involve electro-magnetic interaction, not gravitational interaction. The two are vastly different and the electro-magnetic interactions are up to 1039 times more powerful than gravity! Things can form much more quickly by this mechanism as well, which has other implications in the origins debate.
Well, that should give you some things to digest for the moment. You will have other questions. Get back to me when you want answers, and I will do my best.
Setterfield: Many thanks for keeping us informed on this; it is appreciated!
The whole scenario with dark matter comes from the rotation rate of galaxy arms. If it is gravitational - the dark matter must exist. If galaxies rotate under the influence of plasma physics, nothing is missing - it is all behaving exactly as it should. Unfortunately gravitational astronomers have painted themselves into a corner with this one and cannot escape without losing a lot of credibility. The plasma physicists are having a good laugh.
Setterfield: Original answer -- You ask where the radiation etc disappears to when it gets trapped inside a black hole. The quick answer is that it becomes absorbed into the fabric of space, as that fabric is made up of incredibly tiny Planck particle pairs that are effectively the same density as the black holes. The trapped radiation cannot travel across distances shorter than those between the Planck particles, and so gets absorbed into the vacuum. This brings us to your question of what a black hole really is. It essence, it is a region of space that has the same density as the Planck particle pairs that make up the fabric of space. It seems that these centres of galaxies may have formed as an agglomeration of Planck particle pairs at the inception of the universe and acted as the nucleus around which matter collected to form galaxies.
February, 2012 -- Since the above answer was written, there have been some developments in astronomy as a result of plasma physics and experiments conducted in plasma laboratories. In the laboratories, a spinning disk with polar jets can be readily produced as a result of electric currents and magnetic fields. The principle on which they operate is basically the same as the spinning disk of your electricity meter. What we see with "black holes" is the spinning disk and the polar jets. We measure the rate of spin of the disk and from that gravitational physics allows us to deduce that there is a tremendous mass in the center of the disk. This is where the idea of a black hole came from. However, with plasma physics, the rate of spin of the disk is entirely dependent upon the strength of the electric current involved. Plasma physics shows that there is a galactic circuit of current for every galaxy. As a consequence of this, and the experimental evidence we have, plasma physics does not have the necessity for a black hole to explain what we are seeing. It can all be explained by electric currents and magnetic fields. In reviewing this evidence, I came to the conclusion that plasma physics, applied to astronomy, is the better way to go, and that gravitational astronomy has caused itself some considerable problems in this and other areas. For example, in the case of our own galaxy, and the suspected black hole at its center, it has been deduced from the motion of objects in the vicinity that the mass of the black hole is such that it should cause gravitational lensing of objects in the area. We have searched for gravitational lensing for over two decades and none has been found -- despite extensive searches. This indicates that it is not a black hole we have in the center of our galaxy.
Keep in mind no one has ever seen a black hole. It is simply a necessary construct if gravity is to be considered the main force forming and maintaining objects in the universe. However, gravity is a very weak force, especially when compared to electromagnetism, which is what we have seen out there in the form of vast plasma filaments.
Setterfield: Astronomically, many of the important distant Quasars turn out to be the superluminous, hyper-active centres of galaxies associated with a supermassive black hole. There is a mathematical relationship between the size of the black hole powering the quasar and the size of the nucleus of any given galaxy. As a result of this relationship, there is a debate going on as to whether the quasar/black hole came first and the galaxy formed around it, or vice versa. Currently, I tend to favour the first option. Nevertheless, which ever option is adopted, the main effect of dropping values of c on quasars is that as c decays, the diameter of the black hole powering the quasar will progressively increase. This will allow progressive engulfment of material from the region surrounding the black hole and so should feed their axial jets of ejected matter. This is the key prediction from the cDK model on that matter.
As far as pulsars are concerned, there recently has been some doubt cast on the accepted model for the cause of the phenomenon that we are observing. Until a short time ago, it was thought that the age of a pulsar could be established from the rotation period of the object that was thought to give rise to the precisely timed signal. However, recent work on two fronts has thrown into confusion both the model for the age of the pulsar based on its rotation period, and also actual cause of the signal, and hence what it is that is rotating. Until these issues can be settled, it is difficult to make accurate predictions from the cDK model.
Setterfield: Astronomically, many of the important distant Quasars turn out to be the superluminous, hyper-active centres of galaxies associated with a supermassive black hole. There is a mathematical relationship between the size of the black hole powering the quasar and the size of the nucleus of any given galaxy. As a result of this relationship, there is a debate going on as to whether the quasar/black hole came first and the galaxy formed around it, or vice versa. Currently, I tend to favour the first option. Nevertheless, whichever option is adopted, the main effect of dropping values of c on quasars is that as c slows, the diameter of the black hole powering the quasar will progressively increase. This will allow progressive engulfment of material from the region surrounding the black hole and so should feed their axial jets of ejected matter. This is the key prediction from my variable speed of light model on that matter.
As far as pulsars are concerned, there recently has been some doubt cast on the accepted model for the cause of the phenomenon that we are observing. Until a short time ago, it was thought that the age of a pulsar could be established from the rotation period of the object that was thought to give rise to the precisely timed signal. However, recent work on two fronts has thrown into confusion both the model for the age of the pulsar based on its rotation period, and also actual cause of the signal, and hence what it is that is rotating. Until these issues can be settled, it is difficult to make accurate predictions.
Question: I've been asked why distant pulsars don't show a change in their rotation rate if the speed of light is slowing. Do you have an explanation? If you have any information on whether their observed rate of change should change with a slowing of c I would appreciate it.Setterfield: Thanks for your question about pulsars. There are several aspects to this. First of all, pulsars are not all that distant, the furthest that we can detect are in small satellite galaxies of our own Milky Way system. Second, because the curve of lightspeed is very flat at those distances compared with the very steep climb closer to the origin, the change in lightspeed is small. This means that any pulsar slowdown rate originating with the changing speed of light is also small. The third point is that the mechanism that produces the pulses is in dispute as some theories link the pulses with magnetic effects separate from the star itself, so that the spin rate of the host star may not be involved. Until this mechanism is finally determined, the final word about the pulses and the effects of lightspeed cannot be given. If you have Cepheid variables in mind, a different situation exists. The pulsation, that gives the characteristic curve of light intensity variation, is produced by the behaviour of a thin segment near the star's surface layer. The behaviour of this segment of the star's outer layers is directly linked with the speed of light. This means that any slow-down effect of light in transit will already have been counteracted by a change in the pulsation rate of this layer at the time of emission. The final result will be that any given Cepheid variable will appear to have a constant period for the light intensity curve, no matter where it is in space and no matter how long we observe it.
Comment: Dr. Tom Bridgman website shows in calculus the metric for canceling out the red-shift. Even more important is the metric he shows for kinematic argument using pulsar periods out to a distance of 1,000 parsecs from earth. According to Tom, you have very observable effects in pulsar periods using c-decay that is way off so c-decay is not taking place within 1,000 parsecs of the earth.
Setterfield: Yes. As far as pulsars are concerned within 1000 parsecs of the earth, there should only be very, very minimal effects due to a changing c. In fact, at that distance, the change in c would be so small as to not even give rise to any redshift effects at all. The first redshift effect comes at the first quantum jump which occurs beyond the Magellanic Clouds. This shows that the rate at which c is climbing across our galaxy is very, very small. Consequently, any effect with pulsars within our galaxy is going to be negligible.
Recently the reason for the spindown rate in pulsars has been seriously questioned, and new models for pulsar behaviour will have to be examined. See, for example, New Scientist 28 April, 2001, page 28. There are other references which I do not have on hand at the moment, but which document other problems as well. Until a viable model of pulsar behaviour and the cause of the pulsars themselves has been finalised, it is difficult to make any predictions about how the speed of light is going to affect them. I am curious about the metric that Bridgman is using, because I show the change in wavelength over the wavelength (which is the definition of the redshift) is, in fact, occurring.
Response: Dr. William T. (Tom) Bridgman is using kinematics and a calculus formula to shoot down your c-decay metric. He even uses your predicted c-decay curve based upon the past 300 years or so of measurments for c and incorporates this into the pulsar changes that he claims should be observed. if correct then c-decay did not take place within past 300 years or even out to distances of 1 kiloparsec
Setterfield: The curve that Tom is using for pulsar analysis is outdated and no longer applicable to the situation as it dates from the 1987 paper. The recent work undergoing review indicates a very different curve which includes a slight oscillation. This oscillation means that there is very little variability in light speed out to the limits of our galaxy. Thus, even if the rest of his math were correct, and the behaviour of pulsars were known accurately, Tom's conclusions are not valid. I suggest that he some of the more recent material before attempting fireworks.
Response: If c-decay has predictable and observable side effects like pulsar timing changes, changes in eclipses of Jupiter's and Saturn's moons, and also changes in stellar occultations in the ecliptic, these should be rigourously tested to see if they support or deny c-decay. At the moment Tom's metric denies c-decay as published based upon the 1987 c-decay curve
Setterfield: I have examined pulsar timing changes in detail and responded some years ago to that so-called "problem". Any observed changes are well within the limits predicted by this variable light-speed (Vc) model just introduced. One hostile website used the pulsar argument for a while until I pointed out their conceptual error to a friend and then it was deleted and has not appeared again since. I suspect that Tom is making the same error.
The changes in the eclipse times of Jupiter's and Saturn's moons have in fact been used as basic data in the theory. Stellar occultations along the ecliptic have also been used as data based on the work of Tom van Flandern who studied the interval 1955-1981. He then came to the conclusion: "the number of atomic seconds in a dynamical interval is becoming fewer. Presumably, if the result has any generality to it, this means that atomic phenomena are slowing down with respect to dynamical phenomena..." In this case the eclipse times and the occultations were used to build the original model and as such are not in conflict with it.
Setterfield: In Physical Review Letters published on 27 August 2001 there appeared a report from a group of scientists in the USA, UK and Australia, led by Dr. John K. Webb of the University of New South Wales, Australia. That report indicated the fine-structure constant, a, may have changed over the lifetime of the universe. The Press came up with stories that the speed of light might be changing as a consequence. However, the change that has been measured is only one part in 100,000 over a distance of 12 billion light-years. This means that the differences from the expected measurements are very subtle. Furthermore, the complicated analysis needed to disentangle the effect from the data left some, like Dr. John Bahcall from the Institute for Advanced Study in Princeton, N.J., expressing doubts as to the validity of the claim. This is further amplified because all the measurements and analyses have been done at only one observatory, and may therefore be the result equipment aberration. Other observatories will be involved in the near future, according to current plans. This may clarify that aspect of the situation.
The suggested change in the speed of light in the Press articles was mentioned because light-speed, c, is one of the components making up the fine-structure constant. In fact K. W. Ford, (Classical and Modern Physics, Vol. 3, p.1152, Wiley 1974), among others gives the precise formulation as a = e2/(2ehc) where e is the electronic charge, e is the electric permittivity of the vacuum, and h is Planck’s constant. In this quantity a, the behaviour of the individual terms is important. For that reason it is necessary to make sure that the e term is specifically included instead of merely implied as some formulations do. Indeed, I did not specifically include it in the 1987 Report as it played no part in the discussion at that point. To illustrate the necessity of considering the behaviour of individual terms, the value of light-speed, c, has been measured as decreasing since the 17th or 18th century. Furthermore, while c was measured as declining during the 20th century, Planck’s constant, h, was measured as increasing. However, deep space measurements of the quantity hc revealed this to be invariant over astronomical time. The data obtained from these determinations can be found tabulated in the 1987 Report The Atomic Constants, Light, and Time by T. Norman and B. Setterfield. Since c has been measured as declining, h has been measured as increasing, and hc shown to be invariant, the logical conclusion from this observational evidence is that h must vary precisely as 1/c at all times. If there is any change in a, this observational evidence indicates it can only originate in the ratio e2/e. This quantity is discussed in detail in the paper "General Relativity and the Zero Point Energy."
Setterfield: In answer to the question, the quantity being measured here is the fine structure constant, alpha. This is made up of 4 other atomic quantities in two groups of two each. The first is the product of Planck's constant, h, and the speed of light, c. Since the beginning of this work in the 1980's it has been demonstrated that the product hc is an absolute constant. That is to say it is invariant with all changes in the properties of the vacuum. Thus, if h goes up, c goes down in inverse proportion. Therefore, the fine structure constant, alpha, cannot register any changes in c or h individually, and, as we have just pointed out, the product hc is also invariant.
As a consequence, any changes in alpha must come from the other ratio involved, namely the square of the electronic charge, e, divided by the permittivity of free space. Our work has shown that this ratio is constant in free space. However, in a gravitational field this ratio may vary in such a way that alpha increases very slightly. If there are any genuine changes in alpha, this is the source of the change. The errors in measurement in the article that raised the query cover the predicted range of the variation on our approach. The details of what we predict and a further discussion on this matter are found in our article in the Journal of Theoretics entitled "General Relativity and the Zero Point Energy."
Setterfield: It has been stated on a number of occasions that Supernova 1987A in the Large Magellanic Cloud (LMC) has effectively demonstrated that the speed of light, c, is a constant. There are two phenomena associated with SN1987A that lead some to this erroneous conclusion. The first of these features was the exponential decay in the relevant part of the light-intensity curve. This gave sufficient evidence that it was powered by the release of energy from the radioactive decay of cobalt 56 whose half-life is well-known. The second feature was the enlarging rings of light from the explosion that illuminated the sheets of gas and dust some distance from the supernova. We know the approximate distance to the LMC (about 165,000 to 170,000 light years), and we know the angular distance of the ring from the supernova. It is a simple calculation to find how far the gas and dust sheets are from the supernova.
Consequently, we can calculate how long it should take light to get from the supernova to the sheets, and how long the peak intensity should take to pass.
The problem with the radioactive decay rate is that this would have been faster if the speed of light was higher. This would lead to a shorter half-life than the light-intensity curve revealed. For example, if c were 10 times its current value (c now), the half-life would be only 1/10th of what it is today, so the light-intensity curve should decay in 1/10th of the time it takes today. In a similar fashion, it might be expected that if c was 10c now at the supernova, the light should have illuminated the sheets and formed the rings in only 1/10th of the time at today's speed. Unfortunately, or so it seems, both the light intensity curve and the timing of the appearance of the rings (and their disappearance) are in accord with a value for c equal to c now. Therefore it is assumed that this is the proof needed that c has not changed since light was emitted from the LMC, some 170,000 light years away.
However, there is one factor that negates this conclusion for both these features of SN1987A. Let us accept, for the sake of illustration, that c WAS equal to 10c now at the LMC at the time of the explosion. Furthermore, according to the c decay (cDK) hypothesis, light-speed is the same at any instant right throughout the cosmos due to the properties of the physical vacuum. Therefore, light will always arrive at earth with the current value of c now. This means that in transit, light from the supernova has been slowing down. By the time it reaches the earth, it is only travelling at 1/10th of its speed at emission by SN1987A. As a consequence the rate at which we are receiving information from that light beam is now 1/10th of the rate at which it was emitted. In other words we are seeing this entire event in slow-motion. The light-intensity curve may have indeed decayed 10 times faster, and the light may indeed have reached the sheets 10 times sooner than expected on constant c. Our dilemma is that we cannot prove it for sure because of the slow-motion effect. At the same time this cannot be used to disprove the cDK hypothesis. As a consequence other physical evidence is needed to resolve the dilemma. This is done in "Zero Point Energy and the Redshift" as presented at the NPA Conference, June 2010, where it is shown that the redshift of light from distant galaxies gives a value for c at the moment of emission.
By way of clarification, at NO time have I ever claimed the apparent superluminal expansion of quasar jets verify higher values of c in the past. The slow-motion effect discussed earlier rules that out absolutely. The standard solution to that problem is accepted here. The accepted distance of the sheets of matter from the supernova is also not in question. That is fixed by angular measurement. What IS affected by the slow motion effect is the apparent time it took for light to get to those sheets from the supernova, and the rate at which the light-rings on those sheets grew.
Additional Note: In order to clarify some confusion on the SN1987A issue and light-speed, let me give another illustration that does not depend on the geometry of triangles etc. Remember, distances do not change with changing light-speed. Even though it is customary to give distances in light-years (LY), that distance is fixed even if light-speed is changing.
To start, we note that it has been established that the distance from SN1987A to the sheet of material that reflected the peak intensity of the light burst from the SN, is 2 LY, a fixed distance. Imagine that this distance is subdivided into 24 equal light-months (LM). Again the LM is a fixed distance. Imagine further that as the peak of the light burst from the SN moved out towards the sheet of material, it emitted a pulse in the direction of the earth every time it passed a LM subdivision. After 24 LM subdivisions the peak burst reached the sheet.
Let us assume that there is no substantive change in light-speed from the time of the light-burst until the sheet becomes illuminated. Let us further assume for the sake of illustration, that the value of light-speed at the time of the outburst was 10c now. This means that the light-burst traversed the DISTANCE of 24 LM or 2 LY in a TIME of just 2.4 months. It further means that as the travelling light-burst emitted a pulse at each 1 LM subdivision, the series of pulses were emitted 1/10th month apart IN TIME.
However, as this series of pulses travelled to earth, the speed of light slowed down to its present value. It means that the information contained in those pulses now passes our earth-bound observers at a rate that is 10 times slower than the original event. Accordingly, the pulses arrive at earth spaced one month apart in time. Observers on earth assume that c is constant since the pulses were emitted at a DISTANCE of 1 LM apart and the pulses are spaced one month apart in TIME.
The conclusion is that this slow-motion effect makes it impossible to find the value of c at the moment of emission by this sort of process. By a similar line of reasoning, superluminal jets from quasars can be shown to pose just as much of a problem on the variable c model as on conventional theory. The standard explanation therefore is accepted here.
Setterfield: Thanks for the question, its an old one. You have assumed in your question that other atomic constants have in fact remained constant as c has dropped with time. This is not the case. In our 1987 Report, Trevor Norman and I pointed out that a significant number of other atomic constants have been measured as changing lock-step with c during the 20th century. This change is in such a way that energy is conserved during the cDK process. All told, our Report lists 475 measurements of 11 other atomic quantities by 25 methods in dynamical time.
This has the consequence that in the standard equation [E = mc2] the energy E from any reaction is unchanged (within a quantum interval - which is the case in the example under discussion here). This happens because the measured values of the rest-mass, m, of atomic particles reveal that they are proportional to 1/(c2). The reason why this is so, is fully explored in Reviewing the Zero Point Energy. Therefore in reactions from known transitions, such as occurred in SN1987A with the emission of gamma rays and neutrinos, the emission energy will be unchanged for a given reaction. I trust this reply is adequate. (1/21/99)
Setterfield: It really does appear as if the Professor has not done his homework properly on the cDK (or DSLM) issue that he discussed in relation to Super Nova 1987 A. He pointed out that neutrinos gave the first signal that the star was exploding, and that neutrinos are now known to have mass. He then goes on to state (incorrectly) that neutrinos would have an increasing rest mass (or rest energy) as we go BACK into history. He then asks "if we believe in the conservation of energy, where has all this energy gone?" He concluded that this energy must have been radiated away and so should be detectable. Incredibly, the Professor has got the whole thing round the wrong way. If he had read our 1987 Report, he would have realised that the observational data forced us to conclude that with cDK there is also conservation of energy. As the speed of light DECREASES with time, the rest mass will INCREASE with time. This can be seen from the Einstein relation [E = mc2]. For Energy E to remain constant, the rest mass m will INCREASE with time in proportion to [1/ (c2)] as c is dropping. This INCREASE in rest-mass with time has been experimentally supported by the data as listed in Table 14 of our 1987 Report. There is thus no surplus energy to radiate away at all, contrary to the Professor's suggestion, and the rest-mass problem that he poses will also disappear.
In a similar way, light photons would not radiate energy in transit as their speed drops. According to experimental evidence from the early 20th century when c was measured as varying, it was shown that wavelengths, [w], of light in transit are unaffected by changes in c. Now the speed of light is given by [c = fw] where [f] is light frequency. It is thus apparent that as [c] drops, so does the frequency [f], as [w] is unchanged. The energy of a light photon is then given by [E = hf] where [h] is Planck's constant. Experimental evidence listed in Tables 15A and 15B in the 1987 Report as well as the theoretical development shows that [h] is proportional to [1/c] so that [hc] is an absolute constant. This latter is supported by evidence from light from distant galaxies. As a result, since [h] is proportional to [1/c] and [f] is proportional to [c], then [E = hf] must be a constant for photons in transit. Thus there is no extra radiation to be emitted by photons in transit as light-speed slows down, contrary to the Professor's suggestion, as there is no extra energy for the photon to get rid of.
I hope that this clarifies the matter. I do suggest that the 1987 Report be looked at in order to see what physical quantities were changing, and in what way, so that any misunderstanding of the real situation as given by observational evidence can be avoided. "Reviewing the Zero Point Energy" should also be of help.
Question: What about SN1997ff?
Setterfield: There has been much interest generated in the press lately over the analysis by Dr. Adam G. Riess and Dr. Peter E. Nugent of the decay curve of the distant supernova designated as SN 1997ff. In fact, over the past few years, a total of four supernovae have led to the current state of excitement. The reason for the surge of interest is the distances that these supernovae are found to be when compared with their redshift, z. According to the majority of astronomical opinion, the relationship between an object's distance and its redshift should be a smooth function. Thus, given a redshift value, the distance of an object can be reasonably estimated.
One way to check this is to measure the apparent brightness of an object whose intrinsic luminosity is known. Then, since brightness falls off by the inverse square of the distance, the actual distance can be determined. For very distant objects something of exceptional brightness is needed. There are such objects that can be used as 'standard candles', namely supernovae of Type Ia. They have a distinctive decay curve for their luminosity after the supernova explosion, which allows them to be distinguished from other supernovae.In this way, the following four supernovae have been examined as a result of photos taken by the Hubble Space Telescope. SN 1997ff at z = 1.7; SN 1997fg at z = 0.95; SN 1998ef at z = 1.2; and SN 1999fv also at z = 1.2. The higher the redshift z, the more distant the object should be. Two years ago, the supernovae at z = 0.95 and z = 1.2 attracted attention because they were FAINTER and hence further away than expected. This led to two main competing theories among cosmologists. First, that the faintness was due to dust, or second, that the faintness was due to Einstein’s cosmological constant – a kind of negative gravity expanding the universe progressively faster than if the expansion was due solely to the Big Bang.
The cosmological constant has been invoked sporadically since the time of Einstein as the answer to a number of problems. It is sometimes called the "false vacuum energy." However, in stating this, it should be pointed out, as Haisch and others have done, that the cosmological constant has nothing to do with the zero-point energy. This cosmological constant, lambda, is frequently used in various models of the Big Bang, to describe its earliest moments. It has been a mathematical device used by some cosmologists to inflate the universe dramatically, and then have lambda drop to zero. It now appears that it would be helpful if lambda maintained its prominence in the history of the cosmos to solve more problems. Whether it is the real answer is another matter. Nevertheless, it is a useful term to include in some cosmological equations to avoid embarrassment.
At this point, the saga takes another turn. Recent work reveals that the object SN1997ff, the most distant of the four, turns out to be BRIGHTER than expected for its redshift value. This event has elicited the following comments from Adrian Cho in New Scientist for 7 April, 2001, page 6 in an article entitled "What's the big rush?"
Well, that is one option now that dust has been eliminated as a suspect. However, the answer could also lie in a different direction to that suggested above as there is another option well supported by other observational evidence. For the last two decades, astronomer William Tifft of Arizona has pointed out repeatedly that the redshift is not a smooth function at all but is, in fact, going in "jumps", or is quantised. In other words, it proceeds in a steps and stairs fashion. Tifft's analyses were disputed, so in 1992 Guthrie and Napier did a study to disprove the matter. They ended up agreeing with Tifft. The results of that study were themselves disputed, so Guthrie and Napier conducted an exhaustive analysis on a whole new batch of objects. Again, the conclusions confirmed Tifft's contention. The quantisations of the redshift that were noted in these studies were on a relatively small scale, but analysis revealed a basic quantisation that was at the root of the effect, of which the others were simply higher multiples. However, this was sufficient to indicate that the redshift was probably not a smooth function at all. If these results were accepted, then the whole interpretation of the redshift, namely that it represented the expansion of the cosmos by a Doppler effect on light waves, was called into question. This becomes apparent since there was no good reason why that expansion should go in a series of jumps, anymore than cars on a highway should travel only in multiples of, say, 5 kilometres per hour.
However, a periodicity on a much larger scale has also been noted for very distant objects. In 1990, Burbidge and Hewitt reviewed the observational history of these preferred redshifts. Objects were clumping together in preferred redshifts across the whole sky. These redshifts were listed as z = 0.061, 0.30, 0.60, 0.96, 1.41, 1.96, 2.63 and 3.45 [G. Burbidge and A. Hewitt, Astrophysical Journal, vol. 359 (1990), L33]. In 1992, Duari et al. examined 2164 objects with redshifts ranging out to z = 4.43 in a statistical analysis [Astrophysical Journal, vol. 384 (1992), 35], and confirmed these redshift peaks listed by Burbidge and Hewitt. This sequence has also been described accurately by the Karlsson formula. Thus two phenomena must be dealt with, both the quantisation effect itself and the much larger periodicities which mean objects are further away than their redshifts would indicate. Apparent clustering of galaxies is due to this large-scale periodicity.A straightforward interpretation of both the quantisation and periodicity is that the redshift itself is going in a predictable series of steps and stairs on both a small as well as a very large scale. This is giving rise to the apparent clumping of objects at preferred redshifts. The reason is that on the flat portions of the steps and stairs pattern, the redshift remains essentially constant over a large distance, so many objects appear to be at the same redshift. By contrast, on the rapidly rising part of the pattern, the redshift changes dramatically over a short distance, and so relatively few objects will be at any given redshift in that portion of the pattern.
These considerations are important in the current context. As noted above by Reiss, the objects at z = 0.95 and z = 1.2 are systematically faint for their assumed redshift distance. By contrast, the object at z = 1.7 is unusually bright for its assumed redshift distance. Notice that the object at z = 0.95 is at the middle of the flat part of the step according to the redshift analyses, while z = 1.2 is right at the back of the step, just before the steep climb. Consequently for their redshift value, they will be further away in distance than expected, and will therefore appear fainter. By contrast, the object at z = 1.7 is on the steeply rising part of the pattern. Because the redshift is changing rapidly over a very short distance astronomically speaking, the object will be assumed to be further away than it actually is and will thus appear to be brighter than expected.
When interpreted this way, these recent results support the existence of the redshift periodicities noted by Burbidge and Hewitt, statistically confirmed by Duari et al., and described by the Karlsson formula. In so doing, they also imply that redshift behaviour is not a smooth function, but rather goes in a steps and stairs pattern. If this is accepted, it means that the redshift is not a measure of universal expansion, but must have some other interpretation.
The research that has been conducted on the changing speed of light over the last 10 years has been able to replicate both the basic quantisation picked up by Tifft, and the large-scale periodicities that are in evidence here. On this research, the redshift and light-speed are related effects that mutually derive from changing vacuum conditions. The evidence suggests that the vacuum zero-point energy (ZPE) is increasing as a result of initial expansion of the cosmos. It has been shown by Puthoff [Physical Review D 35:10 (1987), 3266] that the ZPE is maintaining all atomic structures throughout the universe. Therefore, as the ZPE increases, the energy available to maintain atomic orbits increases. Once a quantum threshold has been reached, every atom in the cosmos will assume a higher energy state for a given orbit and so the light emitted from those atoms will be bluer than those in the past. Therefore as we look back to distant galaxies, the light emitted from them will appear redder in quantised steps. At the same time, since the speed of light is dependent upon vacuum conditions, it can be shown that a smoothly increasing ZPE will result in a smoothly decreasing light-speed. Although the changing ZPE can be shown to be the result of the initial expansion of the cosmos, the fact that the quantised effects are not "smeared out" also indicate that the cosmos is now essentially static, just as Narliker and Arp have demonstrated [Astrophysical Journal vol. 405 (1993), 51]. In view of the dilemma that confronts astronomers with these supernovae, this observational alternative may be worth serious examination.
Setterfield: If a redshift is due to motion, it is not quantized; it is a smooth smearing depending on the velocity. There is not a quantized effect. We see this smooth smearing in velocities of stars, rotation of stars, and the movement of stars within galaxies. What happens is that the wavelength of the photon is stretched or contracted due to the velocity at the time of emission. Therefore the fact that the photon originates as a discrete packet of energy is irrelevent. The point that needs to be made is that in distant galaxies, photons of light have been emitted with a range of wavelengths. All these wavelengths are simultaneously shifted in jumps by the same fraction, and it is these jumps which Tifft has noted, and which are not indicative of a Doppler shift. So some other effect is at work.
Slow motion effects?
Setterfield: Since many atomic processes are faster proportional to c, but the slow motion effect at the point of reception is also operating, the combined overall result is that everything seems to proceed at the same unchanged pace.
For example, Supernova 1987A involves the radioactive decay of Cobalt 56. Since this is an atomic process, this was decaying much faster when the speed of light was higher. However, this is exactly offset by the slow motion effect when that signal comes to earth. As a consequence, the decay of Cobalt 56 in Supernova 1987A seems to have the same half-life then as it has now. Therefore no astronomical evidence for a slow motion effect in atomic processes would be expected.
Setterfield: When George Gamow (1949) proposed a beginning to the expansion which was the accepted explanation for the redshift, Hoyle derisively called it the "Big Bang." Nothing exploded or banged in Gamow's idea, though; there was simply a hot, dense beginning from which everything expanded. Hoyle put up a different model in which matter was continuously being originated.
The major objection to Gamow's "Big Bang" was that it was too close to the biblical model of creation! In fact, even up to 1985 the Cambridge Atlas of Astronomy (pp 381,384) referred to the "Big Bang" as the 'theory of sudden creation.' On 10th August, 1989, Dr. John Maddocks, editor of Nature, declared the Big Bang philosophically unacceptable in an article entitled “Down with the Big Bang” (Nature, vol. 340, p. 425)So what is the difference between the BB and the biblical model? Essentially naturalism. The Bible says God did it and secular science says it somehow just happened. However the Bible does say, twelve times, that God stretched the heavens. So that expansion is definitely in the Bible.
In the meantime, the steady state theory was effectively disproved by quasars being discovered in the mid '60's.It is interesting that Gamow's two young colleagues, Ralph Alpher and Robert Herman, predicted in a 1948 letter to Nature that the CBR temperature should found to be about 5 deg K. [ Nature, 162,774.] They predicted 10-15 degrees K. The background radiation was found, but at a much lower temperature.
The glitch for the BB right now in terms of a continuously expanding universe is the presence of the quantized redshift. On May 5th and 7th of this year, two abstracts in astrophysics were published.* In the second one, Morley Bell writes: "Evidence was presented recently suggesting that [galaxy] clusters studied by the Hubble Key Project may contain quantized intrinsic redshift components that are related to those reported by Tifft. Here we report the results of a similar analysis using 55 spiral ... and 36 Type Ia supernovae galaxies. We find that even when many more objects are included in the sample there is still clear evidence that the same quantized intrinsic redshifts are present..."This is indication that the redshift might not be a Doppler effect. Back in 1929, Hubble himself had some doubts about connecting the two. The redshift number is obtained by comparing the light received with the standard laboratory measurements for whatever element is being seen. So this is simply a difference between two measurements and there is no intrinsic connection between the redshift measurement and velocity. In fact it has been noted that at the center of the Virgo cluster the high velocity of the galaxies wipes out the quantization of the redshift.
Interestingly, when the Bible speaks of God stretching the heavens, eleven of the twelve times the verb translated "stretched" is in the past completed tense.There is another possible cause for the quantized redshift, which is explored in Zero Point Energy and the Redshift. Essentially, it has to do with the increasing zero point energy, as measured by Planck's constant. The statistics regarding that are in his earlier major paper, here.
As far as the cosmic background radiation is concerned, the first point to note is that its temperature is much lower than that initially suggested by Gamow. This leaves some room for doubt as to whether or not this is the effect he was predicting. It is possible that even in the creation scenario, with rapid expansion from an initial super-dense state, that this effect would still be seen. However, there is another explanation which has surfaced in the last year or so. It has to do with the zero point energy. In the same way that the ZPE allows for the manifestation of virtual particle pairs, such electron/positron pairs, it is also reasonable to propose that the ZPE would allow the formation of virtual tachyon pairs. Calculation has shown that the energy density of the radiation from these tachyon pairs has the same profile as that of the cosmic microwave background. So that is another point to consider.
As far as the abundance of elements is concerned, there are several anomalies existing with current BB theory. Gamow originally proposed the building up of all elements in his BB scenario. However a blockage was found in that process which was difficult to overcome. As a result, Hoyle, Burbidge, and Fowler examined the possibility of elements being built up within stars, which later exploded and spread them out among the intergalactic clouds. This proposal is now generally accepted. However, it leads to a number of problems, such as the anomalous abundance of iron in the regions around quasars. There are other problems, such as anomalous groups of stars near the center of our galaxy and the Andromeda galaxy which have high metal abundances. Because of the current approach using the production of these elements in the first generation of stars, this process obviously takes time. As a consequence, these anomalous stars can only be accounted for by collisions or cannibalization of smaller star systems by larger galaxies. There is another possible answer, however, which creationists need to consider. It has been shown that in a scenario with small black holes, such as Planck particles, the addition of a proton to the system or to a system with a negatively-charged black hole, the build-up of elements becomes possible. The blockage that element formation in stars was designed to overcome is eliminated, because neutrons can also be involved, as can alpha particles. As a consequence, is it possible to build up other elements than hydrogen and helium in the early phases of the universe. This may happen in local concentrations where negative black holes formed by the agglomeration of Planck Particles exist. Stars that form in those areas would then have apparently anomalous metal abundances. Importantly, in this scenario, if Population II stars were formed on Day 1 of Creation Week, as suggested by Job 38, and Population I stars were formed half-way through day 4, as listed in Genesis 1:14, we have a good reason why the Population I stars contain more metals than the Population II stars, as this process from the agglomeration of black holes would have had time to act.Regarding distance and age of galaxies: There is no argument that distance indicates age. This should be stated first. It was this very fact that the further out we looked, the more different the universe appeared, that caused the downfall of the Steady State model. Specifically, it was the discovery of quasars that produced this result. Importantly, quasars become brighter and more numerous the further out we look. At a redshift of around 1.7, their numbers and luminosity appear to plateau. Closer in from 1.7, their numbers and intensity decline. Furthermore, a redshift of 1.7 is also an important marker for the formation of stars. We notice starburst galaxies of increasing activity as we go back to a redshift of 1.7. At that point, star formation activity appears to reach a maximum where young, hot blue stars of Population I are being formed (therefore emitting higher amounts of UV radiation). At a redshift of 1.7, the redshift/distance relationship also undergoes a major change. The curve steepens up considerably as we go back from that point. This has caused current BB thinking to introduce some extra terms into their equations which would indicate that the rate of expansion of the cosmos has speeded up as we come forward in time from that point. On the lightspeed scenario, a redshift of 1.7 effectively marks the close of Creation Week, and so all of these above effects would be expected to taper off after that time.
* Astrophysics, abstract astro-ph/0305060
and Astrophysics, abstract astro-ph/0305112
Setterfield: The idea of the changes in orbits initially came from the work done by Velikovsky. While his data collection is remarkable, I disagree with his conclusions. For example, he talks about planet Venus in a wandering orbit, causing some catastrophes here on earth. One of the problems with this, as with all similar proposals, is that the planet Venus lies in the plane of the ecliptic -- the same as all other planets except Pluto -- while comets and similar wandering bodies move above and below the plane of the ecliptic. Furthermore, as the orbit of a wandering Venus eventually stabilized around the sun, it would still be highly elliptical. By contrast, the orbit of Venus is the most nearly circular of all the planetary orbits. Consequently, it is the least likely to have been a wanderer in the past.
A similar statement may be made about Mars, although its orbit is somewhat more elliptical than the Earth's, but not nearly as elliptical as Pluto's.
Question: In the current issue of SCIENCE (vol 301 29 August 2003 pages 116 and 117), they say, "--space and time aren't smooth at the smallest scale, but fuzzy and foamimg. Now that tantalizing prospect has vanished in a puff of gamms rays." I don't understand, from their observations, how they can conclude the above.
Setterfield: As far as the article and your question is concerned, the point that is being made is that if space is ‘fuzzy’ and foam-like at the smallest dimensions, this is going to interfere in some ways with the images that we are receiving from very distant galaxies and quasars. In other words, these images should get progressively more fuzzy themselves. The initial experiments to prove this turned out to disprove the proposition, and thereby have thrown one section of science into some confusion. On this basis, the conclusions of the article are, in fact, correct.
The changing speed of light over time is not affected by this. One reason for this is that I do not necessarily consider space to be ‘fuzzy’ or foam-like. In an earlier presentation of this work, I made use of Planck Particle Pairs in a way which may have implied agreement with this idea of fuzziness. However, a re-examination of the basis on which this was used has revealed that the presence of Planck Particle Pairs early in the history of the cosmos does not necessarily imply that they are connected with or form the fuzziness other astronomers and cosmologists have been referring to. In fact, over time, the number of Planck Particle Pairs would have dramatically reduced. As these positive and negative pairs have recombined, they have sent out pulses of radiation, which is what the Zero Point Energy is. Thus, because the Planck Particle Pairs are decreasing, the ZPE is building up as a direct result. As the ZPE builds up, the speed of light drops.
Setterfield: You state that you have not seen a satisfactory accounting for the heavy cratering in the early solar system from creationist sources. Some of this is addressed in my article, A Brief Stellar History, in part II.
Setterfield: Thank you for sending this request. I have needed a spark to get me going to respond to this. As it turns out, there is a very good response printed in New Scientist, 9th Sept. 2006, p. 12. The article is entitled "Dark Matter 'Proof' Called into Doubt."
I know not everyone has access to this article, but I do agree with what it says and will try to summarize it for you here.
Here is the opening, which may help:
"When Douglas Clowe of the University of Arizona in Tucson announced on 21 August that his team has 'direct proof of dark matter's existence,' it seemed that the issue had been settled. Now proponents of the so-called modified theories of gravity, who explain the motion of stars and galaxies without resorting to dark matter, have hit back and are suggesting that Clowe's team has jumped the gun.
"'One should not draw premature conclusions about the existence of dark matter without a careful analysis of alternative gravity theories,' writes John Moffatt, of the University of Waterloo in Ontario, Canada, who has pioneered an alternative theory of gravity known as MOG ...Moffatt claims that his MOG theory can explain the Bullet Cluster without an ounce of dark matter."
The article also mentions a number of other theories of gravity that achieve the same result. In essence, my theory (the ZPE theory) also does this. This is part of a paper which we are seeking to have published currently. If we are unsuccessful in getting this paper published, we will put it here on the web. Basically, gravity is caused by the ZPE acting on charged point-particles that give off a secondary radiation that is attractive. This has been shown by Haisch, Rueda, and Putoff to actually be the source of gravity. As I have a look at the equations that I am dealing with in this context, it turns out that there is an additional term which overcomes all of the problems of galaxy rotation and gravitational lensing. This is a direct result of ZPE theory.
Setterfield: Many thanks for forwarding this article to us; it is deeply appreciated.
Wave-particle duality, and the production of interference patterns from particles going through double slits, because of the particles wave-like behavior, links directly in with SED physics and the ZPE. Here is something from a paper I am writing:
In fact that explanation still holds for large molecules because the ZPE is real, and so is the velocity which gives a Doppler shift. Thus, there is a valid explanation in terms of SED physics. The article in question noted that the scientists involved used the massive molecules with their slowest speed so that the beat wavelength would be as long as possible and so register as a wave when going through the interference screen. All this is in line with what SED physics expects. The actual point of crossover from so-called quantum effects to the macroscopic world is not likely to be a sharp one because particle motion is the actual cause when coupled with the jiggling of the ZPE. Thus the larger the molecule, the slower must be the velocity in order to pick up the effects. In theory it should be possible to increase the size of molecules and still get the effect of interferfence, provided that the molecule's velocity was slow enough.
I trust that helps your understanding of the situation. Thanks again for bringing this article to our attention.
Question: The permeability of free space is an arbitrary number equal to 4π/107. If the permeability of free space is meant to be varying, proportional to the energy density of the ZPE, which of these numbers is varying? It is possible that you are confusing systems of units with fundamental changes since the permeability of free space is a constant in any system of units?
Setterfield: The reviewer states that
permeability has been defined as a constant. For the other readers,
permeability is the term describing the magnetic properties of space.
Interestingly, in the latter part of the nineteenth century and the early
part of the twentieth century permeability was defined in a different way,
which linked in with the speed of light. [S.G. Starling and A. J. Woodall, Physics, Longmans, Green and Co., London, 1958, p. 1262] Any change
in the speed of light, therefore, also meant a change in the permeability of
free space. I believe we have made a retrograde step in accepting the
current definition of permeability being a constant.
As to the comment about confusing changes in permeability with other systems of units several points should be noted. First, the constancy of the permeability and its current numerical value is an artifact of the rationalized MKSA system that gave rise to the SI system now in use. For example, in SI Units by B. Chiswell and E.C.M. Grigg [John Wiley and Sons, Australasia, 1971], in Appendix 1, pages 108-110 the changes that occurred in the systems of units over the last century are discussed. Interestingly, and just as Starling and Woodall noted, they point out that originally the permeability was defined as lightspeed dependent. Any change the speed of light will result from a change in the permeability of free space. As the development of the present system was occurring, the lightspeed dependence was dropped and the numerical value of the permeability underwent several changes which then finalized by the inclusion of a factor of 4π.
The second point that emerges here is that, as one looks at these developments, it becomes apparent that other alternatives may exist to an invariant permeability. It seems desirable to consider these options in a situation where lightspeed is changing cosmologically. One reason for this is that space is a non-dispersive medium. In other words, space does not split a light beam into colours as glass or water might do. In order to maintain this property, it is usually considered that the ratio between the permittivity and permeability must remain fixed. This means that the intrinsic impedance of free space, determined by this ratio, should remain unaltered. This means that it might be necessary to devise a system of units where both permittivity and permeability are lightspeed dependent.
However, the reviewer has pointed out that another way of utilizing current developments might be to consider using equations for refractive index instead. This may be an option. However, the behaviour of light waves in an inhomogeneous medium may be intrinsically different to that of light in a cosmologically changing Zero Point Energy (ZPE). Consideration of this leads on to the second item.
Setterfield: Yes there is some basis for what they are saying. The upsurge in solar storm activity is likely to occur as the sun starts to move into its active phase of the 11 year cycle. This peak is expected to be in 2012 AD. At the present moment that upswing in activity is only just beginning to show. During the solar maximum as it is called, the numbers of sunspots increase remarkably as do solar storms and solar flares. The flares emit X-rays and Ultra Violet (UV) radiation.. These can affect earth's ionisphere and disrupt radio communications, and disturb the action of some radar devices. The associated solar storms send streams of charged particles (protons) out from the sun. Protons can pass through a human body and do biological damage. These particle streams are effectively the same as a very strong electric current. When they hit our upper atmosphere, they cause auroras to occur near the poles. The currents involved in these auroras have been measured as being from 650,000 up to about one million amps. These particle streams also heat up the earth's atmosphere causing it to expand which then affects satellites in low earth orbit. If uncorrected, this causes the satellite's orbit to degrade undergoing aero-braking, so that it eventually burns up in our atmosphere.
When a stream of particles from the sun comes towards the earth, astronauts must get back to earth as quickly as possible. There is also a distinct danger to electronic equipment on satellites in earth orbit. The massive currents involved can wreck sensitive electronics. Under these circumstances, at a solar maximum, there is the possibility that the Global Positioning System of satellites used for guidance world-wide could be knocked out. In addition, any facility which relies on satellite communications can be knocked out in the same way that some bank communications did in Canada at the last maximum. This caused chaos in their banking system. There is an additional danger as long overland transmission lines for electricity (and some telephone communications) can have induced current surges from a solar storm. This has the capability of shutting down the system affected for a significant period. On 13th March 1989 in Quebec, over six million people were without electricity for 9 hours from this cause. Large computer networks can also suffer, and so can individual computers depending on circumstances.
That is the scenario that can occur when particles are sent earth-ward. That requires the particles from the sun to be sent in a specific direction, namely directly towards us. That does not always happen, and a lot of solar storms occur that do not affect earth because the particle beams are sent in a different direction. Astronomers keep a watch on the solar activity and can warn us when the particle streams are expected to impinge upon the earth. Our warning time may last as long as 48 hours or as short as 15 minutes depending on the violence of the explosions on the sun as part of the storm. In addition, not every stream of particles that comes towards us is a problem. Everything depends on the magnetic configuration. If the magnetic field of the earth and the proton stream are aligned with their magnetic fields pointing in the same direction, the chances are that the particle stream will be repelled by our magnetosphere. However, if the magnetic polarities are in opposite directions, the earth's magnetic field connects with the sun's and the particles will be injected into our magnetosphere. This forms a vast electric current, sometimes known as an "electro-jet" through our ionisphere which induces the currents in power lines down of the earth's surface and have the potential to disrupt national power grids, and electronics on the earth's surface.
In summary, yes there is a danger, but it can be prepared for. As far as personal computers are concerned, make sure that you have everything important backed up on disk or have it in printed form.
Question: As I continue to dig into this field, I have been reading Hugh Ross' defenses of the Big Bang and one of the items that caught my attention is the Tolman test for surface brightness. Dr. Ross believes that this is a definitive proof of an expanding universe. As I looked into this test, it appears that it has as an inherent assumption that the red shift is caused by universal expansion. Additionally, if I read the results correctly, theory predicted that the brightness should drop by (z+1)4 for an expanding universe but the evidence seems to point to a different result ((z+1)3). I am not sure if this is significant since a static universe would show a brightness drop of (z+1).
Setterfield: The Tolman test for surface brightness was one reason why Hubble felt that universal expansion was NOT the origin of the redshift right up until a few months before he died in 1953. Sandage's May 2009 paper details Hubble's misgivings. In case you had not caught up with it , it can be found here.
The situation is very much as you state, namely that if we really are living in an expanding universe, then the surface brightness of galaxies should decrease by a factor of (1 + z)4, where z is the redshift. The best experimental data give results for the exponent of (1 + z) to be between 2.6 and 3.4. In other words, this tells us that the brightness falls off as (1 + z)3. In order to salvage the expanding universe model, Sandage and many astronomers assume that this result can be accounted for if all galaxies were simply brighter in the past when they were younger; in other words a galaxy light evolution argument.
It may be wondered how do they get the factor (1 + z)4. The first factor of (1 + z) comes from the fact that the number of photons received per unit time from the source will be less than that at emission because of expansion. That situation also applies on the ZPE-plasma Model because the speed of light is slower now than at the time of emission. Consequently we are receiving fewer photons per second now than were emitted in one second at the source. Thus we have the same effect, but a different cause.
The second factor of (1 + z) comes from the fact that the energy of each photon is reduced by the redshift. This is also the case in the ZPE-plasma Model since the light that atoms emitted in the early universe was also less energetic (redder) because the ZPE strength was lower. Again there is the same effect but a different cause.
In an expanding universe, the apparent change in area of a galaxy results in a form of aberration which causes its brightness to drop off as (1 + z). In addition, the cosmological effects of a variety of space curvatures causes a geometrical aberration which is also proportional to the redshift. Neither of these aberrations occur in static cosmos models. With these two effects of aberration, the result is that there should be a clear (1 + z)4 drop off in surface brightness if the universe is expanding.
Currently this is not observed, only a (1 + z)3 drop off where the exponent 3 has an error in observation of +/- 0.4. In other words, the surface brightness does not decrease as rapidly as the expanding universe model predicts. Hugh Ross is therefore on uncertain ground here as he has to prop up his argument with the additional argument of galaxy brightness evolution, otherwise his argument falls to pieces.
However, the ZPE-Plasma Model is in a different situation. We have already shown above that the (1 + z) drop off rate has an exponent of 2; and there is one further factor that must now be noted. The energy density of photons was also lower at the time of emission when the ZPE had a lower strength. This resulted in a lower wave amplitude for all radiation. On this basis, then, the ZPE-Plasma Model has a drop off in galaxy brightness which is proportional to (1 + z)3. This is in accord with observation.
Addressing your specific question, however, the Tolman test cannot distinguish whether or not an initial expansion has occurred, or whether or not an oscillation of the universe is occurring now. I trust that this helps.
Further correspondence brought this from Barry:
Thanks for your reply. It got me digging around for more information. I found an excellent paper by Eric J, Lerner that is entitled "Tolman test from z = 0.1 to z = 5.5: Preliminary results challange the expanding Universe model". The results of his paper show that a static universe gives better results with the Tolman test than an expanding model. The way he assesses it is as follows. He took galaxies over a broad range of redshifts from close in to far away, just as the title to his paper suggests. He points out that previous Tolman tests have only been performed on objects that were consistently far away. What has not been done is to compare the surface brightness of similar galaxies over a broad range of redshifts. When this is done, he shows that the expanding universe model should give a curve rising up at a steep angle from zero until a redshift of about z = 1.5, following which the curve turns over and essentially becomes a horizontal line. In contrast, the static universe model should have data on a straight line. He then gives the results he obtained, and shows that the data are on a straight line over the entire redshift range. This means that the expanding universe model is not supported when galaxies of an extensive range of redshifts is considered. His paper can be found here.
I trust that this explanation makes what is being done somewhat more understandable.
Setterfield: Arp certainly believes that the redshift is quantized, so we are in agreement on that. However, the position that Arp adopts with quasars and redshifts is something different. Arp claims that quasars, nearly all of which have high redshifts (which are usually taken to represent great distance) are in fact relatively nearby objects which have been ejected from the cores of galaxies. His rationale for doing this is that some quasars are found to be in alignment with jets emanating from relatively nearby galaxies. The high redshift of these quasars is taken by Arp to show that they are very young and have not had time to develop into more mature objects like full galaxies. A number of quasars appear to be "interacting" with relatively nearby galaxies.
For what it is worth, the majority astronomical opinion is that these quasar-galaxy alignments are probably coincidental with the quasar far away and exceptionally brilliant compared with the nearby objects it is aligned with. This would give rise to the impression that they are associated with nearby objects. This contention appears to be true, but I recall several incidents some years back which were even more decisive for me. What had been done was to block out the light of the quasar itself. When that was done, it turned out that the rest of the galaxy surrounding the quasar became visible. The quasar was simply a part of the central core of a genuine galaxy. But the light from the rest of the galaxy was being swamping by the light from the quasar in its core. I have not seen any effective answer to this from Arp.
Setterfield: Many thanks for the URL's. Wal Thornhill is a plasma physicist from Australia and is very active in the field developing a number of approaches and theories. This URL is very good:
So all the problems that Thornhill has outlined initially with the gravitational model exist indeed, and his statements accurate. Even the discussion about brown dwarfs as such is correct. But he goes off the rails in linking our solar system with the Brown dwarf and subsequent capture. I believe that he is reading too much into the mythological accounts. While they do give information as to what was happening physically in the sky and between planets when currents and voltages were higher in the early solar system (because of lower ZPE), I think that the speculation of the ancients as to the cause of these effects is off-base.
I trust that this gives you a feel for what is happening. Remember this is a developing field, and some speculation is inevitable. If In can help further, please let me know.
Setterfield: You are correct in your assessment of the explanation that you included. It is very similar to the light created in transit argument, only in this case it is light coming from a subsurface layer created on Day 4. I have produced another approach consistent with nuclear fusion which can be found here in one of my papers.
However, in the last couple of years, there have been a number of developments in astronomy based on plasma physics which is just emerging from a period of suppression due to scientific politics. On one of the plasma-related models, the light from the sun does not come from nuclear fusion. In fact, plasma theorists point out that sunspots have a dark center, and we are looking deeper into the sun at that point. A darker center means that the temperature in the lower layers is lower than at the surface, so the temperature of the center of the sun may not be higher than the surface after all. If not, then the possibility of nuclear fusion fades into oblivion.
The plasma theorists point out a series of facts about the sun and the solar wind and the solar corona, as well as the shape of the curve describing the color and intensity of light from stars known as the Hertzsprung-Russell (H-R) diagram. These facts all point to the sun and stars being powered by galactic electric currents with the surface of the stars being plasma in arc mode due to the current. The form of the H-R diagram is precisely defined by plasma theory with electric currents. Even the sunspot cycle can be shown to be due to the blocking action of the plasmasphere-magnetosphere of the planet Jupiter. It is up to 7 million miles in diameter, and when Jupiter moves in its orbit to a critical direction for the current, the effects are seen. Jupiter's orbital period closely matches the sunspot cycle.
If this is in fact the case, then it can be easily demonstrated that every star lit up as soon as it was formed because the electric current in the plasma filament which formed the star (and our solar system) has continued to flow and produce the arc mode brilliance in the plasma on the stellar surface. No time delay; no necessity for nuclear fusion; no problems with a creation scenario. For more about the electric sun and the plasma approach, you might like to visit this website on The Sun, and the work (especially The Electric Sky) of Professor of electrical engineering, Donald E. Scott.
If you follow these links to related sites and study the literature, you will see that big possibilities exist for a major paradigm shift in astrophysics on this matter.
I trust that gives you several options to pursue. If there are any problems or further questions, please get back to me.
Setterfield: My reply on petroleum is also valid here, since plasma physics does give us an idea of where the water comes from. Remember the filament that formed our solar system and the daughter filaments which formed the planets? As indicated before, there is sorting of ions in those filaments by the process of Marklund convection, and that process concentrates hydrogen ions and oxygen ions close together. Since hydrogen is reactive, it will combine with the oxygen to form water in concentrations that depend on where the planet formed in the main filament. As a result, the plasma model does not need comets or asteroidal debris or dust particles to bring in the water from the depths of the solar system as the standard model does. It is all much simpler with just two levels of Marklund convection to consider.
Setterfield: The Big Bang is not the only scenario in which a CMBR might be produced. It also results from a plasma universe/ZPE (Zero Point Energy) model that I presented to my astronomy students. The reasoning goes along the following lines, but be a little patient with me as we lead into the scenario.
A number of researchers have pointed out that the vacuum Zero Point Energy (ZPE) maintains atomic orbit stability. This has been done by L. de la Pena "Stochastic Processes Applied to Physics", p.428-581), Ed. B. Gomez (world Scientific 1983); Hal Puthoff in Physical Review D 35(10):3266 (1987), V. Spika et al, "Beyond the Quantum," pp.247-270, Ed. T.M. Nieuwenhuizen, (World Scientific, 2007).
From all these analyses, it becomes apparent that, as the ZPE strength increases, so, too, must atomic orbit energies and the light emitted from atomic orbit transitions is thereby more energetic. This means that light emitted by atoms becomes bluer with time. Thus as we look back in time, emitted light will be redder and this is a likely cause of the redshift. Spicka's analysis allows us to quantify that effect including suspected redshift quantization or periodicities. This was done in my NPA Conference paper in June 2010 entitled "Zero Point Energy and the Redshift." This paper links the physics of the atomic behavior induced by the ZPE with the redshift data.
An alternate way of achieving the same result comes from an entirely theoretical approach, which can be done in two ways. One starts with the standard redshift equation and works back to show the conditions which existed to produce the ZPE, since its production and build-up has a known cause. That was done in "The Redshift and the Zero Point Energy."
At this point several things emerge. If the redshift is an inverse measure of the strength of the ZPE, then initially, when the redshift was very high, the ZPE strength was very low. Second, in the earlier days of our cosmos, the ZPE was so low that atomic orbits for electrons could not be sustained. If electrons could not become associated with atomic nuclei, then the state of matter had to be a plasma, with negative electrons and positive nuclei wandering around independently. During this era, plasma physics alone was in operation and the electrons and nuclei scattered light photons back and forth, just like the interior of a fog. It was only as the ZPE built up with time that atomic orbits could be sustained and neutral atoms form. And, here is the crunch point, it is only once neutral atoms form that light can penetrate or escape from this initial dense plasma. Then, looking back on that event, namely the clearing of the "initial fog," we see what we call the Cosmic Microwave Background Radiation, the CMBR. This origin for the CMBR can thereby be anchored in plasma physics, the ZPE behavior and its effects on atoms, not in the Big Bang. Some recent results suggest that the cooler parts of the CMB can now be traced to voids between the plasma filaments which form the clusters of galaxies.
So this approach is one of a number of possible alternative explanations for the CMBR's existence.
Setterfield: By way of clarification, let me first give you an illustration. When you are traveling in a car through thick fog, the light from your car is scattered by the fog particles so that you essentially cannot see any distance. After traveling like this for a while, you may suddenly come out of the fog and then see clearly. After traveling a small distance, you are able to look back at the fog-bank and see its uniform character.
Let us now upgrade our picture to the universe. The early universe consisted entirely of plasma at a high temperature with a low ZPE. These two facts together ensured that it remained as plasma for some little time. In this situation, plasma acted in the same way as fog. All the ions and electrons in the plasma uniformly scattered light and all radiation so that the result was the same as being in a fog. Now the universe was being expanded out, which cooled the temperature, and at the same time increased the strength of the ZPE. When the temperature cooled sufficiently AND the ZPE strength was high enough, atoms formed out of the plasma. Now remember that plasma will still act as a plasma even if it is only 1% ionized. In other words a gas will still act as a plasma even if 99% has become neutral atoms, that is ions with electrons now orbiting them.
Now until neutral atoms start forming, the plasma is opaque to radiation, just like the fog we travel in. However, once neutral atoms have formed, the plasma becomes transparent to radiation, and the radiation streams out of the plasma. As we move further ahead in time, we can look back at this plasma like looking back at the uniform bank of fog. The uniform bank of fog with the radiation streaming out is the CMBR. This radiation is very smooth and uniform to one part in 100,000, so this CMBR "fog bank" presents an extremely consistent appearance from all parts of the sky.The characteristic feature of this radiation is the same as that of a perfectly black body radiating at a temperature of 2.7 degrees Kelvin.
The next part in the picture comes from the fact that if you are moving towards a radiation source, that radiation is blue-shifted somewhat by an amount dependent upon the velocity at which you are traveling. This is a genuine Doppler shift. The converse is also true. If you are traveling away from a source of radiation, the radiation will appear red-shifted from its normal value. Another way of looking at this is to say that when we travel towards a source of radiation, it appears slightly warmer in that direction (as bluer radiation is warmer radiation). Conversely, if we are moving away from the radiation source it appears slightly cooler (since redder radiation is cooler radiation). In the case of the CMBR, this radiation is coming uniformly from all parts of the sky. So if we are traveling in one direction, the CMBR in that direction is slightly bluer or warmer, while in the opposite direction we are moving away so the radiation in that direction looks redder or cooler.
Therefore, the amount of "blueing" or "warming" with its converse in the opposite direction, then gives us the absolute velocity and direction of motion. From this we can pick up the absolute motion of our solar system through space; our galaxy through space; and our Local Group of galaxies, which turns out to be in the direction of the Virgo cluster. At the same time it can be seen that the CMBR gives us an absolute reference frame for the whole universe. This, then, denies the veracity of Einstein's postulate that there is no absolute reference frame for any motion anywhere in the universe. The comments of Martin Harwit showing that relativity can only apply to the atomic frame of reference, follow from this automatically.
I trust that this helps. If you need further explanation or have further questions, please get back to me
Setterfield: As far as the CMBR is concerned, you are correct...BUT...remember we can only pick up the CMBR by the radiation it has emitted, and that radiation has traveled through space to get to us from the CMBR. Therefore, we are seeing the "bank of fog" back there because of the radiation that was emitted as the light was bursting through the fog, and that radiation forms a continuous stream from the fog-bank to us. In that sense, this radiation is "all around us." It is similar to the light we see from a galaxy at the fromtiers of the cosmos. We see the galaxy "way out there", but only because there is a continuous stream of light waves (or particles) coming from that galaxy to us.
If you need further help, please don't hesitate to ask.
Thanks for your question and thanks for drawing my attention to this website again. I have been mentioned many times on that site, not usually in a good context! This article is based on the majority model for pulsars among astronomers today. If that model is not correct, then neither are the conclusions which that website has drawn. So we need to examine the standard model in detail. On that model we have a rapidly rotating, small and extremely dense neutron star which sends out a flash like a lighthouse every time it rotates. Rotation times are extremely fast on this model. In fact, the star is only dense enough to hold together under the rapid rotation if it is made up of neutrons. Those two facts alone present some of the many difficulties for astronomers holding to the standard model. Yet despite these difficulties, the model is persisted with and patched up as new data comes in. Let me explain.
First a number of university professionals have difficulty with the concept of a star made entirely of neutrons, or neutronium. In the lab, neutrons decay into a proton and electron in something under 14 minutes. Atom-like collections of two or more neutrons disrupt almost instantaneously. Thus the statement has been made that "there can be no such entity as a neutron star. It is a fiction that flies in the face of all we know about elements and their atomic nuclei." [D.E. Scott, Professor & Director of Undergraduate Program & Assistant Dept. Head & Director of Instructional Program, University of Massachusetts/Amherst]. He, and a number of other physicists and engineers remain unconvinced by the quantum/relativistic approach that theoretically proposed the existence of neutronium.They point out that it is incorrect procedure to state that neutronium must exist because of the pulsars behavior; that is circular reasoning. So the existence of neutronium itself is the first problem for the model.
Second, there is the rapid rate of rotation. For example, X-ray pulsar SAX J1808.4-3658 flashes every 2.5 thousanths of a second or about 24,000 revs per minute. This goes way beyond what is possible even for a neutron star. In order for the model to hold, this star must have matter even more densly packed than neutrons, so "strange matter" was proposed. Like neutronium, strange matter has never been actually observed, so at this stage it is a non falsifiable proposition. So the evidence from the data itself suggests that we have the model wrong. If the model is changed, we do not need to introduce either the improbability of neutronium or the even worse scenario of strange matter.
Third, on 27th October, 2010, in Astronomy News, a report from NRAO in Socorro, New Mexico was entitled "Astronomers discover most massive neutron star yet known. This object is pulsar PSR J1614-2230. It "spins" some 317 times per second and, like many pulsars, has a proven companion object, in this case, a white dwarf. This white dwarf orbits in just under 9 days. The orbital characteristics and data associated with this companion shows that the neutron star is twice as massive as our sun. And therein lies the next problem. Paul Demorest from NRAO in Tucson stated: "This neutron star is twice as massive as our Sun. This is surprising, and that much mass means that several theoretical models for the internal composition of neutron stars are now ruled out. This mass measurement also has implications for our understanding of all matter at extremely high densities and many details of nuclear physics." In other words, here is further proof that the model is not in accord with reality. Rather than rethink all of nuclear physics and retain the pulsar model, it would be far better to retain nuclear physics and rethink what is happening with pulsars.
In rethinking the model, the proponents of one alternative that has gained some attention point out some facts about the pulse characteristics that we observe in these pulsars. (1) The duty cycle is typically 5% so that the pulsar flashes like a strobe light. The duration of each pulse is only 5% of the length of time between pulses. (2) Some individual pulses vary considerably in intensity. In other words, there is not a consistent signal strength. (3) The pulse polarization indicates that it has come from a strong magnetic field. Importantly, all magnetic fields require electric currents to generate them. These are some important facts. Item (2) alone indicates that the pulsar model likened to a lighthouse flashing is unrealistic. If it was a neutron star with a fixed magnetic field, the signal intensity should be constant. So other options should be considered. Taken together, all these characteristics are typical of an electric arc (lightning) discharge between two closely spaced objects. In fact electrical engineers have known for many years that all these characteristics are typical of relaxation oscillators. In other words, in the lab we can produce there precise characteristics in an entirely different way. This way suggests a different, and probably more viable model. Here is how D.E. Scott describes it:
The paper outlining these proposals was entitled "Radiation Properties of Pulsar Magnetospheres: Observation, Theory and Experiment" and appeared in Astrophysics and Space Science 227(1995):229-253. Another paper outlining a similar proposal using a white dwarf star and a nearby planet instead of a double star system was published by Li, Ferrario and Wickramasinghe. It was entitled "Planets Around White Dwarfs" and appeared in the Astrophysical Journal 503:L151-L154 (20 August 1998). Figure 1 is pertinent in this case. Another paper by Bhardwaj and Michael, entitled the "Io-Jupiter System: A Unique Case of Moon-Planet Interaction" has a section devoted to exploring this effect in the case of Binary stars and Extra-Solar Systems. An additional study by Bhardwaj et al also appeared in Advances in Space Research vol 27:11 (2001) pp. 1915-1922. The whole community of plasma physicists and electrical engineers in the IEEE accept these models or something similar for pulsars rather than the standard neutron star explanation.
Well, where is this heading? The question involved the slow-down in the speed of light in the context of pulsars and their "rotation rate." If pulsars ar not rotating neutron stars at all, but rather involve a systematic electrical discharge in a double star or star and planet system with electric currents in a plasma or dusty disk, then the whole argument breaks down. In fact if the electric duscharge model is followed, then the paper on "Reviewing a Plasma Universe with Zero Point Energy" is extremely relevant. The reason is that an increasing ZPE not only slows the speed of light, but also reduces voltages and electric current strengths. When that is factored into the plasma model for pulsars the rate of discharge seen from earth will remain constant, as the slow-down of light cancels out the initial faster rate of discharge in the pulsar system when currents were higher.
I hope that this answers your question in a satisfactory manner.