• statistics in Roberts. Was: RAW vs. raw image format

    From Rich Ulrich@21:1/5 to anton.txt@gmail.moc on Mon Feb 27 16:00:20 2023
    XPost: alt.usage.english

    Cross-posted to sci.stat.math
    to see if anyone has comments.

    On Sat, 25 Feb 2023 00:13:53 +0300, Anton Shepelev
    <anton.txt@gmail.moc> wrote:

    Rich Ulrich:

    <snip, about computers>

    Glad to find you here! I vaguely remember you were present
    in a statiscits newsgroup, but I can't find it now. Would
    you be interested in discussing Tom Roberts's statistical
    analysis of the Dayton Miller aether-drift experiments? It
    requires some light preparatory reading, but the analysis
    itself occupies about two pages in Section IV of this
    article:

    https://arxiv.org/vc/physics/papers/0608/0608238v2.pdf

    Since Roberts did not publish his data and code, his
    conclusions have zero reproducibility, but I need help in
    understanding the procedure and validity of this analysis as
    described. If you are interested, could we continue in an
    more appropriate newsgroup.

    I've cross-posted to a .stat group that has a few readers left.

    I read the citation, and I'm not very interested. - I know too little
    about the device, etc., or about the ongoing arguments that
    apparently exist.

    I can say a few things about the paper and the analyses.

    Modern statistical analyses and design sophistication for statistics
    were barely being born in 1933, when the Miller experiment was
    published. In regards to complications and pitfalls, Time series is
    worse than analysis of independent points; and what I think of
    as 'circular series' (0-360 degrees) is worse than time series. I once
    had a passing acquaintance with time series (no data experience)
    but I've never touched circular data.

    Also, 'messy data' (with big sources of random error) remains a
    problem with solutions that are mainly ad-hoc (such as, when
    Roberts offers analyses that drop large fractions of the data).

    Roberts shows me that these data are so messy that it is hard
    to imagine Miller retrieveing a tiny signal from the noise, if Miller
    did nothing more than remove linear trends from each cycle. I
    would want to know how the DEVICE made all those errors possible,
    as a clue to how to exclude their influence on an analysis. If
    Miller's data has something, Miller didn't show it right. Roberts
    offers an alternative analysis, one that I'm too ignorant to fault.

    If you are wondering about how he fit his model, I can say a
    little bit. The usual fitting in clinical research (my area) is with least-squares multiple regression, which minimizes the squared
    residuals of a fit. The main alternative is Maximum Likelihood,
    which finds the maximum likelihood from a Likelihood equation.
    That is evaluated by chi-squared ( chisquared= -2*log(likelihood) ).
    Roberts seems to be using some version of that, though I didn't
    yet figure out what he is fitting.

    I thought it /was/ appropriate that he took the consecutive
    differences as the main unit of analysis, given how much noise
    there was in general. From what I understood of the apparatus,
    those are the numbers that are apt to be somewhat usable.

    Ending up with a chi-squared value of around 300 for around
    300 d.f. is appropriate for showing a suitably fitted model -- the
    expected value of X2 by chance for large d.f. is the d.f. A value
    much larger indicates poor fit; much smaller indicates over-fit.

    --
    Rich Ulrich

    --- SoupGate-Win32 v1.05
    * Origin: fsxNet Usenet Gateway (21:1/5)
  • From David Duffy@21:1/5 to Rich Ulrich on Tue Feb 28 07:47:03 2023
    XPost: alt.usage.english

    In sci.stat.math Rich Ulrich <rich.ulrich@comcast.net> wrote:
    Cross-posted to sci.stat.math
    to see if anyone has comments.

    On Sat, 25 Feb 2023 00:13:53 +0300, Anton Shepelev
    <anton.txt@gmail.moc> wrote:

    Rich Ulrich:

    <snip, about computers>

    you be interested in discussing Tom Roberts's statistical
    analysis of the Dayton Miller aether-drift experiments? It
    https://arxiv.org/vc/physics/papers/0608/0608238v2.pdf

    Since Roberts did not publish his data and code, his
    conclusions have zero reproducibility, but I need help in
    understanding the procedure and validity of this analysis as
    described. If you are interested, could we continue in an
    more appropriate newsgroup.

    This is a quick and dirty analysis in the R stats package. The script
    below should be fed into R, so you can look at the plots, which
    give a feel for what I did. The generalized additive mixed model I
    fitted suggests there is a significant sine wave looking trend
    as the interferometer is rotated, with
    Approximate significance of smooth terms:
    edf Ref.df F p-value
    s(dirs) 4.206 4.206 9.572 1.92e-07

    Here dirs is the 16 directions, the edf is the fitted degree of
    spline, which when you plot it peaks at 0 and 180 degrees, and
    the random effect is a separate intercept for each of the 20
    rotations. I don't see how one can exclude other peculiarities
    of the setup. The more recent meta-analyses of all such experiments
    usually accept this result at face value, but demonstrate it is an outlier
    when compared to more recent experiments that have greater precision.

    require(locfit)
    require(gamm4)

    nrotations <- 20
    nrunlen <- 17
    miller <- c(10,11,10,10,9,7,7,8,9,9,7,7,6,6,5,6,7,
    7,7,6,5,4,4,4,3,2,3,3,4,1,1,1,0,1,
    1,1,0,-1,-2,-3,-2,-2,-2,-1,-1,-2,-3,-3,-5,-4,-4,
    -4,-5,-5,-6,-6,-6,-7,-6,-6,-7,-9,-9,-10,-10,-10,-11,-13,
    -13,-15,-15,-16,-17,-19,-19,-18,-17,-17,-18,-19,-19,-18,-17,-16,-15,
    0,0,0,0,0,0,0,1,4,6,7,8,9,9,10,10,8,
    8,7,5,5,3,3,3,4,5,5,5,4,1,0,-1,-1,-2,
    -2,-2,-3,-3,-2,-2,-1,-1,-2,-3,-5,-7,-9,-9,-11,-12,-11,
    -11,-11,-11,-12,-14,-14,-11, -10, -10,-9,-9,-8,-10,-10,-10,-10,-10,
    8,8,8,7,7,6,6,5,4,4,3,1,0,0,-2,-3,-1,
    -1,-1,-1,-2,-3,-3,-2,-2,-2,-1,0,-1,-2,-1,0,0,0,
    0,1,1,1,1,3,4,6,7,7,7,9,9,9,9,8,9,
    9,10,10,10,10,9,9,9,10,10,9,9,9,8,7,7,7,
    7,8,9,8,9,9,9,10,11,12,12,12,11,11,11,11,10,
    10,10,10,8,5,4,3,3,5,4,3,1,1,0,0,0,0,
    0,0,-1,-1,-2,-3,-3,-5,-5,-5,-5,-5,-6,-6,-6,-5,-4,
    -4,-5,-5,-4,-5,-6,-6,-5,-5,-6,-6,-6,-7,-7,-8,-9,-10,
    -10,-10,-11,-11,-12,-12,-11,-10,-10,-10,-10,-11,-11,-11,-12,-12,-12,
    -12,-13,-14,-15,-15,-16,-15,-16,-15,-15,-16,-17,-18,-19,-18,-20,-21,
    1,1,2,1,1,2,4,5,7,7,8,7,6,5,4,4,4)
    angles <- rep(seq(0,16),20)
    repeats <- rep(seq(1,20),each=17)
    #
    # the four runs after a rezeroing
    #
    rezeroed <- c(86, 154, 324)
    runs <- rep(1,length(miller))
    runs[seq(86,153)] <- 2
    runs[seq(154:323)] <- 3
    runs[seq(324,340)] <- 4
    runs <- factor(runs)

    dups <- setdiff(17*seq(1,19)+1, rezeroed)
    run1 <- setdiff(seq(1,85),dups)
    run2 <- setdiff(seq(86,153),dups)
    run3 <- setdiff(seq(154,323),dups)
    run4 <- seq(324,340)
    nr1 <- length(run1)
    nr2 <- length(run2)
    nr3 <- length(run3)
    nr4 <- length(run4)
    nuniq <- nr1+nr2+nr3+nr4
    par(mfcol=c(2,1))
    plot(miller[-dups])

    fit_pieces <- function(sm=1) {
    res <- list()
    res$m1 <- locfit(miller[run1] ~ seq(1,nr1), alpha=sm)
    res$m2 <- locfit(miller[run2] ~ seq(nr1+1,nr1+nr2), alpha=sm)
    res$m3 <- locfit(miller[run3] ~ seq(nr1+nr2+1,nr1+nr2+nr3), alpha=sm)
    res$m4 <- locfit(miller[run4] ~ seq(nr1+nr2+nr3+1,nuniq), alpha=sm)
    res
    }
    mod <- fit_pieces(0.4)
    lines(mod$m1, lwd=2, col="red")
    lines(mod$m2, lwd=2, col="red")
    lines(mod$m3, lwd=2, col="red")
    lines(mod$m4, lwd=2, col="red")

    raw <- miller[-dups]
    runs <- runs[-dups]
    res <- c(residuals(mod$m1), residuals(mod$m2), residuals(mod$m3), residuals(mod$m4))
    dirs <- c(angles[run1], angles[run2], angles[run3], angles[run4])
    rotations <- c(repeats[run1], repeats[run2], repeats[run3], repeats[run4])

    plot(res, type="l")
    abline(v=seq(8,324,8),col="grey80")
    par(mfcol=c(1,1))
    plot(res ~ dirs, t="p", axes=F, xlab="Markers", ylab="Detrended Residual") box()
    axis(2)
    axis(1, at=c(0,4,8,12,16))
    axis(3, at=c(0,4,8,12,16),
    labels=c(0,expression(pi/2),expression(pi),
    expression(3*pi/2),expression(2*pi)))
    lines(locfit(res ~ dirs), lwd=5, col="red")
    lines(dirs[1:17], cos(4*pi*dirs[1:17]/16), col="blue", lwd=3)

    for(i in 1:nrotations) {
    lines(dirs[rotations==i], res[rotations==i])
    }
    miller2 <- data.frame(raw,res,runs,dirs,rotations)
    miller2$group <- factor(miller2$rotations)
    g1 <- gamm4(raw ~ group + s(dirs), random=~(1|group), data=miller2) summary(g1$mer)
    anova(g1$gam)
    plot(g1$gam)

    Cheers, David Duffy.

    --- SoupGate-Win32 v1.05
    * Origin: fsxNet Usenet Gateway (21:1/5)
  • From David Duffy@21:1/5 to David Duffy on Tue Feb 28 09:12:15 2023
    XPost: alt.usage.english

    In sci.stat.math David Duffy <davidd02@tpg.com.au> wrote:

    This is a quick and dirty analysis in the R stats package.
    Approximate significance of smooth terms:
    edf Ref.df F p-value
    s(dirs) 4.206 4.206 9.572 1.92e-07

    Here dirs is the 16 directions, the edf is the fitted degree of
    spline, which when you plot it peaks at 0 and 180 degrees, and
    the random effect is a separate intercept for each of the 20
    rotations.

    I was too quick quick in writing this - I needed to unpack those
    degrees of freedom into a linear decline over the rotation, due
    to the overall drift, which explains most of that signal,
    and the actual bump at 180 degrees. If I instead fit a polynomial term,
    then we can decompose the chi-square for direction into linear
    (highly statistically significant), and higher terms (weakly significant,
    in physicist parlance say 2-sigma ;)).

    g0: raw ~ group + (1 | group)
    g1: raw ~ group + poly(dirs, 1) + (1 | group)
    g4: raw ~ group + poly(dirs, 4) + (1 | group)
    npar AIC BIC logLik deviance Chisq Df Pr(>Chisq)
    g0 22 1524.7 1607.9 -740.37 1480.7
    g1 23 1498.3 1585.2 -726.13 1452.3 28.476 1 9.486e-08 ***
    g4 26 1494.0 1592.3 -721.00 1442.0 10.248 3 0.01657 *

    You can see this in the last plot, which includes confidence envelopes
    around the GAM fitted curve.

    --- SoupGate-Win32 v1.05
    * Origin: fsxNet Usenet Gateway (21:1/5)
  • From David Jones@21:1/5 to Rich Ulrich on Tue Feb 28 12:42:29 2023
    XPost: alt.usage.english

    Rich Ulrich wrote:

    Cross-posted to sci.stat.math
    to see if anyone has comments.

    On Sat, 25 Feb 2023 00:13:53 +0300, Anton Shepelev
    <anton.txt@gmail.moc> wrote:

    Rich Ulrich:

    <snip, about computers>

    Glad to find you here! I vaguely remember you were present
    in a statiscits newsgroup, but I can't find it now. Would
    you be interested in discussing Tom Roberts's statistical
    analysis of the Dayton Miller aether-drift experiments? It
    requires some light preparatory reading, but the analysis
    itself occupies about two pages in Section IV of this
    article:

    https://arxiv.org/vc/physics/papers/0608/0608238v2.pdf

    Since Roberts did not publish his data and code, his
    conclusions have zero reproducibility, but I need help in
    understanding the procedure and validity of this analysis as
    described. If you are interested, could we continue in an
    more appropriate newsgroup.

    I've cross-posted to a .stat group that has a few readers left.

    I read the citation, and I'm not very interested. - I know too little
    about the device, etc., or about the ongoing arguments that
    apparently exist.

    I can say a few things about the paper and the analyses.

    Modern statistical analyses and design sophistication for statistics
    were barely being born in 1933, when the Miller experiment was
    published. In regards to complications and pitfalls, Time series is
    worse than analysis of independent points; and what I think of
    as 'circular series' (0-360 degrees) is worse than time series. I once
    had a passing acquaintance with time series (no data experience)
    but I've never touched circular data.

    Also, 'messy data' (with big sources of random error) remains a
    problem with solutions that are mainly ad-hoc (such as, when
    Roberts offers analyses that drop large fractions of the data).

    Roberts shows me that these data are so messy that it is hard
    to imagine Miller retrieveing a tiny signal from the noise, if Miller
    did nothing more than remove linear trends from each cycle. I
    would want to know how the DEVICE made all those errors possible,
    as a clue to how to exclude their influence on an analysis. If
    Miller's data has something, Miller didn't show it right. Roberts
    offers an alternative analysis, one that I'm too ignorant to fault.

    If you are wondering about how he fit his model, I can say a
    little bit. The usual fitting in clinical research (my area) is with least-squares multiple regression, which minimizes the squared
    residuals of a fit. The main alternative is Maximum Likelihood,
    which finds the maximum likelihood from a Likelihood equation.
    That is evaluated by chi-squared ( chisquared= -2*log(likelihood) ).
    Roberts seems to be using some version of that, though I didn't
    yet figure out what he is fitting.

    I thought it was appropriate that he took the consecutive
    differences as the main unit of analysis, given how much noise
    there was in general. From what I understood of the apparatus,
    those are the numbers that are apt to be somewhat usable.

    Ending up with a chi-squared value of around 300 for around
    300 d.f. is appropriate for showing a suitably fitted model -- the
    expected value of X2 by chance for large d.f. is the d.f. A value
    much larger indicates poor fit; much smaller indicates over-fit.

    The paper is extremely difficult to understand and I have tried very
    hard.. There seems a possibility that you are over-interpreting what
    the author means by "chi-squared". I have heard some non-statistical
    experts in other fields just using "chi-squared" to mean a sum of
    squared errors. So not a formal test-statistic for comparing two models?

    The various data-manipulations, both in the original paper and this one
    are difficult to follow. My guess is that some of the stuff in this
    paper is throwing-out some information about variability in whatever
    "errors" are here. If this were a simple time series, one mainstream
    approach from "time-series analysis" would be to present a spectral
    analysis of a detrended and prefiltered version of the complete
    timeseries, to try to highlight any remaining periodicities. There
    would seem to be a possibility of extending this to remove other
    systematic effects. I think the key point here is to try to
    separate-out any isolated frequencies that may be of interest, rather
    than to average across a range of neighbouring frequencies, as may be
    going on in this paper.

    To go any further in understanding this one would need to have a
    mathematical description of whatever model is being used for the full
    data-set, together with a proper description of what the various
    parameters and error-terms are supposed to mean.

    One wonders if an attempt has been made to contact the author of the
    Roberts paper, for better information. A straightforward search in a
    few steps finds:

    Tom Roberts at Illinois Institute of Technology
    Research Professor of Physics
    630.840.2424
    tom.roberts@iit.edu

    This appears to be current. The date of the paper is not clear.

    --- SoupGate-Win32 v1.05
    * Origin: fsxNet Usenet Gateway (21:1/5)
  • From Anton Shepelev@21:1/5 to All on Fri Mar 3 23:33:25 2023
    XPost: alt.usage.english

    Rich Ulrich:

    I've cross-posted to a .stat group that has a few readers
    left.

    Sad to hear that. Usenet should be taught in school as one
    of the last heteratchical, accessible, and independent
    communication media.

    I read the citation, and I'm not very interested. - I know
    too little about the device, etc., or about the ongoing
    arguments that apparently exist.

    This little knowledge has its advantages -- you could verify
    the model for internal consistency and then comment whether
    it can be a/the right model for /any/ imaginary experiment.
    But since you are not interested -- good luck with whatever
    occupations fill you with enthusiasm, and feel free to skip
    my comments below:

    Modern statistical analyses and design sophistication for
    statistics were barely being born in 1933, when the Miller
    experiment was published. In regards to complications and
    pitfalls, Time series is worse than analysis of
    independent points; and what I think of as 'circular
    series' (0-360 degrees) is worse than time series. I once
    had a passing acquaintance with time series (no data
    experience) but I've never touched circular data.

    Futhermore, there are no time readings in Miller's data.
    Although he tried to rotate the device at a steady rate,
    irregularities were unavodable. But mark you that Miller's
    original analysis is largely of independent points, so that
    whatever linear correction he might have applied could not
    have affected the harmonical dependency of the fringe shift
    upon device orientation.

    Also, 'messy data' (with big sources of random error)
    remains a problem with solutions that are mainly ad-hoc
    (such as, when Roberts offers analyses that drop large
    fractions of the data).

    Yes. Futhermore, Roberts picked 67 of about 300 data sheets
    from different experiments performed with at different
    locations and dates, instead of the entire data from one or
    two of the best ones from Mt. Wilson, with the most
    prominent positive results. I forget how Roberts acquired
    those sheets. If he had manually to type them into the
    computer, this incompleteness may be excused. But knowing
    the importance of this seminal experiment and of his new
    analysis, he realy should have found the time, resources,
    and help to digitise the entire data. Yet, he has not put
    online even the partial data he has.

    Roberts shows me that these data are so messy that it is
    hard to imagine Miller retrieveing a tiny signal from the
    noise, if Miller did nothing more than remove linear
    trends from each cycle.

    Does he show or tell? Do you comment on the graphs of
    Miller data /after/ processing by his statistical model? It
    is the model that I should like to understand better.

    I would want to know how the DEVICE made all those errors
    possible, as a clue to how to exclude their influence on
    an analysis.

    This is an entirely different task -- an analysis of your
    own -- perhaps more interesting and productive, but
    impossible without Miller's original data. The device was a
    large, super sensitive rotatable interferometer with two
    orghogonal arms. The hypothesis tested was that, if the
    Earth moved though the aether, the speed of light was
    orientation-dependent, so that a half-periodic (in
    orientation, not in time!) signal should be detected.

    If Miller's data has something, Miller didn't show it
    right.

    Why do you think so?

    If you are wondering about how he fit his model, I can say
    a little bit. The usual fitting in clinical research (my
    area) is with least-squares multiple regression, which
    minimizes the squared residuals of a fit. The main
    alternative is Maximum Likelihood, which finds the maximum
    likelihood from a Likelihood equation.

    Exactly, and I bet it is symbolic parametrised funtions that
    you fit, and that your models include the random error
    (noise) with perhaps assumtions about its distribution. No
    so with Roberts's model, which is neither symblic nor has
    noise as an explicit term!

    That is evaluated by chi-squared
    ( chisquared= -2*log(likelihood) ).
    Roberts seems to be using some version of that, though I
    didn't yet figure out what he is fitting.

    I have a conjecture, and will discuss it with whoever agrees
    to help me. With a my friend, a data scientist, we count
    three people who find his explanation unclear.

    I thought it /was/ appropriate that he took the
    consecutive differences as the main unit of analysis,
    given how much noise there was in general. From what I
    understood of the apparatus, those are the numbers that
    are apt to be somewhat usable.

    They /are/ usable in that they still contain the supposed
    signal and less random noise (because of "multisampling").
    But you will be surprised if you look at what that does to
    the systematic error!

    Ending up with a chi-squared value of around 300 for
    around 300 d.f. is appropriate for showing a suitably
    fitted model -- the expected value of X2 by chance for
    large d.f. is the d.f. A value much larger indicates
    poor fit; much smaller indicates over-fit.

    OK. My complaint, however, is about the model that he
    fitted, and the way he did it -- by enumerating the
    combinations of the seven free parameters by sheer brute
    force. Roberts jumped smack dab into the jaws of the curse
    of dimensionality where I think nothing called for it! He
    even had to "fold" the raw data in two -- to halve the
    degrees of freedom. I wonder what he would say to applying
    that technique to an experiment with 360 measurements per
    cycle!

    Thanks for your comments, Rich.

    --
    () ascii ribbon campaign -- against html e-mail
    /\ www.asciiribbon.org -- against proprietary attachments

    --- SoupGate-Win32 v1.05
    * Origin: fsxNet Usenet Gateway (21:1/5)
  • From Anton Shepelev@21:1/5 to All on Sat Mar 4 00:46:50 2023
    XPost: alt.usage.english

    David Jones:

    The paper is extremely difficult to understand and I have
    tried very hard..

    Thank you! That makes four people who have found it
    unclear.

    There seems a possibility that you are over-interpreting
    what the author means by "chi-squared". I have heard some
    non-statistical experts in other fields just using "chi-
    squared" to mean a sum of squared errors. So not a formal
    test-statistic for comparing two models?

    That is the least problematic part. Before fitting anythng
    to anything, one must create a good model -- the
    parametrised function to fit, and make sure that function
    correctly describes physcial process.

    The various data-manipulations, both in the original paper
    and this one are difficult to follow.

    Well, I can help you with those in the original:

    Miller D.C.
    The Ether-Drift Experiment and the Determination of
    the Absolute Motion of the Earth
    Reviews of modern physics, Vol.5, July 1933
    http://ether-wind.narod.ru/Miller_1933/Miller1933_ocr.pdf

    I am sure I understand at least them, and that they are
    really simple. Ask away or just ask me to recap it for you.

    My guess is that some of the stuff in this paper is
    throwing-out some information about variability in
    whatever "errors" are here.

    I beg pardon -- do you mean the paper by Roberts or the one
    by Miller (the original)? I fear that Roberts does it, yes.
    Miller, considering the level of statistical science in
    1933, did a top-notch job. Both his graphs and results of
    mechanical harmonic analysis[1] show a dominance of the
    second harmonic in the signal, albeit at a much lower
    magnitude that initially expected.

    If this were a simple time series, one mainstream approach
    from "time-series analysis" would be to present a spectral
    analysis of a detrended and prefiltered version of the
    complete timeseries, to try to highlight any remaining
    periodicities. There would seem to be a possibility of
    extending this to remove other systematic effects.

    Actually, the sequences of consequtive interferometer "runs"
    may be considered as uninterrupted time series, with the
    reservation that the data has no time readings, because the
    experimenters did not intend it for such analysis. They
    averaged the signal between runs for each of the sixteen
    orientations. The problem of separating the systematic error
    from the signal is quite hard and, in my opinion, requires
    an accurately consturcted model, which Roberts seems to
    lack.

    I think the key point here is to try to separate-out any
    isolated frequencies that may be of interest, rather than
    to average across a range of neighbouring frequencies, as
    may be going on in this paper.

    The second harmonic is of special interest, and I will say
    for Roberts that he does try to meausre it in secions II-
    III. This question of mine, however, is specifically about
    Roberts's statistical model in secion IV.

    To go any further in understanding this one would need to
    have a mathematical description of whatever model is being
    used

    If you think Roberts does not provide even this, you confim
    my low opinion of his analysis. I thought that maybe
    Roberts was simply too clever for me to understand. If
    statisticians fail to understand his article and/or find it
    incomplete, then something may be really wrong with it.

    for the full data-set

    I think we have to separate the model and the data to which
    it is fitted and applied. Since Roberts's data is
    incomplete -- he selected 67 datasheets from different
    experiments accoding to undisclosed criteria! -- and as yet
    unpublished, I propose to focus on the model per se, that is
    the mathematics and method behind it, if any. I will peruse
    futher feedback form statisticians and then share my
    criticisms in more detail.

    together with a proper description of what the various
    parameters and error-terms are supposed to mean.

    Indeed. I too found them rather muddy, if not internally
    contradictory. Robert's model is:

    singnal(orientation) + system_error(time)

    but he seems to be confused about what he means by time. At
    one point he says it is the number of the interferometer
    revolution, at another he seems to imply that the sequence
    of sixteen readings /during/ a revolution is also time. But
    then, this kind of time includes orientation, because,
    naturally, the device rotates in time. I therefore fail to
    comprehend how this model gurrantees that the singal is not
    misinterpreted as part of systematic error. Also -- where
    is random error in the model? All in all, I am utterly
    confused by Roberts's model from the start.

    One wonders if an attempt has been made to contact the
    author of the Roberts paper, for better information. A
    straightforward search in a few steps finds:

    Yes. I had a long, yet emotional and unproductive,
    discussion with him several years ago on in relativity
    newsgroup, where he is still available. Now, I should like
    to discuss his paper in a calmer manner, and with
    statisciticians. Roberts being a physicist, I fear his
    statistics are a bit rusty, which is only too bad because
    the entire article, being a re-analysis of pre-existing
    data, is built primarily upon statisics.

    Futhermore, any decent scientific article should be
    understandable without additional help form the author, and
    contrary to J.J. Lodder -- who absurdly forbids me to
    discuss this paper "behind the author's back" -- everyone is
    entitiled and encourated to discuss published scientific
    articles without the biasing presence of their authors. I
    intended to contact Roberts again, after I had acuqired a
    better understanding of his model, to be better armed. If we
    invite Roberts now, I fear there is going to be much flame
    and little argument. I am going to be labeled a "relativity
    crank" &c. My honest intent now is to forget about
    relativity and discuss statistics.

    Thank you for the feedback, David. I begin to wonder if I
    am going to meet a statistician that understands Roberts's
    re-analysis, let alone validates his model as self-
    consistent and sound. One cannot criticise what one does not
    understand.
    ____________________
    1. E.g. Michelson's harmonic analyser:
    https://archive.org/details/pdfy-z5_uTnE-Kga9HKk6

    --
    () ascii ribbon campaign -- against html e-mail
    /\ www.asciiribbon.org -- against proprietary attachments

    --- SoupGate-Win32 v1.05
    * Origin: fsxNet Usenet Gateway (21:1/5)
  • From J. J. Lodder@21:1/5 to Anton Shepelev on Fri Mar 3 23:40:17 2023
    XPost: alt.usage.english

    Anton Shepelev <anton.txt@gmail.moc> wrote:

    Futhermore, any decent scientific article should be
    understandable without additional help form the author,

    Nonsense. Scientific articles are written for peers,
    that is, those who do not need the author's help.
    They are not intended for amateurs.
    In actual practice the number of peers may be small indeed.

    and contrary to J.J. Lodder -- who absurdly forbids me to discuss this
    paper "behind the author's back" -- everyone is entitiled and encourated
    to discuss published scientific articles without the biasing presence of their authors.

    I don't 'forbid' you, I tell you that you are misbehaving.
    Talking about somebody else behind his back,
    telling others that his work is no good, (fishy)
    when he may actually be within earshot is very bad manners indeed.
    (by standard nettiquette, and everyday manners)
    And FYI, 'fishy' in English means: dodgy, shady, suspicious,
    or even stinking, and it is a denigrating term.
    It shouldn't be used lightly.

    I intended to contact Roberts again, after I had acuqired a better understanding of his model, to be better armed. If we invite Roberts now,
    I fear there is going to be much flame and little argument. I am going to
    be labeled a "relativity crank" &c. My honest intent now is to forget
    about relativity and discuss statistics.

    I did not 'label' you a ralativity crank,
    I noted that you are one, on basis of your postings.
    I don't know whether or not Roberts would agree on that.

    And while we are at it:
    you should take this to the statistics or the relativity group.
    The material is not appropriate for AUE,

    Jan

    --- SoupGate-Win32 v1.05
    * Origin: fsxNet Usenet Gateway (21:1/5)
  • From David Jones@21:1/5 to All on Thu Mar 9 17:07:45 2023
    XPost: alt.usage.english

    <snip all background to avoid a long message>

    I'll give a little explanation for my past discussion and give some
    thoughts on some things not raised in parallel threads, which I haven't followed in detail.

    It will be obvious that I am not particularly interested in the detail
    of all this. But...

    On the statistics newsgroup we were asked for opinions of the 2006
    paper, which we started giving. My own contributions were based
    entirely on the contents of that paper ... it's description of the
    original "experiment", data collection, data analysis, etc., and of the
    "new" work contributed by the paper.

    We were later given a link to the 1933 paper, which I haven't followed
    as my internet-safety stuff blocked the link. I couldn't be bothered to
    avoid the block.

    I did later do an internet search for citations of the paper, and found
    a few. One of these is in

    https://wiki.alquds.edu/?query=Dayton_Miller

    which, being in Wikipedia, arguably places consideration of the paper
    firmly in the public domain.

    To be clear, when I wrote about "data-manipulation" I was referring to
    the whole reduction of 5.2 million data points (as stated in the above
    link) to a few hundred.

    Any data analysis has to be mindful of the potential effects of data-manipulation, and such a large-scale reduction from
    "data-cleaning" and the other manipulations makes one wonder as to the
    point of doing any analysis at all. I am particularly doubtful of the
    apparent struggle to construct a single time-series for analysis, which
    should not be necessary.

    Other threads have brought out certain details of what is unclear in
    this paper. Let me concentrate on something not yet covered.
    Specifically the model-fitting.

    Previous replies have said that the fitting was done using a sum-of-squared-errors type of objective function and that, for some
    reason, this gave something that was a discontinuous function of the
    model parameters. There is an implication that this discontinuity was
    derive from whatever allowance is made for the effect of quantisation,
    but there are no details given.

    This seems very strange. There are obvious ways of accounting for
    quantisation effects within the model fitting that would not yield a least-squares objective function but would give one that is a
    continuous function.

    It may well be that some of the data-manipulations have been applied to
    the already-quantised observations, which makes things difficult and,
    depending on the details of those manipulations, maybe impractical. But
    let's suppose that there is a simple model, with the quantisation
    applied to directly yield the data to be analysed. For example the model-structure may have a sinusoid of known period and a random
    observation error to represent what would have been observed without
    the quantisation. Then, assuming statistical independence of the random
    errors. the likelihood function for the quantised data can be found.
    This gives an objective function (to be maximised) that is a sum of
    logarithms of probabilities, where each probability refers to the
    probability of the quantised observation falling in the bin that it was observed in. These probabilities would be expressed as the difference
    of the values of a cumulative distribution function at two points that
    derive from the quantisation limits for the bin and the model
    parameters. No discontinuities involved in treating the quantisation.

    Of course, statistical independence here is very doubtful, but the
    assumption leads to an objective function for fitting that is entirely reasonable. One just has to avoid the trap of following standard maximum-likelihood theory in constructing tests of significance and
    confidence intervals. There are variants of the theory that allow for statistical dependence while still using the simple objective function,
    but it may not be worthwhile following any of these given their
    difficulty. Instead, the obvious suggestion is to apply either block-jackknifing or block-bootstrapping to get an assessment of
    uncertainty.

    The paper does give some discussion of "error-bars" but gives no
    details of how these are calculated. It may be that the effects of
    quantisation are treated as if they were random errors, which they are
    not.

    There is an obvious scientifically-valid alternative to all this, that
    is feasible in this post-modern-computing world. Depending of course on
    what you are trying to prove or disprove. You have a result from a model-fitting procedure, and that procedure can be as awful as you
    like, where that result supposedly measures the size of some effect
    that may or may not be present. The obvious thing to do is to simulate
    a large collection of sets of data, in this case each having 5.2
    million data-points, where the putative effect is absent but which
    include a good representation of all the supposed effects that your data-manipulations are supposed to remove, and then to apply those data manipulation steps before applying whatever your model-fitting
    procedure is. It would of course help if the model-fitting procedure is
    not written in an interpreted language like Java.

    But is it worth doing any further analysis at all, given that the 1933 conclusions have been disproved by later experiments?

    --- SoupGate-Win32 v1.05
    * Origin: fsxNet Usenet Gateway (21:1/5)
  • From David Duffy@21:1/5 to David Duffy on Mon Mar 13 01:23:52 2023
    In sci.stat.math David Duffy <davidd02@tpg.com.au> wrote:
    In sci.stat.math David Duffy <davidd02@tpg.com.au> wrote:

    This is a quick and dirty analysis in the R stats package.

    I was too quick quick in writing this - I needed to unpack those
    degrees of freedom into a linear decline over the rotation, due
    to the overall drift, which explains most of that signal,
    and the actual bump at 180 degrees. If I instead fit a polynomial term,

    I have put the resulting plots up at

    http://users.tpg.com.au/davidd02/

    I smoothed the trends in the data using localized regression separately
    for each time the inferometer was readjusted, and have plotted the
    resulting residuals. They appear roughly the same as Miller's plot. For
    one formal test, I have fitted a random intercept model for the (20)
    rotations, along with a fixed effects linear decline within the rotation,
    and then added higher degree polynomials to show a weakly significant non-linear term.

    Cheers, David Duffy.

    --- SoupGate-Win32 v1.05
    * Origin: fsxNet Usenet Gateway (21:1/5)
  • From Anton Shepelev@21:1/5 to All on Mon Mar 13 10:58:26 2023
    David Duffy:

    [http://users.tpg.com.au/davidd02/]
    For one formal test, I have fitted a random intercept
    model for the (20) rotations, along with a fixed effects
    linear decline within the rotation, and then added higher
    degree polynomials to show a weakly significant non-linear
    term.

    I have a question about your plot of detrended data: why do
    some rotations start at marker 1 and some at marker 0? This
    may have to do with adjustment rotations, and marker 0 is
    the same orientation as marker 16, but still I think the
    sine should be fitted to a sequences of sixteen observations
    at sixteen markers, not seventeen. The extra marker should
    be used to take adjustments into account.

    --
    () ascii ribbon campaign -- against html e-mail
    /\ www.asciiribbon.org -- against proprietary attachments

    --- SoupGate-Win32 v1.05
    * Origin: fsxNet Usenet Gateway (21:1/5)
  • From Anton Shepelev@21:1/5 to All on Mon Mar 13 10:32:58 2023
    David Duffy:

    I have put the resulting plots up at
    http://users.tpg.com.au/davidd02/

    Thank you very much, David. Your great, clean, and pure-
    HTML blog is an eye-cake.

    I smoothed the trends in the data using localized
    regression separately for each time the inferometer was
    readjusted,

    Mr. Roberts "sewed" the entire run (20 turns) into a single
    sequence of observations by joining the ends of adjustment
    turns. Would it not be a better thing to do, yielding a
    single analysable sequence?

    For one formal test, I have fitted a random intercept
    model for the (20) rotations, along with a fixed effects
    linear decline within the rotation, and then added higher
    degree polynomials to show a weakly significant non-linear
    term.

    How did you determine the phase and amplitude of the signal
    that you write "Miller was hoping for"? Altough it requries
    additional work, the time and amplitude of the expected
    signal may be estimated knowing the time and latitude.

    Observe also that Mr. Roberts has uploaded the enire dataset
    he used for his article:

    https://www.dropbox.com/sh/8z5svuenaabegoq/AAAPrjK9AOqP-yyPRr5wNBwra?dl=0

    --
    () ascii ribbon campaign -- against html e-mail
    /\ www.asciiribbon.org -- against proprietary attachments

    --- SoupGate-Win32 v1.05
    * Origin: fsxNet Usenet Gateway (21:1/5)
  • From David Duffy@21:1/5 to Anton Shepelev on Tue Mar 14 04:13:09 2023
    Anton Shepelev <anton.txt@g{oogle}mail.com> wrote:
    David Duffy:

    [http://users.tpg.com.au/davidd02/]
    For one formal test, I have fitted a random intercept
    model for the (20) rotations, along with a fixed effects
    linear decline within the rotation, and then added higher
    degree polynomials to show a weakly significant non-linear
    term.

    I have a question about your plot of detrended data: why do
    some rotations start at marker 1 and some at marker 0? This
    may have to do with adjustment rotations, and marker 0 is

    Yes, I dropped duplicate datapoints, though this nicety probably makes
    little difference. I didn't stitch the post-adjustment runs together out
    of caution, because eyeballing those you see an initial uptick before
    they return to that overall downward slope - I think the adjustment was actually bending the arm, wasn't it? This choice may discard some
    information, in the same way the random effects model might lose some information on the overall measurement drift (not the ether drift!),
    since it is being modelled as 20 rotation starts (so a piecewise way of
    dealing with nonlinearities) plus a simple linear decline over the 16
    markers. One could test a random slope for this as well, so it varies
    from rotation to rotation.

    Anyway, I think this agrees pretty well with Roberts's overall
    conclusions, though I do not explicitly estimate physical model
    parameters and confidence limits.

    THe other approach to this type of data, it seems to me, is as
    a time series analysis with "seasonality" - this strikes me as
    a very similar setup mathematically, and the software is already
    available.

    Anyway, enough from me.

    --- SoupGate-Win32 v1.05
    * Origin: fsxNet Usenet Gateway (21:1/5)
  • From Anton Shepelev@21:1/5 to All on Sun Mar 19 00:41:03 2023
    David Jones:

    If this were a simple time series, one mainstream approach
    from "time-series analysis" would be to present a spectral
    analysis of a detrended and prefiltered version of the
    complete timeseries, to try to highlight any remaining
    periodicities.

    The Miller data /are/ a time series in a way, with the
    readings as uniform as the rotation of the device. Would it
    be possible to analyse it using the SigSpec algorithm:

    https://en.wikipedia.org/wiki/SigSpec

    using the eponymous program:

    https://arxiv.org/pdf/1006.5081.pdf
    --
    () ascii ribbon campaign -- against html e-mail
    /\ www.asciiribbon.org -- against proprietary attachments

    --- SoupGate-Win32 v1.05
    * Origin: fsxNet Usenet Gateway (21:1/5)
  • From David Jones@21:1/5 to Anton Shepelev on Sun Mar 19 00:08:08 2023
    Anton Shepelev wrote:

    David Jones:

    If this were a simple time series, one mainstream approach
    from "time-series analysis" would be to present a spectral
    analysis of a detrended and prefiltered version of the
    complete timeseries, to try to highlight any remaining
    periodicities.

    The Miller data are a time series in a way,

    They are only a time-series because they have been manipulated in the
    form of a time-series. You should not remove real structure in the form
    of groups of data unless you can sure that doing so
    (a) does not remove or mask effects you are looking
    (b) does not introduce effects of the kind you are looking for.



    with the
    readings as uniform as the rotation of the device. Would it
    be possible to analyse it using the SigSpec algorithm:

    https://en.wikipedia.org/wiki/SigSpec

    using the eponymous program:

    https://arxiv.org/pdf/1006.5081.pdf

    There may be better/more-capable packages available from time-series
    analysis specialists. But you should be aware that any statistical
    tests would depend on the validity of the usual assumptions which would
    need to be given serious consideration, If you were planning on doing
    something depending in a simple way on FFTs you would need to consider
    that there is an inherent assumption that the series being analysed is
    a good representative of a stationary process (in terms of the length
    of the series being analysed). Loosely speaking, can you imagine in a
    general way how the observed time-series would have behaved before and
    after the period supposedly observed. The "time-series" in the 2006
    paper seems to show a distinct change in behaviour part way through.

    One might consider a logical way forward that doesn't place heavy
    reliance on assumptions would be to show that the apparent peak in the
    FFT, such as shown in the 2006 paper, is or is not removed when any
    explanatory effects are removed, perhaps leaving this to be judged on
    an informal basis. Even if this can be done, you could still be left
    with the problem that you are looking for an effect whose cause is indistinguishable from the effects of other causes, as previously
    identified in other literature.

    --- SoupGate-Win32 v1.05
    * Origin: fsxNet Usenet Gateway (21:1/5)
  • From Anton Shepelev@21:1/5 to All on Mon Mar 20 01:08:54 2023
    David Jones:

    The Miller data are a time series in a way,

    They are only a time-series because they have been
    manipulated in the form of a time-series.

    Not at all: each run, consisting of 20 consequtive turns
    with perhaps a few "adjustment" turns, spans about 20
    minutes and represents and single measurement of the aether-
    drift. The manipulation comprises reinstating the
    adjustments and unrolling the 20 turns of 16 observasions
    into a sequence 320 observations.

    You should not remove real structure in the form of groups
    of data

    No such structure was removed. The periodicity of individual
    turns is preserved in the observatsion indices and time
    markings. The data of a "run" is physically a time-series.

    unless you can sure that doing so
    (a) does not remove or mask effects you are looking
    (b) does not introduce effects of the kind you are
    looking for.

    I am sure the serialisation in question does neither.

    with the
    readings as uniform as the rotation of the device. Would it
    be possible to analyse it using the SigSpec algorithm:
    https://en.wikipedia.org/wiki/SigSpec
    using the eponymous program:
    https://arxiv.org/pdf/1006.5081.pdf

    There may be better/more-capable packages available from
    time-series analysis specialists.

    There may be, but SigSpec seems one of the very best, and
    specifically designed to detect significant spectral
    components in time series. It will no doubt find significant
    high-magnitude and low-frequency components in the Miller
    signal, but we are interested in whether full- and half-
    period components are prominent above the others, and by how
    much. "Multisine" analysis (like SigSpec) seems more
    unbiased in this case than the standard Fourier, with its
    fixed set of harmonics.

    But you should be aware that any statistical tests would
    depend on the validity of the usual assumptions which
    would need to be given serious consideration, If you were
    planning on doing something depending in a simple way on
    FFTs you would need to consider that there is an inherent
    assumption that the series being analysed is a good
    representative of a stationary process (in terms of the
    length of the series being analysed).

    The signal sought is stationary within each run, the noise
    is also stationary, whereas the instrumental drift is
    probably not.

    Loosely speaking, can you imagine in a general way how the
    observed time-series would have behaved before and after
    the period supposedly observed. But if we assume the
    instrumental drift to be free of any periodicity in turn,
    we may discrard spectral components whose frequencies are
    not multiples of 1/turn.

    I can imagine that about the hypothetical signal and noise,
    but not about the instrumental drift. The assumption,
    however, that it is of lower frequency than the signal, may
    help to separate one from the other.

    The "time-series" in the 2006 paper seems to show a
    distinct change in behaviour part way through.

    It does.

    One might consider a logical way forward that doesn't
    place heavy reliance on assumptions would be to show that
    the apparent peak in the FFT, such as shown in the 2006
    paper, is or is not removed when any explanatory effects
    are removed, perhaps leaving this to be judged on an
    informal basis.

    I thought that the spectral-significance (SigSpec) measure
    was made to answer such questions.

    Even if this can be done, you could still be left with the
    problem that you are looking for an effect whose cause is
    indistinguishable from the effects of other causes, as
    previously identified in other literature.

    Yes, the instrumental error itself could be periodic, but
    then it would be present with similar parameters in all
    "runs", which is not the case. Mr. Roberts made the same
    assumtion -- that the instrumental error is not periodic in
    a turn.


    --
    () ascii ribbon campaign -- against html e-mail
    /\ www.asciiribbon.org -- against proprietary attachments

    --- SoupGate-Win32 v1.05
    * Origin: fsxNet Usenet Gateway (21:1/5)
  • From David Jones@21:1/5 to Anton Shepelev on Sun Mar 19 23:14:12 2023
    Anton Shepelev wrote:

    David Jones:

    The Miller data are a time series in a way,

    They are only a time-series because they have been
    manipulated in the form of a time-series.

    Not at all: each run, consisting of 20 consequtive turns
    with perhaps a few "adjustment" turns, spans about 20
    minutes and represents and single measurement of the aether-
    drift. The manipulation comprises reinstating the
    adjustments and unrolling the 20 turns of 16 observasions
    into a sequence 320 observations.

    You should not remove real structure in the form of groups
    of data

    No such structure was removed. The periodicity of individual
    turns is preserved in the observatsion indices and time
    markings. The data of a "run" is physically a time-series.

    unless you can sure that doing so
    (a) does not remove or mask effects you are looking
    (b) does not introduce effects of the kind you are
    looking for.

    I am sure the serialisation in question does neither.

    The question will be: will anyone else be sure?

    I think I see a common approach between you and Prof. Roberts:
    "I think I see a problem, this is what I think will solve the problem,
    this is what I have done, therefore I have solved the problem"



    with the
    readings as uniform as the rotation of the device. Would it
    be possible to analyse it using the SigSpec algorithm:
    https://en.wikipedia.org/wiki/SigSpec
    using the eponymous program:
    https://arxiv.org/pdf/1006.5081.pdf

    There may be better/more-capable packages available from
    time-series analysis specialists.

    There may be, but SigSpec seems one of the very best, and
    specifically designed to detect significant spectral
    components in time series. It will no doubt find significant
    high-magnitude and low-frequency components in the Miller
    signal, but we are interested in whether full- and half-
    period components are prominent above the others, and by how
    much. "Multisine" analysis (like SigSpec) seems more
    unbiased in this case than the standard Fourier, with its
    fixed set of harmonics.

    But you should be aware that any statistical tests would
    depend on the validity of the usual assumptions which
    would need to be given serious consideration, If you were
    planning on doing something depending in a simple way on
    FFTs you would need to consider that there is an inherent
    assumption that the series being analysed is a good
    representative of a stationary process (in terms of the
    length of the series being analysed).

    The signal sought is stationary within each run, the noise
    is also stationary, whereas the instrumental drift is
    probably not.

    Loosely speaking, can you imagine in a general way how the
    observed time-series would have behaved before and after
    the period supposedly observed. But if we assume the
    instrumental drift to be free of any periodicity in turn,
    we may discrard spectral components whose frequencies are
    not multiples of 1/turn.

    I can imagine that about the hypothetical signal and noise,
    but not about the instrumental drift. The assumption,
    however, that it is of lower frequency than the signal, may
    help to separate one from the other.

    The "time-series" in the 2006 paper seems to show a
    distinct change in behaviour part way through.

    It does.

    One might consider a logical way forward that doesn't
    place heavy reliance on assumptions would be to show that
    the apparent peak in the FFT, such as shown in the 2006
    paper, is or is not removed when any explanatory effects
    are removed, perhaps leaving this to be judged on an
    informal basis.

    I thought that the spectral-significance (SigSpec) measure
    was made to answer such questions.

    You will need to get someone competent to check all the assumptions
    involved.



    Even if this can be done, you could still be left with the
    problem that you are looking for an effect whose cause is
    indistinguishable from the effects of other causes, as
    previously identified in other literature.

    Yes, the instrumental error itself could be periodic, but
    then it would be present with similar parameters in all
    "runs", which is not the case. Mr. Roberts made the same
    assumtion -- that the instrumental error is not periodic in
    a turn.

    But it is not just "instrumental errors" that need to be thought, it is
    all the possible explanations put forward by people like Shankland.

    --- SoupGate-Win32 v1.05
    * Origin: fsxNet Usenet Gateway (21:1/5)
  • From David Jones@21:1/5 to David Jones on Sun Mar 19 23:48:21 2023
    David Jones wrote:

    Anton Shepelev wrote:

    One might consider a logical way forward that doesn't
    place heavy reliance on assumptions would be to show that
    the apparent peak in the FFT, such as shown in the 2006
    paper, is or is not removed when any explanatory effects
    are removed, perhaps leaving this to be judged on an
    informal basis.

    I thought that the spectral-significance (SigSpec) measure
    was made to answer such questions.

    You will need to get someone competent to check all the assumptions
    involved.


    Let me expand on that. It seems that the "statistical tests" are based
    on asymptotic properties/results that are only valid if there is a
    stationary process to be analysed. You agreed that the observed series
    looks non-stationary. So the basic results cannot be used. However the
    package might contain something to allow some version to be applied.

    You may be hoping that a spectral analysis package will provide all
    your answers, but recall that results of the FFT are just a
    sophisticated version of regression analysis, and you may be better off
    looking to that for a way to proceed.... provided that you don't apply
    the parts of the theory of regression that are not valid here.

    --- SoupGate-Win32 v1.05
    * Origin: fsxNet Usenet Gateway (21:1/5)
  • From Anton Shepelev@21:1/5 to All on Mon Mar 20 11:58:49 2023
    David Jones to Anton Shepelev:
    David Jones to Anton Shepelev:

    The data of a "run" is physically a time-series.

    unless you can sure that doing so
    (a) does not remove or mask effects you are looking
    (b) does not introduce effects of the kind you are
    looking for.

    I am sure the serialisation in question does neither.

    The question will be: will anyone else be sure?

    Indeed, but it is hard to prove the absense of either loss
    you mention.

    If you, or anybody else, think that the serialisation of the
    turns of a run may introduce some distorition or make the
    signal otherwise less noticeable, then please share you
    specific concerns, that we may discuss whether they are
    justified.

    For my part, I can only repeat the each "run" represents
    twenty or more consequent turns of the interferometer within
    a space of 15-20 minutes. It contains 20*16+1=321
    observations made over twenty "observation turns",
    occasionally interrupted by "adjustment turns", during which
    no observations were recorded. The data, therefore, is a
    physical time series with gaps. You can view them in this
    form in the seq_t directory in this archive:

    http://freeshell.de/~antonius/file_host/RobertsMillerData.7z

    Since the signal we seek is half-periodic in a turn,
    adjustment turns do not disrupt it (in any way that I can
    think of).

    I think I see a common approach between you and Prof.
    Roberts: "I think I see a problem, this is what I think
    will solve the problem, this is what I have done,
    therefore I have solved the problem"

    Please note, that I initiated discussion of the statistical
    analysis of the Miller experiemnts in this group,
    specifically because I needed your help and advice as expert
    statisticians. Mr. Roberts, on the other hand, professes no
    such desire...

    I thought that the spectral-significance (SigSpec)
    measure was made to answer such questions.

    You will need to get someone competent to check all the
    assumptions involved.

    Can you help me first to identify those assumtptions? That
    the signal saught is stationary and periodic in a half-turn
    is a fact. Noise is not periodic. The instrumental dirft may
    be assumed to be aperiodic from looking at the measurements,
    but a specific phycial or statistical justification is
    welcome. The key point is to determine whether it may pose
    as signal or not.

    Let me expand on that. It seems that the "statistical
    tests" are based on asymptotic properties/results that are
    only valid if there is a stationary process to be
    analysed. You agreed that the observed series looks non-
    stationary. So the basic results cannot be used. However
    the package might contain something to allow some version
    to be applied.

    How does one determine whether the instrumental drift is a
    stationary process? What do you think can make that process
    non-stationary? The dominance of the basic linear drift
    during the entire run seems to indicate that it is
    statuionary within the period of the run. After consulting
    the definitiona of a stationary process, I retract my
    previous statement to the contrary.

    The SigSpec program performs a multisine analysys of a time
    series, finding its most significant spectral components (in
    no way limited to multiples of a fundamental frequency),
    their respective significance, and the residual data. This
    should work as well if the singal is stationary and the
    error is not.

    You may be hoping that a spectral analysis package will
    provide all your answers, but recall that results of the
    FFT are just a sophisticated version of regression
    analysis, and you may be better off looking to that for a
    way to proceed.... provided that you don't apply the parts
    of the theory of regression that are not valid here.

    With FFT, we know our basis beforehand. With multisine, we
    do not, which makes it less "prejudiced" to what is sought.
    If a significant half-period component appear in multisine,
    it will indicate much more than such a component in the FFT,
    where it is mathematicaly bound to appear, as Mr. Roberts
    correctly observes. Thank you for the advice about
    regression. I will think how I can apply it to the data in a
    way different from that of Mr. Roberts. Basically,j

    But it is not just "instrumental errors" that need to be
    thought, it is all the possible explanations put forward
    by people like Shankland.

    Yes, and that is another direction of research. Can such a
    termperature gradient in the room be imagined as to produce
    a half-period effect? Can this situation occur in reality?
    Is it compaible with the thermometer indications during the
    experiment? Is the themal inerita of the insulated
    interferometer arms sufficient to suppress that effect to a
    magnitude much lower than that of the observed signal?
    Unfortunately, I have not been able to read Shankland's
    original, and am only acquainted with it through the
    criticism of James DeMeo:

    http://www.orgonelab.org/miller.htm

    See there "see: Shankland Team's 1955 Critique of Miller,"
    which is quite interesting, e.g.:

    If the periodic effects observed by Miller were the
    product of temperature variations, as was claimed by
    Shankland and Joos, then why would that variation
    systematically point to the same set of azimuth
    coordinates along the celestial sidereal clock, but not
    to any single terrestrial coordinate linked to civil
    time? Miller repeatedly asked this question of his
    critics, who had no answer for it. The Shankland team
    likewise evaded the question.

    --
    () ascii ribbon campaign -- against html e-mail
    /\ www.asciiribbon.org -- against proprietary attachments

    --- SoupGate-Win32 v1.05
    * Origin: fsxNet Usenet Gateway (21:1/5)
  • From David Jones@21:1/5 to Anton Shepelev on Mon Mar 20 17:23:53 2023
    Anton Shepelev wrote:

    David Jones to Anton Shepelev:
    David Jones to Anton Shepelev:

    The data of a "run" is physically a time-series.

    unless you can sure that doing so
    (a) does not remove or mask effects you are looking
    (b) does not introduce effects of the kind you are
    looking for.

    I am sure the serialisation in question does neither.

    The question will be: will anyone else be sure?

    Indeed, but it is hard to prove the absense of either loss
    you mention.

    If you, or anybody else, think that the serialisation of the
    turns of a run may introduce some distorition or make the
    signal otherwise less noticeable, then please share you
    specific concerns, that we may discuss whether they are
    justified.

    For my part, I can only repeat the each "run" represents
    twenty or more consequent turns of the interferometer within
    a space of 15-20 minutes. It contains 20*16+1=321
    observations made over twenty "observation turns",
    occasionally interrupted by "adjustment turns", during which
    no observations were recorded. The data, therefore, is a
    physical time series with gaps. You can view them in this
    form in the seq_t directory in this archive:

    http://freeshell.de/~antonius/file_host/RobertsMillerData.7z

    Since the signal we seek is half-periodic in a turn,
    adjustment turns do not disrupt it (in any way that I can
    think of).

    I think I see a common approach between you and Prof.
    Roberts: "I think I see a problem, this is what I think
    will solve the problem, this is what I have done,
    therefore I have solved the problem"

    Please note, that I initiated discussion of the statistical
    analysis of the Miller experiemnts in this group,
    specifically because I needed your help and advice as expert
    statisticians. Mr. Roberts, on the other hand, professes no
    such desire...

    I thought that the spectral-significance (SigSpec)
    measure was made to answer such questions.

    You will need to get someone competent to check all the
    assumptions involved.

    Can you help me first to identify those assumtptions? That
    the signal saught is stationary and periodic in a half-turn
    is a fact. Noise is not periodic. The instrumental dirft may
    be assumed to be aperiodic from looking at the measurements,
    but a specific phycial or statistical justification is
    welcome. The key point is to determine whether it may pose
    as signal or not.

    Let me expand on that. It seems that the "statistical
    tests" are based on asymptotic properties/results that are
    only valid if there is a stationary process to be
    analysed. You agreed that the observed series looks non-
    stationary. So the basic results cannot be used. However
    the package might contain something to allow some version
    to be applied.

    How does one determine whether the instrumental drift is a
    stationary process? What do you think can make that process
    non-stationary? The dominance of the basic linear drift
    during the entire run seems to indicate that it is
    statuionary within the period of the run. After consulting
    the definitiona of a stationary process, I retract my
    previous statement to the contrary.

    The SigSpec program performs a multisine analysys of a time
    series, finding its most significant spectral components (in
    no way limited to multiples of a fundamental frequency),
    their respective significance, and the residual data. This
    should work as well if the singal is stationary and the
    error is not.

    You may be hoping that a spectral analysis package will
    provide all your answers, but recall that results of the
    FFT are just a sophisticated version of regression
    analysis, and you may be better off looking to that for a
    way to proceed.... provided that you don't apply the parts
    of the theory of regression that are not valid here.

    With FFT, we know our basis beforehand. With multisine, we
    do not, which makes it less "prejudiced" to what is sought.
    If a significant half-period component appear in multisine,
    it will indicate much more than such a component in the FFT,
    where it is mathematicaly bound to appear, as Mr. Roberts
    correctly observes. Thank you for the advice about
    regression. I will think how I can apply it to the data in a
    way different from that of Mr. Roberts. Basically,j


    <snip>


    Obviously we can’t hope to deal with the whole of statistical theory
    here. But we can look, in some simple cases, at the effects of dealing
    or not dealing with pre-analysis data-manipulations within the data
    analysis.

    Even the most basic statistics work relates to dealing with
    within-analysis manipulations. For example the usual formula for the
    estimated variance contains the divisor (n-1) instead of the divisor n,
    and this can be considered to be an adjustment to take account of the
    fact that you subtract-off the sample mean within the analysis.
    Similarly, in regression, the sum-of-squares is divided by (n-p) to
    take account of fitting a total of p parameters. In both cases the
    adjustment is made to get an unbiased estimate of the variance.

    So, let’s consider some pre-analysis data manipulations. Let’s assume
    you have two pairs of observations (X1,X2) and (Y1,Y2), with
    statistical independence within and between pairs. Let the theoretical
    mean of each observation in the first pair be M1, and let the
    theoretical mean of each observation in the second pair be M2. Suppose
    it is assumed the theoretical variance for each of the four
    observations is the same, and consider two cases where this is either
    known to be 1, or else it needs to be estimated. Then consider four
    versions of analyses with different pre-analysis manipulations as
    follows.

    (a) Separate analysis. Here the data being analysed consists of the two
    pairs (X1,X2) and (Y1,Y2). Then the sample-means with each pair,
    provide unbiased estimates of the two values M1 and M2, and the
    theoretical variance of each estimate is 1/2 if the variance of the observations is assumed known at 1. If the variance of observations is
    unknown, one could get and use two different estimates of that variance
    from the sample variance applied within each pair. Each such estimate
    would be unbiased.

    (b) Separate analysis, but pooled. This is the same as for (a), above,
    except that the variance of the observations is estimated by the
    average the sampling variances from the two pairs. The theoretical
    variances of the means remain the same as in (a), but one gets better
    estimates of those variances. This is achieved by making use of an
    assumed structure across the pairs (that the variances are the same).

    (c) Subtraction of means. To yield a special case of what might be done
    for longer series, suppose that a single dataset of 4 values
    (Q1,Q2,Q3,Q4) is constructed from the two pairs by subtracting the two
    sample means, giving
    Q1=(X1-X2)/2, Q2=(X2-X1)/2, Q3=(Y1-Y2)/2, Q4=(Y2-Y1)/2
    Obviously doing this prevents any estimation of the means M1 and M2.
    Applying the usual formula to get a sample variance from (Q1,Q2,Q3,Q4)
    gives an estimate that has a mean value of 2/3 when the true
    observation variance is known to be 1. To get a good (unbiased)
    estimate you have to know the structure of the pre-analysis data
    manipulation that yielded the data-to-be-analysed (Q1,Q2,Q3,Q4). In
    fact this turns out to be the pooled sample variance from the original
    pairs as in (b). Thus, not all is necessarily lost in doing
    pre-analysis data-manipulations, provided that the actual analysis
    takes account of those manipulations.

    (d) Joining of data. To emulate the data-joining of the paper and of
    your proposed analysis, we can consider dealing with a revised dataset (Z1,Z2,Z3), where
    Z1=X1, Z2=X2, Z3=Y2+X2-Y1
    Then the mean of each value is M1, and it clear that M1 can be
    estimated but not M2. One might use the sample mean of (Z1,Z2,Z3) to
    estimate M1: this estimate has a theoretical variance of 7/9. Thus this estimate is worse than the sample mean of just (Z1,Z2), which is the
    same as the sample mean of (X1,X2), whose variance is 1/2. The usual
    sample variance obtained from (Z1,Z2,Z3) has an expected value of 5/3
    when the theoretical observation variance is 1. If the usual sample
    variance obtained from (Z1,Z2,Z3) is used to estimate the variance of
    the sample mean of (Z1,Z2,Z3), this would have an expected value of 5/9
    rather than the true variance of this sample mean which is 7/9. So
    here, if one ignores the way in which (Z1,Z2,Z3) were obtained and just
    uses the usual sample estimates, we get an estimate for M1 which is
    worse (in terms of variance) than what might have been obtained by just
    using one the one sample pair (X1,X2). Moreover the usual formula would
    give estimated variances which are biased in either case of trying to
    estimate the observation variance or the variance of the sample mean.
    One might consider other estimates here, derived from (Z1,Z2,Z3), but
    whether or not one looked for optimal estimates this would involve
    taking into account the structure by which the dataset was created. To summarise, poor performance will arise from any attempt to analyse the constructed dataset without taking into account the details of how it
    was constructed. In this example, the data-manipulation throws away any
    ability to estimate an important property (M2) of one part the original
    dataset whereas retaining all the original data and the structure
    therein allows everything to be estimated.

    So my conclusion is that you should not try to merge groups of data
    into one supposedly-continuous time-series as you don’t have to do so.
    It is possible to do a combined analysis of all groups within joining
    them. Since there is just one pre-specified frequency there is no need
    to do a spectral analysis. But, if you really wanted to do a spectral
    analysis combining all groups without joining them together, this is
    certainly possible ... you just have to understand the meaning of the quantities produced in the analysis of a single series.

    --- SoupGate-Win32 v1.05
    * Origin: fsxNet Usenet Gateway (21:1/5)
  • From Anton Shepelev@21:1/5 to All on Fri Mar 24 00:11:50 2023
    Thank you for the answer, David Jones:

    Obviously we can't hope to deal with the whole of
    statistical theory here. But we can look, in some simple
    cases, at the effects of dealing or not dealing with pre-
    analysis data-manipulations within the data analysis.

    Even the most basic statistics work relates to dealing
    with within-analysis manipulations.

    Understood.

    For example the usual formula for the estimated variance
    contains the divisor (n-1) instead of the divisor n, and
    this can be considered to be an adjustment to take account
    of the fact that you subtract-off the sample mean within
    the analysis.

    In my understanding, this relates to the summands (and
    degrees of freedom) being one fewer than the elements in the
    sample.

    Similarly, in regression, the sum-of-squares is divided by
    (n-p) to take account of fitting a total of p parameters.
    In both cases the adjustment is made to get an unbiased
    estimate of the variance.

    Understood.

    So, let's consider some pre-analysis data manipulations.
    Let's assume you have two pairs of observations (X1,X2)
    and (Y1,Y2), with statistical independence within and
    between pairs. Let the theoretical mean of each
    observation in the first pair be M1, and let the
    theoretical mean of each observation in the second pair be
    M2. Suppose it is assumed the theoretical variance for
    each of the four observations is the same, and consider
    two cases where this is either known to be 1, or else it
    needs to be estimated. Then consider four versions of
    analyses with different pre-analysis manipulations as
    follows.

    (a) Separate analysis. Here the data being analysed
    consists of the two pairs (X1,X2) and (Y1,Y2). Then the
    sample-means with each pair, provide unbiased estimates of
    the two values M1 and M2, and the theoretical variance of
    each estimate is 1/2 if the variance of the observations
    is assumed known at 1. If the variance of observations is
    unknown, one could get and use two different estimates of
    that variance from the sample variance applied within each
    pair. Each such estimate would be unbiased.

    OK.

    (b) Separate analysis, but pooled. This is the same as for
    (a), above, except that the variance of the observations
    is estimated by the average the sampling variances from
    the two pairs. The theoretical variances of the means
    remain the same as in (a), but one gets better estimates
    of those variances. This is achieved by making use of an
    assumed structure across the pairs (that the variances are
    the same).

    In this case, we calculate the means separately, but then
    pool the samples together to calculate their common
    variance.

    (c) Subtraction of means. To yield a special case of what
    might be done for longer series, suppose that a single
    dataset of 4 values (Q1,Q2,Q3,Q4) is constructed from the
    two pairs by subtracting the two sample means, giving
    Q1=(X1-X2)/2, Q2=(X2-X1)/2, Q3=(Y1-Y2)/2, Q4=(Y2-Y1)/2

    I don't like the method of this subtaction, because it
    produces a reduandant dataset: Q1=-Q2 and Q3=-Q4. Since the
    size of the dataset is equal to the total size of the
    original datasets, half the information has been lost.

    Obviously doing this prevents any estimation of the means
    M1 and M2. Applying the usual formula to get a sample
    variance from (Q1,Q2,Q3,Q4) gives an estimate that has a
    mean value of 2/3 when the true observation variance is
    known to be 1.

    Indeed, but this estimate belongs to a very different
    sample. The difference of random variables is distributed
    quite unlike either variable.

    To get a good (unbiased) estimate you have to know the
    structure of the pre-analysis data manipulation that
    yielded the data-to-be-analysed (Q1,Q2,Q3,Q4).

    Yes.

    In fact this turns out to be the pooled sample variance
    from the original pairs as in (b).

    Hmmmm. I don't see why, but will take it for granted now.
    Will check later.

    Thus, not all is necessarily lost in doing pre-analysis
    data-manipulations, provided that the actual analysis
    takes account of those manipulations.

    Thou shalt know thy data.

    (d) Joining of data. To emulate the data-joining of the
    paper and of your proposed analysis, we can consider
    dealing with a revised dataset (Z1,Z2,Z3), where
    Z1=X1, Z2=X2, Z3=Y2+X2-Y1

    If the (X1,X2) represent the first turn and (Y1,Y2) the
    second turn, then the revised dataset according to the paper
    is: (X1, X2, Y1, Y2). Later Tom Roberts constucts error-
    differences:

    Ed1[0 ] = 0; Ed1[1 ] = Y1-X1
    Ed2[1/2] = 0; Ed2[3/2] = Y2-X2

    and fits their initial levels b1 and b2 to make the error
    function E

    E = b1 * Ed1 + b2 * Ed2

    as smooth as possible in terms of L2 between adjacent values
    weighted by the inverse errorbar. His calculation of the
    errorbars is another story.

    I did not propose subract the subsequences.

    Then the mean of each value is M1, and it clear that M1
    can be estimated but not M2. One might use the sample mean
    of (Z1,Z2,Z3) to estimate M1: this estimate has a
    theoretical variance of 7/9. Thus this estimate is worse
    than the sample mean of just (Z1,Z2), which is the same as
    the sample mean of (X1,X2), whose variance is 1/2. The
    usual sample variance obtained from (Z1,Z2,Z3) has an
    expected value of 5/3 when the theoretical observation
    variance is 1. If the usual sample variance obtained from
    (Z1,Z2,Z3) is used to estimate the variance of the sample
    mean of (Z1,Z2,Z3), this would have an expected value of
    5/9 rather than the true variance of this sample mean
    which is 7/9. So here, if one ignores the way in which
    (Z1,Z2,Z3) were obtained and just uses the usual sample
    estimates, we get an estimate for M1 which is worse (in
    terms of variance) than what might have been obtained by
    just using one the one sample pair (X1,X2). Moreover the
    usual formula would give estimated variances which are
    biased in either case of trying to estimate the
    observation variance or the variance of the sample mean.

    Yes, because the new data is transformed from the original,
    it has a different distribution and, generally, different
    moments.

    One might consider other estimates here, derived from
    (Z1,Z2,Z3), but whether or not one looked for optimal
    estimates this would involve taking into account the
    structure by which the dataset was created. To summarise,
    poor performance will arise from any attempt to analyse
    the constructed dataset without taking into account the
    details of how it was constructed.

    That is /as if/ that dataset were the original,
    untransformed, sample, which it is not.

    In this example, the data-manipulation throws away any
    ability to estimate an important property (M2) of one part
    the original dataset whereas retaining all the original
    data and the structure therein allows everything to be
    estimated.

    Yes.

    So my conclusion is that you should not try to merge
    groups of data into one supposedly-continuous time-series
    as you don't have to do so.

    No, I should not. But in the experiment in question[1], the
    data is truly a continuous times series, arranged in a two-
    dimensional table for presentation. Therefore, when I (and
    Mr. Roberts) merge it back into a single time series, I
    commit no fallacy nor transform the data in a statistically
    significant manner. Of course, I should not join the
    sequences of individual turns if they did not come from the
    same pool, and were not produced in the same uninterruted
    sequence of observations.

    It is possible to do a combined analysis of all groups
    within joining them.

    Without?

    Since there is just one pre-specified frequency there is
    no need to do a spectral analysis.

    Perhaps not, but I think it one of valid approaches (if not
    the best): a comparison of the various spectral components
    may provide insights into whether the one component in
    question is genuine signal or part of the noise and
    systematic device drift.

    But, if you really wanted to do a spectral analysis
    combining all groups without joining them together, this
    is certainly possible ... you just have to understand the
    meaning of the quantities produced in the analysis of a
    single series.

    So, you propose to amend my analysis by performing 20
    separate spectral analyses?
    ____________________
    1. https://freeshell.de//~antonius/file_host/Miller-EtherDrift-1933.pdf

    --
    () ascii ribbon campaign -- against html e-mail
    /\ www.asciiribbon.org -- against proprietary attachments

    --- SoupGate-Win32 v1.05
    * Origin: fsxNet Usenet Gateway (21:1/5)
  • From David Jones@21:1/5 to Anton Shepelev on Thu Mar 23 23:59:06 2023
    Anton Shepelev wrote:

    Thank you for the answer, David Jones:



    But, if you really wanted to do a spectral analysis
    combining all groups without joining them together, this
    is certainly possible ... you just have to understand the
    meaning of the quantities produced in the analysis of a
    single series.

    So, you propose to amend my analysis by performing 20
    separate spectral analyses?
    ____________________

    No, I suggest you do a single combined spoectral analysis, or that you
    do a sine-curve regression, by not pretending you have a single
    time-series. You have to understand that an ordinary spectral analysis
    is just a special case of regression.

    A usual time-series analysis would proceed on the basis that the
    relevant "times" are equally spaced and that the position/index within
    a single array can be used as a time, instead of having the time
    specified explicitly. Then when the time is needed, for example to use
    in a sine or cosine function, it is immediately available, rather than
    deriving it from the position in the array, So, as a starting position
    you would have a dataset that closely corresponds to the actual results
    of the experiment. Thus, in instances where the "joining" approach
    would replace two values with one, the two original values would be
    kept separately, and there would be no related adjustments.
    Additionally where there were actually gaps in the original
    observations, this can be included. Similarly for incomplete sequences.

    Of course this means that the usual time-series packages would not be
    useable and, specifically, not the FFT. BUt the series for your problem
    are not very long, so no strong need for the FFT.

    If you take the view that one of your objectives is to leave a set of
    data that are available for others to re-analyse, then it would be good
    to include as much explicit information as possible, without
    pre-judging what analysis might be done.

    To analyse the data, you need to have a statstical model for the
    original observations. Such a model explicitly represents what the
    modeller thinks explains the variation in the observations. Possibly
    this would be represented as an ordinary regression model, but a mixed
    fixed- and random-effects model might be considered. THere are
    reasonably standard procedures linked to regression that allow checks
    to be made on the various assumptions and to look-out for unexplained
    effects.

    To turn this into a "spectral analysis", YOu would just need to do a
    sequence of regression analyses, where each one would have a sine- and cosine-pair of a single frequency, and where the corresponding value in
    the "spectrogram" would be the sum of the square of the regression
    coefficients of the sine and cosine terms. It may be possible to
    justify reducing the compuations here, by first do a single regression
    with no sinusoid at all, and then doing the spectral analysis on the
    residuals.

    I Suggest you find an experienced statistician to undertake much of
    this. At the very least you need someone with the time and ability to
    think, and someone who will not try to force the analysis to fit within
    the compass of some existing computer package.

    --- SoupGate-Win32 v1.05
    * Origin: fsxNet Usenet Gateway (21:1/5)