Most active commenters
  • mg(6)
  • robwwilliams(6)
  • (4)
  • jrootabega(4)
  • sgc(4)
  • K0balt(3)
  • luqtas(3)
  • BolexNOLA(3)
  • CamperBob2(3)
  • eru(3)

←back to thread

817 points dynm | 93 comments | | HN request time: 2.375s | source | bottom
1. mg ◴[] No.43307263[source]
This is great. The author defines their own metrics, is doing their own A/B tests and publishes their interpretation plus the raw data. Imagine a world where all health blogging was like that.

Personally, I have not published any results yet, but I have been doing this type of experiments for 4 years now. And collected 48874 data points so far. I built a simple system to do it in Vim:

https://www.gibney.org/a_syntax_for_self-tracking

I also built a bunch of tooling to analyze the data.

I think that mankind could greatly benefit from more people doing randomized studies on their own. Especially if we find a way to collectively interpret the data.

So I really applaud the author for conducting this and especially for providing the raw data.

Reading through the article and the comments here on HN, I wish there was more focus on the interpretation of the experiment. Pretty much all comments here seem to be anecdotal.

Let's look at the author's interpretation. Personally, I find that part a bit short.

They calculated 4 p-values and write:

    Technically, I did find two significant results.
I wonder what "Technically" means here. Are there "significant results" that are "better" than just "technically significant results"?

Then they continue:

    Of course, I don’t think this
    means I’ve proven theanine is harmful.
So what does it mean? What was the goal of collecting the data? What would the interpretation have been if the data would show a significant positive effect of Theanine?

It's great that they offer the raw data. I look forward to taking a look at it later today.

replies(14): >>43307304 #>>43307775 #>>43307806 #>>43307937 #>>43308201 #>>43308318 #>>43308320 #>>43308521 #>>43308854 #>>43309271 #>>43310099 #>>43320433 #>>43333903 #>>43380374 #
2. aszantu ◴[] No.43307304[source]
I found, I take Abt. 4 days to react to stuff. Except for those cheap cooking oils, I get migraines within a time-frame of about 3 hrs.
3. brodo ◴[] No.43307775[source]
Your self-tracking syntax looks neat. How do you track if you are not in front of a computer?
replies(1): >>43307816 #
4. kortilla ◴[] No.43307806[source]
Well the issue is that an experiment with 1 person can’t prove much because you can’t have a control group.

Too much other stuff is changing in a single persons life that could account for all observed side effects.

You also have latent side effect issues. A person could smoke for 10 years, not smoke for another 10, and then conclude that smoking doesn’t cause cancer. Then they get lung cancer 20 years later.

Excellent data and statistics is not sufficient for a good experiment

replies(3): >>43307871 #>>43309673 #>>43316721 #
5. mg ◴[] No.43307816[source]
I write into a text file on my phone. And later move those lines into the file on my computer.

It's a bit cumbersome, but not too much. But yeah, I look forward to when I finally get around to writing an app for it.

replies(1): >>43309132 #
6. mg ◴[] No.43307871[source]
An experiment of 1 person can very well produce useful data.

It depends on the setup of the experiment.

Imagine an experiment where a person's thumb gets randomly hit with either a hammer or a feather once per day. And they then subjectively rate the experience. After 1000 days of collecting data, I doubt that we would wrongly come to the conclusion that the hammer treatment leads to the nicer outcome.

The setup of the Theanine experiment which is the basis of this thread looks good on first sight. I have the feeling that the interpretation could use more thought though.

replies(8): >>43307925 #>>43308004 #>>43308014 #>>43308106 #>>43308117 #>>43308394 #>>43309398 #>>43322506 #
7. ◴[] No.43307925{3}[source]
8. kqr ◴[] No.43307937[source]
> So what does it mean? What was the goal of collecting the data? What would the interpretation have been if the data would show a significant positive effect of Theanine?

I think what the author is saying is that for them to bother with theanine on a permanent basis, it would have to have shown an effect large enough to be apparent just from plotting.

In other words, they mean technical significance as opposed to clinical significance. A small effect can be statistically verifiable without being meaningful in practice.

replies(1): >>43308016 #
9. tomalbrc ◴[] No.43308004{3}[source]
What a weird take, in 99.999% of cases you don’t have such a black/white contrast
replies(1): >>43308048 #
10. ◴[] No.43308014{3}[source]
11. mg ◴[] No.43308016[source]

    an effect large enough to be
    apparent just from plotting
And how large is that? Without putting a number on it, how do we come to the conclusion that the effect is not large enough? That it didn't show in their sample of data points could have been just random chance.

But before we take the measured effects at face value, I think it's important think about them more. They report significant p values of their success in predicting if the capsule holds Theanine and also for the effect of the capsule when it holds Theanine. Both negative correlations. My first thought reading this is that the Placebo tasted more like Theanine and thinking they took Theanine had a positive effect on the outcome.

12. mg ◴[] No.43308048{4}[source]
Sure. But even when you add noise to the described experiment, you get useful data.

That is the point I am making: Experiments of a single person can be useful.

The critics of single person experiments usually come up with examples vastly different than the Theanine experiment described here. With long term experiments which are only conducted once. But the Theanine experiment was looking for a short term effect and can be conducted many times. The hammer experiment I made up would be an extreme example of this type of experiment which leads itself well to be conducted by a single person.

What I am trying to point out is that if you are a skeptic, it would be better to try and find weaknesses in the experiment at hand. Not making up completely different experiments.

13. pbhjpbhj ◴[] No.43308106{3}[source]
Then you find {imagining scenarios} that the bruising causes some effect that 'inoculates' against heart disease, or the feather carries a pathogen that later induces dementia??
14. chiefalchemist ◴[] No.43308117{3}[source]
Yes and no. The issue with a self-administered experiment is that becomes part of the experiment. Does the self-administration and the associated thoughts and beliefs affect the results? It’s not like the experimentor can issue a placebo to themselves.
replies(1): >>43309099 #
15. stevage ◴[] No.43308201[source]
They were looking for a clear effect. they did not find one.
16. matthewdgreen ◴[] No.43308318[source]
This is an N=1 trial. Dressing your N=1 trial up with lots of pseudo controls and pseudo blinding and data collection does not make it better. In fact: putting this much effort into any medication trial makes it much more likely that you’re going to be incentivized to find effects that don’t exist. I think it’s nice that the author admits that they found nothing, but statistically, worthless drugs show effects in much better-designed trials than this one: it’s basically a coin toss.
replies(12): >>43308423 #>>43308440 #>>43309081 #>>43309263 #>>43309513 #>>43309704 #>>43309722 #>>43309838 #>>43311651 #>>43313233 #>>43315219 #>>43322581 #
17. robwwilliams ◴[] No.43308320[source]
“Technically” here could imply “not corrected for multiple tests”. But the typical qualifier I use is “nominally significant” when I don’t apply an correction for multiple tests.

Or “not using a one-way t test”.

The most appropriate null hypothesis in this lovely study is “does theanine REDUCE anxiety”, not “does theanine change anxiety either up or down”.

What impressed me most is the suggestion for an improved experimental design to remove his temporal drift by using 100 pre-loaded envelopes and only decoding the results at the end.

replies(2): >>43308383 #>>43310822 #
18. azalemeth ◴[] No.43308383[source]
> The most appropriate null hypothesis in this lovely study is “does theanine REDUCE anxiety”, not “does theanine change anxiety either up or down”.

I disagree with this. You have a prior belief that theanine might reduce anxiety; if you wanted to you could codify that subjective belief and perform some variety of Bayesian hypothesis test [1] and compute a Bayes factor. The main reason that one-sided tests are advocated for is power; that is often the same as having a prior belief in disguise. Why not quantify it?

However, scientifically, if the data conclusively show that "theanine increases anxiety" that is a meaningful, non-artefactual result: it is hugely important to be sensitive to the answer 'you are wrong' and may well ironically spur development in a direction to help understand what is going on. I personally think that one sided tests are best avoided except in the case where it is physically impossible to have an effect in the other direction. Examples of this are rare, but they do occasionally exist.

[1] https://mspeekenbrink.github.io/sdam-book/ch-Bayes-factors.h...

replies(1): >>43308845 #
19. chkgk ◴[] No.43308394{3}[source]
A problem I see is that you can never be sure that it also works for other people. Just because one person reacts to a specific treatment does not mean that people on average react to the treatment in the same way. That’s the problem of having a sample size of 1. In other words: We cannot say much about whether the effect generalizes to a larger population from knowing that it has an affect on just one person.
replies(1): >>43309149 #
20. episteme ◴[] No.43308423[source]
You could argue that this N is the only N that matters though.
replies(1): >>43308561 #
21. robwwilliams ◴[] No.43308440[source]
Complete injustice to this lovely study. Why do you say unblinded? Why do you insult a time series study as “dressing up with lots of data”? Would you rather see less data? Or are you volunteering to be test subject #2? Show us how to do it right Dr. M.!

In my opinion this is an exemplary N=1 study that is well designed and thoughtfully executed. Deserve accolades, not derision. And the author even recognizes possible improvements.

Unlike most large high N clinical trials this is a high resolution longitudinal trial, and it is perfectly “controlled” for genetic difference (none), well controlled for environment, and there is only one evaluator.

Compare this to the messy and mostly useless massive studies of human depression reviewed by Jonathan Flint.

https://pubmed.ncbi.nlm.nih.gov/36702864/

replies(6): >>43309100 #>>43309447 #>>43309575 #>>43312415 #>>43313221 #>>43316813 #
22. K0balt ◴[] No.43308521[source]
Citizens science used to be much more common and there were several publications initially built around the concept. Unfortunately, the idea has all but been obliterated by inadequate public education and the chronic lack of time that now embattles the middle classes.
replies(1): >>43308750 #
23. malfist ◴[] No.43308561{3}[source]
Not too me though. Especially not to me if I'm trying to decide to supplement l-theanine
replies(2): >>43309069 #>>43310268 #
24. luqtas ◴[] No.43308750[source]
fortunately* citzen science was replaced by solid methods... this is a long fancy self reported post on a metric that is easily figured out by biomarkers (stress levels). what's the consequences if the author decided to make some claim about a substance?
replies(2): >>43309902 #>>43315643 #
25. robwwilliams ◴[] No.43308845{3}[source]
Sure. I can see your point, but the most reasonable posterior probability of the null given the biohacker community’s belief is one-tailed. This also gives more power to reject the null.
26. quijoteuniv ◴[] No.43308854[source]
I think is great that folk self observe, and that is a key to a lot, everyone ought to find what works for themselves, however there is a tendency to want to fix things with pills. There is a fundamental error there, if successful you might damp your symptoms but then you start a Whac-A-Mole (whac-a-symptom) but you are not looking at the root. The problem will come in another form. Have you consider that the symptom is a defence mechanism ?What about Taichi, mindfulness ness practices, yoga why not a study on that? It is definitely more work… than taking some pills. No I am not against taking supplements. But ultimately is a workaround you are not fixing the bug
replies(1): >>43308947 #
27. BolexNOLA ◴[] No.43308947[source]
I’m not sure if this is your intention, but I take issue with the implication that “pills” (aka “medication” and “medicine,” “pills” has a generally negative connotation) aren’t often the solution to the root problem.

It is great if you can solve things with diet, mindfulness, etc. But sometimes you need medical intervention and yes sometimes that means you need to take medication in the form of pills. There are millions of people who need that regardless of how they change their lifestyle or regulate their emotions/mental health without them.

Basically I don’t like the idea that you are implying medication is a bandaid and not ever the actual solution. If I misreading your comment my apologies

replies(2): >>43309138 #>>43310010 #
28. ◴[] No.43309069{4}[source]
29. wslh ◴[] No.43309081[source]
In science, an n=1 experiment isn’t discarded; instead, it adds information that can guide future experiments.
30. dublinben ◴[] No.43309099{4}[source]
>It’s not like the experimentor can issue a placebo to themselves.

TFA describes a protocol for doing just that. The author randomly selected between the treatment dose and placebo. They didn't reveal the choice until after the effect should be complete so they could record the results.

31. grafmax ◴[] No.43309100{3}[source]
I think social media discussions of science would be better informed by the concept of a ‘hierarchy of evidence’.

Anecdotal data, n=1 trials of varying quality, correlations studies, double blind studies (with small and large cohorts), studies without attempted replication and studies with heavy replication - they are all provide evidence of varying quality and can inform the holistic scientific picture. They can all serve a purpose such as inspiring further research, providing fodder for meta analyses, etc. It simply isn’t true that gathered evidence ought to be casually discarded if it doesn’t attain the highest levels of the hierarchy of evidence. Neither is it true that some small study showing (or not showing) some supposed effect should drastically change all our lifestyle habits. The truth lies somewhere in the middle. The concept of a hierarchy of evidence can help us navigate these apparently mixed signals so prevalent in popular science discussions.

replies(1): >>43312713 #
32. bosie ◴[] No.43309132{3}[source]
do you use an app and shortcuts (or something similar) to speed this up?
replies(1): >>43309148 #
33. quijoteuniv ◴[] No.43309138{3}[source]
This is good point, thank you, lot of people needs medicine, and i am glad they are accesible. My point goes more towards this: Imagine someone runs a marathon everyday, after a couple of months, start experiencing pain and inflammation. Do they need painkillers or do they need to ask themselves why they are running a marathon everyday? I am not against taking some painkillers as long as you also ask yourself the question. Do i really need to run a marathon everyday? I do not want to cure disease with yoga, no, but you should give it a go, whatever works for you, self observe, train a bit, walk in the forest, surf, meditate. Wellbeing cannot be achieved only with medicine.
replies(1): >>43309614 #
34. mg ◴[] No.43309148{4}[source]
On the phone I don't.

In Vim, I use multiple shortcuts that I made for this use case.

replies(1): >>43309829 #
35. yimby2001 ◴[] No.43309149{4}[source]
It’s very possible that some things can be good for one person and not for another. Should we only ever have drugs that can help everyone?
36. derlvative ◴[] No.43309263[source]
Noooooo you can’t just run independent experiments you need institutions and phds and bureaucracy and gold plating nooooo
replies(2): >>43310620 #>>43311503 #
37. jvanderbot ◴[] No.43309271[source]
Really fantastic.

I've been using theanine for a long time, but never for any of these purported benefits. And the benefits I do use it for would be near impossible to measure. I just use it to make a over-caffinated monkey brain state tend toward a "lock-in" mental state. That's super hard to measure, and just as likely the theanine is a trigger for a mental deep dive that could just as well be sugar. But the ritual works, and that's what's important to me. It just took the intial "It helps mellow out caffeine for deep focus" idea to establish the ritual.

Science? No. Effective? Yeah, I think so.

38. VerdisQuo5678 ◴[] No.43309398{3}[source]
i hope your 1 person isn't a masochist
39. ◴[] No.43309447{3}[source]
40. altcognito ◴[] No.43309513[source]
Why do you say pseudo blinding, it seems like it is blind in that the author doesn’t know if he is taking the test or not.

Now you can argue that there isn’t enough time between samples, or he needs more subjects but he was blind to whether he was taking it that day or not.

replies(1): >>43309627 #
41. jrootabega ◴[] No.43309575{3}[source]
If he said unblinded at some point, it could have been because the study author looked into the cup to determine which substance had been taken too soon. The subject should have had no knowledge of what was taken until the entire 16-month trial was over.

We should avoid extreme polarization of our judgments in general. The study deserves some amount of praise for things it did somewhat well (like the method of blinding which is clever, but not applicable to everyone), and criticism for things it did not do well, such as designing your own study methodology for your own mood. That alone will affect the results. Simply RUNNING an experiment can affect your mood because it's interesting (or even maybe frustrating). The subject probably felt pride and satisfaction whenever they used their pill selection technique, which could improve mood on its own. Neither accolades nor complete derision are appropriate, although trying to claim too strong a result from this study is kinda deserving of derision if you claim to be science-minded.

The study was well-meaning and displayed cleverness.

replies(1): >>43310475 #
42. BolexNOLA ◴[] No.43309614{4}[source]
Fair point! Thanks for clearing that up for me
43. jrootabega ◴[] No.43309627{3}[source]
If the author felt good on a particular day for whatever reason, and then learned they had taken the active substance, their reports are contaminated forever. It works the other way, too. It works any way you slice it.
44. Zak ◴[] No.43309673[source]
The experiment answers a useful question: does theanine have a strong, acute anti-stress effect on the experimenter?

Effect size is the key element of this question. If the substance was alprazolam (Xanax) instead of theanine, we would almost certainly see a strong effect here. The same would be true for heroin, ethanol, or cocaine.

It's not trying to test whether there might be a strong effect for most people or whether there are any side effects. Other experiments have been done with theanine seeking to answer those questions; the answers appear to be no and no.

45. brothrock ◴[] No.43309704[source]
N=1 is addressed, see outcome predictions. N=1 comes with caveats, of course, but a study like this, with a proven harmless supplement, should be welcomed and praised.

It is clearly a step forward from what you can watch about theanine on YouTube or TikTok. I consider this a work of citizen science. While it should not be taken for more than it is, it’s a great example of how someone can experiment without a high burden.

replies(2): >>43310245 #>>43312350 #
46. jrootabega ◴[] No.43309722[source]
Hell, I'd say it's an 0<=N<1 because it involves subjective mood reporting, and there was no participant who was not contaminated by flaws in the methodology.
47. yjftsjthsd-h ◴[] No.43309829{5}[source]
You can use vim via termux, which is nice when you already have things in version control.

Alternatively, may I suggest the DailyLog app in F-Droid?

48. tomrod ◴[] No.43309838[source]
Aye. Completely agree. Path dependence matters, which is why you can't just look at pre/post action.
49. luqtas ◴[] No.43309902{3}[source]
couldn't edit but here it's how science measure stress: https://pmc.ncbi.nlm.nih.gov/articles/PMC8434839/
50. Jerrrrrry ◴[] No.43310010{3}[source]
The strawman sentiment of "no medication ever" is a pretty easy take to herald against.

The nuanced truth is that our medical industry is flawed; since most illnesses are defined by their set of apparent symptoms (with most root causes not fully understood) the standard approach is to treat "symptoms" with medicines with sides effects rivaling the original ailment!

What if the lack of proper diet ('proper' varies wildly with populations), drugs/alcohol, sleep, stress, and exercise were the originating cause? Rushing to the medication treatment without fixing those vitals first eliminates the opportunity.

Our bodies have much more adaptive self-healing resilience properties innate to our species development than our own species hubris seems to acknowledge. And most people severely underestimate the importance big four.

replies(1): >>43380364 #
51. ninetyninenine ◴[] No.43310099[source]
>Pretty much all comments here seem to be anecdotal.

The entire blog post was anecdotal.

52. ryandrake ◴[] No.43310245{3}[source]
That's a pretty low bar though. OK, it's one step up from a monetized YouTube video that boils down to "It works--Trust me, bro." I still wouldn't really call it citizen science.
53. ryandrake ◴[] No.43310268{4}[source]
Right. The only N that you can draw any conclusions about is the author himself. So, why even publish it? The "results" are not applicable to anyone reading it. This is the health version of the software industry's "It works on my system!"
replies(1): >>43311222 #
54. robwwilliams ◴[] No.43310475{4}[source]
And that is exactly the point made in the target post by the author. He explicitly raised that criticism himself. Double kudos for self-criticism. You will not find many conventional science publications pointing out: “Shucks, we could have done this a better”.
replies(1): >>43310551 #
55. jrootabega ◴[] No.43310551{5}[source]
The ancestor post is neither a "Complete injustice" nor "derision" nor an "insult", and it doesn't warrant a hostile mocking reply. Its tone could have been gentler, but it wasn't that bad. And the study doesn't really deserve "accolades", it deserves to be recognized for whatever it does well. Such polarization of tone and vocabulary doesn't accomplish much, and I'll even propose that it actually prevents good things from happening. It is good that the author is aware of, and acknowledges, the problems in the study. What other studies and journals have done wrong doesn't make the author or study more deserving of praise.

Also, you asked why he said "unblinded", and I think you now have the answer to that.

replies(1): >>43316199 #
56. smohare ◴[] No.43310620{3}[source]
It’s not about elitism. It’s that there are so many confounding factors that even a well-informed approach makes such a study comtain very little of value

Comments like yours expose a particularly distasteful amount of hubris.

replies(1): >>43320155 #
57. ac29 ◴[] No.43310822[source]
> The most appropriate null hypothesis in this lovely study is “does theanine REDUCE anxiety”, not “does theanine change anxiety either up or down”.

Neither of those is a hypothesis, which require a prediction not just a question.

The null hypothesis for this experiment would be "Theanine has no effect on stress".

58. episteme ◴[] No.43311222{5}[source]
Publish it is quite a loaded word here. Almost every blog or post you read is from the point of view of a single person. There's nothing wrong with him putting his experience out into the world.
59. CamperBob2 ◴[] No.43311503{3}[source]
These smug pilots have lost touch with the down-to-earth lives and concerns of ordinary passengers like us. Let's see a show of hands: who thinks I should fly the plane?
replies(1): >>43316712 #
60. azeirah ◴[] No.43311651[source]
Any experiment you perform on yourself has N=1. Self-science isn't as robust as gold standard double blind blablabla PhD trials, but come on.

How else are you going to find out whether a particular diet or medication works for you specifically? It's ALWAYS N=1.

replies(1): >>43311770 #
61. hartator ◴[] No.43311770{3}[source]
And a honest n=1 is better than a double blind n=1000 but outcome-interested study. It’s so easy to make data tell the story you want.
62. matthewdgreen ◴[] No.43312350{3}[source]
N=1 studies aren’t evil. They’re just pretty close to the entire history of pre-modern medicine that led us to bad evidence. My concern here is not that someone is sharing their opinions, it’s the fact that the person doing this explicitly heaps derision on the “placebo people” (or some other phrasing) and then heaps praise on other people doing N=1 studies and proceeds to do one. This stuff all needs to be treated with humor, good faith, and then extreme skepticism about any result it produces.
63. levocardia ◴[] No.43312415{3}[source]
Because....

>While I was blinded during each trial, I saw the theanine/D result when I wrote it down. Over time I couldn’t help but notice that my stress dropped even when I took vitamin D, and that I was terrible at predicting what I’d taken

That is not blinding

replies(1): >>43313019 #
64. abirch ◴[] No.43312713{4}[source]
Here’s a relevant xkcd https://xkcd.com/882/ For the small sample
65. sgc ◴[] No.43313019{4}[source]
I would have taken a well-calibrated photo of the cup each time without looking, maybe with a color card in the bottom, and only entered results at the end of the trial.

Given that there is no documentation of whether the events during the hour of test time were more or less stressful than those before it, and no taking the time of day, diet and exercise, sleep, location (quiet island or next to a construction site), etc into account, the data seems useless.

As a note, I have no idea why he bothered trying to guess what he had taken. What possible value could that have in this type of experiment?

Perhaps the correct course of action would be to ask for feedback in the design phase of an N=1 trial, especially a longer one, to avoid some basic mistakes.

replies(2): >>43314257 #>>43316298 #
66. throwup238 ◴[] No.43313221{3}[source]
> Why do you say unblinded?

It’s unblinded because the subject is preparing the concoction under study. There is no way they can create a blind experiment if they’re the ones preparing the control. The placebo effect is nothing if not pernicious and cunning, able to exploit even the most subtle psychological signal - like minuscule differences in the amount of powder in a capsule.

Blinded studies have independent doctors prepare and dispense the candidate drug so they know whether its the real thing or a placebo, but their patients dont. In double blinded studies, neither the doctor nor the patient have any idea about what they’re getting because a third party prepares the drugs.

67. cortesoft ◴[] No.43313233[source]
N=1 trial is great if you are trying to figure out what works for you as an individual
68. kqr ◴[] No.43314257{5}[source]
> As a note, I have no idea why he bothered trying to guess what he had taken. What possible value could that have in this type of experiment?

It gives a hint of how well the blinding worked.

replies(1): >>43315493 #
69. bonestamp2 ◴[] No.43315219[source]
> This is an N=1 trial

Sure, OP addressed that and said it would be especially useful more people did it and "if we find a way to collectively interpret the data".

70. sgc ◴[] No.43315493{6}[source]
Yes, that is a good point. But it also causes him to continuously think about indicators to determine what he has taken. He is constantly trying to punch the veil on this "blind" study instead of doing everything he can to avoid that. Are they really the exact same weight? Do they really feel the same in his hand, mouth and throat? Does one have a slightly different taste - after all, no filling process will leave the outside of the capsule 100% free of powder, etc? Very subtle differences could manifest themselves over 16 months.
71. K0balt ◴[] No.43315643{3}[source]
I think you are ignoring the fact that citizen science was alive and very well up into the 1980s. It was never a replacement for professional science, but rather citizen scientists often observed things that were deemed worth of follow-up by academic researchers.

Citizens scientists rarely post “facts” but rather interesting avenues for research or further investigation. Part of being an educated citizen scientist is to understand the limitations of your knowledge, data, and methods.

Quacks and cranks, on the other hand, are always making grand new “discoveries” lol.

Anyone who has a decent education can make observations apply the scientific method. I say this coming from a family of actual scientists from molecular biology to particle physics, who will tell you the same, and also give credit to the multitude of citizen scientists who have done just that.

As for myself, I’ll stick to engineering.

replies(2): >>43316385 #>>43322467 #
72. robwwilliams ◴[] No.43316199{6}[source]
Yes, perhaps. But please tell me you have read the original post. It is thoughtful, self-deprecatory, careful, well analyzed, and upfront about limitations and possible improvements.

Re-reading such a negative critique of a solid home-brew experiment is unwarranted. There are several word here worth red flags.

>This is an N=1 trial. Dressing your N=1 trial up with lots of pseudo controls and pseudo blinding and data collection does not make it better. In fact: putting this much effort into any medication trial makes it much more likely that you’re going to be incentivized to find effects that don’t exist. I think it’s nice that the author admits that they found nothing, but statistically, worthless drugs show effects in much better-designed trials than this one: it’s basically a coin toss.

73. fc417fc802 ◴[] No.43316298{5}[source]
If your sample is blinded yet you consistently guess correctly, then presumably either 1. you failed at blinding or 2. there is a strong and discernible effect, regardless of what your other metrics might say (your other metrics could always be flawed after all).

> no documentation of whether the events during the hour of test time were more or less stressful than those before it, and no taking the time of day, diet and exercise, sleep, location

Assuming the blinded samples are uniformly randomly distributed, and assuming the study goes on long enough, then you'd expect that stuff to average out.

But I agree, it should be recorded nonetheless. That way you can verify at the end that it did, in fact, average out as you expected. If it didn't then your data is invalid.

replies(1): >>43328032 #
74. luqtas ◴[] No.43316385{4}[source]
i think anyone researching whatever in their home is doing 'science'

but this isn't the case! i linked in an answer to my post what science knows when measuring stress... it's far away from a self-reported subjective question

i think the problem is whatever you post on the internet is politic. it may have a huge reach, and so far so good people trying far from lethal doses of theanine. now, what if an influencer wants to make their stuff look intellectual and go buy pure caffeine? how about the people coming after wards?

one thing is a 3D-print project at your garage on keyboards ergonomics, home automation, another is rock-climbing gear and another is substances with reachable lethal doses where the masses can buy (not the case here but again, i cited caffeine but there's much more). the author has a stellar presentation but they seriously researched what science does when measuring stress? they didn't even considered The Perceived Stress Scale (PSS), which is used on professional research. i stand against public posts with badly methods (or the complete lack of research) suppressed by fancy graphics on substance use/abuse. i also bet isn't that hard to buy empty pills with opaque color

75. eru ◴[] No.43316712{4}[source]
You should link to the source of your quote.
replies(1): >>43317308 #
76. eru ◴[] No.43316721[source]
> Too much other stuff is changing in a single persons life that could account for all observed side effects.

> You also have latent side effect issues. A person could smoke for 10 years, not smoke for another 10, and then conclude that smoking doesn’t cause cancer. Then they get lung cancer 20 years later.

But that's not the actual experiment here. Your criticism is invalid. The author randomised every day which supplement to take.

That's pretty close to a blocked design.

77. eth0up ◴[] No.43316813{3}[source]
If you ever wonder why some folks with a fair amount of potential and something to offer keep to themselves, this isn't the worst example.

I think most people could criticize the carbon out of a corpse if they themselves weren't being criticized into one.

If we devolved from apes, maybe apes devolved from piranhas.

replies(1): >>43319369 #
78. CamperBob2 ◴[] No.43317308{5}[source]
Arguably so, but all I can find are other plagiarized copies. :-P

The New Yorker's paywall has successfully obscured the origin of the joke, so that's on them, as far as I'm concerned.

replies(1): >>43318253 #
79. eru ◴[] No.43318253{6}[source]
Yes.

Well, instead of an actual web link, you could just mention that it's from the New Yorker.

replies(1): >>43322365 #
80. robwwilliams ◴[] No.43319369{4}[source]
hilarious
81. derlvative ◴[] No.43320155{4}[source]
You don’t like it you don’t have to read the blog article. I assure you are not the intended audience. For the rest of us it provided valuable insight.
82. gus_massa ◴[] No.43320433[source]
>> Technically, I did find two significant results.

The problem is that he is comparing two very different things, the level when he took the pill and the level one hour after that. So it's not surprising that they are different. Let's imagine a very very very stupid experiment, where the problem is more obvious.

Does Coca Cola or Pepsi improve luck? N=1000000, double blind randomized controlled trial.

1) Each subject flips a coin. tail=0, head=1.

2) They drink a glas of soda, 50% Coke or 50% Pepsi, that is served in a hidden place and nor the subject or the experimenter know which one.

3) They roll a dice (an usual one, D6)

Results:

* Average before Coke = 0.5002

* Average after Coke = 3.5005

* Average before Pepsi = 0.5004

* Average after Pepsi = 3.5003

So the conclusion is that Coke improves the average (p<1E-a-lot) and Pepsi improves the average (p<1E-a-lot). Both are "technically" statistically significant (but it's caused by a horrible experiment design).

Unsurprisingly, the difference in the average after drinking Coke or Pepsi is not statistically significant (p<.something).

(I'm too lazy to run a simulation now, but it's not difficult to get realistic averages and p values.)

In conclusion, the useful result is the comparison of the anxiety after taking both drugs, not the difference of before and after taking them.

As the article says:

>> So I propose a new rule: Blind trial or GTFO.

83. CamperBob2 ◴[] No.43322365{7}[source]
"It's from the New Yorker. The most important magazine of our time. Probably the most important magazine that ever was."

(From a vaguely-remembered 1990s-era ad campaign that I thought was excruciatingly self-indulgent at the time, but which evidently worked.)

84. Suppafly ◴[] No.43322467{4}[source]
>I think you are ignoring the fact that citizen science was alive and very well up into the 1980s.

Some version of that still exists among people with a naturalist bent recording observations of reptiles, birds, plants, etc. But yeah, we don't really have backyard chemists analyzing things anymore.

replies(1): >>43327428 #
85. Suppafly ◴[] No.43322506{3}[source]
A similar, actually useful, experiment is the one where a guy cracked the knuckles of only one hand for several years and found no ill effects from doing so compared to the other hand with non-cracked knuckles.
86. Enginerrrd ◴[] No.43322581[source]
>Dressing your N=1 trial up with lots of pseudo controls and pseudo blinding and data collection does not make it better.

This is only an appropriate criticism in so far as you want to make conclusions about theanine as an intervention in the broader population.

It is however, perhaps much BETTER than large N trials if the author wishes to draw conclusions about how theanine affects THEM.

87. K0balt ◴[] No.43327428{5}[source]
Astronomy also has an important citizen-science component, and many discoveries are made by amateur astronomers each year.

Another interesting aspect of citizen science is replication of existing scientific research, Often with experimental modifications that make the experiment much more approachable for amateurs. Sometimes this even leads to process improvements that facilitate industrial application.

88. sgc ◴[] No.43328032{6}[source]
He has 94 data points. Not nearly enough to average out so many potentially confounding variables, and there is no way to know they would. That will be the case in almost all N=1 experiments. Perhaps he takes the pill at the onset of stress, and stress almost always tends to build afterwards. This would be a probable case for many people trying to use theanine in this way. I could never accept that we should presume it would average out, and thus I consider logging potentially confounding variables essential to a valid experiment of this type of uncontrolled experiment. Your hypothesis would be much more relevant in a larger experiment of course.
replies(1): >>43328233 #
89. fc417fc802 ◴[] No.43328233{7}[source]
> Perhaps he takes the pill at the onset of stress, and stress almost always tends to build afterwards.

That would not be a problem regarding the averaging I referred to, although it could well pose a problem for measurement depending on how it interacted with the selected metrics.

Note that the averaging I refer to is not regarding all possible values of some metric, but rather any discrepancy in the distribution of metrics which we expected to follow the same distribution between the sample and the control.

I think maybe there's a misunderstanding? It seems that we both agree that a variety of additional variable should be logged. I was not suggesting to omit them, but rather to use discrepancies in them to detect fundamental issues with the data or study design. I would also expect larger studies to do the same where possible.

At 94 data points it is entirely possible that there would be outliers that would have averaged out for a larger N but did not. In such a scenario the presence of such outliers should then be taken to indicate a problem with the data (ie the more discrepancies you observe, the less you should trust the data).

replies(1): >>43331812 #
90. sgc ◴[] No.43331812{8}[source]
I was mainly saying I don't understand why you were indicating we should presume or expect they would average out in an N=1 experiment. Even in much larger experiments that is not reliably the case. Science would be relatively easy if there were not a lot of noise in the real world. So, to be clear, my concern is mainly one of scale - the experiment is far too restricted to overlook this: the smaller the experiment, the more important this type of information. Perhaps you were not saying that such averaging might be possible in an N=1 experiment and I misread, since it seems that you comment here indicates a different point.

I of course agree that logging them is basic scientific methodology - in order to detect issues with the experiment, and even hopefully to see the signal through the noise.

91. silent_cal ◴[] No.43333903[source]
From what I've seen, most scientists don't even publish their raw data.
92. BolexNOLA ◴[] No.43380364{4}[source]
I think calling my comment a strawman is a little unfair given the way he wrote it seemed pretty unequivocal at first. Once they clarified their stance I understood it better and was perfectly fine accepting that’s not what they meant.
93. angg ◴[] No.43380374[source]
> https://www.gibney.org/a_syntax_for_self-tracking

Loved reading this as I have been on a similar self tracking journey of my own.

I am curious about your thoughts on mobile support? Specifically, have you implemented anything to make adding/editing events while away from your computer easier? Or perhaps you feel it is not important for your use case?

Personally, I've found that being able to log an event in an instant, or just on a whim, is invaluable for capturing the sort of data I care about tracking the most (like what I just ate or my mood), and I find too much friction causes many events simply go unlogged as they almost always feel too unimportant or mundane in retrospect to keep a mental note of and track later.

I started out with a text file system similar to yours but I've since begrudgingly resorted to google forms for most of my tracking needs. It's nice that I can arbitrarily add and remove fields while still keeping everything relatively structured and parseable, and it does have (albeit inferior) revision history. But I dislike not having ownership and control over where my data is stored and absolutely *dread* waiting for the form to load every time I want to log an event, especially on mobile.