Is thirst a good predictor of dehydration?
10almonds is reader-supported. We may, at no cost to you, receive a portion of sales if you purchase a product through a link in this article.
Water is essential for daily functioning and health, and we can only survive a few days without it. Yet we constantly lose water through sweat, urination and even evaporation when we breathe.
This is why we have evolved a way to regulate and maintain water in our bodies. Like other animals, our survival relies on a strong biological drive that tells us to find and drink water to balance fluid loss.
This is thirst – a sensation of dryness in the mouth signalling we need to have a drink. This basic physiological mechanism is controlled mainly by part of the brain’s “control centre”, called the hypothalamus. The hypothalamus receives signals from various regions of the body and in return, releases hormones that act as a messenger to signal the thirst sensation.
What is dehydration?
Staying hydrated (having enough water in our bodies) is important for several reasons, including:
- regulating body temperature through sweat and respiration
- lubricating joints and eyes
- preventing infections
- digesting and absorbing nutrients
- flushing out waste (via the kidneys)
- preventing constipation
- brain function (including memory and concentration)
- mood and energy levels
- physical performance and recovery from exercise
- skin health.
Dehydration occurs when our body doesn’t have enough water. Even slight drops in fluid levels have noticeable consequences, such as headaches, feeling dizzy, lethargy and struggling to concentrate.
Chronic dehydration can pose more serious health risks, including urinary tract infections, constipation and kidney stones.
What does the evidence say?
Despite thirst being one of the most basic biological drivers for good hydration, science suggests our feelings of thirst and subsequent fluid intake don’t always correlate with hydration levels.
For example, a recent study explored the impact of thirst on fluid intake and hydration status. Participants attended a lab in the morning and then later in the afternoon to provide markers of hydration status (such as urine, blood samples and body weight). The relationship between levels of thirst in the morning and afternoon hydration status was negligible.
Further, thirst may be driven by environmental factors, such as access to water. For example, one study looked at whether ample access to water in a lab influenced how much people drank and how hydrated they were. The link between how thirsty they felt and how hydrated they were was weak, suggesting the availability of water influenced their fluid intake more than thirst.
Exercise can also change our thirst mechanism, though studies are limited at this stage.
Interestingly, research shows women experience thirst more strongly than men, regardless of hydration status. To understand gender differences in thirst, researchers infused men and women with fluids and then measured their thirst and how hydrated they were. They found women generally reported thirst at a lower level of fluid loss. Women have also been found to respond more to feeling thirsty by drinking more water.
Other ways to tell if you need to drink some water
While acknowledging some people will need to drink more or less, for many people, eight cups (or two litres) a day is a good amount of water to aim for.
But beyond thirst, there are many other ways to tell whether you might need to drink more water.
1. urine colour: pale yellow urine typically indicates good hydration, while darker, concentrated urine suggests dehydration
2. frequency of going to the toilet: urinating regularly (around four to six times a day) indicates good hydration. Infrequent urination can signal dehydration
3. skin turgor test: gently pinching the skin (for example, on the back of the hand) and observing how quickly the skin returns to its normal position can help assess hydration. Slow return may indicate dehydration
4. mouth and lips: a dry mouth or cracked lips can be early signs of dehydration
5. headaches and fatigue: frequent headaches, dizziness, or unexplained fatigue can be signs of inadequate hydration
6. sweating: in physically active people, monitoring how much they sweat during activity can help estimate fluid loss and hydration needs. Higher levels of sweat may predispose a person to dehydration if they are unable to replace the fluid lost through water intake
These indicators, used together, provide a more comprehensive picture of hydration without solely depending on the sensation of thirst.
Of course, if you do feel thirsty, it’s still a good idea to drink some water.
Lauren Ball, Professor of Community Health and Wellbeing, The University of Queensland and Kiara Too, PhD candidate, School of Human Movement and Nutrition Sciences, The University of Queensland
This article is republished from The Conversation under a Creative Commons license. Read the original article.
Don’t Forget…
Did you arrive here from our newsletter? Don’t forget to return to the email to continue learning!
Recommended
Learn to Age Gracefully
Join the 98k+ American women taking control of their health & aging with our 100% free (and fun!) daily emails:
-
Walnuts vs Pecans – Which is Healthier?
10almonds is reader-supported. We may, at no cost to you, receive a portion of sales if you purchase a product through a link in this article.
Our Verdict
When comparing walnuts to pecans, we picked the walnuts.
Why?
It was very close, though, and an argument could be made for pecans! Walnuts are nevertheless always a very good bet, and so far in our This-or-That comparisons, the only nut to beat them so far as been almonds, and that was very close too.
In terms of macros, walnuts have a lot more protein, while pecans have a little more fiber (for approximately the same carbs). Both are equally fatty (near enough; technically pecans have a little more) but where the walnuts stand out in the fat category is that while pecans have mostly healthy monounsaturated fats, walnuts have mostly healthy polyunsaturated fats, including including a good balance of omega-3 and omega-6 fatty acids. So, while we do love the extra fiber from pecans, we’re calling it for walnuts in the macros category, on account of the extra protein and the best lipids profile (not that pecans’ lipids profile is bad by any stretch; just, walnuts have it better).
In the vitamins category, walnuts have more of vitamins B2, B6, B9, and C, while pecans offer more of vitamins A, B1, B3, B5, E, K, and choline. The margins aren’t huge and walnuts are also excellent for all the vitamins that pecans narrowly beat them on, but still, the vitamins category is a win for pecans.
When it comes to minerals, walnuts take back the crown; walnuts offer more calcium, copper, iron, magnesium, phosphorus, potassium, and selenium, while pecans have a little more manganese and zinc. Once again, the margins aren’t huge and pecans are also excellent for all the minerals that walnuts narrowly beat them on, but still, the minerals category is a win for walnuts.
In short: enjoy both of these nuts for their healthy fats, vitamins, minerals, protein, and fiber, but if you’re going to pick one, walnuts come out on top.
Want to learn more?
You might like to read:
Why You Should Diversify Your Nuts!
Take care!
Share This Post
-
The Snooze-Button Controversy
10almonds is reader-supported. We may, at no cost to you, receive a portion of sales if you purchase a product through a link in this article.
To Snooze Or Not To Snooze? (Science Has Answers)
This is Dr. Jennifer Kanaan. She’s a medical doctor with a focus on pulmonary critical care, sleep disorders, and sleep medicine.
What does she want to tell us?
She wants us to be wary of the many news articles that have jumped on a certain recent sleep study, such as:
- Is hitting the snooze button really a bad idea? Study sheds light on the impact of morning alarms on sleep and cognition
- Hitting Snooze May Help You Feel Less Sleepy and More Alert, Research Says
- Is it okay to press the snooze button?
- Hitting Snooze May Help You Feel Less Sleepy and More Alert, Research Says
- Hitting the snooze button on your alarm doesn’t make you more tired
For the curious, here is the paper itself, by Dr. Tina Sundelin et al. It’s actually two studies, by the way, but one paper:
The authors of this study concluded:
❝There were no clear effects of snoozing on the cortisol awakening response, morning sleepiness, mood, or overnight sleep architecture.
A brief snooze period may thus help alleviate sleep inertia, without substantially disturbing sleep, for late chronotypes and those with morning drowsiness.❞
Notably, people tend to snooze because an alarm clock will, if not “smart” about it, wake us up mid sleep-cycle more often than not, and that will produce a short “sleep hangover”. By snoozing, we are basically re-rolling the dice on being woken up between sleep cycles, and thus feeling more refreshed.
What’s Dr. Kanaan’s counterpoint?
Dr. Kanaan says:
❝If you’re coming in and out of sleep for 30 minutes, after the alarm goes off the first time, you’re costing yourself 30 minutes of uninterrupted, quality, restorative sleep. This study doesn’t change that fact.❞
She advises that rather than snoozing, we should prioritize getting good sleep in the first place, and once we do wake up, mid sleep-cycle or not, get sunlight. That way, our brain will start promptly scrubbing melatonin and producing the appropriate wakefulness hormones instead. That means serotonin, and also a spike of cortisol.
Remember: cortisol is only bad when it’s chronically elevated. It’s fine, and even beneficial, to have a short spike of cortisol. We make it for a reason!
If you’d like to hear more from Dr. Kanaan, you might like this interview with her at the University of Connecticut:
Want the best of both worlds?
A great option to avoid getting woken in the middle of a sleep cycle, and also not needing to hit snooze, is a sunrise alarm clock. Specifics of these devices vary, but for example, the kind this writer has starts gently glowing an hour before the set alarm time,and gradually gets brighter and lighter over the course of the hour.
We don’t sell them, but here’s an example sunrise alarm clock on Amazon, for your convenience
Share This Post
-
Rebounding Into The Best Of Health
10almonds is reader-supported. We may, at no cost to you, receive a portion of sales if you purchase a product through a link in this article.
“Trampoline” is a brand-name that’s been popularized as a generic name, and “rebounding”, the name used in this video, is the same thing as “trampolining”. With that in mind, let us bounce swiftly onwards:
Surprising benefits
It’s easy to think “isn’t that cheating?” to the point that such “cheating” could be useless, since surely the device is doing most of the work?
The thing is, while indeed it’s doing a lot of the work for you, your muscles are still doing a lot—mostly stabilization work, which is of course a critical thing for our muscles to be able to do. While it’s rare that we need to do a somersault in everyday life, it’s common that we have to keep ourselves from falling over, after all.
It also represents a kind of gentle resistance exercise, and as such, improves bone density—something first discovered during NASA research for astronauts. Other related benefits pertain to the body’s ability to deal with acceleration and deceleration; it also benefits the lymphatic system, which unlike the blood’s circulatory system, has no pump of its own. Rebounding does also benefit the cardiovascular system, though, as now the heart gets confused (in the healthy way, a little like it gets confused with high-intensity interval training).
Those are the main evidence-based benefits; anecdotally (but credibly, since these things can be said of most exercise) it’s also claimed that it benefits posture, improves sleep and mood, promotes weight loss and better digestion, reduces bloating, improves skin (the latter being due to improved circulation), and alleviates arthritis (most moderate exercise improves immune response, and thus reduces chronic inflammation, so again, this is reasonable, even if anecdotal).
For more details on all of these and more, enjoy:
Click Here If The Embedded Video Doesn’t Load Automatically!
Want to learn more?
You might also like to read:
- Exercise Less, Move More
- How To Do HIIT (Without Wrecking Your Body)
- Resistance Is Useful! (Especially As We Get Older)
- HIIT, But Make It HIRT
- The Lymphatic System Against Cancer & More
Take care!
Share This Post
Related Posts
-
The 3 Phases Of Fat Loss (& How To Do It Right!)
10almonds is reader-supported. We may, at no cost to you, receive a portion of sales if you purchase a product through a link in this article.
Cori Lefkowith, of “Redefining Strength” and “Strength At Any Age” fame, has advice:
As easy as 1, 2, 3?
Any kind of fat loss plan will not work unless it takes into account that the body can and will adapt to a caloric deficit, meaning that constantly running a deficit will only ever yield short term results, followed by regaining weight (and feeling hungry the whole time). So, instead, if fat loss is your goal, you might want to consider doing it in these stages:
1. Lifestyle adjustments (main phase)
Focus on sustainable, gradual improvements in diet and workouts.
- Key strategies:
- Start with small, manageable changes, for example focusing on making your protein intake around 30–35% of your total calories.
- Track your current habits to identify realistic adjustments.
- Balance strength training and cardio, as maintaining your muscle is (and will remain) important.
- Signs of Progress:
- Slow changes in the numbers on the scale (up to 1 lb/week).
- Inches being lost (but probably not many), improved energy levels, and stable performance in workouts.
Caution: avoid feelings of extreme hunger or restriction. This is not supposed to be arduous.
2. Mini cut (short-term intensive)
Used for quick fat loss or breaking plateaus; lasts 7–14 days.
- Key strategies:
- Larger calorie deficit (e.g: 500 calories).
- High protein intake (40–50% of your total calories).
- Focus on strength training and reduce cardio, to avoid muscle loss.
- Signs of Progress:
- Rapid scale changes (up to 5 lbs/week).
- Reduced bloating, potential energy dips, and cravings.
- Temporary performance stagnation in workouts. Don’t worry about this; it’s expected and fine.
Caution: do not exceed 21 days, to avoid the metabolic adaptation that we talked about.
3. Diet break (rest & reset)
A maintenance period to recharge mentally and physically, typically lasting 7–21 days.
- Key strategies:
- Gradually increase calories (200–500) to maintenance level.
- Focus on performance goals and reintroducing foods you enjoy.
- Combine strength training with steady-state cardio.
- Signs of Progress:
- Increased energy, improved workout performance, and feeling fuller.
- Scale may fluctuate initially but stabilize or decrease by the end.
- Inches will be lost as muscle is built and fat is burned.
The purpose of this third stage is to prevent metabolic adaptation, regain motivation, and (importantly!) test maintenance.
For more on these and how best to implement them, enjoy:
Click Here If The Embedded Video Doesn’t Load Automatically!
Want to learn more?
You might also like to read:
Take care!
Don’t Forget…
Did you arrive here from our newsletter? Don’t forget to return to the email to continue learning!
Learn to Age Gracefully
Join the 98k+ American women taking control of their health & aging with our 100% free (and fun!) daily emails:
- Key strategies:
-
How do science journalists decide whether a psychology study is worth covering?
10almonds is reader-supported. We may, at no cost to you, receive a portion of sales if you purchase a product through a link in this article.
Complex research papers and data flood academic journals daily, and science journalists play a pivotal role in disseminating that information to the public. This can be a daunting task, requiring a keen understanding of the subject matter and the ability to translate dense academic language into narratives that resonate with the general public.
Several resources and tip sheets, including the Know Your Research section here at The Journalist’s Resource, aim to help journalists hone their skills in reporting on academic research.
But what factors do science journalists look for to decide whether a social science research study is trustworthy and newsworthy? That’s the question researchers at the University of California, Davis, and the University of Melbourne in Australia examine in a recent study, “How Do Science Journalists Evaluate Psychology Research?” published in September in Advances in Methods and Practices in Psychological Science.
Their online survey of 181 mostly U.S.-based science journalists looked at how and whether they were influenced by four factors in fictitious research summaries: the sample size (number of participants in the study), sample representativeness (whether the participants in the study were from a convenience sample or a more representative sample), the statistical significance level of the result (just barely statistically significant or well below the significance threshold), and the prestige of a researcher’s university.
The researchers found that sample size was the only factor that had a robust influence on journalists’ ratings of how trustworthy and newsworthy a study finding was.
University prestige had no effect, while the effects of sample representativeness and statistical significance were inconclusive.
But there’s nuance to the findings, the authors note.
“I don’t want people to think that science journalists aren’t paying attention to other things, and are only paying attention to sample size,” says Julia Bottesini, an independent researcher, a recent Ph.D. graduate from the Psychology Department at UC Davis, and the first author of the study.
Overall, the results show that “these journalists are doing a very decent job” vetting research findings, Bottesini says.
Also, the findings from the study are not generalizable to all science journalists or other fields of research, the authors note.
“Instead, our conclusions should be circumscribed to U.S.-based science journalists who are at least somewhat familiar with the statistical and replication challenges facing science,” they write. (Over the past decade a series of projects have found that the results of many studies in psychology and other fields can’t be reproduced, leading to what has been called a ‘replication crisis.’)
“This [study] is just one tiny brick in the wall and I hope other people get excited about this topic and do more research on it,” Bottesini says.
More on the study’s findings
The study’s findings can be useful for researchers who want to better understand how science journalists read their research and what kind of intervention — such as teaching journalists about statistics — can help journalists better understand research papers.
“As an academic, I take away the idea that journalists are a great population to try to study because they’re doing something really important and it’s important to know more about what they’re doing,” says Ellen Peters, director of Center for Science Communication Research at the School of Journalism and Communication at the University of Oregon. Peters, who was not involved in the study, is also a psychologist who studies human judgment and decision-making.
Peters says the study was “overall terrific.” She adds that understanding how journalists do their work “is an incredibly important thing to do because journalists are who reach the majority of the U.S. with science news, so understanding how they’re reading some of our scientific studies and then choosing whether to write about them or not is important.”
The study, conducted between December 2020 and March 2021, is based on an online survey of journalists who said they at least sometimes covered science or other topics related to health, medicine, psychology, social sciences, or well-being. They were offered a $25 Amazon gift card as compensation.
Among the participants, 77% were women, 19% were men, 3% were nonbinary and 1% preferred not to say. About 62% said they had studied physical or natural sciences at the undergraduate level, and 24% at the graduate level. Also, 48% reported having a journalism degree. The study did not include the journalists’ news reporting experience level.
Participants were recruited through the professional network of Christie Aschwanden, an independent journalist and consultant on the study, which could be a source of bias, the authors note.
“Although the size of the sample we obtained (N = 181) suggests we were able to collect a range of perspectives, we suspect this sample is biased by an ‘Aschwanden effect’: that science journalists in the same professional network as C. Aschwanden will be more familiar with issues related to the replication crisis in psychology and subsequent methodological reform, a topic C. Aschwanden has covered extensively in her work,” they write.
Participants were randomly presented with eight of 22 one-paragraph fictitious social and personality psychology research summaries with fictitious authors. The summaries are posted on Open Science Framework, a free and open-source project management tool for researchers by the Center for Open Science, with a mission to increase openness, integrity and reproducibility of research.
For instance, one of the vignettes reads:
“Scientists at Harvard University announced today the results of a study exploring whether introspection can improve cooperation. 550 undergraduates at the university were randomly assigned to either do a breathing exercise or reflect on a series of questions designed to promote introspective thoughts for 5 minutes. Participants then engaged in a cooperative decision-making game, where cooperation resulted in better outcomes. People who spent time on introspection performed significantly better at these cooperative games (t (548) = 3.21, p = 0.001). ‘Introspection seems to promote better cooperation between people,’ says Dr. Quinn, the lead author on the paper.”
In addition to answering multiple-choice survey questions, participants were given the opportunity to answer open-ended questions, such as “What characteristics do you [typically] consider when evaluating the trustworthiness of a scientific finding?”
Bottesini says those responses illuminated how science journalists analyze a research study. Participants often mentioned the prestige of the journal in which it was published or whether the study had been peer-reviewed. Many also seemed to value experimental research designs over observational studies.
Considering statistical significance
When it came to considering p-values, “some answers suggested that journalists do take statistical significance into account, but only very few included explanations that suggested they made any distinction between higher or lower p values; instead, most mentions of p values suggest journalists focused on whether the key result was statistically significant,” the authors write.
Also, many participants mentioned that it was very important to talk to outside experts or researchers in the same field to get a better understanding of the finding and whether it could be trusted, the authors write.
“Journalists also expressed that it was important to understand who funded the study and whether the researchers or funders had any conflicts of interest,” they write.
Participants also “indicated that making claims that were calibrated to the evidence was also important and expressed misgivings about studies for which the conclusions do not follow from the evidence,” the authors write.
In response to the open-ended question, “What characteristics do you [typically] consider when evaluating the trustworthiness of a scientific finding?” some journalists wrote they checked whether the study was overstating conclusions or claims. Below are some of their written responses:
- “Is the researcher adamant that this study of 40 college kids is representative? If so, that’s a red flag.”
- “Whether authors make sweeping generalizations based on the study or take a more measured approach to sharing and promoting it.”
- “Another major point for me is how ‘certain’ the scientists appear to be when commenting on their findings. If a researcher makes claims which I consider to be over-the-top about the validity or impact of their findings, I often won’t cover.”
- “I also look at the difference between what an experiment actually shows versus the conclusion researchers draw from it — if there’s a big gap, that’s a huge red flag.”
Peters says the study’s findings show that “not only are journalists smart, but they have also gone out of their way to get educated about things that should matter.”
What other research shows about science journalists
A 2023 study, published in the International Journal of Communication, based on an online survey of 82 U.S. science journalists, aims to understand what they know and think about open-access research, including peer-reviewed journals and articles that don’t have a paywall, and preprints. Data was collected between October 2021 and February 2022. Preprints are scientific studies that have yet to be peer-reviewed and are shared on open repositories such as medRxiv and bioRxiv. The study finds that its respondents “are aware of OA and related issues and make conscious decisions around which OA scholarly articles they use as sources.”
A 2021 study, published in the Journal of Science Communication, looks at the impact of the COVID-19 pandemic on the work of science journalists. Based on an online survey of 633 science journalists from 77 countries, it finds that the pandemic somewhat brought scientists and science journalists closer together. “For most respondents, scientists were more available and more talkative,” the authors write. The pandemic has also provided an opportunity to explain the scientific process to the public, and remind them that “science is not a finished enterprise,” the authors write.
More than a decade ago, a 2008 study, published in PLOS Medicine, and based on an analysis of 500 health news stories, found that “journalists usually fail to discuss costs, the quality of the evidence, the existence of alternative options, and the absolute magnitude of potential benefits and harms,” when reporting on research studies. Giving time to journalists to research and understand the studies, giving them space for publication and broadcasting of the stories, and training them in understanding academic research are some of the solutions to fill the gaps, writes Gary Schwitzer, the study author.
Advice for journalists
We asked Bottesini, Peters, Aschwanden and Tamar Wilner, a postdoctoral fellow at the University of Texas, who was not involved in the study, to share advice for journalists who cover research studies. Wilner is conducting a study on how journalism research informs the practice of journalism. Here are their tips:
1. Examine the study before reporting it.
Does the study claim match the evidence? “One thing that makes me trust the paper more is if their interpretation of the findings is very calibrated to the kind of evidence that they have,” says Bottesini. In other words, if the study makes a claim in its results that’s far-fetched, the authors should present a lot of evidence to back that claim.
Not all surprising results are newsworthy. If you come across a surprising finding from a single study, Peters advises you to step back and remember Carl Sagan’s quote: “Extraordinary claims require extraordinary evidence.”
How transparent are the authors about their data? For instance, are the authors posting information such as their data and the computer codes they use to analyze the data on platforms such as Open Science Framework, AsPredicted, or The Dataverse Project? Some researchers ‘preregister’ their studies, which means they share how they’re planning to analyze the data before they see them. “Transparency doesn’t automatically mean that a study is trustworthy,” but it gives others the chance to double-check the findings, Bottesini says.
Look at the study design. Is it an experimental study or an observational study? Observational studies can show correlations but not causation.
“Observational studies can be very important for suggesting hypotheses and pointing us towards relationships and associations,” Aschwanden says.
Experimental studies can provide stronger evidence toward a cause, but journalists must still be cautious when reporting the results, she advises. “If we end up implying causality, then once it’s published and people see it, it can really take hold,” she says.
Know the difference between preprints and peer-reviewed, published studies. Peer-reviewed papers tend to be of higher quality than those that are not peer-reviewed. Read our tip sheet on the difference between preprints and journal articles.
Beware of predatory journals. Predatory journals are journals that “claim to be legitimate scholarly journals, but misrepresent their publishing practices,” according to a 2020 journal article, published in the journal Toxicologic Pathology, “Predatory Journals: What They Are and How to Avoid Them.”
2. Zoom in on data.
Read the methods section of the study. The methods section of the study usually appears after the introduction and background section. “To me, the methods section is almost the most important part of any scientific paper,” says Aschwanden. “It’s amazing to me how often you read the design and the methods section, and anyone can see that it’s a flawed design. So just giving things a gut-level check can be really important.”
What’s the sample size? Not all good studies have large numbers of participants but pay attention to the claims a study makes with a small sample size. “If you have a small sample, you calibrate your claims to the things you can tell about those people and don’t make big claims based on a little bit of evidence,” says Bottesini.
But also remember that factors such as sample size and p-value are not “as clear cut as some journalists might assume,” says Wilner.
How representative of a population is the study sample? “If the study has a non-representative sample of, say, undergraduate students, and they’re making claims about the general population, that’s kind of a red flag,” says Bottesini. Aschwanden points to the acronym WEIRD, which stands for “Western, Educated, Industrialized, Rich, and Democratic,” and is used to highlight a lack of diversity in a sample. Studies based on such samples may not be generalizable to the entire population, she says.
Look at the p-value. Statistical significance is both confusing and controversial, but it’s important to consider. Read our tip sheet, “5 Things Journalists Need to Know About Statistical Significance,” to better understand it.
3. Talk to scientists not involved in the study.
If you’re not sure about the quality of a study, ask for help. “Talk to someone who is an expert in study design or statistics to make sure that [the study authors] use the appropriate statistics and that methods they use are appropriate because it’s amazing to me how often they’re not,” says Aschwanden.
Get an opinion from an outside expert. It’s always a good idea to present the study to other researchers in the field, who have no conflicts of interest and are not involved in the research you’re covering and get their opinion. “Don’t take scientists at their word. Look into it. Ask other scientists, preferably the ones who don’t have a conflict of interest with the research,” says Bottesini.
4. Remember that a single study is simply one piece of a growing body of evidence.
“I have a general rule that a single study doesn’t tell us very much; it just gives us proof of concept,” says Peters. “It gives us interesting ideas. It should be retested. We need an accumulation of evidence.”
Aschwanden says as a practice, she tries to avoid reporting stories about individual studies, with some exceptions such as very large, randomized controlled studies that have been underway for a long time and have a large number of participants. “I don’t want to say you never want to write a single-study story, but it always needs to be placed in the context of the rest of the evidence that we have available,” she says.
Wilner advises journalists to spend some time looking at the scope of research on the study’s specific topic and learn how it has been written about and studied up to that point.
“We would want science journalists to be reporting balance of evidence, and not focusing unduly on the findings that are just in front of them in a most recent study,” Wilner says. “And that’s a very difficult thing to as journalists to do because they’re being asked to make their article very newsy, so it’s a difficult balancing act, but we can try and push journalists to do more of that.”
5. Remind readers that science is always changing.
“Science is always two steps forward, one step back,” says Peters. Give the public a notion of uncertainty, she advises. “This is what we know today. It may change tomorrow, but this is the best science that we know of today.”
Aschwanden echoes the sentiment. “All scientific results are provisional, and we need to keep that in mind,” she says. “It doesn’t mean that we can’t know anything, but it’s very important that we don’t overstate things.”
Authors of a study published in PNAS in January analyzed more than 14,000 psychology papers and found that replication success rates differ widely by psychology subfields. That study also found that papers that could not be replicated received more initial press coverage than those that could.
The authors note that the media “plays a significant role in creating the public’s image of science and democratizing knowledge, but it is often incentivized to report on counterintuitive and eye-catching results.”
Ideally, the news media would have a positive relationship with replication success rates in psychology, the authors of the PNAS study write. “Contrary to this ideal, however, we found a negative association between media coverage of a paper and the paper’s likelihood of replication success,” they write. “Therefore, deciding a paper’s merit based on its media coverage is unwise. It would be valuable for the media to remind the audience that new and novel scientific results are only food for thought before future replication confirms their robustness.”
Additional reading
Uncovering the Research Behaviors of Reporters: A Conceptual Framework for Information Literacy in Journalism
Katerine E. Boss, et al. Journalism & Mass Communication Educator, October 2022.The Problem with Psychological Research in the Media
Steven Stosny. Psychology Today, September 2022.Critically Evaluating Claims
Megha Satyanarayana, The Open Notebook, January 2022.How Should Journalists Report a Scientific Study?
Charles Binkley and Subramaniam Vincent. Markkula Center for Applied Ethics at Santa Clara University, September 2020.What Journalists Get Wrong About Social Science: Full Responses
Brian Resnick. Vox, January 2016.From The Journalist’s Resource
8 Ways Journalists Can Access Academic Research for Free
5 Things Journalists Need to Know About Statistical Significance
5 Common Research Designs: A Quick Primer for Journalists
5 Tips for Using PubPeer to Investigate Scientific Research Errors and Misconduct
What’s Standard Deviation? 4 Things Journalists Need to Know
This article first appeared on The Journalist’s Resource and is republished here under a Creative Commons license.
Don’t Forget…
Did you arrive here from our newsletter? Don’t forget to return to the email to continue learning!
Learn to Age Gracefully
Join the 98k+ American women taking control of their health & aging with our 100% free (and fun!) daily emails:
-
Kava vs Anxiety
10almonds is reader-supported. We may, at no cost to you, receive a portion of sales if you purchase a product through a link in this article.
Kava, sometimes also called “kava kava” but we’re just going to call it kava once for the sake of brevity, is a heart-shaped herb that
bestows the powers of the Black Pantheris popularly enjoyed for its anxiolytic (anxiety-reducing) effects. Despite the similarity of the name in many languages, it is unrelated to coffee (except insofar as they are both plants), and its botanical name is Piper methysticum.Does it work?
Yes! At least in the short-term; more on that later.
Firstly, you may be wondering how it works; it works by its potentiation of GABA receptors in the brain. GABA (or gamma-aminobutyric acid, to give it its full name), as you may recall, is a neurotransmitter that is associated with feelings of calm; we wrote about it here:
So, what does “potentiation of GABA receptors” mean? It means… Scientists don’t for 100% sure know how it works yet, but it does make GABA receptors fire more. It’s possible that to some degree GABA fits the “molecular lock” of the receptors and causes them to say “GABA is here”; it’s also possible that they just make them more sensitive to the real GABA that is there, or there could be another explanation as yet undiscovered. Either way, it means that taking kava has a similar effect to having increased GABA levels in the brain:
As for how much to use, 20–300mg appears to be an effective dose, and most sources recommend 80–250mg:
Kava as a Clinical Nutrient: Promises and Challenges
This review of clinical trials found that it was more effective than placebo in only 3 of 7 trials; specifically, it was beneficial in the short-term and not in the long-term. For these reasons, the researchers concluded:
❝Kava Kava appears to be a short-term treatment for anxiety, but not a replacement for prolonged anti-anxiety use. Although not witnessed in this review, liver toxicity is especially possible if taken longer than 8 weeks.❞
Another review of clinical trials found better results over the course of 11 clinical trials, though again, short-term treatment only was considered to be where the “safe and effective” claim can be placed:
❝Compared with placebo, kava extract appears to be an effective symptomatic treatment option for anxiety. The data available from the reviewed studies suggest that kava is relatively safe for short-term treatment (1 to 24 weeks), although more information is required. Further rigorous investigations, particularly into the long-term safety profile of kava are warrant❞
Source: Kava extract for treating anxiety
Is it safe?
Nope! It has been associated with liver damage:
The likely main mechanism of toxicity is that it simply monopolizes the liver’s metabolic abilities, meaning that while it’s metabolizing the kava, it’s not metabolizing other things (such as alcohol or other medications), which will then build up, and potentially overwhelm the liver:
Constituents in kava extracts potentially involved in hepatotoxicity: a review
However, traditionally-prepared kava has not had the same effect as modern extracts; at first it seemed the difference was the traditional aqueous extracts vs modern acetonic/ethanolic extracts, but eventually that was found not to be the case, as toxicity occurred with industrial aqueous extracts too. The conclusion so far is that it is about the quality of the source ingredients, and the problems inherent to mass-production:
Meanwhile, short-term use doesn’t seem to have this problem, if you’re not drinking alcohol or taking medications that affect the liver:
Mechanisms/risk factors – kava-associated hepatotoxicity ← you’ll need to scroll down to 4.2.4 to read about this
Want to try it?
If the potential for hepatotoxicity doesn’t put you off, here’s an example product on Amazon ← we do not recommend it, but we are not the boss of you, and maybe you’re confident about your liver and want to use it only very short-term?
Take care!
Don’t Forget…
Did you arrive here from our newsletter? Don’t forget to return to the email to continue learning!
Learn to Age Gracefully
Join the 98k+ American women taking control of their health & aging with our 100% free (and fun!) daily emails: