The debate on how to combat social media misinformation is as relevant as ever. In recent years, we’ve seen medical misinformation spreading alongside COVID, and political misinformation impacting the outcome of elections and national referendums.
Experts have largely placed the impetus to stop misinformation on three groups: social media users, government regulators, and social media platforms themselves.
Tips for social media users include educating themselves on the subject, and being aware of the tactics used to spread misinformation. Government bodies are urged to work together with social media platforms to regulate and prevent the spread of misinformation.
But all of these approaches depend on social media content we can easily recognise as “reliable”. Unfortunately, our new study just published in The Journal of Health Communication shows reliable content is not easy to find. Even we – researchers and subject matter experts – struggled to identify the evidence base of the social media posts we analysed.
Based on our findings, we have developed new guidelines to help experts create engaging content that also clearly communicates the evidence behind the post.
What did our study find?
We collected and analysed 300 mental health research-related X (formerly Twitter) posts from two large Australian mental health research organisations, posted between September 2018 and September 2019. We chose this timeframe to avoid the influence of recent major events like the Australian bushfires (starting November 2019) or the COVID-19 pandemic (starting March 2020) on our findings.
We assessed the written content within each individual post, whether the post contained hashtags, mentions, hyperlinks and multimedia, and whether the post featured an evidence-based source, such as a peer-reviewed journal article, conference presentation, or clinical treatment guidelines.
We found it was challenging to reliably establish the evidence behind the posts. Our team agreed on whether a post contained evidence-based information only 56% of the time. These chances are not much better than a coin flip.
When people with years of relevant training can’t reliably identify evidence-based content, how can we expect everyone else to?
Examples of fake X posts communicating mental health research where the evidence-basis of the information is not clearly communicated in the post. Erin Madden et al.
Although some posts appeared to be evidence-informed (for example, a researcher commenting on their area of expertise), the source of their statement was unclear in the majority of posts.
This raises further questions about the distinction between misinformation or disinformation, and “poor quality information”. We would argue the latter happens when the level of evidence supporting a post is not recognisable even to those trained in the field.
Additional examples of fake X posts on mental health research where the evidence-basis of the information is not clearly communicated in the post. Erin Madden et al.
While some posts contained links to evidence-based sources, such as peer-reviewed papers, most contained expert opinions only, such as news articles. To judge how reliable this content is, a user would need to put in significant extra work – reading resources thoroughly and following up on sources.
It’s not realistic to ask social media users to do this, especially when peer-reviewed sources use complex, technical language and often require payment to access.
How can researchers communicate evidence-based content?
Academics are encouraged to translate their research for the public, but there is limited guidance on how to balance audience engagement with evidence-based information.
As part of our study, we aimed to create simple, evidence-based guidelines for researchers, on how to effectively disseminate mental health research via X. For example, we found that researchers can boost engagement by emphasising the specific population group the research relates to (for example LBGTIQA+ or culturally diverse communities) and by using images and videos.
Our guidelines encourage experts to create engaging content that also clearly communicates “what” the information is, “where” the information is from, and “who” the information is for.
Our guidelines for sharing evidence-based information on X. Erin Madden et al., CC BY-ND
We outlined five main points:
What do I say?
1. tell the audience what the information is, and where it came from
2. tell your audience who you’re speaking to
How do I say it?
3. use media that adds to the post, expanding on what the information is and where it came from
4. use hyperlinks to verify research information in the post
5. minimise over-use of hashtags.
We also created mock “before and after” posts inspired by the posts from the study dataset, which outline the use of the guidelines in practice.
A ‘before and after’ post, taking into account the guidelines on best practice for sharing evidence-based information. Erin Madden et al., CC BY-ND
The recent increased focus on tackling online misinformation points to a hopeful future for information quality on social media platforms. But we still have a long way to go to promote evidence-based research in an accessible manner.
Research organisations and experts have a unique opportunity to show leadership in this area. Academics can add their voices to the public discourse as trusted and reliable sources of content people want to engage with – all while maintaining research integrity.
Additionally, in today’s world where health information is often promoted, shared and re-shared by everyone, these guidelines also provide useful tips to help anyone share clear and reliable information, and empower people to effectively navigate health information on social media.


US Charges Two Men in Alleged Nvidia Chip Smuggling Scheme to China
SpaceX Reportedly Preparing Record-Breaking IPO Targeting $1.5 Trillion Valuation
Australia Enforces World-First Social Media Age Limit as Global Regulation Looms
Trump Criticizes EU’s €120 Million Fine on Elon Musk’s X Platform
Evercore Reaffirms Alphabet’s Search Dominance as AI Competition Intensifies
SK Hynix Labeled “Investment Warning Stock” After Extraordinary 200% Share Surge
EU Court Cuts Intel Antitrust Fine to €237 Million Amid Long-Running AMD Dispute
Adobe Strengthens AI Strategy Ahead of Q4 Earnings, Says Stifel
SpaceX Insider Share Sale Values Company Near $800 Billion Amid IPO Speculation
Trump’s Approval of AI Chip Sales to China Triggers Bipartisan National Security Concerns
Nvidia Develops New Location-Verification Technology for AI Chips
Mizuho Raises Broadcom Price Target to $450 on Surging AI Chip Demand
SoftBank Shares Slide as Oracle’s AI Spending Plans Fuel Market Jitters
China Adds Domestic AI Chips to Government Procurement List as U.S. Considers Easing Nvidia Export Curbs
Australia’s Under-16 Social Media Ban Sparks Global Debate and Early Challenges 







