Disinformation Targeting Chinese-Australian Communities: A Threat to Democracy
The 2025 Australian election cycle has witnessed a concerning rise in disinformation campaigns targeting Chinese-speaking communities, posing significant threats to both these communities and the integrity of the democratic process. A study by the RECapture research team, focusing on the popular platforms WeChat and RedNote (Xiaohongshu), reveals that the primary drivers of this disinformation are not foreign actors, as often assumed, but rather commercial and domestic political interests exploiting the unique vulnerabilities of these platforms. This disinformation manipulates pre-existing anxieties within these communities, misrepresents political stances, and ultimately aims to influence voter behavior. The limited regulatory oversight of these platforms and the persistent nature of these narratives exacerbate the harm caused by these campaigns.
While the traditional understanding of disinformation hinges on the criteria of falsity, intent to deceive, and demonstrable harm, the RECapture study highlights the complexity of disinformation in practice. The 2023 Voice referendum study revealed that disinformation isn’t always easily categorized as true or false, and its impact can be difficult to quantify. This ambiguity, compounded by Australia’s lack of a clear legal definition for online misinformation and disinformation, presents significant challenges for researchers and regulators alike. For the purposes of the RECapture study, the focus has been on deliberate misrepresentations of policy and manipulative political speech designed to sway voters.
The research uncovered several concerning trends in disinformation targeting Chinese-Australian communities. These campaigns often exploit existing anxieties regarding issues like investor visas, undocumented migration, humanitarian programs, and Australia’s diplomatic relations with countries like India, the US, and China. Several tactics are employed, including exaggerating the likelihood of specific policy changes, manipulating timelines and contexts to revive old news stories and present them as current, and misleadingly pairing visuals with text to create false impressions. Interestingly, the research indicates political parties are not directly behind these campaigns, suggesting other actors are benefiting from sowing discord and manipulating public opinion within these communities.
One example highlighted by the research involved a RedNote post misrepresenting Prime Minister Anthony Albanese’s comments on immigration. The post misconstrued his remarks on a "balanced" immigration approach, falsely claiming Labor would grant amnesty to all immigrants. This distortion fueled discussions within the comments section, with many advocating for a class-based immigration system that prioritized wealthier migrants over humanitarian intakes. Another example involved a WeChat article published by AFN Daily with a sensationalized and ambiguous headline designed to attract clicks. The article misrepresented polling data, falsely claiming the Coalition was ahead of Labor, and promoted a racially charged narrative by falsely alleging Labor had naturalized thousands of Indian-origin citizens to influence the election. This narrative played on existing tensions and prejudices within the Chinese-Australian community.
Further compounding the issue is the vulnerability of these communities to disinformation disseminated through WeChat and RedNote. The limited regulatory oversight of these platforms allows disinformation to spread unchecked. While concerns about cybersecurity and foreign interference are legitimate, the current regulatory landscape has inadvertently created a breeding ground for domestically generated disinformation. This lack of oversight, combined with the closed-off nature of these platforms, makes it difficult to effectively identify and counter disinformation campaigns. Moreover, the persistence of specific disinformation narratives, often laced with racial stereotypes and partisan biases, across multiple election cycles demonstrates the insidious and enduring nature of this problem.
Addressing this complex issue requires a multi-pronged approach. Given the finding that domestic actors, rather than foreign interference, are the primary drivers of disinformation on these platforms, regulatory solutions focusing on foreign interference alone will be inadequate. Targeted civic education and media literacy initiatives tailored to the specific concerns and information consumption habits of Chinese-Australian communities are crucial. While grassroots efforts by community members to debunk false information in comment sections are commendable, a more systematic approach to fostering critical thinking and digital literacy is needed.
Traditional methods of automated disinformation detection and debunking face limitations on WeChat and RedNote due to restrictions on external tools and the imperfect functionality of internal flagging systems. Consequently, human intervention remains the most effective means of identifying and countering disinformation on these platforms. This highlights the critical role of community members, researchers, and media organizations in actively engaging with and fact-checking information circulating within these communities. Ultimately, empowering individuals with the tools to critically evaluate information, regardless of the platform they use, is essential for protecting the integrity of the democratic process and fostering a healthy information environment.