- Malaysia’s government has urged social media platforms to intensify efforts against a surge in harmful online content, referring 51,638 cases in Q1 2024 alone.
- It is critical and necessary that social media platforms take on increasing responsibility for content policing, especially in regions with strict regulations.
OUR TAKE
The organisations that regulate Internet behaviour are still limited in what they can do. It is the social platforms that hold the cards, and their consensus of prioritising profit will undoubtedly undermine cybersecurity. Maintaining network security is the responsibility of every Internet participant. Companies that make huge profits from the Internet should also fulfil this obligation.
–Ashley Wang, BTW reporter
What happened
Malaysia’s government has urged social media platforms to intensify their efforts against a surge in harmful online content. The country’s communications minister, Fahmi Fadzil, revealed that in the first quarter of 2024 alone, 51,638 cases were referred to social media companies for action—surpassing the total number of cases from the previous year.
The increase in harmful content, including scams, cyberbullying, child pornography, and sensitive topics related to race, religion, and royalty, has prompted Malaysia to seek stronger cooperation from major platforms like Meta, TikTok, Telegram, and X (formerly Twitter).
Minister Fadzil highlighted the mixed responses from these companies, noting that Meta showed the highest compliance rates, with Facebook, Instagram, and WhatsApp achieving 85%, 88%, and 79%, respectively. TikTok followed with a 76% compliance rate, while Telegram and X had significantly lower rates at 65% and 25%.
Also read: Social media marketing trends in Malaysia: Shinjiru’s key insights for 2024
Also read: Who is Philip Belamant? CEO of Zilch, combining fintech and social causes
Why it’s important
Malaysia’s proactive stance against cybercrime and harmful content, areas that have seen a marked increase, showcases its struggle with cybercrime. The government’s approach includes issuing directives and flagging content that breaches local laws. However, the government is still in a passive stance, as the final decision to remove such content rests with the platforms based on their community guidelines.
This initiative underscores the growing responsibility of social media companies to regulate harmful content, particularly in regions with strict cultural and legal norms. As cyber threats and the dissemination of harmful content rise, the compliance and cooperation of social media firms become crucial in maintaining public safety and social harmony. Social media platforms must be held accountable, not just to their shareholders, but to the communities they serve. If they won’t act responsibly, perhaps stricter regulations should be.