TikTok will now explain why ‘For You’ videos are being recommended

>

In a bid to be more ‘transparent’, TikTok has introduced a new tool that explains why specific videos are being recommended to users. 

The tool, available on the ‘For You’ feed on the TikTok app, brings up a window entitled ‘Why you’re seeing this video’. 

It then gives a list of reasons, which can include ‘this video is popular in your region’ or ‘this video was recently posted by an account you follow’. 

TikTok, which is owned by Chinese company Bytedance, says the new feature brings ‘more context’ and ‘meaningful transparency’ for users. 

The tool, available on the 'For You' page on the TikTok app, brings up a window entitled 'Why you're seeing this video'

The tool, available on the ‘For You’ page on the TikTok app, brings up a window entitled ‘Why you’re seeing this video’

How does it work? 

– In your For You feed, tap on the share panel

– Then, tap the question mark icon called ‘Why this video?’

– From there, you can see reasons why a particular video was recommended 

The platform uses an algorithm to promote videos to users on the For You page from people they don’t even follow – but this has proved controversial for promoting dangerous content.

For example, a recent Daily Mail investigation found teenagers are being bombarded with self-harm and suicide content on TikTok within minutes of joining the platform. 

‘At TikTok, we want people to feel empowered creating, connecting, and engaging on our platform,’ TikTok says in a blog post

‘That’s why we equip creators and viewers with a range of features, tools, and resources so they can stay in control of their experience. 

‘Today we’re adding to that toolbox with a feature that helps bring more context to content recommended in For You feeds.’

Back in June 2020, TikTok published a detailed blog post explaining how the For You page works, as part of its transparency push.  

When TikTok users open the app, they are presented with the For You feed – a stream of videos recommended to users. 

A recent Daily Mail investigation found teenagers are being bombarded with self-harm and suicide content on TikTok within minutes of joining the platform. An account set up by the Daily Mail as 14-year-old Emily was shown posts about suicidal thoughts within five minutes of expressing an interest in depression content (pictured)

A recent Daily Mail investigation found teenagers are being bombarded with self-harm and suicide content on TikTok within minutes of joining the platform. An account set up by the Daily Mail as 14-year-old Emily was shown posts about suicidal thoughts within five minutes of expressing an interest in depression content (pictured)

A recent Daily Mail investigation found teenagers are being bombarded with self-harm and suicide content on TikTok within minutes of joining the platform. An account set up by the Daily Mail as 14-year-old Emily was shown posts about suicidal thoughts within five minutes of expressing an interest in depression content (pictured) 

TikTok’s For You page 

‘For You’ is a section of TikTok that recommends content to users. 

For You is powered by an algorithm that makes personalised recommendations based on several metrics, including viewing behaviour and interactions with videos. 

The algorithm refines the choice of videos as it gathers more information about the user’s preferences and interests. 

​​One issue is that the algorithms that personalise For You can push harmful content if the algorithms do not identify and downgrade it.

Back in June 2020, TikTok published a detailed blog post explaining how the For You page works, as part of its transparency push. 

‘On TikTok, the For You feed reflects preferences unique to each user,’ the platform said at the time. 

‘The system recommends content by ranking videos based on a combination of factors – starting from interests you express as a new user and adjusting for things you indicate you’re not interested in, too.’ 

For You is powered by an algorithm that makes personalised recommendations based on several metrics, including viewing behaviour and interactions with videos. 

Similar to the recommendations offered by search engines, streaming services and other social media platforms, this is designed to give users a more personalised experience. 

Unfortunately, an inherent issue with social media algorithms is that users are directed to a stream of posts related to a video that they’ve viewed. 

Just as a user who sees an explosion of football-related videos after watching a person kick a ball into a goal, self-harm or suicide-related posts can end up bombarding a user’s feed.

This came to a tragic head in 2017 when London school girl Molly Russell ended her life after viewing Instagram content linked to anxiety, depression, self-harm and suicide. 

Molly’s father, Ian, who now campaigns for online safety, has previously said the ‘pushy algorithms’ of social media ‘helped kill my daughter’. 

A recent report from the Center for Countering Digital Hate (CCDH) found TikTok’s For You page is ‘bombarding vulnerable teenagers with dangerous content that might encourage self-harm, suicide, disordered eating and eating disorders’.  

Researchers at the non-profit organisation set up accounts posing as 13-year-old girls and ‘liked’ harmful content whenever it was recommended. 

Videos referencing suicide were served to one account within 2.6 minutes, while eating disorder content was served to one account within eight minutes, they found.

The platform uses an algorithm to promote videos to users on the For You page – but this has proved controversial for promoting dangerous content (file photo)

The platform uses an algorithm to promote videos to users on the For You page – but this has proved controversial for promoting dangerous content (file photo)

The platform uses an algorithm to promote videos to users on the For You page – but this has proved controversial for promoting dangerous content (file photo) 

CCDH said TikTok’s For You page lacks ‘meaningful transparency’ and its algorithm operates ‘in an opaque manner’. 

The platform allows users as young as 13 to create an account by asking for a date of birth, although it doesn’t stop kids below this age from using it. 

‘All platforms need to ensure content being targeted at children is consistent with what parents rightly expect to be safe,’ said Richard Collard, NSPCC policy and regulatory manager for child safety online.

TikTok says it does not allow self-harm and suicide content if it ‘promotes, glorifies or normalises’. 

Read similar stories here… 

TikTok tests landscape videos in bid to ‘move into YouTube territory’ 

TikTok to add 3,000 engineers globally while rivals laid off thousands 

Twitter Blue is relaunched – but iPhone users have to pay more 

Daily Mail investigation shows teenagers are being bombarded with self-harm and suicide content on TikTok within minutes

A recent Daily Mail investigation found teenagers are being bombarded with self-harm and suicide content on TikTok within minutes of joining the platform.  

An account set up by the Daily Mail as 14-year-old Emily was shown posts about suicidal thoughts within five minutes of expressing an interest in depression content.

After 15 minutes of scrolling, its algorithm was demonstrating how to hide content from parents.

Within 24 hours, the account had been bombarded with more than 1,000 videos about depression, self-harm, suicide and eating disorders. Some videos had millions of views. 

TikTok says that the account the Mail created for this experiment ‘liked’ a number of videos which were proactively removed by its team for violating its policies, without being reported. 

Read more