1 00:00:02,000 --> 00:00:07,000 Downloaded from YTS.MX 2 00:00:08,000 --> 00:00:13,000 Official YIFY movies site: YTS.MX 3 00:00:19,300 --> 00:00:22,195 - TikTok has very much become a way 4 00:00:22,220 --> 00:00:23,660 for the young generation 5 00:00:23,700 --> 00:00:25,600 to express ourselves in every way. 6 00:00:27,113 --> 00:00:29,676 You can be yourself, you're at home, you're filming, 7 00:00:29,700 --> 00:00:31,586 and there's always gonna be like millions 8 00:00:31,610 --> 00:00:33,036 of people watching you. 9 00:00:33,540 --> 00:00:35,202 I like this one. Look and see. 10 00:00:35,382 --> 00:00:37,254 You never know when you could blow up. 11 00:00:41,450 --> 00:00:43,036 - I think every young kid's dream 12 00:00:43,060 --> 00:00:45,047 is to be successful online. 13 00:00:45,530 --> 00:00:47,326 The Chinese social media platform, 14 00:00:47,350 --> 00:00:49,963 TikTok, has changed the internet. 15 00:00:51,070 --> 00:00:54,346 It's become the most popular app in the world. 16 00:00:54,370 --> 00:00:56,166 - It's fishing videos, it's cooking. 17 00:00:56,190 --> 00:00:58,056 You can make skits, singing, dancing. 18 00:00:58,080 --> 00:01:00,827 Literally everything you can think of, TikTok have. 19 00:01:05,540 --> 00:01:07,306 - It's not an app on their phone anymore. 20 00:01:07,330 --> 00:01:08,356 It's their livelihood. 21 00:01:08,380 --> 00:01:10,166 It's how they communicate with their friends. 22 00:01:10,190 --> 00:01:11,636 It's how they see their world. 23 00:01:11,660 --> 00:01:13,976 That's a part that I don't think everybody 24 00:01:14,000 --> 00:01:15,273 has adjusted to yet. 25 00:01:17,060 --> 00:01:20,546 - We're really at risk of having generations of young people 26 00:01:20,570 --> 00:01:23,506 that performed identities in response to something 27 00:01:23,530 --> 00:01:27,683 that a technology platform prescribes to be the new normal. 28 00:01:29,250 --> 00:01:31,106 Behind the shiny dance videos, 29 00:01:31,130 --> 00:01:34,673 the platform is leading people down dangerous paths. 30 00:01:36,130 --> 00:01:37,916 - I'd like to think that I wouldn't have struggled 31 00:01:37,940 --> 00:01:40,349 with an eating disorder if I hadn't downloaded TikTok. 32 00:01:41,840 --> 00:01:43,106 - My claim with TikTok 33 00:01:43,130 --> 00:01:47,146 is that they are harvesting huge amounts of data illegally 34 00:01:47,170 --> 00:01:49,552 without the consent of children or their parents. 35 00:01:50,990 --> 00:01:52,746 - If you just look at TikTok in isolation, 36 00:01:52,770 --> 00:01:54,006 it seems innocuous. 37 00:01:54,030 --> 00:01:56,876 But it's really takes place in this much larger context 38 00:01:56,900 --> 00:01:59,656 of data collection, artificial intelligence, 39 00:01:59,680 --> 00:02:00,966 and a real effort by the Chinese 40 00:02:00,990 --> 00:02:04,483 to consolidate influence in the region and across the globe. 41 00:02:06,150 --> 00:02:08,556 - Tonight on four corners, TikTok. 42 00:02:08,580 --> 00:02:11,256 In a joint investigation with Hack on Triple J, 43 00:02:11,280 --> 00:02:13,056 we're going down the rabbit hole 44 00:02:13,080 --> 00:02:15,176 to reveal the dark side of app. 45 00:02:15,200 --> 00:02:17,686 How the platform censors political content 46 00:02:17,710 --> 00:02:20,046 and harvests children's data. 47 00:02:20,070 --> 00:02:23,886 And how the app's powerful algorithm exposes people 48 00:02:23,910 --> 00:02:26,773 to misinformation and dangerous content. 49 00:02:35,660 --> 00:02:37,456 - Hi, my name is Rory Eliza. 50 00:02:37,480 --> 00:02:38,925 And, then what? 51 00:02:40,416 --> 00:02:42,076 And what do you do? 52 00:02:42,100 --> 00:02:43,848 - I am a full-time TikToker. 53 00:02:53,250 --> 00:02:54,626 So in the morning I'll wake up, 54 00:02:54,650 --> 00:02:55,946 maybe eight o'clock, nine o'clock. 55 00:02:55,970 --> 00:02:57,226 I'll check my phone, 56 00:02:57,250 --> 00:02:59,176 check if my videos have done well, 57 00:02:59,200 --> 00:03:01,646 or how my followers are reacting 58 00:03:01,670 --> 00:03:03,723 to the content that I've just posted. 59 00:03:07,060 --> 00:03:09,740 TikTok honestly, I get so much love on there. 60 00:03:09,770 --> 00:03:12,136 It's so weird because that's my biggest platform, 61 00:03:12,160 --> 00:03:13,736 is TikTok with 5 million followers. 62 00:03:13,760 --> 00:03:16,326 It's crazy to think that 5 million people, 63 00:03:16,350 --> 00:03:18,376 that's people, it's not just the number. 64 00:03:18,400 --> 00:03:20,216 And if you really think about it's 5 million people 65 00:03:20,240 --> 00:03:21,640 that have tapped that follow button. 66 00:03:23,230 --> 00:03:24,492 They're all just so friendly and 67 00:03:24,516 --> 00:03:25,656 they're kind of like your family. 68 00:03:25,680 --> 00:03:27,483 It's just weird, like you don't know these people, 69 00:03:27,507 --> 00:03:29,216 but they know so much about you 70 00:03:29,240 --> 00:03:31,481 that they treat you like a family member. 71 00:03:43,750 --> 00:03:46,763 And on that note, welcome to the new Rory Eliza. 72 00:03:47,900 --> 00:03:49,796 Rory Eliza is one of millions 73 00:03:49,820 --> 00:03:53,026 of young Australians recording virtually every moment 74 00:03:53,050 --> 00:03:55,666 of their lives to get famous on TikTok. 75 00:03:55,690 --> 00:03:57,466 - Get ready with me for a date. 76 00:03:57,490 --> 00:03:58,916 Yo, I chose an outfit, let's go. 77 00:03:58,940 --> 00:04:00,686 Transition, yeah. 78 00:04:00,710 --> 00:04:02,486 So I think every young kid's dream 79 00:04:02,510 --> 00:04:04,776 is to, you know, be successful online. 80 00:04:04,800 --> 00:04:08,116 So, I think there's definitely a group 81 00:04:08,140 --> 00:04:10,276 where they all wanna be influencers 82 00:04:10,300 --> 00:04:11,726 'cause it's kind of like the in thing now. 83 00:04:11,750 --> 00:04:13,456 And I think that's because of TikTok. 84 00:04:14,340 --> 00:04:15,700 TikTok has been downloaded 85 00:04:15,710 --> 00:04:18,663 more than 3 billion times around the world. 86 00:04:21,710 --> 00:04:23,717 It's become a cultural phenomenon. 87 00:04:23,742 --> 00:04:24,773 I'm 21 and just 88 00:04:24,798 --> 00:04:26,123 learned how to do my own laundry. 89 00:04:26,210 --> 00:04:28,536 Some tomatoes and some cheese. 90 00:04:28,560 --> 00:04:31,320 Dude, no you're gotta go late like, hey. 91 00:04:32,380 --> 00:04:33,553 Oh, okay, okay. Okay. 92 00:04:38,470 --> 00:04:40,570 Everything is about going viral. 93 00:04:41,850 --> 00:04:43,406 The dance, started by someone 94 00:04:43,430 --> 00:04:46,138 in their living room and uploaded to TikTok, 95 00:04:47,720 --> 00:04:49,876 can turn into a stadium full of people, 96 00:04:49,900 --> 00:04:51,586 performing it in unison. 97 00:04:58,165 --> 00:05:00,096 - I like how creative you can be on it. 98 00:05:00,120 --> 00:05:02,293 Like it's just so fun to go on that app 99 00:05:02,317 --> 00:05:04,138 and just express your real self. 100 00:05:13,970 --> 00:05:16,626 Rory started posting comedic skits on TikTok 101 00:05:16,650 --> 00:05:18,153 and her following snowballed. 102 00:05:19,090 --> 00:05:21,827 - No worries. I'll scan that one right through for you right now. 103 00:05:22,700 --> 00:05:24,436 This was her first viral video. 104 00:05:24,460 --> 00:05:26,746 It got nearly 14 million views. 105 00:05:31,050 --> 00:05:33,637 - How about some Peking duck? 106 00:05:33,870 --> 00:05:36,816 Oh yeah, but we actually don't have the Peking duck. 107 00:05:36,840 --> 00:05:38,586 But we've got this sneaking goose. 108 00:05:38,610 --> 00:05:40,102 What an odd name. 109 00:05:40,126 --> 00:05:42,756 It is pretty normal for a book. 110 00:05:42,780 --> 00:05:46,016 In 2019, Rory decided to leave school 111 00:05:46,040 --> 00:05:48,166 to become a full time TikToker. 112 00:05:48,590 --> 00:05:50,623 - Wait, wait, is this a library? 113 00:05:51,510 --> 00:05:53,870 School was just one of those things. I just was not good at. 114 00:05:54,320 --> 00:05:57,156 I decided to leave school when I was in year 11 115 00:05:57,180 --> 00:05:58,536 and I was never there, you know. 116 00:05:58,560 --> 00:06:00,336 I was always in Sydney doing meetings 117 00:06:00,360 --> 00:06:02,526 or presentations for TikTok ends. 118 00:06:02,550 --> 00:06:04,576 I just wasn't there. And when I would come to school, 119 00:06:04,600 --> 00:06:06,476 I would have no idea what we're doing. 120 00:06:06,500 --> 00:06:09,093 'Cause you know, I've been away for heaps of days. 121 00:06:10,440 --> 00:06:13,076 No worries at all. Thanks for coming to our library. 122 00:06:13,100 --> 00:06:15,051 School, you can go back and do it at TAFE. 123 00:06:15,075 --> 00:06:17,026 You can go back any time and do it if you need it. 124 00:06:17,050 --> 00:06:19,106 But you may never get this opportunity again. 125 00:06:19,130 --> 00:06:21,256 So, we just thought it was worth leaving school 126 00:06:21,280 --> 00:06:22,796 and pursuing all the business opportunities 127 00:06:22,820 --> 00:06:23,846 while they were there for her. 128 00:06:23,870 --> 00:06:26,416 - No worries at all. Thanks for coming to our library. 129 00:06:26,680 --> 00:06:28,648 - How do you feel about the fact that 5 million people are 130 00:06:28,672 --> 00:06:30,446 watching her content? 131 00:06:30,470 --> 00:06:31,927 - It's incredible. It's even, when she 132 00:06:31,951 --> 00:06:33,606 goes live, there was a time she went live 133 00:06:33,630 --> 00:06:36,226 and she had 22,000 people watching her in her room. 134 00:06:36,250 --> 00:06:38,226 And I just sort of, in my mind, goes back 135 00:06:38,250 --> 00:06:39,263 to Elton John concert here. 136 00:06:39,287 --> 00:06:40,756 And she had more people watching her. 137 00:06:40,780 --> 00:06:42,436 Than we had at that Elton John concert. 138 00:06:42,460 --> 00:06:44,226 And it kind of way out that's happening 139 00:06:44,250 --> 00:06:46,336 in my daughter's bedroom at the moment. 140 00:06:46,360 --> 00:06:49,849 It was a bit yeah, different. 141 00:06:54,130 --> 00:06:56,266 - Big fashion and cosmetic brands 142 00:06:56,290 --> 00:06:59,126 started noticing Rory success on TikTok 143 00:06:59,150 --> 00:07:01,553 and wanted to tap into a growing audience. 144 00:07:07,149 --> 00:07:09,306 Companies sponsor influences like Rory. 145 00:07:09,330 --> 00:07:13,083 And businesses pay TikTok to advertise on the platform. 146 00:07:14,680 --> 00:07:17,553 This is central to the app's lucrative business model. 147 00:07:19,110 --> 00:07:21,226 - In this work industry being an influencer, 148 00:07:21,250 --> 00:07:23,436 you have to present yourself as a brand, you know. 149 00:07:23,460 --> 00:07:25,566 We aren't really people anymore, we're brands. 150 00:07:25,590 --> 00:07:26,926 We're selling products for brands. 151 00:07:26,950 --> 00:07:28,750 So, you kind of gotta look the part. 152 00:07:32,070 --> 00:07:34,494 The money involved it's enough to live off. 153 00:07:34,518 --> 00:07:36,835 So, it's a pretty fair amount. 154 00:07:36,860 --> 00:07:38,416 I'm about in the medium to high range 155 00:07:38,440 --> 00:07:39,926 of incomes in Australia. 156 00:07:39,950 --> 00:07:41,549 So yeah. Very, very descent. 157 00:07:42,600 --> 00:07:44,246 - Well, it's hard not to even be jealous sometimes 158 00:07:44,270 --> 00:07:46,016 'cause you look at our life and you know, 159 00:07:46,040 --> 00:07:48,590 we get up and we go to work and we come home. 160 00:07:49,905 --> 00:07:51,245 And she can earn money 161 00:07:51,270 --> 00:07:53,696 that can take us days to earn in minutes. 162 00:07:53,720 --> 00:07:56,526 - I found myself driving and just crying, 163 00:07:56,550 --> 00:07:58,666 having like a total breakdown. 164 00:07:58,690 --> 00:07:59,856 And I found myself having 165 00:07:59,880 --> 00:08:02,206 some really quite nasty thoughts and- 166 00:08:02,230 --> 00:08:05,566 - Rory shares her life with 5 million people. 167 00:08:05,590 --> 00:08:07,438 Even her lowest moments. 168 00:08:07,462 --> 00:08:08,994 - Why am I meant to be on this earth. 169 00:08:09,018 --> 00:08:11,626 Like, why does no one like me? 170 00:08:11,650 --> 00:08:13,263 Why do I have no friends? 171 00:08:14,630 --> 00:08:17,459 But most days she feels very alone. 172 00:08:17,483 --> 00:08:19,286 - Okay. That's an old name for, 173 00:08:19,310 --> 00:08:22,176 Being away from people, it's definitely lonely, you know. 174 00:08:22,200 --> 00:08:24,136 I film, oh, four videos a day. 175 00:08:24,160 --> 00:08:25,778 That's a good three hours outta my day. 176 00:08:25,820 --> 00:08:27,100 And then I've got another eight hours 177 00:08:27,111 --> 00:08:28,476 and I'm like, what the heck am I gonna do 178 00:08:28,500 --> 00:08:30,412 for the rest of the day? Like I can't ring out my friends. 179 00:08:30,436 --> 00:08:32,226 Like, y'all, want to hang out? 'Cause they're at work. 180 00:08:32,250 --> 00:08:34,201 So it definitely gets lonely at times. 181 00:08:34,230 --> 00:08:35,726 And you know, sometimes if you're reading 182 00:08:35,750 --> 00:08:38,006 the hate comments and the stress load, 183 00:08:38,030 --> 00:08:39,386 it can be so much for your body 184 00:08:39,410 --> 00:08:41,856 and you're just overwhelmed and you're lonely. 185 00:08:41,880 --> 00:08:43,984 So that can also creep into depression. 186 00:08:44,008 --> 00:08:45,294 - Catherine hasn't had a question. 187 00:08:45,318 --> 00:08:47,953 I'm happy to return to you but let's just keep it civil. 188 00:08:47,977 --> 00:08:48,977 Andrew? 189 00:08:50,860 --> 00:08:51,860 Catherine. 190 00:08:55,860 --> 00:08:58,106 With people stuck at home during lockdown, 191 00:08:58,130 --> 00:08:59,896 desperate for entertainment. 192 00:08:59,920 --> 00:09:03,660 TikTok became the world's most downloaded app in 2020. 193 00:09:07,010 --> 00:09:09,779 And it's continued to hold that title this year. 194 00:09:13,172 --> 00:09:17,416 - TikTok in Australia has seen the same kind of bump 195 00:09:17,440 --> 00:09:19,663 in 2020 as elsewhere in the world. 196 00:09:23,280 --> 00:09:24,306 In October of 2020, there were 197 00:09:24,330 --> 00:09:26,406 an estimated 2.5 million users on TikTok. 198 00:09:26,430 --> 00:09:29,230 Which was about a 50% growth from earlier on in the year. 199 00:09:41,863 --> 00:09:43,522 Of the popular social media apps, 200 00:09:43,547 --> 00:09:45,545 TikTok is the most addictive. 201 00:09:46,030 --> 00:09:48,966 Late TikTok advertising data shows users spend 202 00:09:48,990 --> 00:09:52,213 an average of an hour and a half on the app each day. 203 00:09:54,270 --> 00:09:56,688 - You know it's like 8:00 PM and I'm watching and watching 204 00:09:56,712 --> 00:09:59,216 and then I look up at my clock and it's 2:00 AM. 205 00:09:59,434 --> 00:10:01,394 And I'm like, where the heck did those hours go? 206 00:10:01,420 --> 00:10:03,380 It's cause this, um, "For You" page is so addictive. 207 00:10:03,390 --> 00:10:05,174 It's just so spot on. 208 00:10:09,290 --> 00:10:12,090 TikTok's algorithm is its most valuable asset. 209 00:10:14,610 --> 00:10:17,296 It's designed to determine your interests 210 00:10:17,320 --> 00:10:19,406 and send you personalized content 211 00:10:19,430 --> 00:10:22,086 to keep you on the app for as long as possible. 212 00:10:22,110 --> 00:10:23,186 - I went and saw my mama. 213 00:10:23,210 --> 00:10:25,226 And I went and got my hair done, as well just to- 214 00:10:25,250 --> 00:10:28,656 - TikTok works by recommending content to you 215 00:10:28,680 --> 00:10:30,976 through your activity on the app. 216 00:10:31,000 --> 00:10:33,506 So the more that you scroll through the app, 217 00:10:33,530 --> 00:10:35,176 the better the recommendations are tailored 218 00:10:35,200 --> 00:10:36,783 to your specific interests. 219 00:10:37,850 --> 00:10:40,806 Rather than selecting content that you want to watch 220 00:10:40,830 --> 00:10:43,231 like you would on YouTube or on Netflix. 221 00:10:43,600 --> 00:10:47,216 You primarily access content through one main feed, 222 00:10:47,240 --> 00:10:49,185 which is called the For You page on TikTok. 223 00:10:49,450 --> 00:10:51,936 Which is essentially just an endlessly scrolling, 224 00:10:51,960 --> 00:10:55,266 algorithmically curated feed of videos 225 00:10:55,290 --> 00:10:57,543 that refreshes each time you open the app. 226 00:10:58,947 --> 00:11:00,696 As soon as you sign up to TikTok, 227 00:11:00,720 --> 00:11:02,476 the app starts collecting data 228 00:11:02,500 --> 00:11:05,676 about you, your location, gender, and age, 229 00:11:05,700 --> 00:11:08,726 and also your facial data to figure out who you are 230 00:11:08,750 --> 00:11:10,863 and what kind of videos you want to see. 231 00:11:11,730 --> 00:11:15,703 - Your face is a form of biometric information. 232 00:11:15,727 --> 00:11:19,475 And your face can be analyzed to distinguish a range 233 00:11:19,500 --> 00:11:21,860 of personality and demographic traits. 234 00:11:24,590 --> 00:11:26,336 TikTok collects your facial data 235 00:11:26,360 --> 00:11:29,746 every time you make a video or use a filter on the app. 236 00:11:29,770 --> 00:11:33,816 And can even access photos and videos saved on your phone 237 00:11:33,840 --> 00:11:35,930 that aren't being used on the platform. 238 00:11:36,680 --> 00:11:40,686 To understand how an app like TikTok interprets that data 239 00:11:40,710 --> 00:11:42,656 scientists in Melbourne have developed 240 00:11:42,680 --> 00:11:44,706 what's called a biometric mirror. 241 00:11:45,160 --> 00:11:46,866 - So biometric mirror for instance, 242 00:11:46,890 --> 00:11:49,486 is trained by way of artificial intelligence 243 00:11:49,510 --> 00:11:51,466 to distinguish how intelligent you are, 244 00:11:51,490 --> 00:11:54,596 how attractive, how weird, how responsible 245 00:11:54,620 --> 00:11:57,723 and how emotionally unstable you are. 246 00:11:58,900 --> 00:12:00,366 The interesting thing there is of course, 247 00:12:00,390 --> 00:12:02,756 is that biometric mirror bases it's assumptions 248 00:12:02,788 --> 00:12:05,115 on a single snapshot of your face. 249 00:12:05,140 --> 00:12:07,936 So all of these assumptions are generated based 250 00:12:07,960 --> 00:12:10,606 on the exact appearance of your face 251 00:12:10,630 --> 00:12:13,653 at that exact microsecond that the photo has been taken. 252 00:12:14,834 --> 00:12:17,356 The TikTok algorithm might read your face 253 00:12:17,380 --> 00:12:19,906 and think that you are dealing 254 00:12:19,930 --> 00:12:22,456 with a significant mental health challenge. 255 00:12:22,480 --> 00:12:25,346 You might be presented with videos 256 00:12:25,370 --> 00:12:27,726 that are created by users with-going through 257 00:12:27,750 --> 00:12:29,946 a similar challenge at that time. 258 00:12:29,970 --> 00:12:34,076 And it might really create a very colored worldview for you 259 00:12:34,100 --> 00:12:35,516 where it's really hard to deal 260 00:12:35,540 --> 00:12:37,716 with your mental health challenge at that time. 261 00:12:44,980 --> 00:12:47,636 Lauren Hemings is studying to be a midwife. 262 00:12:47,660 --> 00:12:49,544 She used to spend her uni breaks, 263 00:12:49,568 --> 00:12:51,265 scrolling through TikTok. 264 00:12:51,880 --> 00:12:53,896 - I think it was quarantine boredom 265 00:12:53,920 --> 00:12:56,863 that kind of motivated me to download it. 266 00:12:57,950 --> 00:12:59,276 It was quite an innocent hope 267 00:12:59,300 --> 00:13:01,376 of just getting a good laugh, really. 268 00:13:01,400 --> 00:13:03,426 You know, like getting funny videos 269 00:13:03,450 --> 00:13:05,633 and seeing what was on it. 270 00:13:06,870 --> 00:13:08,136 I never had the intention 271 00:13:08,160 --> 00:13:09,716 of making TikToks or sharing them. 272 00:13:09,740 --> 00:13:13,045 It was more just kind of from the viewpoint of a viewer. 273 00:13:13,820 --> 00:13:15,046 Lauren started following 274 00:13:15,070 --> 00:13:17,679 a popular fitness influencer on the app. 275 00:13:20,100 --> 00:13:24,436 - There's one woman who had like quite a similar body type 276 00:13:24,460 --> 00:13:28,486 to me and she'd expressed that she was unhappy 277 00:13:28,510 --> 00:13:29,593 with that body type. 278 00:13:30,770 --> 00:13:33,136 And she had started tracking calories over quarantine. 279 00:13:33,160 --> 00:13:36,003 She had lost a really, really significant amount of weight. 280 00:13:39,260 --> 00:13:41,306 The algorithm then flooded her feed 281 00:13:41,330 --> 00:13:44,283 with content promoting unhealthy weight loss. 282 00:13:46,330 --> 00:13:48,516 - I was no longer saying funny dance videos or anything. 283 00:13:48,540 --> 00:13:50,736 It was just like this complete focus 284 00:13:50,760 --> 00:13:54,863 on that like fitness and healthy lifestyle goal. 285 00:13:58,763 --> 00:14:01,366 - TikTok pushed Lauren toward the popular trend 286 00:14:01,390 --> 00:14:05,303 of meticulously tracking how many calories you eat in a day. 287 00:14:06,290 --> 00:14:09,453 Something researches, warn, promotes disordered eating. 288 00:14:10,850 --> 00:14:12,746 The hashtag, What I eat in a day, 289 00:14:12,770 --> 00:14:15,665 has more than 7 billion views on TikTok. 290 00:14:18,480 --> 00:14:20,056 - It turned into like this obsession 291 00:14:20,080 --> 00:14:23,496 and I felt that I could not eat anything 292 00:14:23,520 --> 00:14:26,526 without knowing how many calories it contained 293 00:14:26,550 --> 00:14:28,796 and without meeting, you know, my target number 294 00:14:28,820 --> 00:14:30,336 of calories throughout the day. 295 00:14:30,360 --> 00:14:32,516 There was a few months where I didn't put anything 296 00:14:32,540 --> 00:14:34,673 into my mouth that I had not weighed. 297 00:14:36,150 --> 00:14:38,356 Four months after downloading TikTok, 298 00:14:38,380 --> 00:14:40,636 Lauren admitted to her friends and family 299 00:14:40,660 --> 00:14:42,337 she had an eating disorder. 300 00:14:43,000 --> 00:14:44,806 - I'd like to think that I wouldn't have struggled 301 00:14:44,830 --> 00:14:47,156 with an eating disorder if I hadn't downloaded TikTok. 302 00:14:47,180 --> 00:14:48,236 I think, you know, TikTok 303 00:14:48,260 --> 00:14:51,783 was the main contributor to the development of that. 304 00:14:53,470 --> 00:14:56,156 Young users are increasingly turning to TikTok 305 00:14:56,180 --> 00:15:00,086 to find and spread information on how to restrict food 306 00:15:00,110 --> 00:15:02,893 and hide their disordered, eating from their families. 307 00:15:04,250 --> 00:15:06,916 - What they do is they actually share content 308 00:15:06,940 --> 00:15:10,176 of what they go through and what they have done for the day 309 00:15:10,200 --> 00:15:11,913 in the fascination to become thin. 310 00:15:12,800 --> 00:15:14,286 So they would share recipes. 311 00:15:14,310 --> 00:15:16,526 They would share diet plans. 312 00:15:16,550 --> 00:15:19,549 They would share how you need to be disciplined. 313 00:15:19,980 --> 00:15:23,620 For someone who's vulnerable and desperate, 314 00:15:23,650 --> 00:15:26,286 they would follow anyone's advice. 315 00:15:26,310 --> 00:15:28,306 None of this advice is actually good 316 00:15:28,330 --> 00:15:31,396 because some of these advice is, oh lick a pumpkin 317 00:15:31,420 --> 00:15:33,811 for your lunch, but don't eat. 318 00:15:34,600 --> 00:15:36,973 Drink a liter of water and you should be fine. 319 00:15:37,900 --> 00:15:40,046 - I was super hesitant to get on TikTok 320 00:15:40,070 --> 00:15:41,806 because I'd heard that it was a really bad space 321 00:15:41,830 --> 00:15:42,956 for people with eating disorders. 322 00:15:42,980 --> 00:15:45,006 Because the algorithm knows everything 323 00:15:45,030 --> 00:15:46,916 and then it would curate your feed 324 00:15:46,940 --> 00:15:48,485 to be interested in that kind of stuff. 325 00:15:49,440 --> 00:15:51,906 Claire Benstead has been in and out of hospital 326 00:15:51,930 --> 00:15:54,504 for anorexia for more than five years. 327 00:15:54,590 --> 00:15:56,346 She decided to download TikTok 328 00:15:56,370 --> 00:15:59,613 to find support and to promote her earrings business. 329 00:16:00,090 --> 00:16:01,736 - You want that support 330 00:16:01,760 --> 00:16:03,326 because it's such an isolating illness. 331 00:16:03,350 --> 00:16:04,766 And there's so many people in my life 332 00:16:04,790 --> 00:16:06,926 that don't get it and don't understand it. 333 00:16:06,950 --> 00:16:09,286 Claire says the TikTok algorithm identified 334 00:16:09,310 --> 00:16:11,556 she had an eating disorder and she noticed 335 00:16:11,580 --> 00:16:15,436 an immediate change to the types of videos on her feed. 336 00:16:15,460 --> 00:16:16,686 - So it went from being, you know, 337 00:16:16,710 --> 00:16:18,726 my algorithm was, you know, Australian humor 338 00:16:18,750 --> 00:16:21,536 and musical theater humor, and all of that kind of stuff 339 00:16:21,560 --> 00:16:25,346 to just being eating disorder content all the time. 340 00:16:25,370 --> 00:16:28,330 And as I got sicker and I got more obsessive, 341 00:16:28,840 --> 00:16:30,746 all I could do was just flick through my phone, 342 00:16:30,770 --> 00:16:33,692 and look at this footage. 343 00:16:34,220 --> 00:16:37,076 I spent hours on it and just fixated on it. 344 00:16:37,100 --> 00:16:38,106 I wasn't recovering it all. 345 00:16:38,130 --> 00:16:39,723 I was actively relapsing. 346 00:16:40,520 --> 00:16:42,256 Claire was admitted to hospital. 347 00:16:42,280 --> 00:16:45,566 As part of her treatment, her psychologists worked with her 348 00:16:45,590 --> 00:16:48,976 to remove the toxic content from her TikTok feed 349 00:16:49,000 --> 00:16:52,326 by unfollowing accounts and reporting videos. 350 00:16:52,350 --> 00:16:54,436 How long did it actually take you to get rid 351 00:16:54,460 --> 00:16:56,875 of that eating disorder content from your algorithm. 352 00:16:57,850 --> 00:16:58,858 Ages. 353 00:16:58,882 --> 00:17:01,126 Pretty much being in hospital, so probably two months, 354 00:17:01,150 --> 00:17:02,510 it took me to change the algorithm. 355 00:17:03,200 --> 00:17:04,796 When you're kind of scrolling through like this- 356 00:17:04,820 --> 00:17:06,768 - Even while Claire was showing me her 357 00:17:06,792 --> 00:17:08,436 cleaned up TikTok feed, 358 00:17:08,460 --> 00:17:11,659 videos about eating disorders began reappearing. 359 00:17:12,020 --> 00:17:14,021 Hey, there we go. Here's one right now. 360 00:17:14,540 --> 00:17:17,246 Just every five or six videos. 361 00:17:17,270 --> 00:17:21,246 And so, I'm in a good spot that this doesn't trigger me. 362 00:17:21,270 --> 00:17:22,856 - So even though you're saying not interested, 363 00:17:22,880 --> 00:17:24,586 it's still coming up? - It's still coming up. 364 00:17:24,610 --> 00:17:27,086 If you report TikTok videos, the company says 365 00:17:27,110 --> 00:17:30,046 its moderators then decide whether to ban them. 366 00:17:30,070 --> 00:17:31,486 Which in turn is supposed 367 00:17:31,510 --> 00:17:34,076 to teach the algorithm to stop featuring them. 368 00:17:34,100 --> 00:17:35,736 - I just say that I'm not interested in that- 369 00:17:35,760 --> 00:17:38,676 - TikToks policies, say the app bans content 370 00:17:38,700 --> 00:17:41,963 promoting, normalizing or glorifying eating disorders. 371 00:17:41,980 --> 00:17:43,580 And you can say that it's offensive, 372 00:17:43,588 --> 00:17:44,966 But when users like Claire, 373 00:17:44,990 --> 00:17:47,676 have reported those videos, they were told 374 00:17:47,700 --> 00:17:50,026 they don't breach any guidelines. 375 00:17:50,050 --> 00:17:51,946 - You would think that, you know, something this serious 376 00:17:51,970 --> 00:17:53,476 and it's got the highest mortality rate 377 00:17:53,500 --> 00:17:55,456 of any mental illness, you would think that, 378 00:17:55,480 --> 00:17:57,056 that would be something that you could report. 379 00:17:57,080 --> 00:17:58,556 Because it is promoting those behaviors 380 00:17:58,580 --> 00:18:00,375 and it's making it worse. 381 00:18:05,940 --> 00:18:07,158 TikTok also says it 382 00:18:07,180 --> 00:18:09,295 bans pro eating disorder hashtags 383 00:18:09,320 --> 00:18:11,776 so users can search for those videos. 384 00:18:11,800 --> 00:18:13,496 And if they try to, a number 385 00:18:13,520 --> 00:18:15,356 for eating disorder support service, 386 00:18:15,380 --> 00:18:18,756 the Butterfly Foundation automatically pops up. 387 00:18:18,780 --> 00:18:20,823 But users find ways around it. 388 00:18:21,790 --> 00:18:25,486 - But the issue is now that it's ever evolving. 389 00:18:25,510 --> 00:18:27,836 Like there's a hashtag now that people 390 00:18:27,860 --> 00:18:29,356 with eating disorders use. 391 00:18:29,380 --> 00:18:31,106 And you would never guess that 392 00:18:31,130 --> 00:18:33,096 it was an eating disorder hashtag. 393 00:18:33,120 --> 00:18:34,720 Like it's after a famous singer. 394 00:18:36,700 --> 00:18:40,046 So just changing them to be completely irrelevant 395 00:18:40,070 --> 00:18:42,328 from what an eating disorder is. 396 00:18:42,670 --> 00:18:44,386 And so it's so hard to escape now. 397 00:18:44,410 --> 00:18:46,496 And I think it's really hard for TikTok 398 00:18:46,520 --> 00:18:47,818 to keep up with that all. 399 00:18:48,360 --> 00:18:51,066 - There are mechanisms in place 400 00:18:51,090 --> 00:18:52,446 to screen some of that content. 401 00:18:52,470 --> 00:18:55,646 But a lot of it is also reliant on human moderation. 402 00:18:55,670 --> 00:18:58,026 And when you consider the amount of videos 403 00:18:58,050 --> 00:19:00,039 and the volume that is being uploaded to TikTok, 404 00:19:00,063 --> 00:19:02,176 it's a very difficult task to imagine 405 00:19:02,200 --> 00:19:04,100 human moderators can catch everything. 406 00:19:09,526 --> 00:19:12,286 Last year, TikTok established a council 407 00:19:12,310 --> 00:19:14,366 of outside experts to advise 408 00:19:14,390 --> 00:19:16,773 the company about content moderation. 409 00:19:17,730 --> 00:19:19,553 David Polgar is one of them. 410 00:19:20,940 --> 00:19:23,226 - As we know with great power comes great responsibility. 411 00:19:23,250 --> 00:19:25,566 There's a lot of power in TikToks algorithm. 412 00:19:25,590 --> 00:19:28,196 Therefore you have to constantly be aware 413 00:19:28,220 --> 00:19:31,836 of how it's impacting other individuals 414 00:19:31,860 --> 00:19:33,261 and other communities. 415 00:19:33,980 --> 00:19:37,060 I think comparatively speaking TikTok 416 00:19:37,082 --> 00:19:39,886 has done a pretty decent job 417 00:19:39,910 --> 00:19:44,016 with being more reflective on rabbit holes 418 00:19:44,040 --> 00:19:47,986 and how that can affect individuals. 419 00:19:48,010 --> 00:19:53,010 But at the same time, you're dealing with human behaviour. 420 00:19:53,180 --> 00:19:55,656 You're dealing with bad actors. 421 00:19:55,680 --> 00:19:58,106 You're dealing with major differences 422 00:19:58,130 --> 00:20:02,653 of how people define appropriate versus inappropriate. 423 00:20:02,790 --> 00:20:04,106 And we have this tricky kind 424 00:20:04,130 --> 00:20:07,013 of balancing act that's constantly happening. 425 00:20:09,700 --> 00:20:13,380 - TikTok's business model is built on creating a fun, 426 00:20:13,410 --> 00:20:16,356 glossy and glamorous version of the world. 427 00:20:16,380 --> 00:20:19,526 And the company has been found to strictly control content 428 00:20:19,550 --> 00:20:21,446 that doesn't fit with that image. 429 00:20:21,470 --> 00:20:24,936 In March last year, TikTok policy documents were leaked. 430 00:20:24,960 --> 00:20:27,456 Showing content moderators were instructed 431 00:20:27,480 --> 00:20:29,796 to suppress posts by creators 432 00:20:29,820 --> 00:20:32,663 considered ugly, poor or disabled. 433 00:20:36,460 --> 00:20:38,288 The documents said, "Videos, including 434 00:20:38,312 --> 00:20:40,786 people who had chubby or obese 435 00:20:40,810 --> 00:20:42,096 with ugly facial looks, 436 00:20:42,120 --> 00:20:44,946 like too many wrinkles or facial deformities 437 00:20:44,970 --> 00:20:47,806 and other disabilities should be excluded." 438 00:20:47,830 --> 00:20:51,403 TikTok has said it no longer engages in these practices. 439 00:20:55,899 --> 00:20:57,206 - I don't want to admit it, but looks 440 00:20:57,230 --> 00:20:58,326 have a lot to do with it. 441 00:20:58,350 --> 00:21:00,706 And you know, we're all secretly a bit vain. 442 00:21:00,730 --> 00:21:02,396 As much you don't wanna admit it, 443 00:21:02,420 --> 00:21:04,806 you go for looks over non-looks, you know. 444 00:21:04,830 --> 00:21:06,596 So I think looks definitely have a lot to do with it. 445 00:21:06,620 --> 00:21:09,426 And if you look at all the really big time influencers, 446 00:21:09,450 --> 00:21:10,546 they're all beautiful. 447 00:21:10,570 --> 00:21:12,259 Like, if you look at all these influencers, 448 00:21:12,283 --> 00:21:14,376 they're all stunning, like nothing wrong with them. 449 00:21:14,400 --> 00:21:16,865 So I think looks definitely have a lot to do with it. 450 00:21:30,347 --> 00:21:31,968 Much of TikTok's popularity 451 00:21:31,992 --> 00:21:33,584 is driven by dance trends, 452 00:21:33,620 --> 00:21:35,826 choreographed by black creators 453 00:21:35,850 --> 00:21:38,302 and then copied by white influencers. 454 00:21:40,290 --> 00:21:42,026 But black content makers say 455 00:21:42,050 --> 00:21:45,343 that the platform actively discriminates against them. 456 00:21:47,220 --> 00:21:49,746 Think it's high time we let black women 457 00:21:49,770 --> 00:21:52,126 on this app also be famous for doing the bare minimum. 458 00:21:52,150 --> 00:21:54,836 Like I should be able to just sit here in silence, 459 00:21:54,860 --> 00:21:56,006 and let y'all look at me 460 00:21:56,030 --> 00:21:58,830 and the next thing you know, I have a million followers. 461 00:22:00,180 --> 00:22:01,540 - Petition for black people for the rest 462 00:22:01,560 --> 00:22:03,206 of April to stop talking. 463 00:22:03,230 --> 00:22:04,856 - There have been instances 464 00:22:04,880 --> 00:22:08,116 of black creator led mass walk offs 465 00:22:08,140 --> 00:22:09,756 from the platform called Blackouts. 466 00:22:09,780 --> 00:22:11,166 Where on a certain day, 467 00:22:11,190 --> 00:22:12,726 black creators will stop using the platform 468 00:22:12,750 --> 00:22:15,896 or urge other creators to leave the platform 469 00:22:15,920 --> 00:22:19,466 because of TikToks inaction and failure to respond to 470 00:22:19,490 --> 00:22:21,696 or engage with some of the criticisms 471 00:22:21,720 --> 00:22:25,224 and the discourse that black creators have raised. 472 00:22:25,294 --> 00:22:30,869 So if the company continues to be reactive and responsive, 473 00:22:30,980 --> 00:22:33,826 rather than proactive and really meaningfully engage, 474 00:22:33,850 --> 00:22:36,125 then these issues are gonna continue to occur. 475 00:22:39,260 --> 00:22:41,356 - Often, it makes me quite furious, I guess, 476 00:22:41,380 --> 00:22:42,886 'cause it's like these black creators, 477 00:22:42,910 --> 00:22:45,526 they got talent, they're out here dancing 478 00:22:45,550 --> 00:22:47,566 and showing what they're capable of. 479 00:22:47,880 --> 00:22:50,246 So it's kind of very much disappointing 480 00:22:50,270 --> 00:22:53,516 and hard on us when we're out here expected 481 00:22:53,540 --> 00:22:55,976 to have all of these in order to get the views 482 00:22:56,000 --> 00:22:57,956 in order to get the likes and shares. 483 00:22:57,980 --> 00:23:00,066 But no matter how much we try, 484 00:23:00,090 --> 00:23:02,022 we're just not gonna get that. 485 00:23:08,620 --> 00:23:12,896 - Unice Wani is an 18 year old TikTok creator from Perth. 486 00:23:13,010 --> 00:23:14,800 - I like this one. Look and see. 487 00:23:15,120 --> 00:23:16,456 I feel like the more I go viral, 488 00:23:16,480 --> 00:23:20,496 the more I can basically show the younger generation 489 00:23:20,520 --> 00:23:23,216 and show more colored girls, I guess, 490 00:23:23,240 --> 00:23:27,016 or people out there like I'm okay in my own skin 491 00:23:27,040 --> 00:23:28,726 and I love myself the way I am. 492 00:23:28,750 --> 00:23:31,236 I don't care what social media says about me. 493 00:23:31,260 --> 00:23:34,146 What people on the other side of the screen says about me. 494 00:23:34,170 --> 00:23:36,163 You can be yourself at the end of the day. 495 00:23:36,668 --> 00:23:38,136 Let me quickly address this- 496 00:23:38,160 --> 00:23:39,326 - As her following grew, 497 00:23:39,350 --> 00:23:41,816 so did the hateful comments. 498 00:23:41,840 --> 00:23:44,816 And she decided to confront the issue on the app. 499 00:23:44,840 --> 00:23:46,636 - So a majority of you guys still feel the need 500 00:23:46,660 --> 00:23:49,226 to comment about my skin color and about how dark I am 501 00:23:49,250 --> 00:23:51,616 and about how black, black, black, black, black I am. 502 00:23:51,620 --> 00:23:52,654 Well, guess what? 503 00:23:52,679 --> 00:23:54,523 I'm black and I'm so proud. 504 00:23:57,891 --> 00:24:00,516 - Unice says often her videos are hidden 505 00:24:00,540 --> 00:24:02,356 or muted from the TikTok feed. 506 00:24:02,380 --> 00:24:03,946 Meaning few people see them. 507 00:24:03,970 --> 00:24:05,917 A practice known as Shadow Banning. 508 00:24:05,941 --> 00:24:07,617 - Are you pressed? 509 00:24:07,641 --> 00:24:09,147 Are you mad? 510 00:24:09,171 --> 00:24:10,598 Are you upset? 511 00:24:10,622 --> 00:24:11,622 Are you sad? 512 00:24:12,920 --> 00:24:14,076 Sorry, what? 513 00:24:14,100 --> 00:24:17,236 I guess you tend to get a lot of shadow bans 514 00:24:17,260 --> 00:24:20,556 for speaking up about stuff such as racism. 515 00:24:20,580 --> 00:24:21,866 Stuff you couldn't mention. 516 00:24:21,890 --> 00:24:24,476 One word, black, could say all of this 517 00:24:24,500 --> 00:24:26,876 and your video could get shadow banned. 518 00:24:26,900 --> 00:24:32,300 When you post a video, the video, just it's on the app. 519 00:24:32,330 --> 00:24:34,536 It's just, you're not gonna get any views for it. 520 00:24:34,560 --> 00:24:36,916 So you can see it. It's just other people can't see it 521 00:24:36,940 --> 00:24:39,286 when they go onto your account as well. 522 00:24:39,310 --> 00:24:42,509 So it's up there. It's just, it's not going to get any views. 523 00:24:47,270 --> 00:24:49,506 Last year TikTok creators noticed 524 00:24:49,530 --> 00:24:51,776 the algorithm was suppressing posts 525 00:24:51,800 --> 00:24:55,113 with the hashtag Black Lives Matter or George Floyd. 526 00:24:59,620 --> 00:25:04,316 - So word on the street is that TikTok has banned 527 00:25:04,410 --> 00:25:06,169 the Black Lives Matter hashtag. 528 00:25:12,060 --> 00:25:14,332 One of those creators was Sydney man, 529 00:25:14,357 --> 00:25:16,512 Paniora Nukunuku, who had created 530 00:25:16,540 --> 00:25:18,900 a video using a pool table 531 00:25:18,930 --> 00:25:22,296 to explain the Black Lives Matter issue to Australians. 532 00:25:22,320 --> 00:25:23,886 - This is a white Australia table. 533 00:25:23,910 --> 00:25:25,796 And they pretty much had 200 years head-start 534 00:25:25,820 --> 00:25:27,517 and they had established everything in the country. 535 00:25:27,541 --> 00:25:29,476 So their break looks like this. 536 00:25:29,500 --> 00:25:32,455 Bro, can you get home ownership in business? 537 00:25:32,500 --> 00:25:33,540 Beautiful. 538 00:25:33,552 --> 00:25:35,056 That was spicy. 539 00:25:35,080 --> 00:25:37,161 That blew up bigger than I thought it would. 540 00:25:37,278 --> 00:25:38,635 I just need to put this here. 541 00:25:38,660 --> 00:25:39,806 Boy, what the, 542 00:25:39,830 --> 00:25:42,586 - Don't worry. It's trauma, injustice and discrimination. 543 00:25:42,610 --> 00:25:44,056 But I said, sorry, so it should be fine. 544 00:25:44,080 --> 00:25:45,230 So just go for it, bro. 545 00:25:46,520 --> 00:25:48,526 It was the biggest video at the time that I've done. 546 00:25:48,540 --> 00:25:50,060 I think you're just being lazy, hey. 547 00:25:50,085 --> 00:25:51,596 I don't know why. 548 00:25:51,620 --> 00:25:53,962 Oh, I do know why, because it was good. 549 00:25:53,980 --> 00:25:55,361 I shouldn't look at the camera, but I'm 550 00:25:55,385 --> 00:25:56,981 just really proud right now. 551 00:25:57,220 --> 00:25:58,396 Using these two cue balls, 552 00:25:58,420 --> 00:25:59,616 I'll explain to you, that resulted 553 00:25:59,640 --> 00:26:01,136 in my account getting banned 554 00:26:01,160 --> 00:26:02,776 for like seven days. 555 00:26:02,800 --> 00:26:03,856 I don't know why. 556 00:26:03,880 --> 00:26:07,786 They claimed that my video breached community guidelines, 557 00:26:07,810 --> 00:26:10,926 which is extremely vague because there is no swearing, 558 00:26:10,950 --> 00:26:13,079 there is no explicit language. 559 00:26:13,103 --> 00:26:17,536 There's no nudity or explicit like sexual stuff. 560 00:26:17,560 --> 00:26:20,139 None of that. And my account got banned. 561 00:26:20,900 --> 00:26:22,566 - The Black Lives Matter is trending on TikTok, 562 00:26:22,590 --> 00:26:24,976 which is ironic considering how much time TikTok spends 563 00:26:25,000 --> 00:26:26,696 silencing the voices of black creators. 564 00:26:26,720 --> 00:26:28,966 TikTok apologized for suppressing hashtags, 565 00:26:28,990 --> 00:26:30,786 referring to Black Lives Matter, 566 00:26:30,849 --> 00:26:32,815 blaming a glitch in the algorithm. 567 00:26:33,003 --> 00:26:34,803 - Let's take a moment of silence for this man. 568 00:26:34,939 --> 00:26:37,019 - The company responded with a new initiative 569 00:26:37,050 --> 00:26:40,056 for black creators called the TikTok Black Creator Programme. 570 00:26:40,080 --> 00:26:42,926 I've spoken to creators who had been approached 571 00:26:42,950 --> 00:26:46,566 for that programme, who felt that it was lip service. 572 00:26:46,590 --> 00:26:49,836 It wasn't really a well-meaning effort 573 00:26:49,860 --> 00:26:52,826 to engage with black voices and engage with discourse 574 00:26:52,850 --> 00:26:54,800 that is important to black communities. 575 00:26:56,840 --> 00:27:00,835 Paniora, has more than 180,000 followers on TikTok. 576 00:27:01,500 --> 00:27:03,996 He often posts about living with a disability. 577 00:27:04,020 --> 00:27:05,326 - So growing up with the fake leg, 578 00:27:05,350 --> 00:27:06,876 I always got in trouble every time I park 579 00:27:06,900 --> 00:27:08,216 in my disabled spot. 580 00:27:08,240 --> 00:27:10,396 The first video I did, was me going up 581 00:27:10,420 --> 00:27:13,516 to a pool and telling my friends to record me, 582 00:27:13,540 --> 00:27:16,476 dip my fake leg in the water to test the water out. 583 00:27:16,500 --> 00:27:18,236 It was a really dumb idea. 584 00:27:18,260 --> 00:27:19,816 But for some reason, people loved it. 585 00:27:19,840 --> 00:27:21,606 And in this space of eight hours, 586 00:27:21,630 --> 00:27:24,353 it hit about 780,000 views. 587 00:27:25,190 --> 00:27:29,486 If you have this many followers and that many likes, 588 00:27:29,510 --> 00:27:30,916 it's 'cause you're pretty. 589 00:27:30,940 --> 00:27:32,986 If you have this many followers 590 00:27:33,010 --> 00:27:34,783 and the same amount of likes, 591 00:27:38,410 --> 00:27:39,410 you're just funny. 592 00:27:41,280 --> 00:27:44,106 Paniora ran into trouble with the TikTok censors 593 00:27:44,130 --> 00:27:46,766 when he posted a video of a confrontation 594 00:27:46,790 --> 00:27:48,296 with someone who was telling him, 595 00:27:48,320 --> 00:27:50,896 he shouldn't have a disability permit 596 00:27:50,920 --> 00:27:54,326 - So this old lady had the nerve to ask me 597 00:27:54,350 --> 00:27:57,063 if this is my disability card. 598 00:27:58,900 --> 00:28:00,186 This, 599 00:28:00,210 --> 00:28:02,816 I wonder if this is enough. 600 00:28:02,840 --> 00:28:04,696 The video was taken down. 601 00:28:04,720 --> 00:28:07,906 TikTok said it breached the app's community guidelines. 602 00:28:07,930 --> 00:28:10,606 Paniora appealed and it was put back up. 603 00:28:10,630 --> 00:28:11,886 But he's had other videos 604 00:28:11,910 --> 00:28:14,456 about his disability removed as well. 605 00:28:14,480 --> 00:28:15,796 - You don't need to worry about it. 606 00:28:15,820 --> 00:28:18,020 The video got taken down and I didn't even know it 607 00:28:18,030 --> 00:28:20,056 until I looked back at the hashtags 608 00:28:20,080 --> 00:28:22,386 and decided to see which videos that I've done 609 00:28:22,410 --> 00:28:24,808 have like made it to the top and that wasn't there. 610 00:28:25,260 --> 00:28:28,386 I appealed it and I don't know why that was taken down. 611 00:28:28,420 --> 00:28:29,656 Don't ever do that again. 612 00:28:29,820 --> 00:28:31,906 Do I feel like TikTok is being racist? 613 00:28:31,930 --> 00:28:32,930 I don't know. 614 00:28:34,190 --> 00:28:36,602 Has TikTok been hit up in the past, 615 00:28:36,660 --> 00:28:39,086 around the moderators being told 616 00:28:39,100 --> 00:28:44,580 to limit the exposure of disabled people and ugly people? 617 00:28:44,590 --> 00:28:46,216 Yes. They've been called out on that. 618 00:28:46,240 --> 00:28:47,426 Is this happening again? 619 00:28:47,450 --> 00:28:51,016 I hope not, but it definitely feels like it has. 620 00:28:51,040 --> 00:28:54,186 We know that to decolonize Palestine 621 00:28:54,210 --> 00:28:56,516 means also to decolonize- 622 00:28:56,540 --> 00:28:58,986 - I'll probably keep moving, get some shots. 623 00:28:59,010 --> 00:29:00,016 In may of this year, 624 00:29:00,040 --> 00:29:03,406 Paniora posted a video from a pro-Palestine rally. 625 00:29:03,430 --> 00:29:05,636 But TikTok's algorithm flagged it. 626 00:29:05,660 --> 00:29:07,439 And it was instantly taken down. 627 00:29:09,950 --> 00:29:13,086 Other creators posting TikToks about Palestine 628 00:29:13,110 --> 00:29:15,855 have said they've experienced the same thing. 629 00:29:18,198 --> 00:29:20,856 - When TikTok started removing my videos 630 00:29:20,880 --> 00:29:22,566 about the protests in regards 631 00:29:22,590 --> 00:29:25,506 to the Palestinian situation, I was furious. 632 00:29:25,530 --> 00:29:29,216 I was like, why? There is nothing in these videos 633 00:29:29,240 --> 00:29:33,066 that would justify, like a removal. 634 00:29:33,090 --> 00:29:34,333 There really isn't. 635 00:29:34,710 --> 00:29:37,326 - One of the big problems with TikTok 636 00:29:37,350 --> 00:29:42,046 and the unique nature of its opaque algorithm, 637 00:29:42,070 --> 00:29:45,736 is that it's very difficult to understand 638 00:29:45,760 --> 00:29:49,534 or to recognise when censorship is taking place. 639 00:29:49,540 --> 00:29:51,140 People came together to try to- 640 00:29:51,167 --> 00:29:54,796 - So it is possible for content on the app 641 00:29:54,820 --> 00:29:58,676 to be promoted or demoted without anyone knowing. 642 00:29:58,700 --> 00:30:01,566 - I'm so sick and tired of every social media platform 643 00:30:01,590 --> 00:30:03,366 silencing Palestinian voices. 644 00:30:03,390 --> 00:30:08,066 - But we also see evidence of how content moderation 645 00:30:08,090 --> 00:30:09,616 that takes place in China. 646 00:30:09,640 --> 00:30:13,806 How that type of thinking is still applied 647 00:30:13,830 --> 00:30:15,843 to TikTok outside of China. 648 00:30:18,591 --> 00:30:21,445 TikTok is owned by Chinese start-up, ByteDance, 649 00:30:21,470 --> 00:30:25,098 which is believed to be worth more than $250 billion. 650 00:30:25,490 --> 00:30:28,166 It's heavily regulated by the Chinese government. 651 00:30:28,190 --> 00:30:30,847 And there's a Communist Party Internal Committee 652 00:30:30,872 --> 00:30:34,696 in ByteDance, which ensures the parties political goals 653 00:30:34,720 --> 00:30:37,009 are pursued alongside the company's. 654 00:30:38,880 --> 00:30:40,976 - We have to be extra concerned 655 00:30:41,000 --> 00:30:45,006 about how apps like TikTok can be used 656 00:30:45,030 --> 00:30:47,883 as a vector for censorship and surveillance. 657 00:30:49,990 --> 00:30:52,386 - The Australian Strategic Policy Institute 658 00:30:52,410 --> 00:30:54,566 did the first academic investigation 659 00:30:54,590 --> 00:30:56,346 into censorship on TikTok, 660 00:30:56,370 --> 00:30:59,376 concluding the company actively uses the algorithm 661 00:30:59,400 --> 00:31:02,660 to hide political speech it deems controversial. 662 00:31:03,020 --> 00:31:05,616 The research was funded by the US State Department 663 00:31:05,640 --> 00:31:08,566 and found anti-Russian government videos 664 00:31:08,580 --> 00:31:12,035 as well as hashtags about LGBTQI issues 665 00:31:12,060 --> 00:31:13,820 and the mass detention of Uyghurs 666 00:31:13,860 --> 00:31:15,833 were among those being suppressed. 667 00:31:18,510 --> 00:31:22,636 - The company has cooperated with public security bureaus 668 00:31:22,660 --> 00:31:26,076 all throughout China and including in Xinjiang. 669 00:31:26,100 --> 00:31:30,306 And that means that they work, 670 00:31:30,330 --> 00:31:33,176 they coordinate with government agencies 671 00:31:33,200 --> 00:31:36,676 to ensure that the information space in China 672 00:31:36,700 --> 00:31:39,966 is pumped full of this propaganda. 673 00:31:39,990 --> 00:31:42,596 That shows a very rosy picture 674 00:31:42,620 --> 00:31:44,484 of what's happening in Xinjiang. 675 00:31:48,640 --> 00:31:51,296 In 2018, then CEO of ByteDance 676 00:31:51,320 --> 00:31:53,366 was forced to publicly apologise. 677 00:31:53,390 --> 00:31:55,496 Saying one of the company's platforms 678 00:31:55,520 --> 00:31:59,396 had gone against China's core socialist values. 679 00:31:59,420 --> 00:32:02,986 - We have a very clear public statement 680 00:32:03,010 --> 00:32:04,766 from the founder of ByteDance, 681 00:32:04,790 --> 00:32:08,786 that this is something that he's committed to doing 682 00:32:08,810 --> 00:32:11,426 and to ensuring that the company continues 683 00:32:11,450 --> 00:32:15,316 to push this type of propaganda, certainly inside of China. 684 00:32:15,340 --> 00:32:18,676 Whether that is then extended out to the rest of the world 685 00:32:18,700 --> 00:32:22,916 via apps like TikTok, is another question. 686 00:32:22,940 --> 00:32:25,183 And it's something worth watching. 687 00:32:26,190 --> 00:32:27,626 In a statement TikTok said, 688 00:32:27,650 --> 00:32:30,216 it does not moderate or remove content 689 00:32:30,240 --> 00:32:32,086 based on political sensitivities. 690 00:32:32,110 --> 00:32:34,516 And has never content at the request 691 00:32:34,540 --> 00:32:35,906 of the Chinese government. 692 00:32:35,930 --> 00:32:37,976 It also said it embraces diversity 693 00:32:38,000 --> 00:32:41,006 and denied it discriminates against any creator 694 00:32:41,030 --> 00:32:42,943 or community on our platform. 695 00:32:46,730 --> 00:32:49,186 - We've known for a better part of a decade, 696 00:32:49,210 --> 00:32:51,146 both here in the US and in Australia, 697 00:32:51,170 --> 00:32:53,706 about the concerns raised by the prevalence 698 00:32:53,730 --> 00:32:55,306 of Chinese telecommunications companies. 699 00:32:55,330 --> 00:32:56,726 And so then the next question became, 700 00:32:56,750 --> 00:32:58,996 well, what about all these apps that have, 701 00:32:59,020 --> 00:33:00,906 of companies that are headquartered in China? 702 00:33:00,930 --> 00:33:03,072 They're collecting tremendous amounts of user data. 703 00:33:03,280 --> 00:33:06,233 They have access to the devices of individuals. 704 00:33:07,440 --> 00:33:08,846 Jamil Jaffer is Founder 705 00:33:08,870 --> 00:33:11,996 of the National Security Institute in Washington. 706 00:33:12,020 --> 00:33:16,016 And has advised the us government on cyber-security. 707 00:33:16,040 --> 00:33:18,476 - In China, it's all the central government, 708 00:33:18,500 --> 00:33:19,506 the Communist Party. 709 00:33:19,530 --> 00:33:21,926 There's no separation between the branches. 710 00:33:21,950 --> 00:33:24,426 And so, when these apps have all that data, 711 00:33:24,450 --> 00:33:26,006 it's much easier for the Chinese government 712 00:33:26,030 --> 00:33:28,436 to simply obtain access to that data. 713 00:33:31,320 --> 00:33:32,326 - My understanding is that, 714 00:33:32,350 --> 00:33:34,916 about a quarter of the world's population 715 00:33:34,940 --> 00:33:37,316 is a member of TikTok if I'm not mistaken. 716 00:33:37,340 --> 00:33:39,596 So that's obviously an enormous amount of data 717 00:33:39,620 --> 00:33:40,676 that's being generated. 718 00:33:40,700 --> 00:33:43,006 That's being handed over for free 719 00:33:43,030 --> 00:33:45,016 to that single social network 720 00:33:45,040 --> 00:33:46,856 that has pretty much full control 721 00:33:46,880 --> 00:33:48,876 over what it does to the data. 722 00:33:48,900 --> 00:33:52,875 It might analyze it to generate personalized content 723 00:33:52,900 --> 00:33:55,180 for you, but it might also use that data 724 00:33:55,190 --> 00:33:58,026 to offer technology products and services 725 00:33:58,050 --> 00:34:01,056 to other companies moving forward in the future. 726 00:34:23,830 --> 00:34:25,476 - Hello, it's Avani in Sydney. 727 00:34:25,500 --> 00:34:26,726 How's it going? 728 00:34:26,750 --> 00:34:28,246 - Hi. 729 00:34:28,270 --> 00:34:29,836 Anne Longfield is England's 730 00:34:29,860 --> 00:34:31,764 former children's commissioner. 731 00:34:31,780 --> 00:34:33,650 Anne's interview, take one. 732 00:34:33,660 --> 00:34:36,300 - She's representing millions of kids on TikTok 733 00:34:36,340 --> 00:34:39,650 in the UK and Europe in a class action against the company. 734 00:34:39,950 --> 00:34:43,036 - My claim with TikTok at the moment is that, 735 00:34:43,060 --> 00:34:46,216 they are harvesting huge amounts of data illegally 736 00:34:46,240 --> 00:34:48,856 without the consent of children or their parents. 737 00:34:48,880 --> 00:34:52,056 And they aren't giving the right level 738 00:34:52,080 --> 00:34:55,016 of transparency about what happens to that data, 739 00:34:55,040 --> 00:34:57,993 or actually what that data includes. 740 00:35:01,600 --> 00:35:06,086 Almost a third of TikTok's Australian users are under 14. 741 00:35:06,110 --> 00:35:08,666 Lawyers say TikTok takes personal information 742 00:35:08,690 --> 00:35:11,836 like phone numbers, videos, locations, and facial data 743 00:35:11,860 --> 00:35:13,896 from kids without their consent. 744 00:35:13,920 --> 00:35:17,416 As well as the photos and videos recorded using TikTok, 745 00:35:17,440 --> 00:35:20,033 but not uploaded or saved to the platform. 746 00:35:21,180 --> 00:35:25,216 - Given the level of data and the lack of transparency 747 00:35:25,240 --> 00:35:27,306 around there, it's difficult to imagine 748 00:35:27,330 --> 00:35:28,536 that this isn't just a kind of 749 00:35:28,560 --> 00:35:30,386 information gathering service, 750 00:35:30,410 --> 00:35:34,966 which is thinly veiled as some kind of enjoyable platform, 751 00:35:34,990 --> 00:35:38,006 which appeals to young children. 752 00:35:38,030 --> 00:35:41,516 So the real incentive here, 753 00:35:41,540 --> 00:35:43,813 when you look at it in really cold terms, 754 00:35:43,837 --> 00:35:47,596 seems to be, to gather as much data as possible 755 00:35:47,620 --> 00:35:50,460 to really be able to monetize that. 756 00:35:53,180 --> 00:35:55,676 TikTok's already been fined millions of dollars 757 00:35:55,700 --> 00:35:59,113 in the US and South Korea for harvesting children's data. 758 00:36:01,290 --> 00:36:04,136 The company restricted app access for children 759 00:36:04,160 --> 00:36:08,196 and has taken down millions of under-age users' accounts. 760 00:36:08,220 --> 00:36:10,855 There's been no legal action in Australia. 761 00:36:12,530 --> 00:36:15,156 - I think that governments do have a responsibility 762 00:36:15,180 --> 00:36:18,066 to intervene to ensure that children 763 00:36:18,090 --> 00:36:20,546 are protected in whatever kind of environment they're in. 764 00:36:20,570 --> 00:36:23,106 And you see those protections and measures 765 00:36:23,130 --> 00:36:25,356 in terms of the physical environment, 766 00:36:25,380 --> 00:36:28,376 in terms of their safety, you know, 767 00:36:28,400 --> 00:36:29,856 in the communities they live in, 768 00:36:29,880 --> 00:36:31,726 in the environments they are. 769 00:36:31,750 --> 00:36:34,387 But it hasn't always been the case online. 770 00:36:34,420 --> 00:36:37,420 And some governments have 771 00:36:37,461 --> 00:36:39,529 struggled to see what that means. 772 00:36:42,730 --> 00:36:44,254 If the case is successful, 773 00:36:44,278 --> 00:36:45,946 TikTok could have to pay children 774 00:36:45,970 --> 00:36:49,166 from the UK and Europe, billions in compensation. 775 00:36:49,360 --> 00:36:50,826 TikTok is fighting the case. 776 00:36:50,850 --> 00:36:52,453 In a statement, the company said, 777 00:36:52,477 --> 00:36:55,696 "Privacy and safety are top priorities for TikTok 778 00:36:55,720 --> 00:36:58,026 and we have robust policies, processes, 779 00:36:58,050 --> 00:37:01,736 and technologies in place to help protect all users 780 00:37:01,760 --> 00:37:04,056 and our teenage users in particular. 781 00:37:04,080 --> 00:37:05,836 We believe the claims lack merit 782 00:37:05,860 --> 00:37:08,627 and intend to vigorously defend the action." 783 00:37:11,630 --> 00:37:14,006 The US government is reviewing TikTok 784 00:37:14,030 --> 00:37:15,928 and hasn't ruled out a ban. 785 00:37:16,340 --> 00:37:17,585 - The real question you ask is, what 786 00:37:17,609 --> 00:37:19,106 about the national security implications? 787 00:37:19,130 --> 00:37:21,426 So, okay. Yes, a lot of people are using it, right. 788 00:37:21,450 --> 00:37:22,536 But why does that matter? 789 00:37:22,560 --> 00:37:24,986 And it matters, I think, because of the access, 790 00:37:25,010 --> 00:37:27,110 it gives you to this large amount of data. 791 00:37:28,290 --> 00:37:31,416 You never think about the Chinese government in Beijing, 792 00:37:31,440 --> 00:37:34,136 having videos of you in your home, 793 00:37:34,160 --> 00:37:36,936 outside your home, at the park with your kids, 794 00:37:36,960 --> 00:37:38,333 knowing who your kids play with. 795 00:37:38,760 --> 00:37:40,666 I mean, that's what they have now 796 00:37:40,690 --> 00:37:42,186 potentially with this data set. 797 00:37:42,210 --> 00:37:44,516 We've seen now two consecutive presidents 798 00:37:44,540 --> 00:37:45,616 sign executive orders, 799 00:37:45,640 --> 00:37:47,976 making clear that they are very concerned 800 00:37:48,000 --> 00:37:49,996 about the national security implications 801 00:37:50,020 --> 00:37:51,954 of TikTok's data collection. 802 00:37:52,249 --> 00:37:54,025 As well as the impact it has 803 00:37:54,050 --> 00:37:56,250 on the privacy and civil liberties of Americans. 804 00:38:01,550 --> 00:38:03,574 India has announced a ban on TikTok. 805 00:38:04,050 --> 00:38:06,716 And in July last year, Prime Minister Scott Morrison 806 00:38:06,740 --> 00:38:10,223 ordered a review by intelligence agencies into the app. 807 00:38:12,750 --> 00:38:16,503 We are always very mindful of those risks 808 00:38:16,527 --> 00:38:20,326 and we are always monitoring them very closely. 809 00:38:20,350 --> 00:38:22,246 And if we've considered, there is a need 810 00:38:22,270 --> 00:38:24,836 to take further action than we are taking now, 811 00:38:24,860 --> 00:38:27,073 then I can tell you, we won't be shy about it. 812 00:38:28,720 --> 00:38:31,016 - It's certainly a security concern 813 00:38:31,040 --> 00:38:33,506 that the data of Australian users 814 00:38:33,530 --> 00:38:36,726 is potentially going back to Beijing. 815 00:38:37,260 --> 00:38:39,780 TikTok maintained that this is not the case. 816 00:38:39,810 --> 00:38:44,486 And our analysis showed that 817 00:38:44,510 --> 00:38:46,546 there's certainly not a fire hose 818 00:38:46,570 --> 00:38:50,016 of content that's being sent back to Beijing. 819 00:38:50,040 --> 00:38:51,386 But that doesn't mean that, 820 00:38:51,410 --> 00:38:53,486 that content can't be accessed 821 00:38:53,510 --> 00:38:55,531 from Beijing, if that's required. 822 00:38:55,820 --> 00:38:56,826 How are you going? 823 00:38:56,874 --> 00:38:58,406 - I'm good, thank you. How are you? 824 00:38:58,431 --> 00:39:00,735 TikTok maintains Australian users' data 825 00:39:00,760 --> 00:39:03,426 is held on servers in the US and Singapore. 826 00:39:03,450 --> 00:39:05,456 And that it has never provided data 827 00:39:05,480 --> 00:39:06,923 to the Chinese government. 828 00:39:08,480 --> 00:39:11,036 Staff in the Home Affairs and Defense Departments 829 00:39:11,060 --> 00:39:14,016 have been told not to have TikTok on their phones 830 00:39:14,040 --> 00:39:15,842 because of security risks. 831 00:39:16,470 --> 00:39:19,196 But Scott Morrison said there wasn't enough evidence 832 00:39:19,220 --> 00:39:21,037 to ban TikTok in Australia. 833 00:39:21,780 --> 00:39:24,276 - The scope of the investigation did seem 834 00:39:24,300 --> 00:39:25,996 to be quite limited. 835 00:39:26,020 --> 00:39:31,075 And that scope is not really enough to be able 836 00:39:31,100 --> 00:39:34,110 to tell the rest of Australia 837 00:39:34,180 --> 00:39:36,196 and regular Australian citizens, 838 00:39:36,220 --> 00:39:39,786 whether it's a good idea for them to be using the app. 839 00:39:41,150 --> 00:39:44,556 - There should definitely be another more rigorous 840 00:39:44,580 --> 00:39:47,046 and lengthy review into TikTok 841 00:39:47,070 --> 00:39:51,478 to fully understand the risks that TikTok presents. 842 00:39:52,300 --> 00:39:53,973 - And so if you just look at TikTok in isolation, 843 00:39:53,997 --> 00:39:55,683 and say, well, it's just this one app, 844 00:39:55,707 --> 00:39:57,516 and it's just kids doing dancing videos. 845 00:39:57,540 --> 00:39:58,836 It seems innocuous. 846 00:39:58,860 --> 00:40:01,706 But it's really takes place in this much larger context 847 00:40:01,730 --> 00:40:04,486 of data collection, artificial intelligence, 848 00:40:04,510 --> 00:40:08,046 and a real effort by the Chinese to consolidate influence 849 00:40:08,070 --> 00:40:09,602 in the region and across the globe. 850 00:40:13,450 --> 00:40:14,942 In just two years, 851 00:40:14,980 --> 00:40:17,875 TikTok has cemented itself as the app of choice 852 00:40:17,900 --> 00:40:20,016 for millions of Australians. 853 00:40:20,040 --> 00:40:22,916 - So you guys kept telling me to go on the Voice 2021. 854 00:40:22,940 --> 00:40:25,446 So, I did. 855 00:40:25,470 --> 00:40:26,926 There are serious concerns 856 00:40:26,950 --> 00:40:30,406 that TikToks fun and beautiful version of reality 857 00:40:30,430 --> 00:40:33,103 is distorting the way we see the world. 858 00:40:34,680 --> 00:40:38,967 And questions about whether its users understand the risks. 859 00:40:41,280 --> 00:40:44,136 - So we're really at risk of having generations 860 00:40:44,160 --> 00:40:46,306 of young people that haven't been able 861 00:40:46,330 --> 00:40:49,416 to form their own identity in natural ways. 862 00:40:49,440 --> 00:40:51,396 And instead have formed identities 863 00:40:51,420 --> 00:40:54,026 in response to something that a technology 864 00:40:54,050 --> 00:40:56,456 or a technology platform prescribes 865 00:40:56,480 --> 00:40:58,476 to be the normal or the new normal. 866 00:40:58,500 --> 00:41:00,177 - Last time this lady came up to me and go, 867 00:41:00,201 --> 00:41:01,673 you don't look disabled enough. 868 00:41:01,820 --> 00:41:02,980 I don't look disabled enough? 869 00:41:03,010 --> 00:41:05,296 - I understand that TikTok is trying its very best 870 00:41:05,320 --> 00:41:07,706 to make the platform palatable to everyone 871 00:41:07,730 --> 00:41:12,730 by just having fun dance videos and lip sync videos. 872 00:41:16,290 --> 00:41:18,946 But I know that my content gives value 873 00:41:18,970 --> 00:41:21,156 to so many people who look like me. 874 00:41:21,180 --> 00:41:23,876 Who live the same life like me. 875 00:41:23,900 --> 00:41:25,595 Who are brown like me. 876 00:41:33,750 --> 00:41:37,976 - I ended up cutting off TikTok like after a few months. 877 00:41:38,000 --> 00:41:40,186 But even with that, like it still left me 878 00:41:40,210 --> 00:41:42,526 with the eating disorder, you know. 879 00:41:42,550 --> 00:41:45,336 Like, TikTok kind of led to the development and then 880 00:41:45,360 --> 00:41:48,923 it has taken a really, really long time to fix that. 881 00:41:50,238 --> 00:41:52,275 TikTok isn't out here to help people. 882 00:41:52,300 --> 00:41:53,716 I don't think it's coming to the world 883 00:41:53,740 --> 00:41:55,396 with this intention of helping people. 884 00:41:55,420 --> 00:41:57,226 If they're going to make money off of something, 885 00:41:57,250 --> 00:41:59,646 then they will make money off of something. 886 00:42:00,120 --> 00:42:02,246 I think they maybe need to realise 887 00:42:02,270 --> 00:42:04,170 the impact that is having on people.