The Future, This Week 16 Feb 2018: classy Facebook, the end of the internet, China, and more


We’d like to advise that the
following program may contain real news, occasional philosophy and ideas that may offend some
listeners. This is The Future, This Week on Sydney Business Insights. I’m
Sandra Peter and I’m Kai Riemer. Every week we get together and look at the news of the week. We discuss technology, the future of business, the weird and the wonderful and things that
change the world. Okay let’s start. Let’s start. Today in The Future, This Week:
classy Facebook, the end of the Internet, and China, crypto and space stations in other news. I’m Sandra Peter I’m the Director of Sydney Business Insights. I’m Kai Riemer professor at the
Business School and leader of the Digital Disruption
Research Group. So Kai, what happened in the future this
week? Well even though there’s nuclear facilities, mining crypto,there’s 10 new unicorns in China, the biggest story of the week is Facebook. In fact there’s been so many stories around Facebook this week that we really couldn’t go past
the topic. The article we picked is Engadget and it’s titled Facebook patents tech to determine social class.” So Facebook has just patented a plan for what they call the “socio economic group classification based on user features” which is basically Facebook’s way to patent the technology to decide if its users are upper class, or middle class, or working class without using the usual markers that you would have things like
individuals income. So what Facebook is doing instead is classifying people by using things like whether or not you own a home. What is your level of education? The number of gadgets you own. Do you have a phone? Do you have an iPad? Do you have a phone and two iPads? How much do you use the Internet? Potentially what restaurants you go to or what brand of shoes you wear. If you for instance have only one gadget and you don’t use the
Internet so much, Facebook’s algorithm will determine that you probably are a person who’s not really well off. So the question then becomes why does Facebook care? Well they do care because that’s what they do. So there’s this article and there’s a significant number of other articles looking at
Facebook from different angles that we will
go into. So we will look at this story and then look at the bigger picture. Now first of all, Facebook is in the business of
classifying and characterising its users because that’s what they sell to the
advertising industry. So from that point of view it’s not really surprising that
Facebook would engage in this type of classification. Pun intended. But the author really gets worked up about first of all what it reveals about Facebook itself, the way in which they seem to
understand the world and class, but also what it will lead to when rolled out onto the platform. So given the way this is set up by Facebook and all of the patent documents are there you can have a
look at them online and they detail the types of
features that they would be looking at, things like demographic data or device ownership or even your travel history or other household type of data. It’s not a stretch to think that you would be able to say well this sort of restaurants are
attended by a particular class of people or these brands of phones are owned by a particular type of people. So we can see the way that this
would not only contribute greatly to the way Facebook already makes its money by determining these groups of users and allowing people to advertise but also the way in which this will
reinforce how the algorithm actually creates and shapes this groups of people. So what is significant for me is
that Facebook seems to ignore by and large the huge body of research and literature that exists on milieus and class and social stratification in
society. Instead they come up with their own criteria and they seem to have a ranking of
cities you live in for example. Meaning if you live in Palo Alto, it’s better than if you live in
Mountain View, which is better than living in San
Jose. It seems to me a rather haphazard and naive approach. But more importantly is that if they roll this out into the platform it will have
material consequences on the kind of things that will be put in front of people so
depending on where you live and whether you only have two and not seven devices that
communicate with Facebook, you might be given less opportunities you might not be
shown that job ad, you might be shown ads for predatory loans. So really what happens here is that they come up with this crude
definition of what a class stratification in their user group is which then will actually create that very reality and the author points that
out. So it’s indeed the two points that
you make. One is that in a ridiculous way this will create the simplistic
reality that Facebook describes by assigning people to these categories and then reinforcing those
categories with those ads. And second it’s important to note
that this is not out of Facebook’s malicious intent. The fact that poor people will be shown predatory loans, it’s not something that Facebook is
setting out to do or that Facebook plans or wishes in any way it’s a
byproduct of the way this is set up. The algorithm has to show those ads to people who are most likely to click on them. So we know people who are desperate
will most likely accept higher interest rates simply because they cannot
access those loans so the algorithm just does what it’s supposed to do. Those people anyway have lower
chances to get a fair loan or to get access to normal loans. I will actually show them a ad that will allow them access to
those loans so they’re more likely to
click on it. So this is again not out of any intent that Facebook has but rather a byproduct of the
algorithm. To me what makes this story significant is the timing at which it is released. In a week when we had articles in The Outline talking about how Facebook gave money to and funded most of the experts that vetted its messenger kids app and the backlash around that has been in the media in the
past couple of weeks. There was an article on early employees on Facebook and Google forming a coalition to fight the negative impacts on social media. A long article in Wired magazine on what it was like inside Facebook during the two years that shook them to the core during the election and what that revealed
about how Facebook works. And then revelations that Facebook
is spying on its own employees to try and get them from talking to the media and revealing insider information so it hasn’t really been a good week for Facebook so we
thought we’d take a deeper look at the problem and summarise what the issues are and what it means for us. So what we want to do is have a look
at what are basically four big
conversations that are happening around the issues that social media platforms in
general have and Facebook’s been singled out
this week but we could see this of most of the social media platforms that
we are part of and try to unpack what this means not only for the business model of
companies like Facebook but also what this
means for other organisations/ companies using their services and the consumers
themselves. So the first conversation we want to
look at is this idea of the ability to exploit people who are part of these networks. We’ve known this for a long time
right. So saying that if the service is
free you are the product or the old joke about Facebook that if it was a butchery we’d be the sausages. We’ve been telling these stories for
a long time but it is now becoming quite evident that the business model of Facebook is really to sell users and their information to advertisers and anyone who is prepared to pay money for using Facebook’s information to target content towards users. This idea is not inherently wrong. We are accessing free services for which we should pay something and the way we pay for
some of these services is by disclosing some of our data so for instance if I use
Google Maps the service is free that’s a lot of value to my life but I need to share some of my data in exchange for that free service. But something a little bit different is enabled on a platform like
Facebook. Because it is happening at a larger scale and it is done algorithmically, it’s basically out of control in the sense that even Facebook doesn’t understand 100 percent how it works, who is shown which information, and what are the kind of keywords
that the algorithm will allow people to utilise to discriminate between people in targeting content and that has led to these unfortunate incidents where people
were able to target users on racial grounds for example and that has been in the media. In addition to that what happens is that companies who wish to enlist large numbers of people to achieve their own goals and we’ve seen this with the US
elections can do so based on these algorithms, they can create types of content that will engage a certain category of people, make sure it gets shown to that
particular type of people who then will be
tricked into behaving in ways that they didn’t intend to. Yes so the second issue that is often singled out with Facebook and social media is that it leads to usage addiction and that’s at the heart of the initiative by former Silicon Valley employees who basically say these platforms
are built just like poker machines, they are built in a way that maximises people’s time on the platform, that entices them to click on things that make these platforms sticky and people basically share more information as a result of this but at the same time they can’t really let go of the device so it
becomes an addiction and it starts very
early so we see this in children, we see this in teenagers and it’s linked to studies that show that it has a measurable effect on brain development for example. The third conversation is that around this aspect of social
isolation and again with this conversation we
want to stress that this is not something
that Facebook set out to do. This is not something that they
consider a good thing. No ironically it’s quite the
opposite of what they set out to do I mean
Zuckerberg has been outspoken about that the job of Facebook is bringing people closer together. But by trying to bring people closer together it tries to show people
content that they would want to engage in, news from their friends, news from their loved ones, news about topics that are of
interest to them. The unintended side effect of that is that it creates this echo chambers where I have a certain belief, I click on articles that
reinforce that belief or a news that
reinforces that belief, I befriend people who share those beliefs who then
again share stories that reinforce those beliefs. So I end up in my own bubble of news, in my own bubble of beliefs and of interests that I cannot get out of. And of course this is not a new
argument, one that has been discussed before and people are now starting to be
aware of this but they have to make a
conscious effort to get out of that thought
bubble they have to go to other places to
look for that news which is neither easy to do nor is something that the platform itself can help you do. No first of all you have to be aware
of the way in which what you’re being presented is not the full picture but actually what the algorithm has
deduced is most likely to engage you in clicking on stuff. So that’s the first step and then obviously it’s quite
convenient to just stay in your bubble in your daily
life and just consume whatever shows up in your feed. Actually there was a great article
in Techcrunch a few weeks ago titled “Social media is giving us trypophobia” and don’t Google trypophobia – the pictures that come up you cannot
unsee. But anyway the fear of holes it’s
called. It’s not real. I looked it up on Wikipedia. The pictures are still gross but we digress. The article is excellent because it shows that this echo chamber works on different levels so it leads to social segmentation so it polarises so we end up in these groups that are very homogenous but do not interact in any ways with other homogenous groups so it leads to social
fragmentation. And they find it quite difficult to
have a public conversation where the arguments and the debates that they are having
are quite distinct. And we see this in the political
landscape but it also works at the individual
level where content is more and more targeted and what I am presented is quite different from what Sandra is
presented and we do not actually get the news as they happen but the news as the algorithm decides is relevant to me or Sandra. So the last conversation around
Facebook in particular this week, well social media in general, has been the privacy one and this goes back to conversations we have had before on The Future, This Week where we’ve talked about
Facebook’s uncanny ability to suggest, match up friends and contacts who you might not even know exist. Family members that you might not
know exist, connections that you do not know how
they are made. And here we discussed how Facebook uses up to 100 different data providers so it accurately aggregates your
behaviour and data about you from sources other than your activity on Facebook to have a full sometimes eerily accurate picture of who you are and who you interact with and what you’ve done
previously. Every single minute detail of your
life. So what we’re saying is that
Facebook has been in the news for those four issues: the way in which users are exploited and fake news becomes a problem as a result, that it leads to social media addiction, to echo chambers and social isolation or fragmentation, and that it comes with large scale privacy invasion or eradication that privacy is basically non-existent once you become part of this
ecosystem. So the question is how did we get there, how is this happening? The way this comes about is really the business model that
Facebook is built around. Facebook is built around the ability to provide this service to users, the ability to connect and read news and all of that but funds the way it does this by allowing corporations or individuals or governments or charities access to particular groups of people with particular traits, it allows them to advertise to these people, to put content in
front of them and it is based on the ability to do that really really well. So this harks back to last week’s discussion about fiduciary business models and we discussed there that Facebook
has two different groups it caters for:
it has users that use the platform but the income is generated from another group which are advertisers and corporations so they optimise for two different groups and balancing this out can be
difficult because turning the platform into something that is more
profitable does not necessarily come with making it better in whatever sense for their user group. So how does Facebook generate
income? Well it’s a combination of a number of factors. So the story that we’ve discussed around social class means that at the very heart of the business model is to be able to profile its users by way of complex psychometric analysis which comes hand in hand with the collection of a very broad set of data and then the discrimination of
content that is being presented to users be it news, be it adds, be it opportunities for jobs and things like that. And that is paired with the claim that Facebook is a platform that every content bit is presented equal so this hands off approach to content anything can be presented. So those factors in combination
basically gets us these effects and given that this is done in an automated way and at a large scale gets us the effects that we’ve discussed with fake news and echo chamber so absolutely Facebook has never set out to deliberately do it in that way but what we’ve learned in the past year and Facebook has learned this as well is that this system can be exploited. The system has been exploited by Russia meddling in the US election but also by people who just optimise for income with no regard for the kind of information that will
get engagement from users be it real or be it fake. And the potential for this to
continue so the story we started out with was a story around Facebook and its ability to determine social class. Now we talked about predatory lending but you can think about how
this would enable people to rent out their apartments only to a certain type of people and how that would
segregate neighbourhoods. You might be able to
use the same thing for dating, you could date only a certain type
of people or show certain profiles only to people of a certain social class or social background so the potential to do this is far greater than what we’ve seen so far. And the argument has been made in one of the Wired articles this week is that at the heart of the problem is Facebook’s hands off approach so Facebook has basically tried to be a news company without engaging journalists and without making judgements about content whether it’s real, whether it’s fake, whether it’s
good, whether it’s engaging whether it’s of a particular genre, an opinion piece or a factual article. Facebook is doing none of that and has left it to its algorithm to basically discriminate among content and present it to its users. And this is now what’s coming to
bite back. So where this leaves us at the
moment is with a pretty decent understanding of what the problems are and this is true across the various conversations that we’ve discussed with Facebook across all four
conversations. We have a pretty good understanding
of what the problem is and why we have the
problems that we do. They are baked into the
business model. It doesn’t leave us a lot further however in how we solve this. Facebook has been having go’s
at addressing all of this issues for
the past almost a year. They’ve hired thousands of people to weed out the most obvious problems with content presentation for
example. They’ve tried various models for how to present less fake news one of them was adding a Wikipedia button where you would
look up on Wikipedia whether the news is
trustworthy. They flirted with the option of
promoting content that your family and friends think is relevant
content, a variety of different approaches so it’s not for lack of trying. But if you look at the analysis and the complexity of the problems it’s pretty clear that this is dealing with the symptoms more than addressing the systemic problem that is at the heart of what is
happening. So because this is baked into the
business model this is really just tinkering with the surface of it. But what do we do about it? I mean we could just say that’s just
the way it is right, this is how social media
works. But to be honest this is just the
way it is with Facebook. So if you look up a company like Tencent who owns WeChat they have managed to create a
business model that does not rely as strongly on advertising. They found ways to monetise the
small transactions that occur in WeChat let’s remember WeChat also enables payment. So all those little small transactions enable WeChat to have a variety of sources of
income and not rely so heavily in
advertising. The whole Tencent business model is also spread out around the variety of industries including finance and infrastructure and so on. So their model is not solely dependent on selling users to companies and maximising clickthrough rate. So my personal opinion is that Facebook will have to do
something about it much more substantially
than obfuscating or denying or selling bullshit to the media. They will have to actually look at the business model in order to
recapture this positive image that has been with Facebook in its early days where it was known as a way for friends to connect and for expats to stay in contact with friends and family where the
whole narrative of what Facebook is centred around the front end of what it was doing for its users. Facebook is at a risk at the moment that their narrative shifts and that Facebook becomes known for what it does at its back end, being a large scale discrimination machine that basically sells users to the highest bidder. And if that was the case it might disengage its users and therefore erode the very basis on which it rests and we have seen in the past with MySpace that large scale networks can fall apart and unravel. And just because Facebook is big doesn’t necessarily mean that it is too big to fail if something came in its place and we’ve seen that WeChat is making advances into other countries it’s not just operating in China. So who’s to say that WeChat or something else isn’t
coming along to capture some of what Facebook used to be. So if this wasn’t gloomy enough for
you… Here’s our second story from
BuzzFeed News titled “He predicted the 2016 fake news crisis, now he’s worried about information apocalypse”. So who is he Sandra? So he is Aviv Ovadya and he’s an MIT graduate who’s done a few things at various tech companies around the
Valley companies like Quora. But who has now moved on and is a Chief Technologist for the University of Michigan’s Centre
for Social Media Responsibility. And Aviv has pretty much dropped everything in 2016 where he realised that the business
models behind companies like Facebook and Twitter would enable the kind of onslaught of fake news that we’ve
seen influence the American elections. And he has been trying to raise
attention ever since to what he perceives the inherent problems of the web and the information echo system that has developed around the social
media platforms. So the point that Aviv and this article are making is that on top of the kind of social media platforms that we’ve just discussed with Facebook and the way in which it tends to spread all kinds of fake news and things like that, that on top of this we are seeing a new wave of technological developments that make it easier to not just create but automate the creation of fake content and not just text but video, audio at a large scale. And he thinks that if the stories like the one we’ve had before about
the conversations that Facebook scare us, what comes next should really scare the shit out of us. So the stories collected in the
article are truly worrying. So what he does is ask a very simple question. What happens when anyone can make it
appear as if anything has happened
regardless of whether or not it did. And he presents a collection of examples that show what is already possible and then he imagines of where things
are going in a reasonably short amount of time. So for example: there’s examples now and that has been in the news of fake pornographic videos where algorithms use faces of celebrities sourced from Instagram for example to put it on fairly explicit video material. There is also the research out of Stanford that we’ve discussed here before that combine recorded video footage with real time face tracking to manipulate that footage and for instance puppeteer images of let’s say Barack Obama in a very believable way and make him look like he said
things or he did things that he hasn’t
really done. And we’ve discussed previously
research that is being done here at the University
of Sydney Business School into a photo realistic believable avatars which are piloting the technology to basically create synthetic believable humans that can be puppeteered which can be copies of real people or synthetically created people so the prospect is that we
now enter an era where I can make anyone say anything on video. I can create synthetic humans, we would not even know whether we’re
conversing with a real or a fake human. We’re entering an era where not even video or live video streams can be believed. So why does this spell the end of the Internet for Aviv? So there’s two steps to this. First of all obviously there’s these
scenarios that fake content can create serious problems, mass panic, diplomatic incidents. US presidents declaring war on other countries. And currently war game style disaster scenarios based on this technologies are being run to see
how these would play out. But the more worrying part is that the ability to create any type of fake content at the same time discredits what is real. So I can always turn around and say this piece of information
could be fabricated or this video isn’t real because we now have the technology to create this kind of video synthetically. Who gets to say what is real and what is fake as a result. Furthermore these things will all start competing with real people and real voices for the same type of limited attention. So think about your representatives in Government. They could be flooded with a range of messages from constituents that sounds real, look real, asking them to behave in a certain way or vote for certain policies. These things will compete for real time and real attention from legislators and will come at the expense of other voices. And even in the absence of large scale doomsday scenarios it starts small. What happens when you can no longer distinguish between real messages from colleagues in your
inbox and fake messages that are there as phishing attempts, as spam essentially. There’s just this week been an
article about Google rolling out a new service called Reply which is
essentially a bot that allows you to have friend messages answered automatically. So say someone contacts me while I’m doing this podcast with Kai, this bot would see in my calendar that I’m
scheduled to record this, would send the message
saying I’m just recording this with Kai, please call back between
these and these hours when I’m free and so on. But you can imagine the
complexity of these things growing and the accuracy with which they will be able to
conduct conversations on our behalf growing. So the real issue is that when you can no longer discern what is true, what is real, what is fake, what is fabricated in social conversations online, on the Internet, or news platforms on social media this will basically mean the end of the Internet as we know it as a source of information, as a source of social interaction. So where does this leave us in a
world with a lot of fake content where social platforms like Facebook act as accelerators for distributing this content and amplify the effects that they would have. Whilst we are trying to find
solutions to this and we’ve discussed this with our previous story. In the interim we have very few alternatives left. One of them arguably could be turning to trusted institutions where we know that certain news outlets have been trusted institutions for over 100 years, turning to universities as sources of knowledge. Yes absolutely I think we will see a move back to the future if you wish in the sense that for a long time we have decoupled content from its source, social media Facebook trades in presenting all kinds of content decoupled from its originators. I think we will have to see a big move back to trusted brands such as The New York Times or consumer brands Amazon, Apple to institutions who have a proven interest in presenting real and trusted information and services. So I think the Internet as a global democratic space where every voice is equal, I think we’re about to see destroyed precisely because it’s very hard to know what and who we can trust to tell us something about the world or something that is entirely
fabricated to serve some ulterior motive and self-interest. Well that is slightly depressing. That’s more than slightly depressing but lucky enough we have Future
Bites so our quick fire round to finish off this podcast. So what did you learn this week
Sandra? Other than that we’re doomed. I’ve learned some very interesting things about how to mine for cryptocurrencies. An interesting case from Manchester where if you went to the city
council’s website, so let’s say trying to check on when you should put your
bins out for your garbage to be
collected, your computer would be automatically recruited to mine for
crypocurrencies and if you think that’s weird
apparently something similar happened that the
Russian nuclear facility where scientists actually used these super powerful computers
to mine for cryptocurrency. And predictably got in trouble for it. Oh and apparently this whole crypto craze also challenges our search for extraterrestrial life. The SETI program cannot buy powerful graphics cards because
apparently they’re all being enlisted to mine for cryptocurrencies so they’re
unable to purchase them. Well speaking of extra terrestrial life, the space station apparently is up for sale. Donald Trump and his government are
changing the funding for NASA and apparently they are no longer
prepared to foot the bill for the space station and they are putting
up for sale, that’s some exclusive private real estate, comes in at a handy 3 to 4 billion dollars. So who’s going to buy that one? Well we had a think about this and given that Elon Musk has just
launched his Tesla into space we think this could be a great
opportunity for him to acquire it. Build a charging station for it or use it as a garage. But there’s been another article
about people like US billionaire Peter Thiel who increasingly buy property around the world to build bunkers to hide out from the doomsday scenarios that we’re so fond of discussing on The Future, This Week and hey wouldn’t that be a great
hideout spot. So space stations and earth bunkers. But you want to pick them wisely
right. While an earth bunker is great to hide out from alien invasions, the space station might be better in case of a zombie apocalypse so… Get one of each? Hedge your bets. Speaking of bets and things to come, there’s been a whole bunch of
articles about China’s tech revolution that has only just begun and they talk about the ascendance of this super power through entrepreneurship through innovation through a number of start ups that have reached unicorn status this year alone. There have been about 15 of them and effectively about a third of the world’s billion dollar companies were created in
China last year alone. China’s ride-sharing company Didi has surpassed Uber, we’ve talked a lot about Uber on
this podcast, we haven’t spoken that much about similar companies, bigger companies, coming out of China. And we want to stress that Didi has
not just surpassed Uber because of the scale at which China does business, but Didi has ventured into many other countries such as India, Brazil, Mexico, and also opening a lab in Silicon Valley. So really starting to
compete with Uber and other ride-sharing in other jurisdictions. So any last stories Kai? So remember we talked about fake
reviews previously? No. Amazon apparently has a feature
where you can only leave a review if you actually buy stuff. Turns out that doesn’t stop fake reviewers. There’s been an article in The
Outline that reports on a retired couple in Massachusetts who found that they were receiving a lot of free deliveries from Amazon and they started to investigate and it turns out people who want to
leave a fake review for a product buy a gift card to leave no trace and then purchase the item and purchase a data set and just send their stuff to random
people who then find that they receive a never ending stream of free stuff that they neither ordered nor want. And on that very strange note that’s all we have time for this week. Thanks for listening. Thanks for listening. Well that was the depressing. Yeah you thought Facebook was shit. The whole Internet is gone. And you can’t even afford to buy the
space station. Gonna wait for Musk to take it to Mars now. Better luck next week. He’s going up there you know. To charge his Telsa. This was The Future, This Week made awesome by the Sydney Business Insights team and members of the Digital
Disruption Research Group. And every week right here with us our sound editor Megan Wedge who makes us sound good and keeps us honest. Our theme music was composed and played live from a set of garden
hoses by Linsey Pollak. You can subscribe to this podcast on iTunes, Soundcloud, Stitcher, or
wherever you get your podcasts. You can follow us online on
Flipboard, Twitter or sbi.sydney.edu.au. If you have any news that you want us to discuss please send them to [email protected]

, , , , , , , , , , , , , , ,

Post navigation

Leave a Reply

Your email address will not be published. Required fields are marked *