As we’ve argued previously, here, here, and here, algorithms are no magic wand for sorting subjective content, artistic or otherwise. Here the top brass at Apple admits to the fact in a criticism of one of its competitors, Spotify.
Two take-aways from Mr. Cook’s argument. One, he claims that Apple uses the human creativity of its users to create playlists, but tuka uses it’s entire network of users to reward human curation of all content. Yeah, web 3.0 is about the human, not the machine.
Second, the race among the streamers, Apple Music, Amazon Music, Google Play, and Spotify, Pandora, etc. is another loss-leading attempt to build out a user base to monetize the data flow. In other words, streaming doesn’t pay unless you can monetize the network in some other way. As I suspected, the pricing model of streaming is likely financially unsustainable in the long-run as the true price of streaming content is several times what consumers are now paying. Direct ownership of content may be far more economical than renting it without ad support.
Apple CEO Tim Cook recently revealed the company’s streaming music service, Apple Music, has clearly surpassed Spotify in subscribers in the US, Canada, and Spotify.
Instead of gloating, he humbly downplayed the achievement.
“The key thing in music is not the competition between the companies that are providing music, the real challenge is to grow the market. If we put our emphasis on growing the market, which we’re doing, we’ll be the beneficiaries of that, as will others.”
In a recently published article with monthly business magazine Fast Company, Cook revealed his true feelings about his streaming music rivals. Taking a clear swipe at Spotify, he said,
“We worry about the humanity being drained out of music, about it becoming a bits-and-bytes kind of world instead of the art and craft.”
I would add that the major problems for artists in the digital age stem from the explosion of new supply of content. This drives the price down and the search costs of discovery up. The failure then becomes that artists can’t find their audiences and consumers can’t find the content they desire. For poets this means finding an audience not necessarily to sell poetry; rather more important is to find readers and appreciators of their poetry.
Large centralized network servers based on algorithms can’t solve this problem without commoditizing content and delivering the most popular but mundane content churned out by those metrics.
We need to empower the human by connecting the creative.
Recent reports that streaming is now the ‘biggest money-maker’ for the music biz have prompted hyperbolic claims that Spotify and co have ‘saved the music industry’. In reality, this could not be further from the truth.
Progressive music that goes against the aesthetics of whatever the mainstream might be at any given point by its very nature does not cater to the whims of a Spotify algorithm. Now that streaming is the industry’s biggest money-maker it has become the overriding force in music consumption. This dominance will only increase as time goes on, and for artists to gain anything, as a result, requires them to conform or die. There are exceptions, most notably in zeitgeist-seizing movements like grime that are both artistically essential and buoyed by the kind of mass appeal that in effect bypasses the need for a leg-up from the algorithms, but such a lethal combination is rare indeed. Not everything that is great is as popular.
Yes, not everything that is great is popular, and not everything that is popular is great! We need human subjectivity, and that’s more complicated than a complex algorithm.
If streaming platforms keep growing more and more influence over how music is curated and marketed by those in charge, while the revenue for those not mundane enough to fit their algorithms remains so pitifully minute, it is not that impossible to envisage the blandest landscape the industry has ever seen. Great music will continue being made, of course, but getting that music out to people outside of the algorithms will be so much harder. “I hope I am wrong,” says Reeder. “I hope the revenue from streaming does improve, because if it doesn’t, well, who knows how positive the future will be for the majority of music makers and labels out there?”
This is not only the case for music, but for literature, poetry, video, and photography.
Americans are depressed and suicidal because something is wrong with our culture
Excerpt from an article examining the rise of celebrity suicides such as Anthony Bourdain and Kate Spade. This gets to the heart of what makes us fulfilled as human beings. Not fame and fortune, but ascending Maslow’s hierarchy in our own ways:
…why are so many more Americans getting to this level of emotional despair than in the past? As journalist Johann Hari wrote in his best-selling book Lost Connections: Uncovering the Real Causes of Depression — and the Unexpected Solutions, the epidemic of depression and despair in the Western World isn’t always caused by our brains. It’s largely caused by key problems in the way we live.
We exist largely disconnected from our extended families, friends and communities — except in the shallow interactions of social media — because we are too busy trying to “make it” without realizing that once we reach that goal, it won’t be enough.
In an interview this year, the comedian and actor Jim Carrey talked about “getting to the place where you have everything everybody has ever desired and realizing you are still unhappy. And that you can still be unhappy is a shock when you have accomplished everything you ever dreamt of and more….”
If only we get that big raise, or new house or have children we will finally be happy. But we won’t. In fact, as Carrey points out, in many ways achieving all your goals provides the opposite of fulfillment: it lays bare the truth that there is nothing you can purchase, possess or achieve that will make you feel fulfilled over the long term.
Rather than pathologizing the despair and emotional suffering that is a rational response to a culture that values people based on ever escalating financial and personal achievements, we should acknowledge that something is very wrong. We should stop telling people who yearn for a deeper meaning in life that they have an illness or need therapy. Instead, we need to help people craft lives that are more meaningful and built on a firmer foundation than personal success.
Excerpt from a review of books in The New Yorker: “Are Liberals on the Wrong Side of History?”
The algorithms of human existence are not like the predictable, repeatable algorithms of a computer, or people would not have a history, and Donald Trump would not be President. In order to erase humanity as a special category—different from animals, on the one hand, and robots, on the other—Harari [author of Sapiens] points to the power of artificial intelligence, and the prospect that it will learn to do everything we can do, but better. Now, that might happen, but it has been predicted for a long time [there’s more to the human than we assume] and the arrival date keeps getting postponed.
The A.I. that Harari fears and admires doesn’t, on inspection, seem quite so smart. He mentions computer-generated haiku, as though they were on a par with those generated by Japanese poets. Even if such poems exist, they can seem plausible only because the computer is programmed to imitatestylistic tics that we have already been instructed to appreciate, something akin to the way the ocean can “create” a Brancusi—making smooth, oblong stones that our previous experience of art has helped us to see as beautiful—rather than to how artistsmake new styles, which involves breaking the algorithm, not following it.
This argument is relevant not only to the creation of art, but also the appreciation of same. And the appreciation of art is expressed in reviews, aesthetic appeal, and novelty, as well as popularity. Algorithms trigger mostly off various proxies for popularity, such as “likes” or sales. We need human aesthetic judgments to support true artistic creation.
Google? Facebook? These two firms alone control roughly 2/3s of digital media advertising revenues.
That’s power. That’s knowledge.
But knowledge of what?
Mostly of how to program computers and deploy algorithms to sort through, organize, cluster, rank, and order vast quantities of data. In the case of Facebook, Zuckerberg obviously also understood something simple but important about how human beings might enjoy interacting online. That’s not nothing. Actually, it’s a lot. An enormous amount. But it’s not everything — or anything remotely close to what Silicon Valley’s greatest innovators think it is.
When it comes to human beings — what motivates them, how they interact socially, to what end they organize politically — figures like Page and Zuckerberg know very little. Almost nothing, in fact. And that ignorance has enormous consequences for us all.
Very interesting article by Joel Kotkin, who researches the economics and politics of cities. It portrays a future that resembles feudalism more than free market democratic capitalism. I’d optimistically venture there will eventually be a more humanist backlash against the future dominance of technology.
The drive to redesign our cities, however, is not really the end of the agenda of those who Aldous Huxley described as the top of the “scientific caste system.” The oligarchy has also worked to make our homes, our personal space, “connected” to their monitoring and money machines. This may be a multibillion-dollar market soon, but many who have employed such devices at home—appliances that track our activities and speak to us like loyal servants—find them “creepy,” as they should, given that their daily activities are fed back to enrich the high-tech hive mind. Both the city and house the future may owe more to Brave New World than Better Homes and Gardens.
This is a vision of the urban future in which the tech companies’ own workers and whatever other people with skills the machines haven’t yet replaced are a new class of urban serfs living in small apartments, along with a much larger class of dependent persons living on “income maintenance” and housing or housing subsidies provided by the state. “Bees exist on Earth to pollinate flowers, and maybe humans are here to build the machines,” observes professor Andrew Hudson-Smith, from University College London’s Centre for Advanced Spatial Analysis. “The city will be one big joined-up urban machine, and humans’ role on Earth will be done.”
Below are excerpts from a fascinating series of articles by The Guardian (with links). The articles address many of the ways that Internet 2.0 network media models such as Google, Facebook, YouTube, Instagram, etc. are transforming, and in many cases undermining, the foundations of a democratic humanistic society. These issues motivate us at tuka to design solutions to the great question of life’s meaning.
Personally, I don’t believe this dystopia will come to pass because humans are quite resilient as a species and eventually our humanist qualities will dominate our biological urges and economic imperatives. We have free will and ultimately, we choose correctly.
Perhaps that is an overly optimistic opinion, but Internet (or Web) 3.0 technology is rewriting the script with applications that reassert human control over the data universe. We will build more humanistic social communities that employ technology, with the emphasis always on the human. We see this now with the growing refusal to surrender to Web 2.0 by tech insiders.
See excerpts and comments below.
“If politics is an expression of our human will, on individual and collective levels, then the attention economy is directly undermining the assumptions that democracy rests on.” If Apple, Facebook, Google, Twitter, Instagram, and Snapchat are gradually chipping away at our ability to control our own minds, could there come a point, I ask, at which democracy no longer functions?
There are 1.5 billion YouTube users in the world, which is more than the number of households that own televisions. What they watch is shaped by this algorithm, which skims and ranks billions of videos to identify 20 “up next” clips that are both relevant to a previous video and most likely, statistically speaking, to keep a person hooked on their screen.
Company insiders tell me the algorithm is the single most important engine of YouTube’s growth. In one of the few public explanations of how the formula works – an academic paper that sketches the algorithm’s deep neural networks, crunching a vast pool of data about videos and the people who watch them – YouTube engineers describe it as one of the “largest scale and most sophisticated industrial recommendation systems in existence”.
We see here the power of AI data algorithms to filter content. The Google response has been to “expand the army of human moderators.” That’s a necessary method of reasserting human judgment over the network.
The primary focus of the article then turns to politics and the electoral influences of disinformation:
Much has been written about Facebook and Twitter’s impact on politics, but in recent months academics have speculated that YouTube’s algorithms may have been instrumental in fuelling disinformation during the 2016 presidential election. “YouTube is the most overlooked story of 2016,” Zeynep Tufekci, a widely respected sociologist and technology critic, tweeted back in October. “Its search and recommender algorithms are misinformation engines.”
Apparently, the sensationalism surrounding the Trump campaign caused YT’s AI algorithms to push more video feeds favorable to Trump and damaging to Hillary Clinton. One doesn’t need to be a partisan to recognize this was probably true for this particular media channel and its business model that values more views more than anything else.
However, this reality can also be distorted to present a particular conspiracy narrative of its own:
Trump won the electoral college as a result of 80,000 votes spread across three swing states. There were more than 150 million YouTube users in the US. The videos contained in Chaslot’s database of YouTube-recommended election videos were watched, in total, more than 3bn times before the vote in November 2016.
This, unfortunately, is cherry-picking statistical inferences concerning the margin of voting support. What was significant in determining the 2016 election outcome was not 80,000 votes across three states, but a run of popular vote wins in 2,623 of 3,112 counties across the U.S. This 85% share could not be an accident, nor could it be due to the single influence of disinformation, Russian or otherwise. The true difference in the election was not revealed by the popular vote total or the Electoral College vote, but by the geographical distribution of support. One can argue about which is more critical to democratic governance, but this post is about electronic media content, not political analysis.
The next article further addresses how technology is influencing our individual behaviors.
Justin Rosenstein had tweaked his laptop’s operating system to block Reddit, banned himself from Snapchat, which he compares to heroin, and imposed limits on his use of Facebook. But even that wasn’t enough. In August, the 34-year-old tech executive took a more radical step to restrict his use of social media and other addictive technologies.
A decade after he stayed up all night coding a prototype of what was then called an “awesome” button, Rosenstein belongs to a small but growing band of Silicon Valley heretics who complain about the rise of the so-called “attention economy”: an internet shaped around the demands of an advertising economy.
The extent of this addiction is cited by research that shows people touch, swipe or tap their phone 2,617 times a day!
There is growing concern that as well as addicting users, technology is contributing toward so-called “continuous partial attention”, severely limiting people’s ability to focus, and possibly lowering IQ.One recent study showed that the mere presence of smartphones damages cognitive capacity – even when the device is turned off. “Everyone is distracted,” Rosenstein says. “All of the time.”
“The technologies we use have turned into compulsions, if not full-fledged addictions,” Eyal writes. “It’s the impulse to check a message notification. It’s the pull to visit YouTube, Facebook, or Twitter for just a few minutes, only to find yourself still tapping and scrolling an hour later.” None of this is an accident, he writes. It is all “just as their designers intended”.
Tristan Harris, a former Google employee turned vocal critic of the tech industry points out that… “All of us are jacked into this system. All of our minds can be hijacked. Our choices are not as free as we think they are.”
“I don’t know a more urgent problem than this,” Harris says. “It’s changing our democracy, and it’s changing our ability to have the conversations and relationships that we want with each other.”
Harris believes that tech companies never deliberately set out to make their products addictive. They were responding to the incentives of an advertising economy, experimenting with techniques that might capture people’s attention, even stumbling across highly effective design by accident.
“Smartphones are useful tools,” says Loren Brichter, a product designer. “But they’re addictive. Pull-to-refresh is addictive. Twitter is addictive. These are not good things. When I was working on them, it was not something I was mature enough to think about.”
The two inventors listed on Apple’s patent for “managing notification connections and displaying icon badges” are Justin Santamaria and Chris Marcellino. A few years ago Marcellino, 33, left the Bay Area and is now in the final stages of retraining to be a neurosurgeon. He stresses he is no expert on addiction but says he has picked up enough in his medical training to know that technologies can affect the same neurological pathways as gambling and drug use. “These are the same circuits that make people seek out food, comfort, heat, sex,” he says.
“The people who run Facebook and Google are good people, whose well-intentioned strategies have led to horrific unintended consequences,” he says. “The problem is that there is nothing the companies can do to address the harm unless they abandon their current advertising models.
But how can Google and Facebook be forced to abandon the business models that have transformed them into two of the most profitable companies on the planet?
This is exactly the problem – they really can’t. Newer technology, such as distributed social networks tracked by blockchain technology, must be deployed to disrupt the dysfunctional existing technology. New business models will be designed to support this disruption. Human behavioral instincts are crucial to successful new designs that make us more human, rather than less.
James Williams does not believe talk of dystopia is far-fetched. …He says his epiphany came a few years ago when he noticed he was surrounded by technology that was inhibiting him from concentrating on the things he wanted to focus on. “It was that kind of individual, existential realization: what’s going on?” he says. “Isn’t technology supposed to be doing the complete opposite of this?”
The question we ask at tuka is: “What do people really want from technology and social interaction? Distraction or meaning? And how do they find meaning?” Our answer is self-expression through creativity, sharing it, and connecting with communities.
Williams and Harris left Google around the same time and co-founded an advocacy group, Time Well Spent, that seeks to build public momentum for a change in the way big tech companies think about design.
“Eighty-seven percent of people wake up and go to sleep with their smartphones,” he says. The entire world now has a new prism through which to understand politics, and Williams worries the consequences are profound.
The same forces that led tech firms to hook users with design tricks, he says, also encourage those companies to depict the world in a way that makes for compulsive, irresistible viewing. “The attention economy incentivizes the design of technologies that grab our attention,” he says. “In so doing, it privileges our impulses over our intentions.”
That means privileging what is sensational over what is nuanced, appealing to emotion, anger, and outrage. The news media is increasingly working in service to tech companies, Williams adds, and must play by the rules of the attention economy to “sensationalize, bait and entertain in order to survive”.
In the wake of Donald Trump’s stunning electoral victory, many were quick to question the role of so-called “fake news” on Facebook, Russian-created Twitter bots or the data-centric targeting efforts that companies such as Cambridge Analytica used to sway voters. But Williams sees those factors as symptoms of a deeper problem.
It is not just shady or bad actors who were exploiting the internet to change public opinion. The attention economy itself is set up to promote a phenomenon like Trump, who is masterly at grabbing and retaining the attention of supporters and critics alike, often by exploiting or creating outrage.
Orwellian-style coercion is less of a threat to democracy than the more subtle power of psychological manipulation, and “man’s almost infinite appetite for distractions”.
“The dynamics of the attention economy are structurally set up to undermine the human will,” Williams says. “If politics is an expression of our human will, on individual and collective levels, then the attention economy is directly undermining the assumptions that democracy rests on.” If Apple, Facebook, Google, Twitter, Instagram, and Snapchat are gradually chipping away at our ability to control our own minds, could there come a point, I ask, at which democracy no longer functions?
Our politics will survive and democracy is only one form of governance. The bigger question is how does human civilization survive if our behavior becomes self-destructive and meaningless?