This is a fascinating story of psychological manipulation and internal politics. It leaves me feeling queasy about the amount of power wielded by individuals in one single organisation.
I have no doubt that showing just the top outrageous tweets leads to more engagement. If you’re constantly hitting people with outlandish news stories they’ll open the app more often and interact and post about what they think so the cycle continues.
Speculative fiction as a tool for change:
We need to think harder about the future and ask: What if our policies, institutions, and societies didn’t have to be organized as they are now? Good science fiction taps us into a rich seam of radical answers to this question.
An Interview with Nick Harkaway: Algorithmic Futures, Literary Fractals, and Mimetic Immortality - Los Angeles Review of Books
Nick Harkaway on technology in fiction:
Humans without tools are not magically pure; they’re just unvaccinated, cold, and wet.
SF is how we get to know ourselves, either who we are or who we might be. In terms of what is authentically human, SF has a claim to be vastly more honest and important than a literary fiction that refuses to admit the existence of the modern and goes in search of a kind of essential humanness which exists by itself, rather than in the intersection of people, economics, culture, and science which is where we all inevitably live. It’s like saying you can only really understand a flame if you get rid of the candle. Good luck with that.
And on Borges:
He was a genius, and he left this cryptic, brilliant body of work that’s poetic, incomplete, astonishing. It’s like a tasting menu in a restaurant where they let you smell things that go to other tables and never arrive at yours.
James describes an ancient Greek machine called the kleroterian:
The method of governance embodied in the kleroterion, which dates back to the very establishment of democracy, is called sortition, meaning selection by lot, as opposed to election by vote.
A brilliantly written piece by Laurie Penny. Devestating, funny, and sad, featuring journalistic gold like this:
John McAfee has never been convicted of rape and murder, but—crucially—not in the same way that you or I have never been convicted of rape or murder.
This instance of collective action from inside a tech company is important, not just for the specifics of Google, but in acting as an example to workers in other companies.
And of all the demands, this is the one that could have the biggest effect in the US tech world:
An end to Forced Arbitration.
Chris is putting his examination of interfaces in science fiction on pause while he examines a more pressing matter for today’s political climate—an examination of depictions of fascism in science fiction:
Facebook doesn’t have a mind-control problem, it has a corruption problem. Cambridge Analytica didn’t convince decent people to become racists; they convinced racists to become voters.
Prompted by his time at Clearleft’s AI gathering in Juvet, Chris has been delving deep into the stories we tell about artificial intelligence …and what stories are missing.
And here we are at the eponymous answer to the question that I first asked at Juvet around 7 months ago: What stories aren’t we telling ourselves about AI?
What’s happening right now at the US border is heartbreaking and inexcusable. We’re donating 25% of all profits today and tomorrow (June 19 & 20) to RAICES, to help reunite detained immigrant parents and children.
A hand-wringing, finger-pointing litany of hindsight, published with 11 tracking scripts attached.
- Start With Hippie Good Intentions …
- … Then mix in capitalism on steroids.
- The arrival of Wall Streeters didn’t help …
- … And we paid a high price for keeping it free.
- Everything was designed to be really, really addictive.
- At first, it worked — almost too well.
- No one from Silicon Valley was held accountable …
- … Even as social networks became dangerous and toxic.
- … And even as they invaded our privacy.
- Then came 2016.
- Employees are starting to revolt.
- To fix it, we’ll need a new business model …
- … And some tough regulation.
- Maybe nothing will change.
- … Unless, at the very least, some new people are in charge.
During the Industrial Revolution, as new machines were invented to increase output, business owners often dreamed of an entirely automated workforce—of a factory without workers. I assume their workers had different dreams.
Ethan thinks through the ethical implications of increasing automation and efficiency über alles:
I can’t stop thinking about how much automation has changed our industry already. And I know the rate of automation is only going to accelerate from here.
At the very least, maybe it’s worth asking ourselves what might happen next.
A smart look back at historical examples of regulation and what we can learn from them today, by Justine Leblanc:
- Railways in the UK: Public interest as a trigger for regulation
- Engineering in Canada: Accountability as a trigger for regulation
- The automotive industry in the USA: Public outrage as a trigger for regulation
The transcript of a terrific talk by Paul, calling for a more thoughtful, questioning approach to digital design. It covers the issues I’ve raised about Booking.com’s dark patterns and a post I linked to a while back about the shifting priorities of designers working at scale.
Drawing inspiration from architectural practice, its successes and failures, I question the role of design in a world being eaten by software. When the prevailing technocratic culture permits the creation of products that undermine and exploit users, who will protect citizens within the digital spaces they now inhabit?
The word “leak” is right. Our sense of control over our own destinies is being challenged by these leaks. Giant internet platforms are poisoning the commons. They’ve automated it.
I’d love to see some change, and some introspection. A culture of first, do no harm. A recognition that there are huge dangers if you just do what’s possible, or build a macho “fail fast” culture that promotes endangerment. It’s about building teams that know they’ll make mistakes but also recognize the difference between great businesses opportunities and gigantic, universe-sized fuck ups.
I recently received this very instruction about speaking at an upcoming event. I honestly don’t know how I could talk about universality, progressive enhancement, and user experience without it being political. So I interpreted the request to be about partisanship rather than politics:
Sometimes when people hear the term “political,” they understand it as “partisan.” To be political is to acknowledge the lived experiences of people outside of yourself. To be partisan is to advocate for the beliefs or propaganda of a specific party affiliation.
Ethan adds his thoughts to my post about corporations using their power to influence the direction of the web.
Heck, one could even argue the creation of AMP isn’t just Google’s failure, but our failure. More specifically, perhaps it’s pointing to a failure of governance of our little industry. Absent a shared, collective vision for what we want the web to be—and with decent regulatory mechanisms to defend that vision—it’s unsurprising that corporate actors would step into that vacuum, and address the issues they find. And once they do, the solutions they design will inevitably benefit themselves first—and then, after that, the rest of us.
If at all.
A catalogue of design decisions that have had harmful effects on users. This is a call for more inclusive design, but also a warning on the fetishisation of seamlessness:
The focus on details and delight can be traced to manifestos like Steve Krug’s Don’t Make Me Think, which propose a dogmatic adherence to cognitive obviousness and celebrates frictionless interaction as the ultimate design accomplishment.