A gorgeous visualisation of satellites in Earth orbit. Click around to grasp the scale of the network.
A proposed syllabus for critical thinking: Calling Bullshit in the Age of Big Data.
Our aim in this course is to teach you how to think critically about the data and models that constitute evidence in the social and natural sciences.
You, the software engineers and leaders of technology companies, face an enormous responsibility. You know better than anyone how best to protect the millions who have entrusted you with their data, and your knowledge gives you real power as civic actors. If you want to transform the world for the better, here is your moment. Inquire about how a platform will be used. Encrypt as much as you can. Oppose the type of data analysis that predicts people’s orientation, religion, and political preferences if they did not willingly offer that information.
As always with sci-fi interfaces, the important part is telling the story, not realism or accuracy. Personally, I liked the way that the World War II trappings of Rogue One extended to communications and networking technologies.
Watching this data visualisation on its high speed setting is quite hypnotic.
This document provides Best Practices related to the publication and usage of data on the Web designed to help support a self-sustaining ecosystem. Data should be discoverable and understandable by humans and machines.
Run from data-driven companies. In thrall to semi-science and blinded by their dogma, they’ve lost the ability to see intelligent alternative perspectives on their business, their products, and the world. Embrace instead data-informed companies. This isn’t mere grammatical pedantry – a company genuinely informed by data understands the risks of datafication and adopts sophisticated, balanced approaches to strategy that blend quant, qual, and even some of that unfashionable prediction and intuition.
A series of quick’n’dirty prototypes to illustrate some of the design challenges involved in handling personal data:
- Data access tracker
- Data minimisation
- Guardian for digital identity
- Home privacy settings
- Portable shopping list
- Single trip insurance checker
If we don’t start exploring what the General Data Protection Regulation means for people, the same thing that happened with the cookie law will happen again.
These new rights have the potential to improve how our digital products and services work.
Here’s a fun cosmic hypothesis on the scale of an Olaf Stapeldon story. There are even implications for data storage:
By storing its essential data in photons, life could give itself a distributed backup system. And it could go further, manipulating new photons emitted by stars to dictate how they interact with matter. Fronts of electromagnetic radiation could be reaching across the cosmos to set in motion chains of interstellar or planetary chemistry with exquisite timing, exploiting wave interference and excitation energies in atoms and molecules.
If you’re in need of some long-term perspective right now—because, let’s face it, the short-term outlook is looking pretty damn bleak—then why not explore some of Max Roser’s data visualisations? Have a look at some of the global trends in inequality, disease, hunger, and conflict.
The view that more information uncritically produces better decisions is visibly at odds with our contemporary situation.
A superb piece of research and writing by James, skewering the technological determinism that underlies the current faith in “big data.” At best, this misplaced trust is inaccurate; at worst, it is deadly.
To the algorithmic imagination, the practice of journalism and the practice of terrorism appear to be functionally identical.
The Digital Transition: How the Presidential Transition Works in the Social Media Age | whitehouse.gov
Kori Schulman describes the archiving of social media and other online artefacts of the outgoing US president. It’s a shame that a lot of URLs will break, but I’m glad there’s going to be a public backup available.
Best of all, you can get involved:
In the interim, we’re inviting the American public – from students and data engineers, to artists and researchers – to come up with creative ways to archive this content and make it both useful and available for years to come. From Twitter bots and art projects to printed books and query tools, we’re open to it all.
Excellent guidelines from GDS on providing services that work well on mobile. The watchwords are:
- responsive design,
- progressive enhancement,
- open data, and
- emerging technology (service workers, notifications, etc.).
Native and hybrid apps are rarely justified.
Another typically excellent talk from Maciej, this time to the Library of Congress. Digital preservation, surveillance, machine learning …it’s all in there, and it makes for grim reading, but there’s also optimism:
My dream for the web is for it to feel like big city. A place where you rub elbows with people who are not like you. Somewhere a little bit scary, a little chaotic, full of everything you can imagine and a lot of things that you can’t. A place where there’s room for chain stores, room for entertainment conglomerates, but also room for people to be themselves, to create their own spaces, and to learn from one another.
I’ve been thinking about this a lot lately. It feels like a user’s browser history is an incredibly rich seam of valuable information just waiting to be presented in a more interesting way.
Fortunately there’s a back-up on the Internet Archive, but this tale of Google’s overnight destruction of fourteen years of writing is truly infuriating.
When we use their services, we trust that companies like Google will preserve some of the most personal things we have to share. They trust that we will not read the fine print.
When you pitch your tent in someone else’s walled garden, they can tear down your home whenever they want.
Here’s an interesting proposal from Google for a user-initiated way of declaring a site’s offline assets should be prioritised (and not wiped out in a clean-up). Also interesting: the way that this idea is being tried out is through a token that you can request …sure beats prefixes!
If you’re going to make a manifest file for an existing site, start with this very handy tool. You give it the URL of your site and it then parses the content for existing metadata to create a best first stab at a manifest JSON file.
I need to wrap my head around the details of this approach, but it sounds like it might be something I could do here on my site (where I feel nervous about my current dependency on a database).