Found via Kottke, here’s an excellent timeline of the idea of remote work and digital nomadism. The idea has been around for a long time – the timeline starts in 1964, with Arthur C. Clarke predicting it (well before it was broadly feasible – it’s sort of remarkable how much some of those 50’s and 60’s futurists managed to nail it). It’s been a long time coming, and while it’s not for everyone, the pandemic certainly gave many more people the chance to try it out. (It’s an imperfect trial, since being in quarantine and many places in lockdown isn’t necessarily indicative of what it would be like if you didn’t have that restriction and background stress.)
It’s unsurprising but sad that many companies are already insisting people come back to the office, despite it: a) arguably being too early given vaccination rates, new infections, and variants; b) not being necessary, based on general productivity gains and losses compared to in-office; c) not being what their employees want, many of whom seem to prefer either remote or a hybrid of in-office and remote. (Personally, I’m quite happy working remotely 90% of the time, but recognize that it’s useful to get some real face time, too. Anecdotally, I seem to do best when I’m off remote most of the time, then go into the office maybe once or twice a week. I’d be interested in trying out something like being primarily remote and then coming to work from the office for a week or two maybe once a quarter or a few times a year.)
Anyway, definitely some food for thought, and interesting to see the sort of evolution and adoption of digital nomad lifestyles across the past few decades.
Over at Garbage Day, Ryan Broderick discusses A Unified Theory of Online Anger, noting how algorithmic social media has effectively been weaponized (notably by the right, but let’s be honest, not just by them). They’re not wrong.
Continue reading “The Outrage Machine”
As these trending main characters go viral on Twitter, hundreds of online outlets race to turn this into content. And there’s a real financial incentive for covering these stories. As most people working at various content mines can tell you, the thing Facebook readers love the most is getting mad about stuff that’s happening on Twitter.Ryan Broderick
If you are like me and love the National Parks, but hate the crowds, you might appreciate A Night Under the Stars. It’s a look at when people are spending the night the most in each national park. It’s pretty interesting to see when each park is most popular (and by extension, which periods should be low population but still nice to visit). Found via the inimitable Kottke.
Over at the Atlantic, Daniel Markovits discusses how Meritocracy harms everyone. While many of the examples he cites are centered around the wealthy elite, that’s kind of the point – it’s already clear to people not in that caste that the notion of meritocracy, while good in concept, is a sham in practice. As soon as the criteria for “merit” become understood, it is gamed by those who are in a position to do so, further entrenching the already wealthy (but with a slightly more palatable veneer than the up-front nepotism of previous systems).
Hardworking outsiders no longer enjoy genuine opportunity. According to one study, only one out of every 100 children born into the poorest fifth of households, and fewer than one out of every 50 children born into the middle fifth, will join the top 5 percent. Absolute economic mobility is also declining—the odds that a middle-class child will outearn his parents have fallen by more than half since mid-century—and the drop is greater among the middle class than among the poor. Meritocracy frames this exclusion as a failure to measure up, adding a moral insult to economic injury.Daniel Markovits
I agree that it’s a problem, and I don’t really know a good solution. It sounds like the author has some ideas (some of which sound a little… optimistic to me). It’s clear that things are coming to a head, though, and something is going to have to change.
If you’re just not a morning person, science says you may never be, over at Vox. This is from a few years ago, but I only came across it relatively recently: there’s increasing evidence that whether you’re a night person or a morning person has a genetic component, and fighting it can have significant impacts on your health. A salient bit:
Even people who are slightly more oriented to the evening — people who would like to sleep between 1 am and 9 am, say — may be faced with a difficult choice: Listen to your body, or force it to match the sleep habits of most everyone else?
Research has been gaining insight on that question. It turns out our internal clocks are influenced by genes and are incredibly difficult to change. If you’re just not a morning person, it’s likely you’ll never be, at least until the effects of aging kick in.
And what’s more, if we try to live out of sync with these clocks, our health likely suffers. The mismatch between internal time and real-world time has been linked to heart disease, obesity, and depression.Brian Resnick, Vox
For the record, I’ve always been a night owl. When left to my own devices (extended periods where I was setting my own schedule), I tend to go to bed around 2am and then get up around 10am. That’s still just 8 hours, but because it’s offset from the schedule most of society runs at, it still comes across as oversleeping. (These days, a mixture of work and the dog keep me getting up earlier… but I also find myself feeling like I need a nap more often, as well.)
It reminds me of something from The Devil’s Dictionary:
DAWN, n. The time when men of reason go to bed. Certain old men prefer to rise at about that time, taking a cold bath and a long walk with an empty stomach, and otherwise mortifying the flesh. They then point with pride to these practices as the cause of their sturdy health and ripe years; the truth being that they are hearty and old, not because of their habits, but in spite of them. The reason we find only robust persons doing this thing is that it has killed all the others who have tried it.Ambrose Bierce, The Devil’s Dictionary
From the Guardian: Snapping point: how the world’s leading architects fell under the Instagram spell. We’ve always had architecture-as-spectacle – if anything, Instagram is just the latest in a series of driving forces. But it’s still worth thinking about. Something the author notes:
Configuring buildings and public spaces as selfie sets may well work for tourism promotion and the buzz of a launch, but once the novelty factor has worn off, the whimsy can grate and the flimsiness become all too apparent. The urge for quick, affordable spectacle often leads to stick-on, paper-thin cladding materials that look good in photographs, but weather terribly. The stained, peeling facades of the last decade stand as a grim testament to prioritising photographability over function.Oliver Wainwright
Art is wonderful, and public art is essential. It enhances and enriches the spaces we inhabit. But when we insert an overtly capitalist motivation of cashing in on a craze, we sacrifice the care and consideration in how we craft this art.
Something this also gets me thinking about is replication. That is, our penchant for seeing an interesting photo, and striving to replicate it. This isn’t new, and in many cases is designed (consider things like Yosemite, where you round a corner and get a reveal of El Capitan or Half Dome. You stop, and say “Wow,” and take a picture. This experience was designed, over a century ago, and is part of why we have 400 million nearly identical shots of these locations). It just bums me out, for some vague reasoning I can’t quite put a finger on, that rather than being inspired by an interesting photo, there is this impetus to simply replicate it. I should probably chew on that a bit.
Over at the Guardian: From lizarding to lingering: how we really behave in public spaces. It’s pretty fascinating work done by the SWA Group. As noted in their summary:
This research project revisited the primary city of writer William H. Whyte’s Street Life Project and seminal study The Social Life of Small Urban Spaces (1980). It sought to understand how the types of new public spaces have changed some 40 years after he published his book and companion film, what has changed in how people use public realm spaces, and what makes well used spaces.SWA Group
I think it’s interesting to juxtapose their findings with the expectations created by architects, urban planners, and similar when making mockups of their planned spaces. (Expectations versus reality.) Also interesting to think about how this could be applied not just for building more effective spaces, but for presenting more realistic crowd scenes in film or other types of performance.
My heart goes out to the journalists at the multiple organizations laid off this week (and more). Something like a thousand laid off in the space of a week. Fast Company has a solid (and scathing) article about the recent Buzzfeed layoffs: BuzzFeed’s layoffs and the false promise of “unions aren’t for us”. It paints a pretty bleak picture of where things are at, why, and what we can expect more of in the future.
But as an outlet largely dependent on social platforms like Facebook, BuzzFeed was forced to follow platform trends. When Facebook announced it was focusing on video content, BuzzFeed turned its resources just to that. Brands like Tasty were born, which force-fed ubiquitous birds’-eye view videos of generally unappetizing food to the masses. And for a while, this seemed to work. Videos were performing well, thanks to Facebook’s algorithmic push, and BuzzFeed once again looked like a digital trailblazer. But this bet was predicated on the whim of a social network known for pendulum strategy shifts at the expense of its clients; this pivot didn’t take into account what would happen if Facebook changed course. It shouldn’t come as a shock that Facebook did precisely that.
Over at Lost Garden, Daniel Cook has a fantastic piece looking at how to create “human-scale” online games, and why that’s a better approach to MMOs. This is some really fantastic, well thought out stuff, and not just for games: what they’re really talking about is how to build community.
One way of thinking about the constraints suggested by Dunbar’s Layers is to imagine you have a budget of cognitive resources that can be spent on relationships. The physical limits of your human brain mean that you only have enough mental budget for a total of roughly 150 relationships.Daniel Cook
Humans have developed a few tools that have expanded our ability to organize into groups well past our primate cousins—most notably language—but also large-scale systems of government and economics. In the early 2000s, people assumed that new technologies like online social networks could help break past Dunbar’s Number; by offloading the cost of remembering our friendships to a computer, we could live richer, more social lives, with strong relationships to even more people.
We now have copious data that this is not the case. Studies suggest that there’s still a limited budget of cognitive resources at play and even in online platforms we see the exact same distribution of relationships.
If anything, social networks damage our relationships. By making it possible for us to cheaply form superficial relationships (and invest our limited energy in maintaining them), such systems divert cognitive resources from smaller, intimate groups out towards larger, less-intimate groups. The result is that key relationships with best friends and loved ones suffer. And, unfortunately, it is the strength of these high-trust relationships that are most predictive of mental health and overall happiness.
It’s a long read (which you might have guessed by the size of my pull quote), but well worth it.
Over at The Hedgehog Review, Charlie Tyson writes about the Virtuosos of Idleness, and the nature of leisure (and its loss in modern society). It’s an interesting read. There are a number of articles about the coming work-pocalypse of increased automation and the massive inequalities introduced by the “gig economy,” but it’s also worthwhile to look at how we spend our off-time. One choice bit that struck a chord:
Most Americans today find work drudgery and leisure anxiously vacant. In our hours off work, we rarely achieve thrilling adventure, deliberate self-education, or engage in Whitmanian loafing. At the same time, faith is eroding in the idea that paid work can offer pleasure, self-discovery, a means for improving the world, or anything more than material subsistence.Charlie Tyson
I mean, they’re not wrong. I’m lucky enough to have a decent job with some flexibility to learn and grow, but jobs like that are decidedly not the majority of jobs out there. And while the work side might not be terrible at the moment, the “vacant leisure” is real. The author continues:
Recreational pursuits more demanding than fleeting digital absorption are, increasingly, acts of consumption. Leisure is not something you “do” but something you “buy,” whether in the form of hotels and cruises or Arianna Huffington–vetted mindfulness materials. The leisure industry provides work for some while promising relaxation to others, for a fee.
The sorry state of leisure is partly a consequence of an economy in which we are never fully detached from the demands of work. The category of “free” time is not only defined by its opposite (time “free” of work); it is subordinated to it. Free time, Theodor Adorno warns, “is nothing more than a shadowy continuation of labor.” Free time is mere recovery time. Spells of lethargy between periods of labor do little but prepare us for the resumption of work. Workers depleted by their jobs and in need of recuperation turn to escapist entertainment and vacuous hobbies. And the problem of figuring out when work is “over,” in an economy in which knowledge workers spend their job hours tweeting and their evening hours doing unpaid housework and child care, has never seemed more perplexing.Charlie Tyson
Yep. The conversation continues from there, and is worth the time to read.