Three Algorithm Reads
Three Algorithm Reads

“How algorithms rule our working lives,” by Cathy O’Neil, in the Guardian:
These algorithmic “solutions” are targeted at genuine problems. School principals cannot be relied upon to consistently flag problematic teachers, because those teachers are also often their friends. And judges are only human, and being human they have prejudices that prevent them from being entirely fair — their rulings have been shown to be harsher right before lunch, when they’re hungry, for example — so it’s a worthy goal to increase consistency, especially if you can rest assured that the newer system is also scientifically sound.
The difficulty is that last part. Few of the algorithms and scoring systems have been vetted with scientific rigour, and there are good reasons to suspect they wouldn’t pass such tests. For instance, automated teacher assessments can vary widely from year to year, putting their accuracy in question. Tim Clifford, a New York City middle school English teacher of 26 years, got a 6 out of 100 in one year and a 96 the next, without changing his teaching style. Of course, if the scores didn’t matter, that would be one thing, but sometimes the consequences are dire, leading to teachers being fired.
“Why Facebook should kill Trending News once and for all,” by Gillian Branstetter, the Daily Dot
Even if Facebook wanted out of managing a team of editors, it should certainly put less faith in whatever secretive algorithm is managing the list now. It’s a common saying in the study of artificial intelligence that it’s easier to design a software that can run a restaurant than it is a robot that can bus a table. Jobs that require relatively little mental processing can be difficult for a machine lacking the coordination and physical memory of a human. Although editing is not quite a physical task, it still requires a level of complex understanding and contextual observance that machines have often not provided.