More human than human is our motto

From a research standpoint, GPT2 is groundbreaking in two ways. One is its size, says Dario Amodei, OpenAI’s research director. The models “were 12 times bigger, and the dataset was 15 times bigger and much broader” than the previous state-of-the-art AI model. It was trained on a dataset containing about 10m articles, selected by trawling the social news site Reddit for links with more than three votes. The vast collection of text weighed in at 40 GB, enough to store about 35,000 copies of Moby Dick.

The amount of data GPT2 was trained on directly affected its quality, giving it more knowledge of how to understand written text. It also led to the second breakthrough. GPT2 is far more general purpose than previous text models. By structuring the text that is input, it can perform tasks including translation and summarisation, and pass simple reading comprehension tests, often performing as well or better than other AIs that have been built specifically for those tasks. ~ New AI fake text generator may be too dangerous to release, say creators by Alex Hern

Tip of the hat to Chris Aldrich.

2018-12-02 07.24.48

Parliament seizes cache of Facebook internal papers by Carole Cadwalladr (the Guardian)

Parliament has used its legal powers to seize internal Facebook documents in an extraordinary attempt to hold the US social media giant to account after chief executive Mark Zuckerberg repeatedly refused to answer MPs’ questions.

 

The cache of documents is alleged to contain significant revelations about Facebook decisions on data and privacy controls that led to the Cambridge Analytica scandal. It is claimed they include confidential emails between senior executives and correspondence with Zuckerberg.

Years from now, will we consider this period of time to the enrichment of a few at the expense of society?