Face recogintion, machine learning

As part of my third year at Camberwell I'll be doing research into machine learning, computer vision etc. These are my working notes.

January 2, 2018

I managed to work my way through Hands-On Machine Learning with Scikit-Learn & Tensorflow by Aurélien Géron. Some of the more advanced math is still beyond me (remembering how vectors work was hard enough), but I feel like I've now got an actual understanding of some of the acronymns that get thrown around a lot: Deep Learing, Neural Networks, MLP, TensorFlow and so on.

An important point that's made early in the book is that machine learning isn't the same thing as neural networks. Géron quotes Tom Mitchell (1997):

A computer program is said to learn from experience E with respect to some task T and some performance measure P, if its performance on T, as measured by P, improves with experience E.

Basic methods like linear regression fall under this definition as well as neural networks.

Chapter 14, which deals with Recurrent Neural Networks is particularly exciting. As Géron points out,

RNNs [are] a class of nets that can predict the future. [...] RNN's ability to anticipate also makes them capable of surprising creativity.

Evidently this is the technology behind some of these Google Magenta Experiments. A later chapter in the book describes how you can train a neural network in such a way that given a set of source images, it can generate new images that look as real as the input images - exciting stuff. I'm hoping to do this with images of faces - generating portraits of people that don't exist.

However, I do suspect that the laptop I'm typing this on will have nearly enough processing power to do all of that. Finding enough source images will also be a concern. This describes the main problem with advanced machine learning: While the math is well established and relatively accessible, access to the vast amounts of processing power and training data required to build useful software is limited to large organisations.

January 13, 2018

I got my hands on something called the FERET Database. This is a collection of images of faces that the U.S military comissioned in the mid-nineties, containing about 11.000 images of roughly 800 individuals from different angles, wearing different clothes etc. It's what much of modern research into facial recognition algorithms has been based on. Here's the relevant government website.

The way you get this database is emailing the US department of defence. Once you do, they give you login details to download the database. It comes in a weird 90s format, so I had to spend some time extracting and converting the images so I could look at them.

FERET Images Example images from the FERET database

I'm not sure what I'm going to do with these images. I could use them to train a neural network, but they're also an interesting artifact in themselves. They're essentially a time capsule from the campus of George Mason University in the 1990s - 90s haircuts etc. I also like the idea that these are images only ever intended for machines to look at. Also the fact that these are basically scientific documents created for a government agency, yet some of them are surprisingly artisitic.

Evidence Evidence (1977) by Larry Sultan and Mike Mandel. Image Source

It reminds me of Evidence (1977) by Larry Sultan an Mike Mandel, where they took NASA research photographs, took them out of their original context and put them in a new order that tells a story.

January 15, 2018

Turns out Trevor Paglen did some work on the FERET images very recently. The exhibition also includes machine-generated images and some original photography - all very succesful. I'll try and get an exhibition catalogue.

Paglen has been doing this work for a while. Other projects of his include Invisible : covert operations and classified landscapes, a book on restricted government sites. Also Blank Spots on the Map, which is about how governments manipulate maps to hide what they're doing.

January 16, 2018

Readings

Tracey suggests I go see an exhibition called Metadata - How we relate to images at CSM - I've scheduled for Saturday.

I've spent some more time with the FERET database, going through the images, printing some of them and reading some of the related government reports:

The 1996 paper points out:

Some questions were rasied about the age, racial, and sexual distribution of the database. However, at this stage of the program, the key issue was algorithm performance on a database of a large number of individuals.

This might be an area worth exploring. The photos were collected by GMU, suggesting that most of the volunteers are probably students and university staff (not military emplyees as is sometimes suggested). In some sense the whole history of institutional recism and sexism might be baked into this database?

Might be good to run some analytics on gender / age / race distribution of the databse.

I'm still interested in how exactly these photography sessions were conducted - how did they recruit volunteers, whose office was turned into a studio, what did people at the time say about the program etc.

January 17, 2018

Readings

Segune suggests two additional readings on photographic archives (after seeing the FERET images):

Installation view of 48 Portraits by Gerhardt Richter Tate Modern

She also points out 48 Portraits (1971-98) by Gerhardt Richter.

Notes on "Invisible Images (Your Pictures are Looking at You)"

On a basic level, Paglen argues that existing models of visual culture are becoming less relevant because the vast majority of images are now created by machines for other machines. This has to do with the fact that a digital image is primarily machine-readable. You can only make it visible to human eyes for a brief moment using additional software, screens etc.

The second main point is that images are no longer primarily used as representations. Instead, machines use images to make predictions, activate mechanisms and generally actively change the real world. In his words:

Images have begun to intervene in everyday life, their functions changing from representation and mediation, to activations, operations, and enforcement. Invisible images are actively watching us, poking and prodding, guiding our movements, inflicting pain and inducing pleasure. But all of this is hard to see.

Paglen cites a number of examples of this that have been in operation for years. These included cases where license plates are recognised and used to track people's movements and retail companies that analyse customers' facial expressions

He makes the point that places like Facebook are closely modelled on traditional notions of sharing images (using skeumorphic terms like albums, slideshows) but this is only true on the surface. Underneath, your photos are feeding highly developed machine learning algorithms designed to extract value from your images (now or in the future). As Paglen points out, you could easily imagine the license plate recognition case being expanded to include images people share on social media.

He closes by saying that the long-term solution to this needs to be regulation - "hacks" that might be effective against recognition algorithms today will loose their effectiveness over time.

We no longer look at images - images look at us. They no longer simply represent things, but actively intervene in everyday life. We must begin to understand these changes if we are to challenge the exceptional forms of power flowing through the invisible visual culture that we find ourselves emeshed within.

January 20, 2018

Notes on Segune's Readings

(She suggested these a few days ago)

Archive Fever: Photography between History and the Monument

This cites an essay called The Body and the Archive (1986) by Allan Sekula, which talks about how photographic archives have been used as "an instrument of social control an differentiation underwritten by dubious scientific principles".

Bertillon Archive The Metropolitan Museum of Art

Sekula talks about Alphonse Bertillon, a French policeman who created a huge bullshit system to classify criminals based on their photographs of their faces. The Met seems to have a good collection of his stuff. The Science Museum has some of the instruments he used to measure various facial features.

Similar archival projects to classify people along racial lines (The nazis were big fans).

Their projects, Sekula writes, "constitute two methoological poles of the positivist attempts to define and regulate social deviance" The criminal (for Bertillon) and the racially inferior (for Galton) exist in the netherworld of the photographic archive, and when they do assume a prominent place in that archive, it is only to dissociate them, to insist on and illuminate their difference, their archival apartness from normal society

Enwezor goes on to describe a number of examples where archives are used as a way to conserve power, present existing systems of oppression as natural etc.

An Archival Impulse

January 24, 2018

MetaData at the Lethaby Gallery

## January 25, 2018

TODO Spoke to segune about feret images

January 26, 2018

TODO Jak tutorial, discussed ways of presenting face images

January 27, 2018

TODO decided to print feret images, looks like its expenive, need to talk to techinician, emailed tracey

January 29, 2018

TODO Peer assesment

Febuary 14, 2018

Eigenfaces are a way to represent images used in facial recognition software. First introduced by Turk and Pentland (1991). Below is figure 2 from that paper:

Eigenfaces Turk, Pentland (1991)

Something intruiging about the aesthetics of research papers.

More Eigenfaces OpenCV

Febuary 18, 2018

Another Face Database

The National Institue for Standards and Technology (which provides the FERET Database) also has something called the Multiple Encounter Dataset (MED). This is a database containing 683 mugshots of deceased people used to develop facial recognition software. This is starting to get much closer to Berillion. I'm assuming by using photographs of dead people allows them to get around some privacy concerns. They've also removed (in some cases blacked out) any reference to the person's name or reason of arrest. So what you're left with is this archive of black and white photographs of people from the 60s, 70s and 80s (judging by the haircuts).

Mugshots National Institute of Standards and Technology

With the images comes a datafile describing the photographs:

Subject Encounter Record DOB WGT SEX HGT RAC HAI EYE PHD IMT POS VLL HLL
...

Interestingly this contains fields for height (ie. 5'11) weight (in lbs.) and date of birth of the detainee.

Febuary 27, 2018

Eigenfaces

Some more face databases. I'm thinking the reason these are all from the 90s is that research doesn't need this sort of standardised database anymore - People are now working with images collected from the internet. Labelled Faces in the Wild is an example. This has the benefit of being much cheaper than taking original photographs - you can create a database that is orders of magnitudes larger for the same amount of money. Examples:

Facebook research uses internal databases with millions of faces. Maybe there's something to this idea: Back in the day, collecting a database had to be a dedicated effort. Now, we're all contributing to face recognition algorithms (and other machine learning applications by way of our behaviour, movements, writing) involuntarily.

AT&T Laboratory Database of Faces AT&T Laboratories Cambridge

[University of Surrey](http://www.ee.surrey.ac.uk/CVSSP/xm2vtsdb/)

March 6, 2018: RNNS

This might be a fun project to get into generating things with neural networks: The New York Times has an API that makes it really easy to get their content programatically. I pulled every article headline from January 2016 to present - about 4MB of text. This Tensorflow setup makes it trivial to train a character-based RNN on the data, and eventually generate new headlines that (somewhat) match the language of the New York Times. It's pretty amazing to see the network learn English from scratch in a few hours of training.

The Dutch Polders by Bike and Schooner The Royals Take the Title ‘The Affair’ Season 2 Episode 5: Never Read the Book ‘The Walking Dead’ Season 6, Episode 4 Recap: The Making of Morgan ‘Homeland’ Recap, Season 5, Episode 5: Can Carrie Figure Out What’s Going On With Allison? Long Lines for Story Time The Best Moments in College Football This Week Dangers for the Unwary Q. and A.: Chan Koonchung on Imagining a Non-Communist China Report on Bella Vista Health Center Inside the Trial of Sheldon Silver Jeb Bush Says He Was Unaware of Rubio PowerPoint Deck

This sort of automated writing is already widely used at mainstram outlets. The Washington Post seems to be leading the pack.

April 16 Tutorial Notes

Newspaper Clippings

Fake Letterpress Newspaper Clippings

Large Scale Drawing Machine

Continues to be a health and safety nightmare.

Machine Learning Dataset Book

ML Book Spread ML Book Spread