Bookstore day: Deep Neural Network learns the bookcovers – MIT paper Review



I will disclose today one of my dearest passion: bookstores. I just love bookstores, period. Honestly when I was living in London one of my favorites ways ( I think it was the main way, excessively maybe…) of passing my free time was to go to a Waterstone’s bookstore – I was specially fond of Gower Street’s store right next to University College London, and the routine of going there is one of nostalgic memories of my London. It just is that kind of sensation of pleasure and respect for books, which I think is one of my lucky traits.


This is the way to introduce the motive for today’s post. I will re-post a link from the excellent MIT Review newsletter about scientific and technological subjects that I receive every day in my mailbox. The quality of the scientific paper review section is an inspiration for improvements in the way this blog wants it to be reviewing papers. So I will do more often this kind of post, and I hope that it serves well the purpose of achieving here an approximation to that quality in future posts on reviews by The Information Age. Today’s paper is about the application of a machine vision system in bookstores to judge the content of a book by its cover. To this purpose it is used a deep neural network, and it goes like this:



Deep Neural Network Learns to Judge Books by Their Covers



The idiom “never judge a book by its cover” warns against evaluating something purely by the way it looks. And yet book covers are designed to give readers an idea of the content, to make them want to pick up a book and read it. Good book covers are designed to be judged.

And humans are quite good at it. It’s relatively straightforward to pick out a cookery book or a biography or a travel guide just by looking at the cover.

And that raises an interesting question: can machines judge books by their covers, too? We already know they judge people by their faces


Today we get an answer thanks to the work of Brian Kenji Iwana and Seiichi Uchida at Kyushu University in Japan. These guys have trained a deep neural network to study book covers and determine the category of book they come from.


Their method is straightforward. Iwana and Uchida downloaded 137,788 unique book covers from along with the genre of book. There are 20 possible genres but where a book was listed in more than one category, the researchers used just the first.


Next, the pair used 80 percent of the data set to train a neural network to recognize the genre by looking at the cover image.  Their neural network has four layers, each with up to 512 neurons, which together learn to recognize the correlation between cover design and genre. The pair used a further 10 percent of the dataset to validate the model and then tested the neural network on the final 10 percent to see how well it categorizes covers it has never seen.


The results make for interesting reading. The algorithm listed the correct genre in its top 3 choices over 40 percent of the time and found the exact genre more than 20 percent of the time. That’s significantly better than chance. “This shows that classification of book cover designs is possible, although a very difficult task,” say Iwana and Uchida.





Some categories turn out to be easier to recognize than others. For example, travel books and books about computer and technology are relatively easy for the neural network to spot because book designers consistently use similar images and design for these genres.


The neural net also found that cookbooks were easy to recognize if they used pictures of food but were entirely ambiguous if they used a different design such as a picture of the chef.


Biographies and memoires were also problematic with the algorithm often selecting history as the category. Interestingly, for many of these books, history is the secondary genre listed on Amazon, suggesting that the algorithm wasn’t entirely bamboozled.


The algorithm also confused children’s books with comics and graphic novels as well as medical books and science books. Perhaps that’s also understandable given the similarities between these categories.


There is one shortcoming in this work. Iwana and Uchida have not compared the performance of their neural network against humans’ ability to recognize book genres by their covers. That would be an interesting experiment and one that would be relatively straightforward to do with an online crowdsourcing service such as Amazon’s Mechanical Turk.


Until that work is done, there is no way of knowing whether machines are any better at this task than humans.  Although, no matter how good humans are at this task, it is surely only a matter of time before machines outperform them.


Nevertheless, this is interesting work that could help designers improve their skills when it comes to book covers. A more likely outcome, however, is that it could be used to train machines to design book covers without the need for human input. And that means book cover design is just another job that is set to be consigned to the history books.



Ref: Judging a Book by Its Cover


Fascinating read indeed. And I just wonder what on earth would think of the last two paragraphs some good giants of the book industry history such as Guttenberg or Michael Faraday. Yes the great English scientist and engineer – who was also a great Londoner – started his job life as a bookbinder. But Faraday found reading the books much more pleasurable, and would certainly look at these developments with a nice sense of irony, I would say the least.

Body text image: UCL Images 

Featured Image: From the MIT Review Link provided

Leave a Reply

Fill in your details below or click an icon to log in: Logo

You are commenting using your account. Log Out / Change )

Twitter picture

You are commenting using your Twitter account. Log Out / Change )

Facebook photo

You are commenting using your Facebook account. Log Out / Change )

Google+ photo

You are commenting using your Google+ account. Log Out / Change )

Connecting to %s