ImageNet: Analysis

Haneen Fathy
2 min readSep 14, 2020

--

The most surprising thing about the ImageNet dataset was probably the sheer volume of it. I was also very curious about the choice of words used to describe each dataset, particularly since the images are human-annotated. There are definitely ethical considerations when it comes to the fact that ImageNet does not own any of the images. Even though those images are free to view on the web anyways, the owners of the images have the right to know in what ways are their images being used and if they’re being used for machine learning algorithms.

Image Classification:

Plant → Shower Curtain

Glass → Water Jug

Plate → Cowboy Hat

Napkin → Handkerchief

Photo Frame → Desktop Computer

Highlighter → Lighter

Playing around with MobileNet felt frustrating as it did a poor job of recognizing any of the objects. I had to fix the angles and orientations several times before getting an answer similar to the object. As I could tell MobileNet is very dependent on the circumstances surrounding the object in the photo. I tried my best to clear everything else and it made a difference. The more the object took space from the background the more likely it was to be recognized. Another interesting observation I had was that there is a very narrow scope of pictures from other cultures. I added a picture of the Taj Mahal and it wasn’t able to recognize it even though it’s a quite infamous building. I then added a picture of the pyramids and it said it is a wreck. I reached the conclusion that MobileNet probably only contains generic images.

Sign up to discover human stories that deepen your understanding of the world.

Free

Distraction-free reading. No ads.

Organize your knowledge with lists and highlights.

Tell your story. Find your audience.

Membership

Read member-only stories

Support writers you read most

Earn money for your writing

Listen to audio narrations

Read offline with the Medium app

--

--

No responses yet

Write a response