Publication: Analysis of Human Interpretability in Document Classification
DOI
Type:
Thesis
Date
2018
Authors
Journal Title
Journal ISSN
Volume Title
Publisher
Abstract
With high use of computers, the collection of textual data generated,
exchanged, stored and accessed increased in massive amount and became one of
the richest sources of data for the organization. As a result, people are tending to
use natural language processing application to categorize this large volume of data
efficient and accurate manner. Their application of machine learning models.
When it comes to Natural Language processing (NLP) applications where most
of them follows supervised learning techniques, automatic document
classification models developed to do content-based assignment here the
materials are assigned into predefined categories. This makes it easier to find the
relevant information at the right time and for filtering and routing documents
directly to correct users.
Mostly these learning models are operating in black-box manner wh re there is
no way to interpret how the model has decided which class an instarce should
assigned. understanding the reason behind how learning makes these redictions
~7
are very important to trust such learning models in real application. [his thesis
presents the work related to the experimental work been carried with set of text
classifiers to interpret text classifiers predictions, so any classifier can be
evaluated based on how well they support classification purpose.
