Online Public Access Catalogue (OPAC)
Library,Documentation and Information Science Division

“A research journal serves that narrow

borderland which separates the known from the unknown”

-P.C.Mahalanobis


Dimensionality Reduction for Data Visualization and Classification/ (Record no. 436547)

MARC details
000 -LEADER
fixed length control field 04422nam a22002657a 4500
003 - CONTROL NUMBER IDENTIFIER
control field ISI Library, Kolkata
005 - DATE AND TIME OF LATEST TRANSACTION
control field 20241003171500.0
008 - FIXED-LENGTH DATA ELEMENTS--GENERAL INFORMATION
fixed length control field 241001b |||||||| |||| 00| 0 eng d
040 ## - CATALOGING SOURCE
Original cataloging agency ISI Library
Language of cataloging English
082 04 - DEWEY DECIMAL CLASSIFICATION NUMBER
Edition number 23
Classification number 006.31
Item number Su942
100 1# - MAIN ENTRY--PERSONAL NAME
Personal name Das, Suchismita
Relator term author
245 ## - TITLE STATEMENT
Title Dimensionality Reduction for Data Visualization and Classification/
Statement of responsibility, etc Suchismita Das
260 ## - PUBLICATION, DISTRIBUTION, ETC. (IMPRINT)
Place of publication, distribution, etc Kolkata:
Name of publisher, distributor, etc Indian Statistical Institute,
Date of publication, distribution, etc 2023
300 ## - PHYSICAL DESCRIPTION
Extent xxii, 169 pages;
502 ## - DISSERTATION NOTE
Dissertation note Thesis (Ph.D.)- Indian statistical Institute, 2023
504 ## - BIBLIOGRAPHY, ETC. NOTE
Bibliography, etc Includes bibliography
505 0# - FORMATTED CONTENTS NOTE
Formatted contents note Explicit maps for manifold learning -- Nonlinear dimensionality reduction for visualization -- Feature selection preserving class & cluster structures -- Class-specific and Rule-specific feature selection
508 ## - CREATION/PRODUCTION CREDITS NOTE
Creation/production credits note Guided by Prof. Nikhil R. Pal
520 ## - SUMMARY, ETC.
Summary, etc In this thesis, we identify a few gaps in the existing methods of dimensionality reduction for data visualization and classification and propose some solutions to those as summarized below. Most of the data visualization methods do not learn any explicit function to project high dimensional data to a lower dimension. To overcome the difficulty associated with the absence of an explicit map, in Chapter 2, we propose a framework to estimate explicit maps for data visualization in a supervised setting. The quality of output of any regression-type system depends on the quality of the target data. However, even for simple data, sometimes the target data for visualization may be severely distorted. We present a framework that can significantly correct such distortions in the output for data visualization. For any supervised data visualization method the availability of target data is indispensable, which limits the applicability of such methods. Another problem with most of the methods is that they always produce some output given any input, even when the test input is far from the “sampling window” of the training data. In Chapter 3, using a fuzzy rule-based system (FRBS), we propose an unsupervised approach to learn explicit maps for data visualization that addresses the previously mentioned issues. The proposed method can project out-of-sample instances in a straightforward manner. It can also refuse to project an out-of-sample instance when it is far away from the sampling window of the training data. We have demonstrated the generality of the proposed framework using different objective functions for learning the FRBS. When a data set has significant differences between its class and cluster structure, features selected considering only the discrimination between classes would lead to poor clustering performance. Similarly, features selected considering only the preservation of cluster structures would lead to poor classification performance. To address this issue, in Chapter 4, we propose a neural network-based feature selection method that focuses both on class discrimination and structure preservation. For large datasets, to reduce the computational overhead we propose an effective sample-based method. When a data set has class-specific characteristics, selecting a single feature subset for the entire data set may not characterize the data correctly, although the classifier performance may be satisfactory. To address this, in Chapter 5, we have proposed class-specific feature selection (CSFS) schemes using feature modulators embedded in a fuzzy rule-based classifier. The parameters of the modulators are tuned by minimizing a loss function comprising classification error and a regularizer to make the modulators completely select or reject features in a class-specific manner. Our method is free from the hazards of most of the existing CSFS methods, which suffer due to the use of onevs- all strategy. We have extended the CSFS scheme so that it can monitor class-specific redundancy between selected features. We note here that data from a particular class may have multiple clusters and different clusters may be effectively defined by different subsets of features. To address this, finally, our CSFS framework is generalized to a rule-specific feature selection framework.
650 #4 - SUBJECT ADDED ENTRY--TOPICAL TERM
Topical term or geographic name as entry element Artificial Intelligence
650 #4 - SUBJECT ADDED ENTRY--TOPICAL TERM
Topical term or geographic name as entry element Machine Learning
650 #4 - SUBJECT ADDED ENTRY--TOPICAL TERM
Topical term or geographic name as entry element Systems, schools, theories
856 ## - ELECTRONIC LOCATION AND ACCESS
Uniform Resource Identifier <a href="http://dspace.isical.ac.in:8080/jspui/handle/10263/7427">http://dspace.isical.ac.in:8080/jspui/handle/10263/7427</a>
Link text Full Text
942 ## - ADDED ENTRY ELEMENTS (KOHA)
Source of classification or shelving scheme Dewey Decimal Classification
Koha item type THESIS
Holdings
Lost status Not for loan Home library Current library Date acquired Full call number Accession Number Koha item type Public note
    ISI Library, Kolkata ISI Library, Kolkata 01/10/2024 006.31 Su942 TH606 THESIS E-Thesis
Library, Documentation and Information Science Division, Indian Statistical Institute, 203 B T Road, Kolkata 700108, INDIA
Phone no. 91-33-2575 2100, Fax no. 91-33-2578 1412, ksatpathy@isical.ac.in