Initial commit
This commit is contained in:
commit
89621232bd
6
.gitignore
vendored
Normal file
6
.gitignore
vendored
Normal file
@ -0,0 +1,6 @@
|
|||||||
|
*.pdf
|
||||||
|
*.aux
|
||||||
|
*.fdb_latexmk
|
||||||
|
*.fls
|
||||||
|
*.log
|
||||||
|
*.synctex.gz
|
||||||
6347
IEEEtran.cls
Normal file
6347
IEEEtran.cls
Normal file
File diff suppressed because it is too large
Load Diff
171
sim.tex
Normal file
171
sim.tex
Normal file
@ -0,0 +1,171 @@
|
|||||||
|
\documentclass[conference]{IEEEtran}
|
||||||
|
\usepackage{cite}
|
||||||
|
\usepackage{amsmath,amssymb,amsfonts}
|
||||||
|
\usepackage{algorithmic}
|
||||||
|
\usepackage{graphicx}
|
||||||
|
\usepackage{textcomp}
|
||||||
|
\usepackage{xcolor}
|
||||||
|
\def\BibTeX{{\rm B\kern-.05em{\sc i\kern-.025em b}\kern-.08em
|
||||||
|
T\kern-.1667em\lower.7ex\hbox{E}\kern-.125emX}}
|
||||||
|
\begin{document}
|
||||||
|
|
||||||
|
\title{Similarity Modeling 1/2 Abstracts}
|
||||||
|
|
||||||
|
\author{\IEEEauthorblockN{Tobias Eidelpes}
|
||||||
|
\IEEEauthorblockA{\textit{TU Wien}\\
|
||||||
|
Vienna, Austria \\
|
||||||
|
e1527193@student.tuwien.ac.at}
|
||||||
|
}
|
||||||
|
|
||||||
|
\maketitle
|
||||||
|
|
||||||
|
% \begin{abstract}
|
||||||
|
% \end{abstract}
|
||||||
|
|
||||||
|
% \begin{IEEEkeywords}
|
||||||
|
% component, formatting, style, styling, insert
|
||||||
|
% \end{IEEEkeywords}
|
||||||
|
|
||||||
|
\section{Setting}
|
||||||
|
|
||||||
|
To understand the term \emph{Similarity Modeling} and what it encompasses, it is
|
||||||
|
first important to know how we as humans perceive and understand the things we
|
||||||
|
pick up. An illustrative example for this process is the process of seeing
|
||||||
|
(\emph{detecting}) a face, \emph{recognizing} it and deriving the emotion
|
||||||
|
attached to it. These three steps are placed on a figurative \emph{semantic
|
||||||
|
ladder}, where detecting a face sits on the bottom and recognizing emotion on
|
||||||
|
the top. Face detection thus carries a relatively low semantic meaning, whereas
|
||||||
|
recognizing emotion is a much more sophisticated process. All three steps are
|
||||||
|
only possible to be carried out by humans, because they have internal models for
|
||||||
|
the faces they see, whether they have seen them before and if there is a known
|
||||||
|
emotion attached to how the face looks. These models are acquired from a young
|
||||||
|
age through the process of learning. Visual stimuli and models alone are not
|
||||||
|
enough to be able to conclude that a certain face appears similar or not. The
|
||||||
|
process connecting stimuli and models is comparing the two, also called
|
||||||
|
\emph{looking for similarities}. Together, modeling and looking for
|
||||||
|
similarities, they can be summarized under the term \emph{Similarity Modeling}.
|
||||||
|
|
||||||
|
The goal of Similarity Modeling is usually to find a \emph{class} for the object
|
||||||
|
of interest. The flow of information thus starts with the stimulus, continues on
|
||||||
|
to the modeling part, where we derive a model of the stimulus and—after finding
|
||||||
|
similarities to existing knowledge—ends in a class or label. As mentioned
|
||||||
|
previously, the existing knowledge is fed during the modeling process which
|
||||||
|
describes the feedback loop we call learning. The difficult part lies in
|
||||||
|
properly modeling the input stimulus. It is impossible to store every stimulus
|
||||||
|
verbatim into our existing knowledge base, because it would be too much data if
|
||||||
|
every variety of a stimulus would have to be saved. Therefore, classification
|
||||||
|
systems need the modeling step to \emph{break down} the stimulus into small
|
||||||
|
components which generalize well. The similarity part is generally the same for
|
||||||
|
various domains. Once a proper model of a stimulus exists, checking for
|
||||||
|
similarities in the preexisting knowledge base follows the same patterns,
|
||||||
|
regardless of the type of stimulus.
|
||||||
|
|
||||||
|
Common problems that arise when engineers try to model and classify stimuli come
|
||||||
|
from the fact that there is a wide variety of input signals. This variety is
|
||||||
|
represented by signals which can be local and have large and sudden increases or
|
||||||
|
drops. Others are smooth and the defining characteristic is the absence of
|
||||||
|
sudden variations. Still different signals can have recurring patterns (e.g.
|
||||||
|
EEG) or none at all (e.g. stocks). After detection the most crucial problem
|
||||||
|
remains, which is understanding semantics (also known as the \emph{semantic
|
||||||
|
gap}). The next problem is getting away from the individual samples to be able
|
||||||
|
to construct a model. This is known as the \emph{gravity of the sample}. Another
|
||||||
|
problem is commonly referred to as the \emph{curse of dimensionality}, where we
|
||||||
|
end up with a huge parameter space and have to optimize those parameters to find
|
||||||
|
good models. The last problem is bad data. This can be missing data, misleading
|
||||||
|
data or noisy data.
|
||||||
|
|
||||||
|
\section{Similarity Measurement}
|
||||||
|
|
||||||
|
The artificial process of measuring similarity in computers is shaped by the
|
||||||
|
same rules and fundamentals which are governing similarity measurements in
|
||||||
|
humans. Understanding how similarity measurements work in humans is thus
|
||||||
|
invaluable for any kind of measurement done using computers. A concept which
|
||||||
|
appears in both domains is the \emph{feature space}. An example for a feature
|
||||||
|
space is one where we have two characteristics of humans, gender and age, which
|
||||||
|
we want to explore with regards to their relation to each other. Gender exists
|
||||||
|
on a continuum which goes from male to female. Age, on the other hand, goes from
|
||||||
|
young to old. Because we are only concerned with two characteristics, we have a
|
||||||
|
\mbox{two-dimensional} feature space. Theoretically, a feature space can be
|
||||||
|
$n$-dimensional, where increasing values for $n$ result in increasing
|
||||||
|
complexity. In our brains processing of inputs happens in neurons which receive
|
||||||
|
weighted signals from synapses. The neuron contains a summarization operation
|
||||||
|
and a comparison to a threshold. If the threshold is exceeded, the neuron fires
|
||||||
|
and sends the information to an axon. The weights constitute the dimensions of
|
||||||
|
the feature space. In computers we can populate the feature space with samples
|
||||||
|
and then do either a distance (negative convolution) or a cosine similarity
|
||||||
|
measurement (positive convolution). Since the cosine similarity measurement uses
|
||||||
|
the product of two vectors, it is at its maximum when the two factors are the
|
||||||
|
same. It is much more discriminatory than the distance measurement. Distance
|
||||||
|
measurements are also called \emph{thematic} or \emph{integral}, whereas cosine
|
||||||
|
similarity measurements are called \emph{taxonomic} or \emph{separable}. Due to
|
||||||
|
the latter exhibiting highly taxonomic traits, questions of high semantics such
|
||||||
|
as ``is this person old?'', which require a \emph{true} (1) or \emph{false} (0)
|
||||||
|
answer, fit the discriminatory properties of cosine similarity.
|
||||||
|
|
||||||
|
The relationship between distance and similarity measurements is described by
|
||||||
|
the \emph{Generalization} function. Whenever the distance is zero, the similarity
|
||||||
|
measurement is one. Conversely, similarity is at its lowest when the distance is
|
||||||
|
at its highest. The relationship in-between the extremes is nonlinear and
|
||||||
|
described by the function $g(d)=s=e^{-d}$, which means that only small increases
|
||||||
|
in distance disproportionately affect similarity. Generalization allows us to
|
||||||
|
convert distance measurements to similarity measurements and vice-versa.
|
||||||
|
|
||||||
|
\begin{equation}
|
||||||
|
\label{eq:dpm}
|
||||||
|
\mathrm{dpm} = \alpha\cdot\vec{s} + (1-\alpha)\cdot
|
||||||
|
g(\vec{d})\quad\mathrm{with}\quad\alpha\in[0,1]
|
||||||
|
\end{equation}
|
||||||
|
|
||||||
|
Both, cosine similarity and distance measurements, can be combined to form
|
||||||
|
\emph{Dual Process Models of Similarity} (DPMs). One such example is given in
|
||||||
|
\eqref{eq:dpm} where both measurements are weighted and the distance
|
||||||
|
measurement is expressed as a similarity measure using the generalization
|
||||||
|
function. DPMs model humans' perception particularly well, but are not widely
|
||||||
|
used in the computer science domain.
|
||||||
|
|
||||||
|
\section{Feature Engineering 500 words}
|
||||||
|
|
||||||
|
\section{Classification 500 words}
|
||||||
|
|
||||||
|
\section{Evaluation 200 words}
|
||||||
|
|
||||||
|
\section{Perception and Psychophysics 600 words}
|
||||||
|
|
||||||
|
\section{Spectral Features 600 words}
|
||||||
|
|
||||||
|
\section{Semantic Modeling 200 words}
|
||||||
|
|
||||||
|
\section{Learning over Time 600 words}
|
||||||
|
|
||||||
|
\section*{References}
|
||||||
|
|
||||||
|
Please number citations consecutively within brackets \cite{b1}. The
|
||||||
|
sentence punctuation follows the bracket \cite{b2}. Refer simply to the reference
|
||||||
|
number, as in \cite{b3}---do not use ``Ref. \cite{b3}'' or ``reference \cite{b3}'' except at
|
||||||
|
the beginning of a sentence: ``Reference \cite{b3} was the first $\ldots$''
|
||||||
|
|
||||||
|
Number footnotes separately in superscripts. Place the actual footnote at
|
||||||
|
the bottom of the column in which it was cited. Do not put footnotes in the
|
||||||
|
abstract or reference list. Use letters for table footnotes.
|
||||||
|
|
||||||
|
Unless there are six authors or more give all authors' names; do not use
|
||||||
|
``et al.''. Papers that have not been published, even if they have been
|
||||||
|
submitted for publication, should be cited as ``unpublished'' \cite{b4}. Papers
|
||||||
|
that have been accepted for publication should be cited as ``in press'' \cite{b5}.
|
||||||
|
Capitalize only the first word in a paper title, except for proper nouns and
|
||||||
|
element symbols.
|
||||||
|
|
||||||
|
For papers published in translation journals, please give the English
|
||||||
|
citation first, followed by the original foreign-language citation \cite{b6}.
|
||||||
|
|
||||||
|
\begin{thebibliography}{00}
|
||||||
|
\bibitem{b1} G. Eason, B. Noble, and I. N. Sneddon, ``On certain integrals of Lipschitz-Hankel type involving products of Bessel functions,'' Phil. Trans. Roy. Soc. London, vol. A247, pp. 529--551, April 1955.
|
||||||
|
\bibitem{b2} J. Clerk Maxwell, A Treatise on Electricity and Magnetism, 3rd ed., vol. 2. Oxford: Clarendon, 1892, pp.68--73.
|
||||||
|
\bibitem{b3} I. S. Jacobs and C. P. Bean, ``Fine particles, thin films and exchange anisotropy,'' in Magnetism, vol. III, G. T. Rado and H. Suhl, Eds. New York: Academic, 1963, pp. 271--350.
|
||||||
|
\bibitem{b4} K. Elissa, ``Title of paper if known,'' unpublished.
|
||||||
|
\bibitem{b5} R. Nicole, ``Title of paper with only first word capitalized,'' J. Name Stand. Abbrev., in press.
|
||||||
|
\bibitem{b6} Y. Yorozu, M. Hirano, K. Oka, and Y. Tagawa, ``Electron spectroscopy studies on magneto-optical media and plastic substrate interface,'' IEEE Transl. J. Magn. Japan, vol. 2, pp. 740--741, August 1987 [Digests 9th Annual Conf. Magnetics Japan, p. 301, 1982].
|
||||||
|
\bibitem{b7} M. Young, The Technical Writer's Handbook. Mill Valley, CA: University Science, 1989.
|
||||||
|
\end{thebibliography}
|
||||||
|
|
||||||
|
\end{document}
|
||||||
Loading…
x
Reference in New Issue
Block a user