Changeset - 7babd96c03f9
[Not reviewed]
0 1 0
Gebrekirstos Gebremeskel - 11 years ago 2014-06-12 05:21:03
destinycome@gmail.com
updated
1 file changed with 4 insertions and 2 deletions:
0 comments (0 inline, 0 general)
mypaper-final.tex
Show inline comments
 
@@ -681,100 +681,102 @@ name-variant partial over name-variant is 8.3\%.
 
%saturation: documents have already been extracted by the different
 
%name variants and thus using their partial names do not bring in many
 
%new relevant documents.
 
For Wikipedia entities, canonical
 
partial achieves better recall than name-variant in both the cleansed and
 
the raw corpus.  %In the raw extraction, the difference is about 3.7.
 
In Twitter entities, recall of canonical matching is very low.%
 
\footnote{Canonical
 
and canonical partial are the same for Twitter entities because they
 
are one word strings. For example in https://twitter.com/roryscovel,
 
``roryscovel`` is the canonical name and its partial is identical.}
 
%The low recall is because the canonical names of Twitter entities are
 
%not really names; they are usually arbitrarily created user names. It
 
%shows that  documents  refer to them by their display names, rarely
 
%by their user name, which is reflected in the name-variant recall
 
%(67.9\%). The use of name-variant partial increases the recall to
 
%88.2\%.
 
 
 
 
The tables in \ref{tab:name} and \ref{tab:source-delta} show a higher recall
 
for Wikipedia than for Twitter entities. Generally, at both
 
aggregate and document category levels, we observe that recall
 
increases as we move from canonicals to canonical partial, to
 
name-variant, and to name-variant partial. The only case where this
 
does not hold is in the transition from Wikipedia's canonical partial
 
to name-variant. At the aggregate level (as can be inferred from Table
 
\ref{tab:name}), the difference in performance between  canonical  and
 
name-variant partial is 31.9\% on all entities, 20.7\% on Wikipedia
 
entities, and 79.5\% on Twitter entities. 
 
 
Section \ref{sec:analysis} discusses the most plausible explanations for these findings.
 
%% TODO: PERHAPS SUMMARY OF DISCUSSION HERE
 
 
\section{Impact on classification}\label{sec:impact}
 
In the overall experimental setup, classification, ranking, and
 
evaluation are kept constant. Following \cite{balog2013multi}
 
settings, we use
 
WEKA's\footnote{\url{http://www.cs.waikato.ac.nz/~ml/weka/}} Classification
 
Random Forest. However, we use fewer numbers of features which we
 
found to be more effective. We determined the effectiveness of the
 
features by running the classification algorithm using the fewer
 
features we implemented and their features. Our feature
 
implementations achieved better results.  The total numbers of
 
features we used are 13 and are listed below.
 
  
 
\paragraph*{Google's Cross Lingual Dictionary (GCLD)}
 
 
This is a mapping of strings to Wikipedia concepts and vice versa
 
\cite{spitkovsky2012cross}. 
 
The GCLD corpus estimates two probabilities:
 
(1) the probability with which a string is used as anchor text to
 
a Wikipedia entity 
 
%thus distributing the probability mass over the different entities that it is used as anchor text;
 
and (2) the 
 
probability that indicates the strength of co-reference of an anchor with respect to other anchors to  a given Wikipedia entity.  We use the product of both for each string.
 
 
\paragraph*{jac} 
 
  Jaccard similarity between the document and the entity's Wikipedia page
 
\paragraph*{cos} 
 
  Cosine similarity between the document and the entity's Wikipedia page
 
\paragraph*{kl} 
 
  KL-divergence between the document and the entity's Wikipedia page
 
  
 
  \paragraph*{PPR}
 
For each entity, we computed a PPR score from
 
a Wikipedia snapshot  and we kept the top 100  entities along
 
with the corresponding scores.
 
 
 
\paragraph*{Surface Form (sForm)}
 
For each Wikipedia entity, we gathered DBpedia name variants. These
 
are redirects, labels and names.
 
 
 
\paragraph*{Context (contxL, contxR)}
 
From the WikiLink corpus \cite{singh12:wiki-links}, we collected
 
all left and right contexts (2 sentences left and 2 sentences
 
right) and generated n-grams between uni-grams and quadro-grams
 
for each left and right context. 
 
Finally,  we select  the 5 most frequent n-grams for each context.
 
 
\paragraph*{FirstPos}
 
  Term position of the first occurrence of the target entity in the document 
 
  body 
 
\paragraph*{LastPos }
 
  Term position of the last occurrence of the target entity in the document body
 
 
\paragraph*{LengthBody} Term count of document body
 
\paragraph*{LengthAnchor} Term count  of document anchor
 
  
 
\paragraph*{FirstPosNorm} 
 
  Term position of the first occurrence of the target entity in the document 
 
  body normalised by the document length 
 
\paragraph*{MentionsBody }
 
  No. of occurrences of the target entity in the  document body
 
 
 
 
  
 
  Features we use incude similarity features such as cosine and jaccard, document-entity features such as docuemnt mentions entity in title, in body, frequency  of mention, etc., and related entity features such as page rank scores. In total we sue  The features consist of similarity measures between the KB entiities profile text, document-entity features such as  
 
  In here, we present results showing how  the choices in corpus, entity types, and entity profiles impact these latest stages of the pipeline.  In tables \ref{tab:class-vital} and \ref{tab:class-vital-relevant}, we show the performances in max-F. 
 
\begin{table*}
 
\caption{vital performance under different name variants(upper part from cleansed, lower part from raw)}
0 comments (0 inline, 0 general)