We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF or read online on Scribd
You are on page 1/ 20
SEMANTIC PARSING
What is Semantic parsing?
The process of understanding the meaning and interpretation of words, signs and
sentence structure is called semantic parsing.
+ Using semantic parsing, the computers can understand Natural Language the
way humans do
+ Itis the toughest phase and not fully solved.
+ semantic -----> study of meaning
+ parsing --------> identify and relate pieces of information.Requirements for achieving a semantic representation:
1. Structural Ambiguity
2. Word sense
3. Entity and Event resolution
4. Predicate-Argument Structure
5. Meaning RepresentationInterpretation
+ Semantic interpretation is a large process as it involves various components (representation of text
that can be fed into a computer to allow other computational manipulations in order to understand
any language).
+ Katz and Fodor in their 1963 paper “The Structure of a Semantic Theory” put forward what
properties a semantic theory should posses,
Asemantic theory should be able to :
1. Explain sentences having ambiguous meanings. For example, it should account for the fact that the
‘word bilfin the sentence The bil is Jargeis ambiguous in the sense that it could represent money or
the beak of the bird.
2. Resolve the ambiguities of words in context. For example, ifthe same sentence is extended to form
The bill is large but need to be paid, then the theory should be able to disambiguate the monetary
meaning of the bill.
3. Identify meaningless but but syntactically well-formed sentences, such as the famous example by
Chomsky: Colorless green ideas sleep furiously.
4, Identify syntactically or transformationally unrelated paraphrases of a concept having the same
semantic content1, Structural Ambiguity
2. Word sense
3. Entity and Event resolution
4, Predicate-Argument Structure
5, Meaning Representation
STRUCTURAL AMBIGUITY
Structural ambiguity arises when a sentence has more than one meaning due to the way words are
arranged in that sentence,
For example,
The sentence S70 caught the butterfly by the trees structurally ambiguous because it has 2 meanings:
1, Sara caught the butterfly while she was standing by the tree.
2, Sara caught the butterfly which was fluttering near the tree,WORD SENSE
The same word type, or word lemma, is used in different morphological variants to represent different
entities or concepts in the world,
For example,
Take a word nail (human anatomy & metallic object)
1. He nailed the loose arm of the chair.
2. He got a box of metallic nails.
3. This nails are growing too fast.
4. He went to manicure to remove his nails.EVENT AND ENTITY RESOLUTION
Entity: The process of identifying entities Le people, organization, location and more.
Example:
Steve Jobs was the co-founder of Apple, which is headquartered in Cupertino.
Event: The actions described with associated entities.
Example:
Elon Musk announced that Tesla will build a new factory in Texas,
+ Identifying and linking references to the same entity across the text.(MEANING REPRESENTATION.
‘The final process of the semantic interpretation is to build a semantic representation or meaning
representation that can be manipulated by algorithms.
‘This process is sometimes called as deep representation.
For example,
1. Ifour player 2 has the ball, then position our player 5 in the midfield.
(oowleriplayer 2))(doiplayer 5) (posimistfield))
2, Which river isthe longest?
answer (x1, longest (xt rivertx1))SYSTEM PARADIGMS
+ Researches have examined meaning representations and methods to recover from different levels of
granularity (the level of detail) and generality (how broad or general the information is).
+ For many of the potential experimental conditions, no hand-annotated data is available.
+ Therefore, it is important to get a perspective on the various primary dimensions on which the problem of
semantic interpretation has been tackled.
The historic approaches which are more prevalent and successful generally fall into 3 categories:
1, SYSTEM ARCHITECTURES
(a) Knowledge based: As the name suggests, these systems use a predefined set of rules or knowledge base
to obtain a solution to a new problem.
(b) Unsupervised: These systems tend to require minimal human intervention to be functional by using
‘existing resources that can be bootstrapped for a particular application or domain.(c) Supervised: These systems require some manual annotation. Typically, researches create feature
functions, A model is trained to use these features to predict labels, and then itis applied to unseen data,
(d) Semi-supervised: Manual annotation is very expensive and does not yield enough data. In such instances,
researches can automatically expand the dataset on which their models are trained either by employing
machine-generated output directly or by bootstrapping off of an existing model by having humans correct its
output. In many cases, a model from one domain is used to quickly adapt to a new domain.2. SCOPE
(a) Domain Dependent: These systems are specific to certain domains.
(b) Domain Independent: These systems are general enough that the techniques can be applicable to
multiple domains.
A covernce
{0) Shallow: These systems tend to produce an intermediate representation that can then be converted to
one that machine can base its actions on.
(b) Deep: These systems usually create a terminal representation that is directly consumed by a machine or
application.Word Sense System
Researchers have explored various system architectures to address the sense disambiguation problem.
We can classify these systems into four main categories:
1. Rule based or Knowledge based
2. Supervised
3. Unsupervised
4, Semi-supervised1, Rule-based,
-Rule-based systems for word sense disambiguation are among the earliest methods developed to tackle
the
problem of determining the correct meaning of a word based on its context.
“These systems rely heavily on dictionaries, thesauri, and handcrafted rules.
Algorithms and Techniques:
( Lesk Algorithm:
-One of the simplest and oldest dictionary-based algorithms.