BibTeX file (contains some UTF-8 characters)
Description: The second edition of The Handbook of
Contemporary Semantic Theory presents a comprehensive introduction to
cutting-edge research in contemporary theoretical and computational
semantics: Features completely new content from the first edition of
The Handbook of Contemporary Semantic Theory; Features contributions
by leading semanticists, who introduce core areas of contemporary semantic
research, while discussing current research; Suitable for graduate students
for courses in semantic theory and for advanced researchers as an
introduction to current theoretical work. Description: This comprehensive reference work provides
an overview of the concepts, methodologies, and applications in computational
linguistics and natural language processing (NLP). Features contributions by
the top researchers in the field, reflecting the work that is driving the
discipline forward; Includes an introduction to the major theoretical issues
in these fields, as well as the central engineering applications that the
work has produced; Presents the major developments in an accessible way,
explaining the close connection between scientific understanding of the
computational properties of natural language and the creation of effective
language technologies; Serves as an invaluable state-of-the-art reference
source for computational linguists and software engineers developing NLP
applications in industrial research and development labs of software
companies. We present Property Theory with Curry Typing (PTCT), an
intensional first-order logic for natural language semantics. PTCT permits
fine-grained specifications of meaning. It also supports polymorphic types
and separation types. We develop an intensional number theory within PTCT in
order to represent proportional generalized quantifiers like âmost.â We
use the type system and our treatment of generalized quantifiers in natural
language to construct a type-theoretic approach to pronominal anaphora that
avoids some of the difficulties that undermine previous type-theoretic
analyses of this phenomenon. This monograph is concerned with exploring various ontological
assumptions, and whether they can be eliminated. It examines the basic
notions of proposition and property, as adopted by property theory, and then
goes on to explore what other ontological assumptions may be necessary for a
semantic theory of natural language, covering plurals, mass terms,
intensional individuals and discourse representation.Books (4)
This chapter considers the various philosophical and
methodological questions that arise in the formal analysis of the semantics
of language. Formal semantics aims to provide a systematic account of the
meaning of language in a rigorous formal framework. It is typically a
rule-based analysis of the relevant data and intuitions. This is a broad and
complex problem, given the nuances in the use and meaning of everyday
language. In practice, this means that a given analysis will confine itself
to some specific aspect of meaning, an appropriate sample of the language,
and some constrained context of use. Some issues in the analysis of imperatives and a sketch of a
proof-theoretic analysis. What is it that semanticists think they are doing when using
formalisation? What kind of endeavour is the formal semantics of natural
language: scientific; linguistic; philosophical; logical; mathematical? If
formal semantics is a scientific endeavour, then there ought to be empirical
criteria for determining whether such a theory is correct, or an improvement
on an alternative account. The question then arises as to the nature of the
evidence that is being accounted for. It could be argued that the empirical
questions are little different in kind to other aspects of linguistic
analysis, involving questions of performance and competence (Chomsky 1965;
Saussure 1916). But there are aspects of formal accounts of meaning that
appear to sit outside this scientific realm. One key issue concerns the
precise nature of the formalisation that is adopted; what criteria are to be
used to decide between accounts that are founded on different formal systems,
with different ontological assumptions? Indeed, is it necessary to judge
semantic frameworks on such grounds? In other words, are two theoretical
accounts to be treated as equivalent for all relevant purposes if they
account for exactly the same linguistic data? Broadly speaking, there are two
related perspectives on the analysis of propositional statements, one
âtruth conditionalâ â reducing sentence meaning to the conditions under
which the sentence is judged to be true (e.g. Montague 1973) â the other
âproof theoreticâ â reducing sentence meanings to patterns of
entailments that are supported (e.g. Ranta 1994, Fox & Lappin 2005 ,Fox
2000). Variations of these perspectives might be required in the case of
non-assertoric utterances. We may wonder what critieria might be used to
decide between these approaches. This brings us back to the nature of the
data itself. If the data is (merely) about which arguments, or truth
conditions, subjects agree with, and which they disagree with, then the terms
in which the theory is expressed may be irrelevant. But it may also be
legitimate to be concerned with either (a) the intuitions that people have
when reasoning with language, or (b) some technical or philosophical issues
relating to the chosen formalism. The truth-conditional vs proof-theoretic
dichotomy might broadly be characterised as model-theoretic vs axiomatic,
where the model-theoretic tend to be built around a pre-existing formal
theory, and the axiomatic involves formulating rules of behaviour âfrom
scratchâ. In some sense, fitting an analysis of a new problem into an
existing framework could be described as providing some kind of
âexplanatoryâ power, assuming that the existing framework has some
salient motivation that is independent of the specific details of the
phenomena in question. In contrast, building a new theory that captures the
behaviour might then be characterised as âdescriptiveâ, as â
superficially at least â it does not show how an existing theory
âalreadyâ accounts for the data in some sense. Here we observe instead
that the argument can be run in the other direction. that a reductive
model-theoretic account merely âdescribesâ how some aspects of a problem
can be reduced to some formalisation, but may fail to capture a subjectâs
understanding or intuitions about meaning. It is surely appropriate for the
formal theory itself to be at least sympathetic to the ontological concerns
and intuitions of its subjects â if not inform them (Dummett 1991). The
alternative amounts to little more than carving out otherwise arbitrary
aspects of a system that mimics the required behaviour, without a coherent
explanation of why some aspects of a formal theory characterise, or capture,
the intended meaning, but not others (cf. Benacerraf 1965). That seems an
impoverished approach, weakening any claim to âexplainâ. Any constraint
this imposes on what is it to be an explanatory account of meaning then faces
the same problem as naive notions of compositionality (Zadrozny 1994) â
that is, what appears to be a meaningful restriction is, in reality, a
mirage. Various questions arise in semantic analysis concerning the
nature of types. These questions include whether we need types in a semantic
theory, and if so, whether some version of simple type theory (STT, Church
1940) is adequate or whether a richer more flexible theory is required to
capture our semantic intuitions. Propositions and propositional attitudes can
be represented in an essentially untyped first-order language, provided a
sufficiently rich language of terms is adopted. In the absence of rigid
typing, care needs to be taken to avoid the paradoxes, for example by
constraining what kinds of expressions are to be interpreted as propositions
(Turner 1992). But the notion of type is ontologically appealing. In some
respects, STT seems overly restrictive for natural language semantics. For
this reason it is appropriate to consider a system of types that is more
flexible than STT, such as a Curry-style typing (Curry & Feys 1958). Care
then has to be taken to avoid the logical paradoxes. Here we show how such an
account, based on the Property Theory with Curry Types (PTCT, Fox & Lappin
2005), can be formalised within Typed Predicate Logic (TPL, Turner 2009).
This presentation provides a clear distinction between the classes of types
that are being used to (i) avoid paradoxes (ii) allow predicative polymorphic
types. TPL itself provides a means of expressing PTCT in a uniform
language. Accounts of the formal semantics of natural language often adopt
a pre-existing framework. Such formalisations rely upon informal narrative to
explain the intended interpretation of an expression â an expression that
may have different interpretations in different circumstances, and may
supports patterns of behaviour that exceed what is intended. This ought to
make us question the sense in which such formalisations capture our
intuitions about semantic behaviour. In the case of theories of questions and
answers, a question might be interpreted as a set (of possible propositional
answers), or as a function (that yields a proposition given a term that is
intended to be interpreted as a phrasal answer), but the formal theory itself
provides no means of distinguishing such sets and functions from other cases
where they are not intended to represent questions, or their answers. Here we
sketch an alternative approach to formalisation a theory of questions and
answers that aims to be sensitive to such ontological
considerations. We may wonder about the status of logical accounts of the
meaning of language. When does a particular proposal count as a theory? How
do we judge a theory to be correct? What criteria can we use to decide
whether one theory is âbetterâ than another? Implicitly, many accounts
attribute a foundational status to set theory, and set-theoretic
characterisations of possible worlds in particular. The goal of a semantic
theory is then to find a translation of the phenomena of interest into a
set-theoretic model. Such theories may be deemed to have âexplanatoryâ or
âpredictiveâ power if a mapping can found into expressions of set-theory
that have the appropriate behaviour by virtue of the rules of set-theory (for
example Montague 1973; Montague1974). This can be contrasted with an approach
in which we can help ourselves to ânewâ primitives and ontological
categories, and devise logical rules and axioms that capture the appropriate
inferential behaviour (as in Turner 1992). In general, this alternative
approach can be criticised as being mere âdescriptivismâ, lacking
predictive or explanatory power. Here we will seek to defend the axiomatic
approach. Any formal account must assume some normative interpretation, but
there is a sense in which such theories can provide a more honest
characterisation (cf. Dummett 1991). In contrast, the set-theoretic approach
tends to conflate distinct ontological notions. Mapping a pattern of semantic
behaviour into some pre-existing set-theoretic behaviour may lead to certain
aspects of that behaviour being overlooked, or ignored (Chierchia & Turner
1988; Bealer 1982). Arguments about the explanatory and predictive power of
set-theoretic interpretations can also be questioned (see Benacerraf 1965,
for example). We aim to provide alternative notions for evaluating the
quality of a formalisation, and the role of formal theory. Ultimately, claims
about the methodological and conceptual inadequacies of axiomatic accounts
compared to set-theoretic reductions must rely on criteria and assumptions
that lie outside the domain of formal semantics as
such. At the time of writing, summarisation systems for Arabic are not
as sophisticated and as reliable as those developed for languages like
English. In this paper we discuss two summarisation systems for Arabic and
report on a large user study performed on these systems. The first system,
the Arabic Query-Based Text Summarisation System (AQBTSS), uses standard
retrieval methods to map a query against a document collection and to create
a summary. The second system, the Arabic Concept-Based Text Summarisation
System (ACBTSS), creates a query-independent document summary. Five groups of
users from different ages and educational levels participated in evaluating
our systems. A brief introduction to Computational
Semantics. Cooper (1983) pioneered underspecified scope representation in
formal and computational semantics through his introduction of quantifier
storage into Montague semantics as an alternative to the syntactic operation
of quantifying-in. In this paper we address an important issue in the
development of an adequate formal theory of underspecified semantics. The
tension between expressive power and computational tractability poses an
acute problem for any such theory. Ebert (2005) shows that any reasonable
current treatment of underspecified semantic representation either suffers
from expressive incompleteness or produces a combinatorial explosion that is
equivalent to generating the full set of possible scope readings in the
course of disambiguation. In previous work we have presented an account of
underspecified scope representations within Property Theory with Curry Typing
(PTCT), an intensional first-order theory for natural language semantics.
Here we show how filters applied to the underspecified-scope terms of PTCT
permit both expressive completeness and the reduction of computational
complexity in a significant class of non-worst case
scenarios. It is reasonably well-understood that natural language displays
polymorphic behaviour in both its syntax and semantics, where various
constructions can operate on a range of syntactic categories, or semantic
types. In mathematics, logic and computer science it is appreciated that
there are various ways in which such type-general behaviours can be
formulated. It is also known that natural languages are highly ambiguous with
respect to scoping artifacts, as evident with quantifiers, negation and
certain modifier expressions. To deal with such issues, formal frameworks
have been explored in which the polymorphic nature of natural language can be
expressed, and theories of underspecified semantics have been proposed which
seek to separate the process of pure compositional interpretation from the
assignment of scope. To date, however, there has been no work on bringing
these two aspects together; there is no semantic treatments of scope
ambiguity and underspecification which explicitly takes into account the
polymorphic nature of natural language quantifiers. In this paper, we extend
an existing treatment of underspecification and scope ambiguity in Property
Theory with Curry Typing (PTCT) to deal with arbitrary types of
quantification by adopting a form of polymorphism. In this theory of
underspecification, all of the expressions in the theory are terms of the
logic; there is no âmeta-semanticâ machinery. For this reason all aspects
of the theory must be able to deal with polymorphism
appropriately. This chapter is concerned with representing the semantics of
natural language plurals and mass terms in property theory; a weak
first-order theory of Truth, Propositions and Properties with fine-grained
intensionality (Turner 1990, Turner 1992, Aczel 1980). The theory allows
apparently coreferring items to corefer without inconsistency. This is
achieved by using property modifiers which keep track of the property used to
refer to a term, much like Landmanâs roles (Landman 1989). We can thus
predicate apparently contradictory properties of âthe judgeâ and âthe
cleaner,â for example, even if they turn out to be the same individual. The
same device can also be used to control distribution into mereological terms:
when we say âthe dirty water is liquid,â we can infer that those parts
that are dirty water are liquid without inferring that the dirt is liquid.
The theory shows how we can formalise some aspects of natural language
semantics without being forced to make certain ontological commitments. This
is achieved in part by adopting an axiomatic methodology. Axioms are proposed
that are just strong enough to support intuitively acceptable inferences,
whilst being weak enough for some ontological choices to be avoided (such as
whether or not the extensions of mass terms should be homogeneous or atomic).
The axioms are deliberately incomplete, just as in basic PT, where incomplete
axioms are used to avoid the logical paradoxes. The axioms presented are
deliberately too weak to say much aboutânon-denotingâ definite
descriptors. For example, we cannot erroneously prove that they are all
equal. Neither can we prove that predication of such definites results in a
proposition. This means that we cannot question the truth of sentences such
as âthe present king of France is bald.âChapters (11)
Effective information management has long been a problem in
organisations that are not of a scale that they can afford their own
department dedicated to this task. Growing information overload has made this
problem even more pronounced. On the other hand we have recently witnessed
the emergence of intelligent tools, packages and resources that made it
possible to rapidly transfer knowledge from the academic community to
industry, government and other potential beneficiaries. Here we demonstrate
how adopting state-of-the-art natural language processing (NLP) and
crowdsourcing methods has resulted in measurable benefits for a human rights
organisation by transforming their information and knowledge management using
a novel approach that supports human rights monitoring in conflict zones.
More specifically, we report on mining and classifying Arabic Twitter in
order to identify potential human rights abuse incidents in a continuous
stream of social media data within a specified geographical region. Results
show deep learning approaches such as LSTM allow us to push the precision
close to 85% for this task with an F1-score of 75%. Apart from the
scientific insights we also demonstrate the viability of the framework which
has been deployed as the Ceasefire Iraq portal for more than three years
which has already collected thousands of witness reports from within Iraq.
This work is a case study of how progress in artificial intelligence has
disrupted even the operation of relatively small-scale
organisations. In this paper, we investigate the âought implies canâ (OIC)
thesis, focusing on explanations and interpretations of OIC, with a view to
clarifying its uses and relevance to legal philosophy. We first review
various issues concerning the semantics and pragmatics of OIC; then we
consider how OIC may be incorporated in Hartian and Kelsenian theories of the
law. Along the way we also propose a taxonomy of OIC-related
claims. Information systems that utilise contextual information have the
potential of helping a user identify relevant information more quickly and
more accurately than systems that work the same for all users and contexts.
Contextual information comes in a variety of types, often derived from
records of past interactions between a user and the information system. It
can be individual or group based. We are focusing on the latter, harnessing
the search behaviour of cohorts of users, turning it into a domain model that
can then be used to assist other users of the same cohort. More specifically,
we aim to explore how such a domain model is best utilised for profile-biased
summarisation of documents in a navigation scenario in which such summaries
can be displayed as hover text as a user moves the mouse over a link. The
main motivation is to help a user find relevant documents more quickly. Given
the fact that the Web in general has been studied extensively already, we
focus our attention on Web sites and similar document collections. Such
collections can be notoriously difficult to search or explore. The process of
acquiring the domain model is not a research interest here; we simply adopt a
biologically inspired method that resembles the idea of ant colony
optimisation. This has been shown to work well in a variety of application
areas. The model can be built in a continuous learning cycle that exploits
search patterns as recorded in typical query log files. Our research explores
different summarisation techniques, some of which use the domain model and
some that do not. We perform task-based evaluations of these different
techniquesâthus of the impact of the domain model and profile-biased
summarisationâin the context of Web site navigation. Language resources are important for those working on
computational methods to analyse and study languages. These resources are
needed to help advancing the research in fields such as natural language
processing, machine learning, information retrieval and text analysis in
general. We describe the creation of useful resources for languages that
currently lack them, taking resources for Arabic summarisation as a case
study. We illustrate three different paradigms for creating language
resources, namely: (1) using crowdsourcing to produce a small resource
rapidly and relatively cheaply; (2) translating an existing gold-standard
dataset, which is relatively easy but potentially of lower quality; and (3)
using manual effort with appropriately skilled human participants to create a
resource that is more expensive but of high quality. The last of these was
used as a test collection for TAC-2011. An evaluation of the resources is
also presented. A reformulation of Curry-Typed Property Theory within Typed
Predicate Logic, and some discussion of an operational interpretation of
intensional distinctions. This paper proposes a framework for formalising intuitions about
the behaviour of imperative commands. It seeks to capture notions of
satisfaction and coherence. Rules are proposed to express key aspects of the
general logical behaviour of imperative constructions. A key objective is for
the framework to allow patterns of behaviour to be described while avoiding
making any commitments about how commands, and their satisfaction criteria,
are to be interpreted. We consider the status of some conundrums of
imperative logic in the context of this proposal. Utterances and statements that are concerned with obligations
and permissions are known as âdeonticâ expressions. They can present
something of a challenge when it comes to formalising their meaning and
behaviour. The content of these expressions can appear to support entailment
relations similar to those of classical propositions, but such behaviour can
sometimes lead to counter-intuitive outcomes. Historically, much of the
descriptive work in this area has been philosophical in outlook,
concentrating on questions of morality and jurisprudence. Some additional
contributions have come from computer science, in part due to the need to
specify normative behaviour. There are a number of formal proposals that seek
to account for obligations and permissions, such as âStandard Deontic
Logicâ. In the literature, there has also been discussion of various
conundrums and dilemmas that need to be resolved, such as âthe Good
Samaritanâ, âthe Knowerâ, âthe Gentle Murdererâ, âContrary to
Duty Obligationsâ, âRossâs Paradoxâ, âJĂžrgensenâs Dilemmaâ,
âSartreâs Dilemmaâ, and âPlatoâs Dilemmaâ. Despite all this work,
there still appears to be no definite consensus about how these kinds of
expressions should be analysed, or how all the deontic dilemmas should be
resolved. It is possible that obligations themselves, as opposed to their
satisfaction criteria, do not directly support a conventional logical
analysis. It is also possible that a linguistically informed analysis of
obligations and permissions may help to resolve some of the deontic dilemmas,
and clarify intuitions about how best to formulate a logic of deontic
expressions. In this paper we address an important issue in the development
of an adequate formal theory of underspecified semantics. The tension between
expressive power and computational tractability poses an acute problem for
any such theory. Generating the full set of resolved scope readings from an
underspecified representation produces a combinatorial explosion that
undermines the efficiency of these representations. Moreover, Ebert (2005)
shows that most current theories of underspecified semantic representation
suffer from expressive incompleteness. In previous work we present an account
of underspecified scope representations within Property Theory with Curry
Typing (PTCT), an intensional first-order theory for natural language
semantics. We review this account, and we show that filters applied to the
underspecified-scope terms of PTCT permit expressive completeness. While they
do not solve the general complexity problem, they do significantly reduce the
search space for computing the full set of resolved scope readings in
non-worst cases. We explore the role of filters in achieving expressive
completeness, and their relationship to the complexity involved in producing
full interpretations from underspecified representations. This paper is
dedicated to Jim Lambek. Program conditioning consists of identifying and removing a set
of statements which cannot be executed when a condition of interest holds at
some point in a program. It has been applied to problems in maintenance,
testing, reÂuse and reÂengineering. Program conditioning relies upon both
symbolic execution and reasoning about symbolic predicates. Automation of the
process therefore requires some form of automated theorem proving. However,
the use of a full-power âheavyweightâ theorem prover would impose
unrealistic performance constraints. This paper reports on a lightweight
approach to theorem proving using the FermaT simplify decision procedure.
This is used as a component to ConSUS, a program conditioning system for the
Wide Spectrum Language WSL. The paper describes the symbolic execution
algorithm used by ConSUS, which prunes as it conditions. The paper also
provides empirical evidence that conditioning produces a significant
reduction in program size and, although exponential in the worst case, the
conditioning system has low degree polynomial behaviour in many cases,
thereby making it scalable to unit level applications of program
conditioning. Program slicing is an automated source code extraction technique
that has been applied to a number of problems including testing, debugging,
maintenance, reverse engineering, program comprehension, reuse and program
integration. In all these applications the size of the slice is crucial; the
smaller the better. It is known that statement minimal slices are not
computable, but the question of dataflow minimal slicing has remained open
since Weiser posed it in 1979. This paper proves that static slicing
algorithms produce dataflow minimal end slices for programs which can be
represented as schemas which are free and liberal. Test data generation by hand is a tedious, expensive and
error-prone activity, yet testing is a vital part of the development process.
Several techniques have been proposed to automate the generation of test
data, but all of these are hindered by the presence of unstructured control
flow. This paper addresses the problem using testability transformation.
Testability transformation does not preserve the traditional meaning of the
program, rather it deals with preserving test-adequate sets of input data.
This requires new equivalence relations which, in turn, entail novel proof
obligations. The paper illustrates this using the branch coverage adequacy
criterion and develops a branch adequacy equivalence relation and a
testability transformation for restructuring. It then presents a proof that
the transformation preserves branch adequacy. Abstract In previous work we have developed Property Theory with
Curry Typing (PTCT), an intensional first-order logic for natural language
semantics. PTCT permits fine-grained specifications of meaning. It also
supports polymorphic types and separation types. We develop an intensional
number theory within PTCT in order to represent proportional generalized
quantifiers like "most", and we suggest a dynamic type-theoretic approach to
anaphora and ellipsis resolution. Here we extend the type system to include
product types, and use these to define a permutation function that generates
underspecified scope representations within PTCT. We indicate how filters can
be added to encode constraints on possible scope readings. Our account offers
several important advantages over other current theories of
underspecification. We present Property Theory with Curry Typing (PTCT), an
intensional first-order logic for natural language semantics. PTCT permits
fine-grained specifications of meaning. It also supports polymorphic types
and separation types. We develop an intensional number theory within PTCT in
order to represent proportional generalized quantifiers like âmost.â We
use the type system and our treatment of generalized quantifiers in natural
language to construct a type-theoretic approach to pronominal anaphora that
avoids some of the difficulties that undermine previous type-theoretic
analyses of this phenomenon. Conditioned slicing is a source code extraction technique. The
extraction is performed with respect to a slicing criterion which contains a
set of variables and conditions of interest. Conditioned slicing removes the
parts of the original program which cannot affect the variables at the point
of interest, when the conditions are satisfied. This produces a conditioned
slice, which preserves the behaviour of the original with respect to the
slicing criterion. Conditioned slicing has applications in source code
comprehension, reuse, restructuring and testing. Unfortunately,
implementation is not straightforward because the full exploitation of
conditions requires the combination of symbolic execution, theorem proving
and traditional static slicing. Hitherto, this difficultly has hindered
development of fully automated conditioning slicing tools. This paper
describes the first fully automated conditioned slicing system, ConSIT,
detailing the theory that underlies it, its architecture and the way it
combines symbolic execution, theorem proving and slicing technologies. The
use of ConSIT is illustrated with respect to the applications of testing and
comprehension. This paper reports on work done in the INSPIRE project on
developing the Process Representation Module (PRM). The major aim of INSPIRE
is the development of a tool for intelligent, human-orientated business
process re-engineering. Our task was to develop the PRM, a core module of the
INSPIRE tool. The main responsibility of the PRM is to provide an
all-encompasing and consistent representation of business processes. The
paper describes the architecture and data-models of the system, plus
discussion of business prcess modelling and the formalisms used in INSPIRE,
and the details of the design and implementation of the
PRM. This paper introduces a notation for business process modeling
and shows how it can be formally interpreted in terms of Petri nets. Petri
nets have a quite respectable research community, which is 35 years old.
However, they were only recently proposed for business process modeling. This
is probably due to the fact they are often claimed to be âtoo complexâ
for this task. Nevertheless, they are quite well understood and the theory
behind them is well developed, so we think they have a good potential for
business process modeling, but more work needs to be done. In this paper we
show that Petri nets can help in understanding formally the business process
modeling notation developed in the Inspire project. This understanding can
act as a basis for a future work on formal analysis of business process
models developed with the INSPIRE tool. The Inspire project (IST-10387-1999)
aims to develop an integrated tool-set to support a systematic and more
human-oriented approach to business process
re-engineering. This paper describes the use of conditioned slicing to assist
partition testing, illustrating this with a case study. The paper shows how a
conditioned slicing tool can be used to provide confidence in the uniformity
hypothesis for correct programs, to aid fault detection in incorrect programs
and to highlight special cases. Review of a once popular reference book for the GNU/Linux
operating system. This paper discusses a property-theoretic approach to the
existence presuppositions that occur with definite descriptors and some
examples of anaphoric reference. Property theory can avoid the logical
paradoxes (such as the Liar: âthis sentence is false.â) by taking them to
involve a category mistake, so they do not express felicitous propositions.
It is suggested that this approach might be extended to other cases of
infelicitous discourse, such as those due to a false presupposition (as in:
âThe present queen of France is bald.â) or due to a missing antecedent
(as in: âEvery man walks in. He whistles.â). These examples may be
represented by terms that embody category mistakes, so semantically they do
not express propositions. Felicity of discourse then corresponds with the
propositionhood of the representation. This paper describes an extension to a system for parsing
English a relational calculus (SQL), by way of Property-theoretic semantics,
that paraphrases the relational query back into English, to ensure the query
has been interpreted as intended. This paper describes an extension to a system for parsing
English into a relational calculus (SQL), by way of Property-theoretic
semantics, that allows such a system to answer modal questions, such as
âCan Sally earn 12,000?â The system answered such queries by performing a
trial update to see if any of the constraints on the database would then be
broken.Articles (21)
This special issue was a spin off of the second workshop on
Lambda Calculus, Type Theory and Natural Language. The Workshop on Lambda
Calculus, Type Theory, and Natural Language (LCTTNL) was held in London in
September 2005. This special issue was a spin off of the first workshop on
Lambda Calculus, Type Theory and Natural Language, which was held in London
in December, 2003.Special Issues (2)
Privacy concerns are becoming a dominant focus in search
applications, thus there is a growing need to understand implications of
efforts to address these concerns. Our research investigates a search system
with privacy warning labels, an approach inspired by decision making research
on food nutrition labels. This approach is designed to alert users to
potential privacy threats in their search for information as one possible
avenue to address privacy concerns. Our primary goal is to understand the
extent to which attitudes towards privacy are linked to behaviors that
protect privacy. In the present study, participants were given a set of
fact-based decision tasks from the domain of health search. Participants were
rotated through variations of search engine results pages (SERPs) including a
SERP with a privacy warning light system. Lastly, participants completed a
survey to capture attitudes towards privacy, behaviors to protect privacy,
and other demographic information. In addition to the comparison of
interactive search behaviors of a privacy warning SERP with a control SERP,
we compared self-report privacy measures with interactive search behaviors.
Participants reported strong concerns around privacy of health information
while simultaneously placing high importance on the correctness of this
information. Analysis of our interactive experiment and self-report privacy
measures indicate that 1) choice of privacy-protective browsers has a
significant link to privacy attitudes and privacy-protective behaviors in a
SERP and 2) there are no significant links between reported concerns towards
privacy and recorded behavior in an information retrieval system with
warnings that enable users to protect their privacy. From their impacts to potential threats, privacy and
misinformation are a recurring top news story. Social media platforms (e.g.
Facebook) and information retrieval (IR) systems (e.g. Google), are now in
the public spotlight to address these issues. Our research investigates an
approach, known as Nudging, applied to the domain of IR, as a potential means
to minimize impacts and threats surrounding both matters. We perform our
study in the space of health search for two reasons. First, encounters with
misinformation in this space have potentially grave outcomes. Second, there
are many potential threats to personal privacy as a result of the data
collected during a search task. Adopting methods and a corpus from previous
work as the foundation, our study asked users to determine the effectiveness
of a treatment for 10 medical conditions. Users performed the tasks on 4
variants of a search engine results page (SERP) and a control, with 3 of the
SERPâs being a Nudge (re-ranking, filtering and a visual cue) intended to
reduce impacts to privacy with minimal impact to search result quality. The
aim of our work is to determine the Nudge that is least impactful to good
decision making while simultaneously increasing privacy protection. We find
privacy impacts are significantly reduced for the re-ranking and filtering
strategies, with no significant impacts on quality of decision
making. © LREC 2018 - 11th International Conference on Language
Resources and Evaluation. All rights reserved. Hate speech has become a major
issue that is currently a hot topic in the domain of social media.
Simultaneously, current proposed methods to address the issue raise concerns
about censorship. Broadly speaking, our research focus is the area human
rights, including the development of new methods to identify and better
address discrimination while protecting freedom of expression. As neural
network approaches are becoming state of the art for text classification
problems, an ensemble method is adapted for usage with neural networks and is
presented to better classify hate speech. Our method utilizes a publicly
available embedding model, which is tested against a hate speech corpus from
Twitter. To confirm robustness of our results, we additionally test against a
popular sentiment dataset. Given our goal, we are pleased that our method has
a nearly 5 point improvement in F-measure when compared to original work on a
publicly available hate speech evaluation dataset. We also note difficulties
encountered with reproducibility of deep learning methods and comparison of
findings from other work. Based on our experience, more details are needed in
published work reliant on deep learning methods, with additional evaluation
information a consideration too. This information is provided to foster
discussion within the research community for future
work. © 2016, CEUR-WS. All rights reserved. This invited talk
provided an overview of some of our work in relation to extracting meaningful
knowledge from social media feeds to help in addressing human rights issues
highlighting the potential that the rise of âbig dataâ offers in this
respect looking at both sides of the coin regarding big data and human
rights: how big data can help human rights work, but also the potential
dangers that can originate from the ability to analyse massive amounts of
data very quickly. The primary focus of our work is on applying natural
language processing methods to turn large-scale unstructured and partially
structured data streams into actionable knowledge. Query expansion techniques have proved to have an impact on
retrieval performance across many retrieval tasks. This paper reports
research on query expansion in the entity finding domain. We used a number of
methods for query formulation including thesaurus-based, relevance feedback,
and exploiting NLP structure. We incorporated the query expansion component
as part of our entity finding pipeline and report the results of the
aforementioned models on the CERC collection. This was the second year of the C@merata task [16,1] which
relates natural language processing to music information retrieval.
Participants each build a system which takes as input a query and a music
score and produces as output one or more matching passages in the score. This
year, questions were more difficult and scores were more complex.
Participants were the same as last year and once again CLAS was the best with
a Beat F-Score of 0.620. There is a vast body of musicological literature containing
detailed analyses of musical works. These texts make frequent references to
musical passages in scores by means of natural language phrases. Our
long-term aim is to investigate whether these phrases can be linked
automatically to the musical passages to which they refer. As a first step,
we have organised for two years running a shared evaluation in which
participants must develop software to identify passages in a MusicXML score
based on a short noun phrase in English. In this paper, we present the
rationale for this work, discuss the kind of references to musical passages
which can occur in actual scholarly texts, describe the first two years of
the evaluation and finally appraise the results to establish what progress we
have made. With the continuous attention of modern search engines to
retrieve entities and not just documents for any given query, we introduce a
new method for enhancing the entity-ranking task. An entity-ranking task is
concerned with retrieving a ranked list of entities as a response to a
specific query. Some successful models used the idea of association discovery
in a window of text, rather than in the whole document. However, these
studies considered only fixed window sizes. This work proposes a way of
generating an adaptive window size for each document by utilising some of the
document features. These features include document length, average sentence
length, number of entities in the document, and the readability index.
Experimental results show a positive effect once taking these document
features into consideration when determining window
size. This paper summarises the C@merata task in which participants
built systems to answer short natural language queries about classical music
scores in MusicXML. The task thus combined natural language processing with
music information retrieval. Five groups from four countries submitted eight
runs. The best submission scored Beat Precision 0.713 and Beat Recall
0.904. Text summarisation is the process of distilling the most
important information from a source to produce an abridged version for a
particular user or task. This poster investigates the use of profile-based
summarisation to provide contextualisation and interactive support for
enterprise searches. We employ log analysis to acquire continuously updated
profiles to provide profile-based summarisations of search results. These
profiles could be capturing an individualâs interests or (as discussed
here) those of a group of users. Here we report on a first pilot
study. The goal of expert-finding is to retrieve a ranked list of
people as a response to a user query. Some models that proved to be very
successful used the idea of association discovery in a window of text rather
than the whole document. So far, all these studies only considered fixed
window sizes. We pro-pose an adaptive window-size approach for
expert-finding. For this work we use some of the document attributes, such as
document length, average sentence length, and number of candidates, to adjust
the window size for the document. The experimental results indicate that
taking document features into consideration when determining the window size,
does have an effect on the retrieval outcome. The results shows an
improvement over a range of baseline approaches. Text summarisation is the process of distilling the most
important information from a source to produce an abridged version for a
particular user or task. This demo presents the use of profile-based
summarisation to provide contextualisation and interactive support for site
search and enter- prise search. We employ log analysis to acquire
continuously updated profiles to provide profile-based summarisations of
search results. These profiles could be capturing an individualâs interests
or those of a group of users. Here we look at acquiring profiles for groups
of users. In this paper we explore clustering for multi-document Arabic
summarisation. For our evaluation we use an Arabic version of the DUC-2002
dataset that we previously generated using Google Translate. We explore how
clustering (at the sentence level) can be applied to multi-document
summarisation as well as for redundancy elimination within this process. We
use different parameter settings including the cluster size and the selection
model applied in the extractive summarisation process. The automatically
generated summaries are evaluated using the ROUGE metric, as well as
precision and recall. The results we achieve are compared with the top five
systems in the DUC-2002 multi-document summarisation
task. In this paper we present our generic extractive Arabic and
English multi-document summarisers. We also describe the use of machine
translation for evaluating the generated Arabic multi-document summaries
using English extractive gold standards. In this work we first address the
lack of Arabic multi-document corpora for summarisation and the absence of
automatic and manual Arabic gold-standard summaries. These are required to
evaluate any automatic Arabic summarisers. Second, we demonstrate the use of
Google Translate in creating an Arabic version of the DUC-2002 dataset. The
parallel Arabic/English dataset is summarised using the Arabic and English
summarisation systems. The automatically generated summaries are evaluated
using the ROUGE metric, as well as precision and recall. The results we
achieve are compared with the top five systems in the DUC-2002 multi-document
summarisation task. We present the results of our Arabic and English runs at the TAC
2011 Multilingual summarisation (MultiLing) task. We participated with
centroid-based clustering for multi-document summarisation. The automatically
generated Arabic and English summaries were evaluated by human participants
and by two automatic evaluation metrics, ROUGE and AutoSummENG. The results
are compared with the other systems that participated in the same track on
both Arabic and English languages. Our Arabic summariser performed
particularly well in the human evaluation. This paper describes the creation of a human-generated corpus of
extractive Arabic summaries of a selection of Wikipedia and Arabic newspaper
articles using Mechanical Turkâan online workforce. The purpose of this
exercise was two-fold. First, it addresses a shortage of relevant data for
Arabic natural language processing. Second, it demonstrates the application
of Mechanical Turk to the problem of creating natural language resources. The
paper also reports on a number of evaluations we have performed to compare
the collected summaries against results obtained from a variety of automatic
summarisation systems. When developing formal theories of the meaning of language, it
is appropriate to consider how apparent paradoxes and conundrums of language
are best resolved. But if we base our analysis on a small sample of data then
we may fail to take into account the influence of other aspects of meaning on
our intuitions. Here we consider the so-called Good Samaritan Paradox (Prior,
1958), where we wish to avoid any implication that there is an obligation to
rob someone from âYou must help a robbed manâ. We argue that before
settling on a formal analysis of such sentences, we should consider examples
of the same form, but with intuitively different entailmentsâsuch as âYou
must use a clean knifeââand also actively seek other examples that
exhibit similar contrasts in meaning, even if they do not exemplify the
phenomena that is under investigation. This can refine our intuitions and
help us to attribute aspects of interpretation to the various facets of
meaning. The volume of information available on the Web is increasing
rapidly. The need for systems that can automatically summarise documents is
becoming ever more desirable. For this reason, text summarisation has quickly
grown into a major research area as illustrated by the DUC and TAC conference
series. Summarisation systems for Arabic are however still not as
sophisticated and as reliable as those developed for languages like English.
In this paper we discuss two summarisation systems for Arabic and report on a
large user study performed on these systems. The first system, the Arabic
Query-Based Text Summarisation System (AQBTSS), uses standard retrieval
methods to map a query against a document collection and to create a summary.
The second system, the Arabic Concept-Based Text Summarisation System
(ACBTSS), creates a query-independent document summary. Five groups of users
from different ages and educational levels participated in evaluating our
systems. This paper proposes a logic of transgressions for obligations
and permissions. A key objective of this logic is to allow deontic conflicts
(Lemmon, 1962) but without appealing to defeasible or paraconsistent
reasoning, or multiple levels of obligation. This logic of transgressions can
be viewed as conceptually related to those approaches that formulate
obligations in terms of âescapingâ from a sanction (Prior, 1958;
Nowell-Smith and Lemmon, 1960), and its modal variants (Anderson, 1958;
Kanger, 1971), but where the notion of a transgression is more fine-grained
than a single âsanctionâ The paper discusses how to reduce a statechart model by slicing.
We start with the discussion of control dependencies and data dependencies in
statecharts. The and-or dependence graph is introduced to represent control
and data dependencies for statecharts. We show how to slice statecharts by
using this dependence graph. Our slicing approach helps systems analysts and
system designers in understanding system specifications, maintaining software
systems, and reusing parts of systems models. This paper explores the influence of text pre-processing
techniques on plagiarism detection. We examine stop-word removal,
lemmatization, number replacement, synonymy recognition, and word
generalization. We also look into the influence of punctuation and word-order
within N-grams. All these techniques are evaluated according to their impact
on F_1-measure and speed of execution. Our experiments were performed on a
Czech corpus of plagiarized documents about politics. At the end of this
paper, we propose what we consider to be the best combination of text
pre-processing techniques. The construction of an And-Or dependence graphs is illustrated,
and its use in slicing statecharts is described. The additional structure
allows for more precise slices to be constructed in the event of additional
information, such as may be provided by static analysis and model checking,
and with constraints on the global state and external
events. Business process modeling is an important phase during
requirements collection. Usually functional, dynamic and role models are
needed. We propose to integrate Role Activity Diagrams and Hybrid IDEF for
business process modeling within Model Driven Architecture. Our proposal is
demonstrated with a sample implementation. Ebert (2005) points out that most current theories of
underspecified semantic representation either suffer from expressive
incompleteness or do not avoid generating the full set of possible scope
readings in the course of disambiguation. In previous work we have presented
an account of underspecified scope representations within an intensional
first-order property theory enriched with Curry Typing for natural language
semantics. Here we show how filters applied to the underspecified scope terms
of this theory permit both expressive completeness and the reduction of the
search space of possible scope interpretations. A description of the process dimension of a notation for
business process modelling that integrates aspects from IDEF0 and IDEF3 in a
novel way is presented. The features of this notation include black box
modelling of activities in the style of IDEF0 and glass box refinements of
activities using connectors for specifying process branching in the style of
IDEF3. The semantics of the notation is given by a mapping to a
place/transition net. The notation is shown to be as expressive as a Standard
Workflow Model. In many business process modelling situations using Petri nets,
the resulting model does not have a single input place and a single output
place. Therefore, the correctness of the model cannot be assessed within the
existing frameworks, which are devised for workflow nets â a particular
class of Petri nets with a single input place and a single output place.
Moreover, the existing approaches for tackling this problem are rather
simplistic and they do not work even for simple examples. This paper shows
that, by an appropriate reduction of a multiple input/multiple output Petri
net, it is possible to use the existing techniques to check the correctness
of the original process. The approach is demonstrated with an appropriate
example. The aim of this paper is to present a model for the
interpretation of imperative sentences in which reasoning agents play the
role of speakers and hearers. A requirement is associated with both the
person who makes and the person who receives the order which prevents the
hearer coming to inappropriate conclusions about the actions s/he has been
commanded to do. By relating imperatives with the actions they prescribe, the
dynamic aspect of imperatives is captured. Further, by using the idea of
âencapsulationâ, it is possible to distinguish what is demanded by an
imperative from the inferential consequences of the imperative. These two
ingredients provide agents with the tools to avoid inferential problems in
interpretation. These two ingredients provide agents with the tools to avoid
inferential problems in interpretation. The aim of this paper is to present a model for the
interpretation of imperative sentences in which reasoning agents play the
role of speakers and hearers. A requirement is associated with both the
person who makes and the person who receives the order which prevents the
hearer coming to inappropriate conclusions about the actions s/he has been
commanded to do. By relating imperatives with the actions they prescribe, the
dynamic aspect of imperatives is captured and by using the idea of
encapsulation, it is possible to distinguish what is demanded from what is
not. These two ingredients provide agents with the tools to avoid inferential
problems in interpretation. We present Property Theory with Curry Typing (PTCT), an
intensional first-order logic for natural language semantics. PTCT permits
fine-grained specifications of meaning. It also supports polymorphic types
and separation types. (Separation types are also known as sub-types.)
We develop an intensional number theory within PTCT in order to represent
proportional generalized quantifiers like most. We use the type system
and our treatment of generalized quantifiers in natural language to construct
a type-theoretic approach to pronominal anaphora that avoids some of the
difficulties that undermine previous type-theoretic analyses of this
phenomenon. This paper presents an axiomatisation of imperatives using Hoare
logic. It accounts for some inferential pragmatic aspects of imperatives.
Unlike Jorgensenâs, Ross, and Chellas, proposals, rather than assigning
truth-values to imperatives, imperatives are evaluated as a relation between
the state demanded and the state or circumstances in which the imperative is
uttered. We present a dynamic deontic model for the interpretation of
imperative sentences in terms of Obligation (O) and Permission (P). Under the
view that imperatives prescribe actions and unlike the so-called "standard
solution" (Huntley, 1984) these operators act over actions rather that over
statements. Then by distinguishing obligatory from non obligatory actions we
tackle the paradox of Free Choice Permission (FCP). We present an approach to anaphora and ellipsis resolution in
which pronouns and elided structures are interpreted by the dynamic
identification in discourse of type constraints on their semantic
representations. The content of these conditions is recovered in context from
an antecedent expression. The constraints define separation types (sub-types)
in Property Theory with Curry Typing (PTCT), an expressive first-order logic
with Curry typing that we have proposed as a formal framework for natural
language semantics Conditioned slicing can be applied to reverse engineering
problems which involve the extraction of executable fragments of code in the
context of some criteria of interest. This paper introduces ConSUS, a
conditioner for the Wide Spectrum Language, WSL. The symbolic executor of
ConSUS prunes the symbolic execution paths, and its predicate reasoning
system uses the FermaT simplify transformation in place of a more
conventional theorem prover. We show that this combination of pruning and
simplification as-reasoner leads to a more scalable approach to
conditioning. presents Property Theory with Curry Typing (PTCT) where the
language of terms and well-formed formulĂŠ are joined by a language of
types. In addition to supporting fine-grained intensionality, the basic
theory is essentially first-order, so that implementations using the theory
can apply standard first-order theorem proving techniques. The paper sketches
a system of tableau rules that implement the theory. Some extensions to the
type theory are discussed, including type polymorphism, which provides a
useful analysis of conjunctive terms. Such terms can be given a single
polymorphic type that expresses the fact that they can conjoin phrases of any
one type, yielding an expression of the same type. Variable dependence is an analysis problem in which we seek to
determine the set of input variables which can affect the values stored in a
chosen set of intermediate program variables. Traditionally the problem is
studied as a dataflow analysis problem, and the answers are computed in terms
of solutions to data and control flow relations. This paper shows the
relationship between the variable dependence analysis problem and slicing and
describes a system, VADA, which implements variable dependence analysis for
C. In order to cover the full range of C constructs and features, a
transformation to a core language is employed. Thus, the full analysis is
only required for the core language, which is relatively simple. This reduces
the overall effort required. The transformations used need only preserve the
variable dependence relation, and therefore need not be meaning preserving in
the traditional sense. We show how this relaxed meaning further simplifies
the transformation phase of the approach. Finally, we present the results of
an empirical study into the performance of the system. Evolutionary testing is a search based approach to the automated
generation of systematic test data, in which the search is guided by the test
data adequacy criterion. Two problems for evolutionary testing are the large
size of the search space and structural impediments in the implementation of
the program which inhibit the formulation of a suitable fitness function to
guide the search. In this paper we claim that slicing can be used to narrow
the search space and transformation can be applied to the problem of
structural impediments. The talk presents examples of how these two
techniques have been successfully employed to make evolutionary testing both
more efficient and more effective. This paper extends a previously developed intraproce- dural
denotational program slicer to handle procedures. Using the denotational
approach, slices can be defined in terms of the abstract syntax of the object
language without the need of a control flow graph or similar intermediate
structure. The algorithm presented here is capable of correctly handling the
interplay between function and procedure calls, side-effects, and
short-circuit expression evaluation. The ability to deal with these features
is required in reverse engineering of legacy systems, where code often
contains side-effects. This paper describes a higher-order logic with fine-grained
intensionality (FIL). Unlike traditional Montogovian type theory,
intensionality is treated as basic. rather than derived through possible
worlds. This allows for fine-grained intensionality without impossible
worlds. Possible worlds and modalities are defined algebraically. The proof
theory for FIL is given as a set of tableau rules, and an algebraic model
theory is specified. The proof theory is shown to be sound relative to this
model theory. FIL avoids many of the problems created by classical
course-grained intensional logics that have been used in formal and
computational semantics. presents Property Theory with Curry Typing (PTCT) where the
language of terms and well-formed formulĂŠ are joined by a language of
types. In addition to supporting fine-grained intensionality, the basic
theory is essentially first-order, so that implementations using the theory
can apply standard first-order theorem proving techniques. Some extensions to
the type theory are discussed, type polymorphism, and enriching the system
with sufficient number theory to account for quantifiers of number, such as
âmost.â This paper introduces a notation for business process modelling
based on flownominal expressions, and shows how it can be used for static
verification of business processes, under the assumption of single instance
executions, by evaulationg them over boolean relations. Its main advantage is
simplicity, but it is also more restrictive than other approaches because it
can only indiciate those input patterns that can cause it to enter an
infinite loop, or resource starvation. Nevertheless, this is useful because
it can help isolate problems at an early stage, prior to running any dynamic
simulations. This paper introduces backward conditioning. Like forward
conditioning (used in conditioned slicing), backward conditioning consists of
specialising a program with respect to a condition inserted into the program.
However, whereas forward conditioning deletes statements which are not
executed when the initial state satisfies the condition, backward
conditioning deletes statements which cannot cause execution to enter a state
which satisfies the condition. The relationship between backward and forward
conditioning is reminiscent of the relationship between backward and forward
slicing. Forward conditioning addresses program comprehension questions of
the form âwhat happens if the program starts in a state satisfying
condition c?â, whereas backward conditioning addresses questions of the
form âwhat parts of the program could potentially lead to the program
arriving in a state satisfying condition c?â. The paper illustrates the use
of backward conditioning as a program comprehension assistant and presents an
algorithm for constructing backward conditioned
programs. In this paper we present a framework for constructing
hyperintensional semantics for natural language. On this approach, the axiom
of extensionality is discarded from the axiom base of a logic. Weaker
conditions are specified for the connection between equivalence and identity
which prevent the reduction of the former relation to the latter. In
addition, by axiomatising an intensional number theory we can provide an
internal account of proportional cardinality quantifiers, like most.
We use a (pre-)lattice defined in terms of a (pre-)order that models the
entailment relation. Possible worlds/situations/indices are then prime
filters of propositions in the (pre-)lattice. Truth in a world/situation is
then reducible to membership of a prime filter. We show how this approach can
be implemented within (i) an intensional higher-order type theory, and (ii)
first-order property theory. Slicing has been shown to be a useful program abstraction
technique, with applications at many points in the software development
life-cycle, particularly as a tool to assist software evolution.
Unfortunately, slicing algorithms scale up rather poorly, diminishing the
applicability of slicing in practise. In applications where many slices are
required from a largely unchanging system, incremental approaches to the
construction of dependence information can be used, ensuring that slices are
constructed speedily. However, for some applications, the only way to compute
slices within effective time constraints will be to trade precision for
speed. This approach has been successfully applied to a number of other
computationally expensive source code analysis techniques, most notably
point-to analysis. This paper introduces a theory for trading precision for
speed in slicing based upon âblobbing togetherâ, or âcoarseningâ,
several individual Control Flow Graph nodes. The theory defines the
properties which should be possessed by a logical calculus for
âcoarseningâ (coalescing several nodes in a region into a single
representative of the region). The theory is illustrated with a case study
which presents a calculus for R-coarsening, and a consistent and complete set
of inference rules which compromise precision for
speed. shows how analysis of programs in terms of pre- and post-
conditions can be improved using a generalisation of conditioned program
slicing called pre/post conditioned slicing. Such conditions play an
important role in program comprehension, reuse, verification and
re-engineering. Fully automated analysis is impossible because of the
inherent undecidability of pre- and post- conditions. The method presented
here reformulates the problem to circumvent this. The reformulation is
constructed so that programs which respect the pre- and post-conditions
applied to them have empty slices. For those which do not respect the
conditions, the slice contains statements which could potentially break the
conditions. This separates the automatable part of the analysis from the
human analysis. Conditioned slicing is a powerful generalisation of static and
dynamic slicing which has applications to many problems in software
maintenance and evolution, including re-use, re-engineering and program
comprehension. However, there has been relatively little work on the
implementation of conditioned slicing. Algorithms for implementing
conditioned slicing necessarily involve reasoning about the values of program
predicates in certain sets of states derived from the conditioned slicing
criterion, making implementation particularly demanding. This paper
introduces ConSIT, a conditional slicing system which is based upon
conventional static slicing, symbolic execution and theorem proving. ConSIT
is the first fully automated implementation of conditioned
slicing. In this paper, an approach is described which mixes testing,
slicing, transformation and program verification to investigate speculative
hypotheses concerning a program formulated during program comprehension
activity. Our philosophy is that such hypotheses (which are typically
undecidable) can, in some sense, be âansweredâ by a partly automated
system which returns neither âtrueâ nor âfalseâ, but a program (the
âtest programâ) which computes the answer. The motivation for this
philosophy is the way in which, as we demonstrate, static analysis and
manipulation technology can be applied to ensure that the resulting program
is significantly simpler than the original program, thereby simplifying the
process of investigating the original hypothesis. It is possible to use a combination of classical logic and
dependent types to represent natural language discourse and singular anaphora
(Fox 1994b). In this paper, these ideas are extended to account for some
cases of plural anaphora. In the theory described universal quantification
and conditionals give rise to a context in which singular referents within
its scope are transformed into plurals. These ideas are implemented in
axiomatic Property Theory (Turner 1992) extended with plurals (Fox 1993),
giving a treatment of some examples of singular and plural anaphora in a
highly intensional, weakly typed, classical, first-order
logic. Since Aristotle, it has been accepted that the appropriate
interpretation of sentences is as propositions, and that general terms should
be interpreted as properties, distinct from propositions. Recent proposals
for natural language semantics have used constructive type theories such as
Martin-Löfâs Type Theory MLTT (Martin-Löf 1982, 1984) which treat
anaphora and âdonkeyâ sentences using dependent types (Sundholm 1989,
Ranta 1991, Davila 1994). However, MLTT conflates the notions of proposition
and property. This work shows how, within Property Theory, dependent-type
expressions representing natural language discourse can be mapped
systematically into first-order expressions with a classical notion of
propositionhood, distinct from that of properties. This paper reappraises Landmanâs formal theory of intensional
individualsâindividuals under roles, or guises (Landman 1989)âwithin
property theory (PT) (Turner 1992). As many of Landmanâs axioms exist to
overcome the strong typing of his representation, casting his ideas in weakly
typed PT produces a simpler theory. However, there is the possibility of an
even greater simplification: if roles, or guises, are represented with
property modifiers then there is no need for Landmanâs intensional
individuals. Landmanâs argument against the use of property modifiers is
re-examined, and shown to be mistaken. This paper describes a system for parsing English into
Property-theoretic semantics, using an attribute-value grammar and a
bi-directional chart parser, which then translates this representation into a
relational calculus (SQL) query which can be presented to a database
( textsc Ingres) for evaluation. The Property Theory used is a highly
intensional first-order theory, which avoids some of the problems of
higher-order intensional logics. The paper describes a system in which a chart parser translates
a question in English into a propositional representation in a non-monotonic
logic. A âcontext machineâ uses this representation to extract salient
statements from a knowledge base. A tableau theorem prover then takes these
statements and attempts to prove the proposition associated with the original
question. If the proof fails, the reason for failure can be used to provide a
relevant helpful answer. This paper describes a system for parsing English into
Property-theoretic semantics, using an attribute-value grammar and a
bi-directional chart parser, which then translates this representation into a
relational calculus (SQL) query which can be presented to a database INGRES
for evaluation. The query was optimised using techniques from resolution
theorem proving.Conference Proceedings (52)
Conference presentation. The paper of the same name is based on
this talk. Conference slides: some thoughts on the Ought Implies Can
puzzle, and a meta-level proposal. Conference talk on which âIn Defense of Axiomatic Semanticsâ
is based. When developing formal theories of the meaning of language, it
is appropriate to consider how apparent paradoxes and conundrums of language
are best resolved. But if we base our analysis on a small sample of data then
we may fail to take into account the influence of other aspects of meaning on
our intuitions. Here we consider the so-called Good Samaritan Paradox (Prior,
1958), where we wish to avoid any implication that there is an obligation to
rob someone from âYou must help a robbed manâ. We argue that before
settling on a formal analysis of such sentences, we should consider examples
of the same form, but with intuitively different entailmentsâsuch as âYou
must use a clean knifeââand also actively seek other examples that
exhibit similar contrasts in meaning, even if they do not exemplify the
phenomena that is under investigation. This can refine our intuitions and
help us to attribute aspects of interpretation to the various facets of
meaning. In the case of indicative sentences, broadly speaking there is
some consensus about how to approximate a range of phenomena by appeal to
truth conditional semantics and various forms of predicate logic (modulo
differences in theoretical framework and philosophical taste). Even when
these approximations fall short, they can help provide a background against
which the behaviour of more recalcitrant data can be better understood. In
the case of imperatives, there have been various proposals for their formal
semantics. Unfortunately, these theories are often presented without all the
relevant formal details, and may rely on complex and problematic notions,
such as actions, cause and effect. This can make it difficult to compare and
evaluate such theories, or discern any general consensus about how to address
a given phenomena. The current proposal seeks to capture the formal logical
behaviour of core imperatives by way of inference rules over propositional
satisfaction criteria. One objective is to find a level of abstraction which
avoids troublesome notions such as actions and causality, but with sufficient
expressive power to capture key intuitions about the meaning of imperatives.
In addition to giving an informative analysis, the hope is that this will
provide a baseline of clearly formulated and generally accepted patterns of
behaviour that can be used to evaluate other proposals, and help us
understand more recalcitrant data. An introduction to the notions of program slicing and program
conditioning. In previous work we have developed Property Theory with Curry
Typing (PTCT), an intensional first-order logic for natural language
semantics. PTCT permits fine-grained specifications of meaning. It also
supports polymorphic types and separation types. We develop an intensional
number theory within PTCT in order to represent proportional generalized
quantifiers like most, and we suggest a dynamic type-theoretic approach to
anaphora and ellipsis resolution. Here we extend the type system to include
product types, and use these to define a permutation function that generates
underspecified scope representations within PTCT. We indicate how filters can
be added to encode constraints on possible scope readings. Our account offers
several important advantages over other current theories of
underspecification. In this paper we show that by adding Curry typing to a
first-order property theory it is possible to represent the full range of
generalized quantifiers (GQs) corresponding to natural language determiners.
We characterize GQs as property terms that specify cardinality relations
between properties (or separation types). We also generate underspecified
quantifier scope representations within the representation language, rather
than through meta-language devices, as in most current treatments of
underspecification (Reyle, 1993; Bos, 1995; Blackburn & Bos, 2003;
Copestake, Flickinger, & Sag, 1997). In this talk I shall present Property Theory with Curry Typing
(PTCT), an intensional first-order theory for natural language semantics
developed by myself and Shalom Lappin. PTCT permits fine-grained
specifications of meaning. It also supports polymorphic types and separation
types. We have developed an intensional number theory within PTCT in order to
represent proportional generalized quantifiers like "most". We use the type
system and our treatment of generalized quantifiers in natural language to
construct a type-theoretic approach to pronominal anaphora and ellipsis. We
have also developed a theory of underspecification that is expressed within
the term language of the theory. The talk will focus on the basics of PTCT
itself, and outline the treatment of anaphora and ellipsis. If there is time,
a sketch of our treatment of underspecification may also be
given. We present an approach to anaphora and ellipsis resolution in
which pronouns and elided structures are interpreted by the dynamic
identification in discourse of type constraints on their semantic
representations. The content of these conditions is recovered in context from
an antecedent expression. The constraints define separation types (sub-types)
in Property Theory with Curry Typing (PTCT), an expressive first-order logic
with Curry typing that we have proposed as a formal framework for natural
language semantics. A highly intensional first-order logic will be presented which
incorporates an expressive type system, including general function spaces,
separation types and type polymorphism. Although first-order in power, the
logic is sufficiently expressive to capture aspects of natural language
semantics that are often characterised as requiring a higher-order analysis.
Aspects of the model theory for this logic will also be
discussed. We present Property Theory with Curry Typing (PTCT), an
intensional first-logic for natural language semantics. PTCT permits
fine-grained specifications of meaning. It also supports polymorphic,
separation, and dependent types. We develop an intensional number theory with
PTCT in order to represent proportional generalized quantifiers like
most. We use the type system and our treatment of generalized
quantifiers in natural language to construct a type-theoretic approach to
pronominal anaphora that avoids some of the difficulties that undermine
previous type-theoretic analyses of this phenomenon.Selected Talks (19)
Episodic Logic (EL), as described in Chung Hee Hwang and Lenhart
K. Schubertâs paper âEpisodic Logic: a Situational Logic for Natural
Language Processingâ (Hwang & Schubert1993), is a formal theory of natural
language semantics which has an extensive coverage of phenomena. The theory
has been applied effectively in various software implementations of natural
language systems. This paper is not intended to undermine this theoretical
and applied work. It aims merely to illustrate some problems with the
informal intuitions that purport to explain and justify the formal theory of
EL. In particular, this paper criticises the view that we should think of
events as situations (episodes) which can be completely characterised by
natural language sentences. I argue that: (1) there are no genuine natural
language examples which require it; (2) it results in a loss of
expressiveness; and (3) it leads to problems when giving the logical form of
causal statements. I suggest that the motivating example can be dealt with
adequately using a (neo-)Davidsonian approach. That these arguments do not
undermine the formal theory of EL and its application in various systems can
be seen from the fact (discussed at the end of Section II) that the formal
theory appears to make no use of the problematic notions; they only appear in
its informal motivation. In effect, EL can be seen to provide a
neo-Davidsonian theory. This paper is structured as follows: Section I
introduces those aspects of EL relevant for the discussion; Section II
presents detailed criticisms; Section III re-appraises the (neo-)Davidsonian
approach to events, and shows how it can cope with Hwang and Schubertâs
motivating example; and Section IV makes some concluding
remarks.Selected Reports & Whitepapers (13)
Stand-alone âcomputer aided learningâ resources, based upon
an existing printed volume. Subject guide LaTeX classfile (software) for automating the
production of distance learning materials for the University of London in
their house style. A Subject Guide for the University of Londonâs
undergraduate programme for external students. A Subject Guide for the University of Londonâs
undergraduate programme for external students.Distance Learning (4)
The Thesis Mass Terms and Plurals in Property Theory is
concerned with extending a weak axiomatic theory of Truth Propositions, and
Properties, with fine grained intensionality (PT), to represent the semantics
of natural language (NL) sentences involving plurals and mass terms. The use
of PT as a semantic theory for NL eases the problem of modeling the
behaviours of these phenomena by removing the artificial burdens imposed by
strongly typed, model-theoretic semantic theories. By deliberately using
incomplete axioms, following the example set by basic PT, it is possible to
remain uncommitted about: the existence of atomic mass terms; the existence
of a âbottom elementâ (a part of all terms) as a denotation of NL
nominals; and the propositionhood (and hence truth of) sentences such as
âthe present King of France is bald.â Landmanâs theory concerning the
representation of individuals under roles, or guises is reappraised in terms
of property modifiers. This is used to offer a solution to the problem of
distinguishing predication of equal fusions of distinct properties, and the
control of distribution into mereological terms, when used to represent NL
mass nominals, without assuming an homogeneous extension. The final theory
provides a uniform framework for representing sentences involving both mass
and count nominals.PhD Thesis (1!)
Music (7)
Created: 2021-11-19 Fri 08:17