liu.seSearch for publications in DiVA
Change search
Refine search result
123 1 - 50 of 135
CiteExportLink to result list
Permanent link
Cite
Citation style
  • apa
  • harvard1
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Rows per page
  • 5
  • 10
  • 20
  • 50
  • 100
  • 250
Sort
  • Standard (Relevance)
  • Author A-Ö
  • Author Ö-A
  • Title A-Ö
  • Title Ö-A
  • Publication type A-Ö
  • Publication type Ö-A
  • Issued (Oldest first)
  • Issued (Newest first)
  • Created (Oldest first)
  • Created (Newest first)
  • Last updated (Oldest first)
  • Last updated (Newest first)
  • Disputation date (earliest first)
  • Disputation date (latest first)
  • Standard (Relevance)
  • Author A-Ö
  • Author Ö-A
  • Title A-Ö
  • Title Ö-A
  • Publication type A-Ö
  • Publication type Ö-A
  • Issued (Oldest first)
  • Issued (Newest first)
  • Created (Oldest first)
  • Created (Newest first)
  • Last updated (Oldest first)
  • Last updated (Newest first)
  • Disputation date (earliest first)
  • Disputation date (latest first)
Select
The maximal number of hits you can export is 250. When you want to export more records please use the 'Create feeds' function.
  • 1.
    Bialek, Lukasz
    et al.
    Institute of Informatics, University of Warsaw, Warsaw, Poland.
    Szalas, Andrzej
    Linköping University, Department of Computer and Information Science, Artificial Intelligence and Intergrated Computer systems. Linköping University, Faculty of Science & Engineering.
    Lightweight Reasoning with Incomplete and Inconsistent Information: a Case Study2014In: 2014 IEEE/WIC/ACM International Joint Conferences on  (Volume:3 ) Web Intelligence (WI) and Intelligent Agent Technologies (IAT),, IEEE , 2014, Vol. 3, p. 325-332Conference paper (Refereed)
    Abstract [en]

    Dealing with heterogeneous information sources and reasoning techniques allowing for incomplete and inconsistent information is one of current challenges in the area of knowledge representation and reasoning. We advocate for 4QL, a rule-based query language, as a proper tool allowing one to address these challenges. To justify this point of view we discuss a rescue robotics scenario for which a simulator has been developed and tested. In particular, we present a planner using 4QL and, therefore, capable to deal with lack of knowledge and inconsistencies. Through the case study we show that our approach allows one to use lightweight knowledge representation tools: due to the use of 4QL tractability of modeling and reasoning is guaranteed and high usability is achieved.

  • 2.
    Bolc, Leonard
    et al.
    Institute of Computer Science, Polish Academy of Sciences.
    Dziewicki, Krzysztof
    Rychlik, Piotr
    Szalas, Andrzej
    University of Warsaw.
    Wnioskowanie w logikach nieklasycznych: Automatyzacja wnioskowania1998Book (Other academic)
  • 3.
    Bolc, Leonard
    et al.
    Institute of Computer Science, Polish Academy of Sciences.
    Dziewicki, Krzysztof
    Rychlik, Piotr
    Szalas, Andrzej
    University of Warsaw.
    Wnioskowanie w logikach nieklasycznych: Podstawy teoretyczne1995Book (Other academic)
    Abstract [pl]

    Prezentowana książka stanowi drugi tom dwutomowej monografii poświeconej wnioskowaniu w logikach nieklasycznych. Opracowanie to obejmuje takie formalizmy jak logiki modalne, logiki wielowartościowe i mechanizmy wnioskowania niemonotonicznego. W pierwszym tomie przedstawione zostały podstawy teoretyczne wnioskowania w wybranych logikach nieklasycznych. Tom drugi poświęcony jest zagadnieniom automatyzacji procesu wnioskowania.

  • 4.
    Bolc, Leonard
    et al.
    Institute of Computer Science, Polish Academy of Sciences.
    Szalas, AndrzejUniversity of Warsaw.
    Time and Logic: A Computational Approach1995Collection (editor) (Other academic)
    Abstract [en]

    Time and logic are central driving concepts in science and technology. In this book, some of the major current developments in our understanding and application of temporal logic are presented in computational terms. "Time and Logic: A Computational Approach" should be a useful sourcebook for those within the specific field of temporal logic, as well as providing valuable introductory material for those seeking an entry into this increasingly important area of theoretical computing.; The emphasis of the book is on presenting a broad range of approaches to computational applications. The techniques used will also be applicable in many cases to formalize beyond temporal logic alone, and it is hoped that adaptation to many different logics of programmes will be facilitated. Throughout, the authors have kept implementation-oriented solutions in mind.; The book begins with an introduction to the basic ideas of temporal logic. Successive chapters then examine particular aspects of the temporal theoretical computing domain, relating their applications to familiar areas of research, such as stochastic process theory, automata theory, established proof systems, model checking, relational logic and classical predicate logic. This should be a useful addition to the library of all theoretical computer scientists, providing a synthesis of well established results in temporal logic with the most up-to-date findings of some of the world's leading theoreticians.

  • 5.
    Cao, Son Thanh
    et al.
    Vinh University, Vietnam.
    Nguyen, Anh Linh
    Institute of Informatics, Warsaw University.
    Szalas, Andrzej
    Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab. Linköping University, The Institute of Technology.
    On the Web Ontology Rule Language OWL 2 RL2011In: Proceedings of the 3rd International Conference on Computational Collective Intelligence, Technologies and Applications (ICCCI) / [ed] Piotr Jedrzejowicz, Ngoc Thanh Nguyen and Kiem Hoang, Springer Berlin/Heidelberg, 2011, p. 254-264Conference paper (Refereed)
    Abstract [en]

    It is known that the OWL 2 RL Web Ontology Language Profile has PTime data complexity and can be translated into Datalog. However, a knowledge base in OWL 2 RL may be unsatisfiable. The reason is that, when translated into Datalog, the result may consist of a Datalog program and a set of constraints in the form of negative clauses. In this paper we first identify a maximal fragment of OWL 2 RL called OWL 2 RL + with the property that every knowledge base expressed in this fragment can be translated into a Datalog program and hence is satisfiable. We then propose some extensions of OWL 2 RL and OWL 2 RL +  that still have PTime data complexity.

  • 6.
    Cao, Son Thanh
    et al.
    Vinh University, Nghe An, Vietnam .
    Nguyen, Linh Anh
    University of Warsaw, Poland and VNU University of Engineering and Technology, Hanoi, Vietnam .
    Szalas, Andrzej
    Linköping University, Department of Computer and Information Science, Artificial Intelligence and Integrated Computer Systems. Linköping University, The Institute of Technology. University of Warsaw, Poland.
    WORL: a nonmonotonic rule language for the semantic web2014In: Vietnam Journal of Computer Science, ISSN 2196-8888, Vol. 1, no 1, p. 57-69Article in journal (Refereed)
    Abstract [en]

    We develop a new Web ontology rule language, called WORL, which combines a variant of OWL 2 RL with eDatalog ¬ . We allow additional features like negation, the minimal number restriction and unary external checkable predicates to occur at the left-hand side of concept inclusion axioms. Some restrictions are adopted to guarantee a translation into eDatalog ¬ . We also develop the well-founded semantics and the stable model semantics for WORL as well as the standard semantics for stratified WORL (SWORL) via translation into eDatalog ¬ . Both WORL with respect to the well-founded semantics and SWORL with respect to the standard semantics have PTime data complexity. In contrast to the existing combined formalisms, in WORL and SWORL negation in concept inclusion axioms is interpreted using nonmonotonic semantics.

  • 7.
    Cao, Son Thanh
    et al.
    Vinh Univ, Vietnam.
    Nguyen, Linh Anh
    Institute of Informatics, Warsaw University.
    Szalas, Andrzej
    Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab. Linköping University, The Institute of Technology.
    WORL: A Web Ontology Rule Language2011In: Proceedings of the 3rd International Conference on Knowledge and Systems Engineering (KSE), IEEE , 2011, p. 32-39Conference paper (Refereed)
    Abstract [en]

    We develop a Web ontology rule language, called WORL, which combines a variant of OWL 2 RL with eDatalog-with-negation. We disallow the features of OWL 2 RL that play the role of constraints (i.e., the ones that are translated to negative clauses), but allow additional features like negation, the minimal number restriction and unary external checkable predicates to occur in the left hand side of concept inclusion axioms. Some restrictions are adopted to guarantee a translation into eDatalog-with-negation. We also develop the well-founded semantics for WORL and the standard semantics for stratified WORL (SWORL) via translation into eDatalog-with-negation. Both WORL and SWORL have PTime data complexity. In contrast to the existing combined formalisms, in WORL and SWORL negation in concept inclusion axioms is interpreted using nonmonotonic semantics.

  • 8.
    Cao, S.T.
    et al.
    Faculty of Information Technology, Vinh University, 182 Le Duan street, Vinh Nghe An, Viet Nam; Institute of Informatics, University of of Warsaw, Banacha 2, 02-097 Warsaw, Poland.
    Nguyen, L.A.
    Institute of Informatics, University of of Warsaw, Banacha 2, 02-097 Warsaw, Poland; Faculty of Information Technology, VNU University of of Engineering and Technology, 144 Xuan Thuy, Hanoi, Viet Nam.
    Szalas, Andrzej
    Linköping University, Department of Computer and Information Science, Artificial Intelligence and Integrated Computer Systems. Linköping University, The Institute of Technology. Institute of Informatics, University of of Warsaw, Banacha 2, 02-097 Warsaw, Poland.
    The web ontology rule language OWL 2 RL+ and its extensions2014In: Transactions on Computational Collective Intelligence XIII / [ed] Gerhard Goos, Juris Hartmanis, and Jan van Leeuwen, Springer Verlag (Germany) , 2014, Vol. 8342, p. 152-175Conference paper (Refereed)
    Abstract [en]

    It is known that the OWL 2RL Web Ontology Language Profile has PTime data complexity and can be translated into Datalog. However, the result of translation may consist of a Datalog program and a set of constraints in the form of negative clauses. Therefore, a knowledge base in OWL 2RL may be unsatisfiable. In the current paper we first identify a maximal fragment of OWL 2RL, called OWL 2RL+, with the property that every knowledge base expressed in OWL2RL+ can be translated to a Datalog program and hence is satisfiable. We then propose some extensions of OWL 2RL and OWL 2RL + that still have PTime data complexity. © 2014 Springer-Verlag Berlin Heidelberg.

  • 9. Cunningham, R. J.
    et al.
    Nonnengart, Andreas
    Max-Planck-Institut Informatik.
    Szalas, Andrzej
    University of Warsaw.
    A Compositional Method for the Design and Proof of Asynchronous Processes1987In: Proceedings of the 4th Annual ESPRIT Conference (ESPRIT), Amsterdam: North-Holland , 1987, p. 566-580Conference paper (Refereed)
  • 10.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Dunin-Keplicz, Barbara
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Dynamics of approximate information fusion2007In: Proceedings of the International Conference on Rough Sets and Emerging Intelligent Systems Paradigms (RSEISP), Heidelberg, Germany: Springer Berlin/Heidelberg, 2007, p. 668-677Conference paper (Refereed)
    Abstract [en]

    The multi-agent system paradigm has proven to be a useful means of abstraction when considering distributed systems with interacting components. It is often the case that each component may be viewed as an intelligent agent with specific and often limited perceptual capabilities. It is also the case that these agent components may be used as information sources and such sources may be aggregated to provide global information about particular states, situations or activities in the embedding environment. This paper investigates a framework for information fusion based on the use of generalizations of rough set theory and the use of dynamic logic as a basis for aggregating similarity relations among objects where the similarity relations represent individual agents perceptual capabilities or limitations. As an added benefit, it is shown how this idea may also be integrated into description logics.

  • 11.
    Doherty, Patrick
    et al.
    Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab. Linköping University, The Institute of Technology.
    Dunin-Keplicz, Barbara
    Institute of Informatics, Warsaw University.
    Szalas, Andrzej
    Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab. Linköping University, The Institute of Technology.
    Tractable model checking for fragments of higher-order coalition logic2011In: Proceedings of the 10th International Conference on Autonomous Agents and Multiagent Systems - Volume 2 / [ed] Liz Sonenberg, Peter Stone, Kagan Tumer, Pinar Yolum, Richland: AAAI Press, 2011, p. 743-750Conference paper (Refereed)
    Abstract [en]

    A number of popular logical formalisms for representing and reasoning about the abilities of teams or coalitions of agents have been proposed beginning with the Coalition Logic (CL) of Pauly. Ågotnes et al introduced a means of succinctly expressing quantification over coalitions without compromising the computational complexity of model checking in CL by introducing Quantified Coalition Logic (QCL). QCL introduces a separate logical language for characterizing coalitions in the modal operators used in QCL. Boella et al, increased the representational expressibility of such formalisms by introducing Higher-Order Coalition Logic (HCL), a monadic second-order logic with special set grouping operators. Tractable fragments of HCL suitable for efficient model checking have yet to be identified. In this paper, we relax the monadic restriction used in HCL and restrict ourselves to the diamond operator. We show how formulas using the diamond operator are logically equivalent to second-order formulas. This permits us to isolate and define well-behaved expressive fragments of second-order logic amenable to model-checking in PTime. To do this, we appeal to techniques used in deductive databases and quantifier elimination. In addition, we take advantage of the monotonicity of the effectivity function resulting in exponentially more succinct representation of models. The net result is identification of highly expressible fragments of a generalized HCL where model checking can be done efficiently in PTime.

  • 12.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Grabowski, M
    Lukaszewicz, Witold
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Towards a framework for approximate ontologies2003In: Fundamenta Informaticae, ISSN 0169-2968, E-ISSN 1875-8681, Vol. 57, no 2-4, p. 147-165Article in journal (Refereed)
    Abstract [en]

    Currently, there is a great deal of interest in developing tools for the generation and use of ontologies on the WWW. These knowledge structures are considered essential to the success of the semantic web, the next phase in the evolution of the WWW. Much recent work with ontologies assumes that the concepts used as building blocks are crisp as opposed to approximate. It is a premise of this paper that approximate concepts and ontologies will become increasingly more important as the semantic web becomes a reality. We propose a framework for specifying, generating and using approximate ontologies. More specifically, (1) a formal framework for defining approximate concepts, ontologies and operations on approximate concepts and ontologies is presented. The framework is based on intuitions from rough set theory, (2) algorithms for automatically generating approximate ontologies from traditional crisp ontologies or from large data sets together with additional knowledge are presented. The knowledge will generally be related to similarity measurements between individual objects in the data sets, or constraints of a logical nature which rule out particular constellations of concepts and dependencies in generated ontologies. The techniques for generating approximate ontologies are parameterizable. The paper provides specific instantiations and examples.

  • 13.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Kachniarz, Jaroslaw
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Meta-queries on deductive databases1999In: Fundamenta Informaticae, ISSN 0169-2968, E-ISSN 1875-8681, Vol. 40, no 1, p. 17-30Article in journal (Refereed)
    Abstract [en]

    We introduce the notion of a meta-query on relational databases and a technique which can be used to represent and solve a number of interesting problems from the area of knowledge representation using logic. The technique is based on the use of quantifier elimination and may also be used to query relational databases using a declarative query language called SHQL (Semi-Horn Query Language), introduced in [6]. SHQL is a fragment of classical first-order predicate logic and allows us to define a query without supplying its explicit definition. All SHQL queries to the database can be processed in polynomial time (both on the size of the input query and the size of the database). We demonstrate the use of the technique in problem solving by structuring logical puzzles from the Knights and Knaves domain as SHQL meta-queries on relational databases. We also provide additional examples demonstrating the flexibility of the technique. We conclude with a description of a newly developed software tool, The Logic Engineer, which aids in the description of algorithms using transformation and reduction techniques such as those applied in the meta-querying approach.

  • 14.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Kachniarz, Jaroslaw
    Soft Computer Consultants, USA.
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Using Contextually Closed Queries for Local Closed-World Reasoning in Rough Knowledge Databases2004In: Rough-Neural Computing: Techniques for Computing with Words / [ed] Andrzej Skowron,Lech Polkowski ,Sankar K Pal, Springer , 2004, p. 219-250Chapter in book (Other academic)
    Abstract [en]

    Soft computing comprises various paradigms dedicated to approximately solving real-world problems, e.g., in decision making, classification or learning; among these paradigms are fuzzy sets, rough sets, neural networks, and genetic algorithms.

    It is well understood now in the soft computing community that hybrid approaches combining various paradigms provide very promising attempts to solving complex problems. Exploiting the potential and strength of both neural networks and rough sets, this book is devoted to rough-neurocomputing which is also related to the novel aspect of computing based on information granulation, in particular to computing with words. It provides foundational and methodological issues as well as applications in various fields.

  • 15.
    Doherty, Patrick
    et al.
    Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab. Linköping University, The Institute of Technology.
    Kertes, Steven
    Stanford University.
    Magnusson, Martin
    Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab. Linköping University, The Institute of Technology.
    Szalas, Andrzej
    Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab. Linköping University, The Institute of Technology.
    Towards a logical analysis of biochemical pathways2004In: Proceedings of the 9th European Conference on Logics in Artificial Intelligence (JELIA) / [ed] José Júlio Alferes and João Alexandre Leite, Springer , 2004, Vol. 3229, p. 667-679Conference paper (Refereed)
    Abstract [en]

    Biochemical pathways or networks are generic representations used to model many different types of complex functional and physical interactions in biological systems. Models based on experimental results are often incomplete, e.g., reactions may be missing and only some products are observed. In such cases, one would like to reason about incomplete network representations and propose candidate hypotheses, which when represented as additional reactions, substrates, products, would complete the network and provide causal explanations for the existing observations. In this paper, we provide a logical model of biochemical pathways and show how abductive hypothesis generation may be used to provide additional information about incomplete pathways. Hypothesis generation is achieved using weakest and strongest necessary conditions which represent these incomplete biochemical pathways and explain observations about the functional and physical interactions being modeled. The techniques are demonstrated using metabolism and molecular synthesis examples.

  • 16.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Kertes, Steven
    Stanford University.
    Magnusson, Martin
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Towards a Logical Analysis of Biochemical Reactions (Extended abstract)2004In: Proceedings of the 16th European Conference on Artificial Intelligence (ECAI) / [ed] Ramon López de Mántaras, Lorenza Saitta, Amsterdam: IOS Press, 2004, p. 997-998Conference paper (Refereed)
    Abstract [en]

    We provide a logical model of biochemical reactions and show how hypothesis generation using weakest sufficient and strongest necessary conditions may be used to provide additional information in the context of an incomplete model of metabolic pathways.

  • 17.
    Doherty, Patrick
    et al.
    Linköping University, Department of Computer and Information Science, Artificial Intelligence and Integrated Computer Systems. Linköping University, Faculty of Science & Engineering.
    Kvarnström, Jonas
    Linköping University, Department of Computer and Information Science, Artificial Intelligence and Integrated Computer Systems. Linköping University, Faculty of Science & Engineering.
    Szalas, Andrzej
    Linköping University, Department of Computer and Information Science, Artificial Intelligence and Integrated Computer Systems. Linköping University, Faculty of Science & Engineering.
    Iteratively-Supported Formulas and Strongly Supported Models for Kleene Answer Set Programs2016In: Logics in Artificial Intelligence: 15th European Conference, JELIA 2016, Larnaca, Cyprus, November 9-11, 2016, Proceedings, Springer Publishing Company, 2016, p. 536-542Conference paper (Refereed)
    Abstract [en]

    In this extended abstract, we discuss the use of iteratively-supported formulas (ISFs) as a basis for computing strongly-supported models for Kleene Answer Set Programs (ASPK). ASPK programs have a syntax identical to classical ASP programs. The semantics of ASPK programs is based on the use of Kleene three-valued logic and strongly-supported models. For normal ASPK programs, their strongly supported models are identical to classical answer sets using stable model semantics.  For disjunctive ASPK programs, the semantics weakens the minimality assumption resulting in a classical interpretation for disjunction. We use ISFs to characterize strongly-supported models and show that they are polynomially bounded.

  • 18.
    Doherty, Patrick
    et al.
    Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab. Linköping University, Department of Computer and Information Science, UASTECH - Autonomous Unmanned Aircraft Systems Technologies. Linköping University, The Institute of Technology.
    Kvarnström, Jonas
    Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab. Linköping University, The Institute of Technology.
    Szalas, Andrzej
    Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab. Linköping University, The Institute of Technology.
    Temporal Composite Actions with Constraints2012In: Proceedings of the 13th International Conference on Principles of Knowledge Representation and Reasoning (KR), AAAI Press, 2012, p. 478-488Conference paper (Refereed)
    Abstract [en]

    Complex mission or task specification languages play a fundamentally important role in human/robotic interaction.  In realistic scenarios such as emergency response, specifying temporal, resource and other constraints on a mission is an essential component due to the dynamic and contingent nature of the operational environments. It is also desirable that in addition to having a formal semantics, the language should be sufficiently expressive, pragmatic and abstract. The main goal of this paper is to propose a mission specification language that meets these requirements. It is based on extending both the syntax and semantics of a well-established formalism for reasoning about action and change, Temporal Action Logic (TAL), in order to represent temporal composite actions with constraints.  Fixpoints are required to specify loops and recursion in the extended language. The results include a sound and complete proof theory for this extension. To ensure that the composite language constructs are adequately grounded in the pragmatic operation of robotic systems, Task Specification Trees (TSTs) and their mapping to these constructs are proposed. The expressive and pragmatic adequacy of this approach is demonstrated using an emergency response scenario.

  • 19.
    Doherty, Patrick
    et al.
    Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab. Linköping University, Department of Computer and Information Science, UASTECH - Autonomous Unmanned Aircraft Systems Technologies.
    Lukaszewicz, W.Andrzej, SkowronSzalas, AndrzejLinköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab. Linköping University, The Institute of Technology.
    Knowledge Representation and Approximate Reasoning2003Conference proceedings (editor) (Other academic)
  • 20.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Lukaszewicz, Witold
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Skowron, Andrzej
    Institute of Mathematics Warsaw University.
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Approximation Transducers and Trees: A Technique for Combining Rough and Crisp Knowledge2004In: Rough-Neural Computing: Techniques for Computing with Words / [ed] Andrzej Skowron,Lech Polkowski ,Sankar K Pal, Berlin, Heidelberg, New York: Springer , 2004, p. 189-218Chapter in book (Other academic)
    Abstract [en]

    Soft computing comprises various paradigms dedicated to approximately solving real-world problems, e.g., in decision making, classification or learning; among these paradigms are fuzzy sets, rough sets, neural networks, and genetic algorithms.

    It is well understood now in the soft computing community that hybrid approaches combining various paradigms provide very promising attempts to solving complex problems. Exploiting the potential and strength of both neural networks and rough sets, this book is devoted to rough-neurocomputing which is also related to the novel aspect of computing based on information granulation, in particular to computing with words. It provides foundational and methodological issues as well as applications in various fields.

  • 21.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Lukaszewicz, Witold
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Skowron, Andrzej
    Faculty of Mathematics, Informatics and Mechanics Warsaw University.
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Knowledge Representation Techniques.: a rough set approach2006Book (Other academic)
    Abstract [en]

    The basis for the material in this book centers around a long term research project with autonomous unmanned aerial vehicle systems. One of the main research topics in the project is knowledge representation and reasoning. The focus of the research has been on the development of tractable combinations of approximate and nonmonotonic reasoning systems. The techniques developed are based on intuitions from rough set theory. Efforts have been made to take theory into practice by instantiating research results in the context of traditional relational database or deductive database systems. This book contains a cohesive, self-contained collection of many of the theoretical and applied research results that have been achieved in this project and for the most part pertain to nonmonotonic and approximate reasoning systems developed for an experimental unmanned aerial vehicle system used in the project. This book should be of interest to the theoretician and applied researcher alike and to autonomous system developers and software agent and intelligent system developers.

  • 22.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Lukaszewicz, Witold
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    A characterization result for circumscribed normal logic programs. Revised version accepted for publication: Special issue of honor of H. Rasiowa, Fundamenta Informaticae.1995Report (Other academic)
  • 23.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Lukaszewicz, Witold
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    A reduction result for circumscribed semi-horn formulas1996In: Fundamenta Informaticae, ISSN 0169-2968, E-ISSN 1875-8681, Vol. 28, no 3,4, p. 261-272Article in journal (Refereed)
    Abstract [en]

    Circumscription has been perceived as an elegant mathematical technique for modeling nonmonotonic and commonsense reasoning, but difficult to apply in practice due to the use of second-order formulas. One proposal for dealing with the computational problems is to identify classes of first-order formulas whose circumscription can be shown to be equivalent to a first-order formula. In previous work, we presented an algorithm which reduces certain classes of second-order circumscription axioms to logically equivalent first-order formulas. The basis for the algorithm is an elimination lemma due to Ackermann. In this paper, we capitalize on the use of a generalization of Ackermann's Lemma in order to deal with a subclass of universal formulas called semi-Horn formulas. Our results subsume previous results by Kolaitis and Papadimitriou regarding a characterization of circumscribed definite logic programs which are first-order expressible. The method for distinguishing which formulas are reducible is based on a boundedness criterion. The approach we use is to first reduce a circumscribed semi-Horn formula to a fixpoint formula which is reducible if the formula is bounded, otherwise not. In addition to a number of other extensions, we also present a fixpoint calculus which is shown to be sound and complete for bounded fixpoint formulas.

  • 24.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Lukaszewicz, Witold
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Approximate Databases and Query Techniques for Agents with Heterogenous Perceptual Capabilities2004In: Proceedings of the 7th International Conference on Information Fusion, Mountain View, CA: ISIF , 2004, p. 175-182Conference paper (Refereed)
    Abstract [en]

    In this paper, we propose a framework that provides software and robotic agents with the ability to ask approximate questions to each other in the context of heterogeneous and contextually limited perceptual capabilities. The framework focuses on situations where agents have varying ability to perceive their environments. These limitations on perceptual capability are formalized using the idea of tolerance spaces. It is assumed that each agent has one or more approximate databases where approximate relations are represented using intuitions from rough set theory. It is shown how sensory and other limitations can be taken into account when constructing approximate databases for each respective agent. Complex relations inherit the approximativeness inherent in the sensors and primitive relations used in their definitions. Agents then query these databases and receive answers through the filters of their perceptual limitations as represented by tolerance spaces and approximate queries. The techniques used are all tractable.

  • 25.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Lukaszewicz, Witold
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Approximative Query Techniques for Agents with Heterogeneous Ontologies and Perceptive Capabilities2004In: Proceedings of the 9th International Conference on the Principles of Knowledge Representation and Reasoning / [ed] Didier Dubois, Christopher A. Welty, Mary-Anne Williams, Menlo Park, California: AAAI Press , 2004, p. 459-468Conference paper (Refereed)
    Abstract [en]

    In this paper, we propose a framework that provides software and robotic agents with the ability to ask approximate questions to each other in the context of heterogeneous ontologies and heterogeneous perceptive capabilities.The framework combines the use of logic-based techniques with ideas from approximate reasoning. Initial queries by an agent are transformed into approximate queries using weakest sufficient and strongest necessary conditions on the query and are interpreted as lower and upper approximations on the query. Once the base communication ability is provided, the framework is extended to situations where there is not only a mismatch between agent ontologies, but the agents have varying ability to perceive their environments. This will affect each agent’s ability to ask and interpret results of queries. Limitations on perceptive capability are formalized using the idea of tolerance spaces.

  • 26.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Lukaszewicz, Witold
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    CAKE: A computer aided knowledge engineering technique.2002In: Proceedings of the 15th European Conference on Artificial Intelligence,2002 / [ed] Frank van Harmelen, IOS Press, 2002, p. 220-224Conference paper (Refereed)
    Abstract [en]

    Introduction: Logic engineering often involves the development of modeling tools and inference mechanisms (both standard and non-standard) which are targeted for use in practical applications where expressiveness in representation must be traded off for efficiency in use. Some representative examples of such applications would be the structuring and querying of knowledge on the semantic web, or the representation and querying of epistemic states used with softbots, robots or smart devices. In these application areas, declarative representations of knowledge enhance the functionality of such systems and also provide a basis for insuring the pragmatic properties of modularity and incremental composition. In addition, the mechanisms developed should be tractable, but at the same time, expressive enough to represent such aspects as default reasoning, or approximate or incomplete representations of the environments in which the entities in question are embedded or used, be they virtual or actual. [...]

  • 27.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Lukaszewicz, Witold
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Communication between agents with heterogeneous perceptual capabilities2007In: Information Fusion, ISSN 1566-2535, E-ISSN 1872-6305, Vol. 8, no 1, p. 56-69Article in journal (Refereed)
    Abstract [en]

    In real world applications robots and software agents often have to be equipped with higher level cognitive functions that enable them to reason, act and perceive in changing, incompletely known and unpredictable environments. One of the major tasks in such circumstances is to fuse information from various data sources. There are many levels of information fusion, ranging from the fusing of low level sensor signals to the fusing of high level, complex knowledge structures. In a dynamically changing environment even a single agent may have varying abilities to perceive its environment which are dependent on particular conditions. The situation becomes even more complex when different agents have different perceptual capabilities and need to communicate with each other. In this paper, we propose a framework that provides agents with the ability to fuse both low and high level approximate knowledge in the context of dynamically changing environments while taking account of heterogeneous and contextually limited perceptual capabilities. To model limitations on an agent's perceptual capabilities we introduce the idea of partial tolerance spaces. We assume that each agent has one or more approximate databases where approximate relations are represented using lower and upper approximations on sets. Approximate relations are generalizations of rough sets. It is shown how sensory and other limitations can be taken into account when constructing and querying approximate databases for each respective agent. Complex relations inherit the approximativeness of primitive relations used in their definitions. Agents then query these databases and receive answers through the filters of their perceptual limitations as represented by (partial) tolerance spaces and approximate queries. The techniques used are all tractable. © 2005 Elsevier B.V. All rights reserved.

  • 28.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Lukaszewicz, Witold
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Computing circumscription revisited.1995In: Proceedings of the 14th International Joint Conference on Artificial Intelligence (IJCAI), 1995, p. 1502-1508Conference paper (Refereed)
  • 29.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Lukaszewicz, Witold
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Computing circumscription revisited: A reduction algorithm.1994Report (Other academic)
    Abstract [en]

    In recent years, a great deal of attention has been devoted to logics of "commonsense" reasoning. Among the candidates proposed, circumscription has been perceived as an elegant mathematical technique for modeling nonmonotonic reasoning, but difficult to apply in practice. The major reason for this is the nd-order nature of circumscription axioms and the difficulty in finding proper substitutions of predicate expressions for predicate variables.  One solution to this problem is to compile, where possible, nd-order formulas into equivalent 1st-order formulas. Although some progress has been made using this approach, the results are not as strong as one might desire and they are isolated in nature. In this article, we provide a general method which can be used in an algorithmic manner to reduce circumscription axioms to 1st-order formulas. The algorithm takes as input an arbitrary 2nd-order formula and either returns as output an equivalent 1st-order formula, or terminates with failure. The class of 2nd-order formulas, and analogously the class of circumscriptive theories which can be reduced, provably subsumes those covered by existing results. We demonstrate the generality of the algorithm using circumscriptive theories with mixed quantifiers (some involving Skolemization), variable constants, non-separated formulas, and formulas with n-ary predicate variables. In addition, we analyze the strength of the algorithm and compare it with existing approaches providing formal subsumption results.

  • 30.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Lukaszewicz, Witold
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Computing circumscription revisited: A reduction algorithm1997In: Journal of automated reasoning, ISSN 0168-7433, E-ISSN 1573-0670, Vol. 18, no 3, p. 297-336Article in journal (Refereed)
    Abstract [en]

    In recent years, a great deal of attention has been devoted to logics of common-sense reasoning. Among the candidates proposed, circumscription has been perceived as an elegant mathematical technique for modeling nonmonotonic reasoning, but difficult to apply in practice. The major reason for this is the second-order nature of circumscription axioms and the difficulty in finding proper substitutions of predicate expressions for predicate variables. One solution to this problem is to compile, where possible, second-order formulas into equivalent first-order formulas. Although some progress has been made using this approach, the results are not as strong as one might desire and they are isolated in nature. In this article, we provide a general method that can be used in an algorithmic manner to reduce certain circumscription axioms to first-order formulas. The algorithm takes as input an arbitrary second-order formula and either returns as output an equivalent first-order formula, or terminates with failure. The class of second-order formulas, and analogously the class of circumscriptive theories that can be reduced, provably subsumes those covered by existing results. We demonstrate the generality of the algorithm using circumscriptive theories with mixed quantifiers (some involving Skolemization), variable constants, nonseparated formulas, and formulas with n-ary predicate variables. In addition, we analyze the strength of the algorithm, compare it with existing approaches, and provide formal subsumption results.

  • 31.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Lukaszewicz, Witold
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Computing strongest necessary and weakest sufficient conditions of first-order formulas2001In: 17th International Joint Conference on Artificial Intelligence,2001, San Francisco, CA, USA: Morgan Kaufmann Publishers Inc. , 2001, p. 145-151Conference paper (Refereed)
    Abstract [en]

    A technique is proposed for computing the weakest sufficient (wsc) and strongest necessary (snc) conditions for formulas in an expressive fragment of first-order logic using quantifier elimination techniques. The efficacy of the approach is demonstrated by using the techniques to compute snc's and wsc's for use in agent communication applications, theory approximation and generation of abductive hypotheses. Additionally, we generalize recent results involving the generation of successor state axioms in the propositional situation calculus via snc's to the first-order case. Subsumption results for existing approaches to this problem and a re-interpretation of the concept of forgetting as a process of quantifier elimination are also provided.

  • 32.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Lukaszewicz, Witold
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Declarative PTIME queries for relational databases using quantifier elimination1999In: Journal of logic and computation (Print), ISSN 0955-792X, E-ISSN 1465-363X, Vol. 9, no 5, p. 737-758Article in journal (Refereed)
    Abstract [en]

    In this paper, we consider the problem of expressing and computing queries on relational deductive databases in a purely declarative query language, called SHQL (Semi-Horn Query Language). Assuming the relational databases in question are ordered, we show that all SHQL queries are computable in PTIME (polynomial time) and the whole class of PTIME queries is expressible in SHQL. Although similar results have been proven for fixpoint languages and extensions to datalog, the claim is that SHQL has the advantage of being purely declarative, where the negation operator is interpreted as classical negation, mixed quantifiers may be used and a query is simply a restricted first-order theory not limited by the rule-based syntactic restrictions associated with logic programs in general. We describe the PTIME algorithm used to compute queries in SHQL which is based in part on quantifier elimination techniques and also consider extending the method to incomplete relational databases using intuitions related to circumscription techniques.

  • 33.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Lukaszewicz, Witold
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Declarative ptime queries to relational databases.1996Report (Other academic)
    Abstract [en]

      

  • 34.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Lukaszewicz, Witold
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Efficient reasoning using the local closed-world assumption2000In: Proceedings of the 9th International Conference on Artificial Intelligence: Methodology, Systems and Applications (AIMSA), Springer Berlin/Heidelberg, 2000, p. 49-58Conference paper (Refereed)
    Abstract [en]

    We present a sound and complete, tractable inference method for reasoning with localized closed world assumptions (LCWA’s) which can be used in applications where a reasoning or planning agent can not assume complete information about planning or reasoning states. This Open World Assumption is generally necessary in most realistic robotics applications. The inference procedure subsumes that described in Etzioni et al [9], and others. In addition, it provides a great deal more expressivity, permitting limited use of negation and disjunction in the representation of LCWA’s, while still retaining tractability. The ap- proach is based on the use of circumscription and quantifier elimination techniques and inference is viewed as querying a deductive database. Both the preprocessing of the database using circumscription and quan- tifier elimination, and the inference method itself, have polynomial time and space complexity.

  • 35.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Lukaszewicz, Witold
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Explaining explanation closure1996In: Proceedings of the 9th International Symposium on Methodologies for Intelligent Systems,1996 / [ed] Zbigniew W. Ras, Maciek Michalewicz, Springer Berlin/Heidelberg, 1996, p. 521-530Conference paper (Refereed)
    Abstract [en]

    Recently, Haas, Schubert, and Reiter, have developed an alternative approach to the frame problem which is based on the idea of using explanation closure axioms. The claim is that there is a monotonic solution for characterizing nonchange in serial worlds with fully specified actions, where one can have both a succinct representation of frame axioms and an effective proof theory for the characterization. In the paper, we propose a circumscriptive version of explanation closure, PMON, that has an effective proof theory and works for both context dependent and nondeterministic actions. The approach retains representational succinctness and a large degree of elaboration tolerance, since the process of generating closure axioms is fully automated and is of no concern to the knowledge engineer. In addition, we argue that the monotonic/nonmonotonic dichotomy proposed by others is not as sharp as previously claimed and is not fully justified.

  • 36.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Lukaszewicz, Witold
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    General domain circumscription and its effective reductions.1998In: Fundamenta Informaticae, ISSN 0169-2968, E-ISSN 1875-8681, Vol. 36, no 1, p. 23-55Article in journal (Refereed)
    Abstract [en]

    We first define general domain circumscription (GDC) and provide it with a semantics. GDC subsumes existing domain circumscription proposals in that it allows varying of arbitrary predicates, functions, or constants, to maximize the minimization of the domain of a theory. We then show that for the class of semi-universal theories without function symbols, that the domain circumscription of such theories can be constructively reduced to logically equivalent first-order theories by using an extension of the DLS algorithm, previously proposed by the authors for reducing second-order formulas. We also show that for a certain class of domain circumscribed theories, that any arbitrary second-order circumscription policy applied to these theories is guaranteed to be reducible to a logically equivalent first-order theory. In the case of semi-universal theories with functions and arbitrary theories which are not separated, we provide additional results, which although not guaranteed to provide reductions in all cases, do provide reductions in some cases. These results are based on the use of fixpoint reductions.

  • 37.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Lukaszewicz, Witold
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    General domain circumscription and its first-order reduction.1996Report (Other academic)
  • 38.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Lukaszewicz, Witold
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    General domain circumscription and its first-order reduction.1996In: Proceedings of the 1st International Conference on Formal and Applied Practical Reasoning (FAPR) / [ed] Dov Gabbay, Hans Olbach, Springer Berlin/Heidelberg, 1996, p. 93-109Conference paper (Refereed)
    Abstract [en]

    We first define general domain circumscription (GDC) and provide it with a semantics. GDC subsumes existing domain circumscription proposals in that it allows varying of arbitrary predicates, functions, or constants, to maximize the minimization of the domain of a theory We then show that for the class of semi-universal theories without function symbols, that the domain circumscription of such theories can be constructively reduced to logically equivalent first-order theories by using an extension of the DLS algorithm, previously proposed by the authors for reducing second-order formulas. We also isolate a class of domain circumscribed theories, such that any arbitrary second-order circumscription policy applied to these theories is guaranteed to be reducible to a logically equivalent first-order theory. In the case of semi-universal theories with functions and arbitrary theories which are not separated, we provide additional results, which although not guaranteed to provide reductions in all cases, do provide reductions in some cases. These results are based on the use of fixpoint reductions.

  • 39.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Lukaszewicz, Witold
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Information Granules for Intelligent Knowledge Structures2003In: Proceedings of the 9th International Conference on Rough Sets, Fuzzy Sets, Data Mining and Granular Computing (RSFDGrC) / [ed] Guoyin Wang, Qing Liu, Yiyu Yao, Andrzej Skowron, Springer , 2003, p. 405-412Conference paper (Refereed)
    Abstract [en]

    The premise of this paper is that the acquisition, aggregation, merging and use of information requires some new ideas, tools and techniques which can simplify the construction, analysis and use of what we call ephemeral knowledge structures. Ephemeral knowledge structures are used and constructed by granular agents. Each agent contains its own granular information structure and granular information structures of agents can be combined together. The main concept considered in this paper is an information granule. An information granule is a concise conceptual unit that can be integrated into a larger information infrastructure consisting of other information granules and dependencies between them. The novelty of this paper is that it provides a concise and formal definition of a particular view of information granule and its associated operators, as required in advanced knowledge representation applications.

  • 40.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Lukaszewicz, Witold
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    On mutual understanding among communicating agents2003In: Proceedings of the International Workshop on Formal Approaches to Multi-Agent Systems (FAMAS) / [ed] B. Dunin-Keplicz and R. Verbrugge, 2003, p. 83-97Conference paper (Refereed)
  • 41.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Lukaszewicz, Witold
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Tolerance Spaces and Approximative Representational Structures2003In: Proceedings of the 26th German Conference on Artificial Intelligence (KI), Springer , 2003, p. 475-489Conference paper (Refereed)
    Abstract [en]

    In traditional approaches to knowledge representation, notions such as tolerance measures on data, distance between objects or individuals, and similarity measures between primitive and complex data structures are rarely considered. There is often a need to use tolerance and similarity measures in processes of data and knowledge abstraction because many complex systems which have knowledge representation components such as robots or software agents receive and process data which is incomplete, noisy, approximative and uncertain. This paper presents a framework for recursively constructing arbitrarily complex knowledge structures which may be compared for similarity, distance and approximativeness. It integrates nicely with more traditional knowledge representation techniques and attempts to bridge a gap between approximate and crisp knowledge representation. It can be viewed in part as a generalization of approximate reasoning techniques used in rough set theory. The strategy that will be used is to define tolerance and distance measures on the value sets associated with attributes or primitive data domains associated with particular applications. These tolerance and distance measures will be induced through the different levels of data and knowledge abstraction in complex representational structures. Once the tolerance and similarity measures are in place, an important structuring generalization can be made where the idea of a tolerance space is introduced. Use of these ideas is exemplified using two application domains related to sensor modeling and communication between agents.

  • 42.
    Doherty, Patrick
    et al.
    Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab. Linköping University, The Institute of Technology.
    Magnusson, Martin
    Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab. Linköping University, The Institute of Technology.
    Szalas, Andrzej
    Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab. Linköping University, The Institute of Technology.
    Approximate Databases: A support tool for approximate reasoning2006In: Journal of applied non-classical logics, ISSN 1166-3081, Vol. 16, no 1-2, p. 87-118Article in journal (Refereed)
    Abstract [en]

    This paper describes an experimental platform for approximate knowledge databases called the Approximate Knowledge Database (AKDB), based on a semantics inspired by rough sets. The implementation is based upon the use of a standard SQL database to store logical facts, augmented with several query interface layers implemented in JAVA through which extensional, intensional and local closed world nonmonotonic queries in the form of crisp or approximate logical formulas can be evaluated tractably. A graphical database design user interface is also provided which simplifies the design of databases, the entering of data and the construction of queries. The theory and semantics for AKDBs is presented in addition to application examples and details concerning the database implementation.

  • 43.
    Doherty, Patrick
    et al.
    Linköping University, Department of Computer and Information Science. Linköping University, The Institute of Technology.
    Michalak, Tomasz
    University of Southampton.
    Sroka, Jacek
    Institute of Informatics, Warsaw University.
    Szalas, Andrzej
    Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab. Linköping University, The Institute of Technology.
    Contextual Coalitional Games2011In: Proceedings of the 4th Indian Conference on Logic and its Applications (ICLA) / [ed] Mohua Banerjee, Anil Seth, Springer Berlin/Heidelberg, 2011, p. 65-78Conference paper (Refereed)
    Abstract [en]

    The study of cooperation among agents is of central interest in multi-agent systems research. A popular way to model cooperation is through coalitional game theory. Much research in this area has had limited practical applicability as regards real-world multi-agent systems due to the fact that it assumesdeterministic payoffs to coalitions and in addition does not apply to multi-agent environments that arestochastic in nature. In this paper, we propose a novel approach to modeling such scenarios where coalitional games will be contextualized through the use of logical expressions representing environmental and other state, and probability distributions will be placed on the space of contexts in order to model the stochastic nature of the scenarios. More formally, we present a formal representation language for representing contextualized coalitional games embedded in stochastic environments and we define and show how to compute expected Shapley values in such games in a computationally efficient manner. We present the value of the approach through an example involving robotics assistance in emergencies.

  • 44.
    Doherty, Patrick
    et al.
    Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab. Linköping University, The Institute of Technology.
    Skowron, Andrew
    University of Warsaw, Poland.
    Lukaszewicz, Witold
    Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab. Linköping University, The Institute of Technology.
    Szalas, Andrzej
    Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab. Linköping University, The Institute of Technology.
    Preface2003In: Fundamenta Informaticae, ISSN 0169-2968, E-ISSN 1875-8681, Vol. 57, no 2-4, p. i-iiiArticle in journal (Other academic)
  • 45.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    A correspondence framework between three-valued logics and similarity-based approximate reasoning2007In: Fundamenta Informaticae, ISSN 0169-2968, E-ISSN 1875-8681, Vol. 75, no 1-4, p. 179-193Article in journal (Refereed)
    Abstract [en]

    This paper focuses on approximate reasoning based on the use of similarity spaces. Similarity spaces and the approximated relations induced by them are a generalization of the rough set-based approximations of Pawlak [17, 18]. Similarity spaces are used to define neighborhoods around individuals and these in turn are used to define approximate sets and relations. In any of the approaches, one would like to embed such relations in an appropriate logic which can be used as a reasoning engine for specific applications with specific constraints. We propose a framework which permits a formal study of the relationship between approximate relations, similarity spaces and three-valued logics. Using ideas from correspondence theory for modal logics and constraints on an accessibility relation, we develop an analogous framework for three-valued logics and constraints on similarity relations. In this manner, we can provide a tool which helps in determining the proper three-valued logical reasoning engine to use for different classes of approximate relations generated via specific types of similarity spaces. Additionally, by choosing a three-valued logic first, the framework determines what constraints would be required on a similarity relation and the approximate relations induced by it. Such information would guide the generation of approximate relations for specific applications.

  • 46.
    Doherty, Patrick
    et al.
    Linköping University, Department of Computer and Information Science, Artificial Intelligence and Integrated Computer Systems. Linköping University, Faculty of Science & Engineering.
    Szalas, Andrzej
    Linköping University, Department of Computer and Information Science, Artificial Intelligence and Integrated Computer Systems. Linköping University, Faculty of Science & Engineering.
    An Entailment Procedure for Kleene Answer Set Programs2016In: Multi-disciplinary Trends in Artificial Intelligence. MIWAI 2016. / [ed] Sombattheera C., Stolzenburg F., Lin F., Nayak A., Springer, 2016, Vol. 10053, p. 24-37Conference paper (Refereed)
    Abstract [en]

    Classical Answer Set Programming is a widely known knowledge representation framework based on the logic programming paradigm that has been extensively studied in the past decades. Semantic theories for classical answer sets are implicitly three-valued in nature, yet with few exceptions, computing classical answer sets is based on translations into classical logic and the use of SAT solving techniques. In this paper, we introduce a variation of Kleene three-valued logic with strong connectives, R3" role="presentation">R3, and then provide a sound and complete proof procedure for R3" role="presentation">R3 based on the use of signed tableaux. We then define a restriction on the syntax of R3" role="presentation">R3 to characterize Kleene ASPs. Strongly-supported models, which are a subset of R3" role="presentation">R3 models are then defined to characterize the semantics of Kleene ASPs. A filtering technique on tableaux for R3" role="presentation">R3 is then introduced which provides a sound and complete tableau-based proof technique for Kleene ASPs. We then show a translation and semantic correspondence between Classical ASPs and Kleene ASPs, where answer sets for normal classical ASPs are equivalent to strongly-supported models. This implies that the proof technique introduced can be used for classical normal ASPs as well as Kleene ASPs. The relation between non-normal classical and Kleene ASPs is also considered.

  • 47.
    Doherty, Patrick
    et al.
    Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab. Linköping University, The Institute of Technology.
    Szalas, Andrzej
    Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab. Linköping University, The Institute of Technology.
    On the Correctness of Rough-Set Based Approximate Reasoning2010In: Proceedings of the 7th International Conference on Rough Sets and Current Trends in Computing (RSCTC) / [ed] M. Szczuka, M. Kryszkiewicz, S. Ramanna, R. Jensen, Q. Hu, Springer , 2010, Vol. 6086, p. 327-336Conference paper (Refereed)
    Abstract [en]

    There is a natural generalization of an indiscernibility relation used in rough set theory, where rather than partitioning the universe of discourse into indiscernibility classes, one can consider a covering of the universe by similarity-based neighborhoods with lower and upper approximations of relations defined via the neighborhoods. When taking this step, there is a need to tune approximate reasoning to the desired accuracy. We provide a framework for analyzing self-adaptive knowledge structures. We focus on studying the interaction between inputs and output concepts in approximate reasoning. The problems we address are: -given similarity relations modeling approximate concepts, what are similarity relations for the output concepts that guarantee correctness of reasoning? -assuming that output similarity relations lead to concepts which are not accurate enough, how can one tune input similarities?

  • 48.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    On the Correspondence between Approximations and Similarity2004In: Proceedings of the International Conference on Rough Sets and Current Trends in Computing (RSCTC) / [ed] Shusaku Tsumoto, Roman Slowinski, Jan Komorowski and Jerzy W. Grzymala-Busse, Springer , 2004, p. 143-152Conference paper (Refereed)
    Abstract [en]

    This paper focuses on the use and interpretation of approximate databases where both rough sets and indiscernibility partitions are generalized and replaced by approximate relations and similarity spaces. Similarity spaces are used to define neighborhoods around individuals and these in turn are used to define approximate sets and relations. There is a wide spectrum of choice as to what properties the similarity relation should have and how this affects the properties of approximate relations in the database. In order to make this interaction precise, we propose a technique which permits specification of both approximation and similarity constraints on approximate databases and automatic translation between them. This technique provides great insight into the relation between similarity and approximation and is similar to that used in modal correspondence theory. In order to automate the translations, quantifier elimination techniques are used.

  • 49.
    Doherty, Patrick
    et al.
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Szalas, Andrzej
    Linköping University, The Institute of Technology. Linköping University, Department of Computer and Information Science, KPLAB - Knowledge Processing Lab.
    Reasoning with Qualitative Preferences and Cardinalities Using Generalized Circumscription2008In: Proceedings of the 11th International Conference on Principles of Knowledge Representation and Reasoning (KR) / [ed] Gerhard Brewka, Jérôme Lang, AAAI Press , 2008, p. 560-570Conference paper (Refereed)
    Abstract [en]

    The topic of preference modeling has recently attracted the interest of a number of sub-disciplines in artificial intelligence such as the nonmonotonic reasoning and action and change communities. The approach in these communities focuses on qualitative preferences and preference models which provide more natural representations from a~commonsense perspective. In this paper, we show how generalized circumscription can be used as a highly expressive framework for qualitative preference modeling. Generalized circumscription proposed by Lifschitz allows for predicates (and thus formulas) to be minimized relative to arbitrary pre-orders (reflexive and transitive). Although it has received little attention, we show how it may be used to model and reason about elaborate qualitative preference relations. One of the perceived weaknesses with any type of circumscription is the 2nd-order nature of the representation. The paper shows how a large variety of preference theories represented using generalized circumscription can in fact be reduced to logically equivalent first-order theories in a constructive way. Finally, we also show how preference relations represented using general circumscription can be extended with cardinality constraints and when these extensions can also be reduced to logically equivalent first-order theories.

  • 50.
    Doherty, Patrick
    et al.
    Linköping University, Department of Computer and Information Science, Artificial Intelligence and Integrated Computer Systems. Linköping University, The Institute of Technology.
    Szalas, Andrzej
    Linköping University, Department of Computer and Information Science, Artificial Intelligence and Integrated Computer Systems. Linköping University, The Institute of Technology.
    Stability, Supportedness, Minimality and Kleene Answer Set Programs2015In: Advances in Knowledge Representation, Logic Programming, and Abstract Argumentation: Essays Dedicated to Gerhard Brewka on the Occasion of His 60th Birthday / [ed] Thomas Eiter, Hannes Strass, Mirosław Truszczynski, Stefan Woltran, Springer, 2015, p. 125-140Chapter in book (Refereed)
    Abstract [en]

    Answer Set Programming is a widely known knowledge representation framework based on the logic programming paradigm that has been extensively studied in the past decades. The semantic framework for Answer Set Programs is based on the use of stable model semantics. There are two characteristics intrinsically associated with the construction of stable models for answer set programs. Any member of an answer set is supported through facts and chains of rules and those members are in the answer set only if generated minimally in such a manner. These two characteristics, supportedness and minimality, provide the essence of stable models. Additionally, answer sets are implicitly partial and that partiality provides epistemic overtones to the interpretation of disjunctiver ules and default negation. This paper is intended to shed light on these characteristics by defining a semantic framework for answer set programming based on an extended first-order Kleene logic with weak and strong negation. Additionally, a definition of strongly supported models is introduced, separate from the minimality assumption explicit in stable models. This is used to both clarify and generate alternative semantic interpretations for answer set programs with disjunctive rules in addition to answer set programs with constraint rules. An algorithm is provided for computing supported models and comparative complexity results between strongly supported and stable model generation are provided.

123 1 - 50 of 135
CiteExportLink to result list
Permanent link
Cite
Citation style
  • apa
  • harvard1
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf