Hi Everyone,
We are pleased to announce that we have a paper accepted for presentation at the 2020 Trends in Functional Programming Symposium (http://www.cse.chalmers.se/~rjmh/tfp/).
The corresponding paper is available reading at https://www.lrde.epita.fr/dload/papers/newton.20.tfp.pdf
The article is entitled: Performance Comparison of Several Folding Strategies
Abstract: In this article we examine the computation order and consequent
performance of three different conceptual implementations of the
fold function. We explore a set of performance based
experiments on different implementations of this function. In
particular, we contrast the fold-left implementation with two
other implements we refer to as pair-wise-fold and
tree-like-fold. We explore two application areas: ratio
arithmetic and Binary Decisions Diagram construction. We demonstrate
several cases where the performance of certain algorithms is very
different depending on the approach taken. In particular iterative
computations where the object size accumulates are good candidates for
the tree-like-fold.
I hope you enjoy reading the paper, and I’m happy to have anyone’s feedback.
Kind regards
Jim NEWTON
Chers collègues,
La prochaine session du séminaire Performance et Généricité du LRDE
(Laboratoire de Recherche et Développement de l'EPITA) aura lieu le
Mardi 17 décembre 2019 (10h -- 11h), IP12A.
Vous trouverez sur le site du séminaire [1] les prochaines séances,
les résumés, captations vidéos et planches des exposés précédents [2],
le détail de cette séance [3] ainsi que le plan d'accès [4].
[1] http://seminaire.lrde.epita.fr <http://seminaire.lrde.epita.fr/>
[2] http://seminaire.lrde.epita.fr/Archives <http://seminaire.lrde.epita.fr/Archives>
[3] http://seminaire.lrde.epita.fr/2019-12-17 <http://seminaire.lrde.epita.fr/2019-12-17>
[4] http://www.lrde.epita.fr/wiki/Contact <http://www.lrde.epita.fr/wiki/Contact>
Au programme du Mardi 17 décembre 2019 :
* 10h -- 11h: Learning the relationship between neighboring pixels for some vision tasks
-- Yongchao Xu, Associate Professor at the School of Electronic Information and Communications, HUST, China
http://www.vlrlab.net/~yxu/ <http://www.vlrlab.net/~yxu/>
The relationship between neighboring pixels plays an important role in
many vision applications. A typical example of a relationship between
neighboring pixels is the intensity order, which gives rise to some
morphological tree-based image representations (e.g., Min/Max tree and
tree of shapes). These trees have been shown useful for many
applications, ranging from image filtering to object detection and
segmentation. Yet, these intensity order based trees do not always
perform well for analyzing complex natural images. The success of deep
learning in many vision tasks motivates us to resort to convolutional
neural networks (CNNs) for learning such a relationship instead of
relying on the simple intensity order. As a starting point, we propose
the flux or direction field representation that encodes the relationship
between neighboring pixels. We then leverage CNNs to learn such a
representation and develop some customized post-processings for several
vision tasks, such as symmetry detection, scene text detection, generic
image segmentation, and crowd counting by localization. This talk is
based on [1] and [2], as well as extension of those previous works that
are currently under review.
[1] Xu, Y., Wang, Y., Zhou, W., Wang, Y., Yang, Z. and Bai, X., 2019.
Textfield: Learning a deep direction field for irregular scene text
detection. IEEE Transactions on Image Processing. [2] Wang, Y., Xu, Y.,
Tsogkas, S., Bai, X., Dickinson, S. and Siddiqi, K., 2019. DeepFlux for
Skeletons in the Wild. In Proceedings of the IEEE Conference on Computer
Vision and Pattern Recognition.
-- Yongchao Xu received in 2010 both the engineer degree in electronics &
embedded systems at Polytech Paris Sud and the master degree in signal
processing & image processing at Université Paris Sud, and the Ph.D.
degree in image processing and mathematical morphology at Université
Paris Est in 2013. After completing his Ph.D. study at LRDE, EPITA,
ESIEE Paris, and LIGM, He worked at LRDE as an assistant professor
(Maître de Conférences). He is currently an Associate Professor at the
School of Electronic Information and Communications, HUST. His research
interests include mathematical morphology, image segmentation, medical
image analysis, and deep learning.
L'entrée du séminaire est libre. Merci de bien vouloir diffuser cette
information le plus largement possible. N'hésitez pas à nous faire
parvenir vos suggestions d'orateurs.
-------------- section suivante --------------
Une pièce jointe HTML a été nettoyée...
URL: <http://lists.lrde.epita.fr/pipermail/seminaire/attachments/20191211/c06b5da…>
_______________________________________________
Seminaire mailing list
Seminaire(a)lrde.epita.fr
https://lists.lrde.epita.fr/listinfo/seminaire
Bonjour à tous,
J'ai le plaisir de vous inviter à la soutenance de ma thèse intitulée :
"A new minimum barrier distance for multivariate images with
applications to salient object detection, shortest path finding, and
segmentation."
L'exposé se déroulera en anglais.
Vous êtes également cordialement invité au pot qui suivra.
Date : le mardi 17 Décembre à 14h
Lieu : en salle KB604 à l'EPITA, 14-16 rue Voltaire, Le Kremlin-Bicêtre
94270
Vous trouverez également toutes les informations sur la page web
suivante :
https://www.lrde.epita.fr/wiki/Affiche-these-MOVN
Le jury sera composé de :
Rapporteurs :
* Nicole VINCENT, Pr., Université Paris Descartes, LIPADE
* Jean-Christophe BURIE, Pr., Université La Rochelle, L3I
Examinateurs :
* Benoît NAEGEL, MdC, Université de Strasbourg, ICube
* Béatriz MARCOTEGUI, Pr., Mines ParisTech, CMM
Encadrants :
* Thierry GÉRAUD, Pr., EPITA, LRDE
* Jonathan FABRIZIO, MdC, EPITA, LRDE
Mots clés : Arbre de formes, morphologie mathématique, représentation
hiérarchique, images multivariées, pseudo-distance de Dahu, distance de
barrière minimale, saillance visuelle, document détection, segmentation
de l’image.
Résumé de la thèse :
Les représentations hiérarchiques d’images sont largement utilisées dans
le traitement d’images
pour modéliser le contenu d’une image par un arbre. Une hiérarchie bien
connue est l’arbre des formes (AdF)
qui encode la relation d’inclusion entre les composants connectés à
partir de différents niveaux de seuil.
Ce genre d’arbre est auto-dual et invariant au changement de contraste;
il est utilisé dans de
nombreuses applications de vision par ordinateur. En raison de ses
propriétés, dans cette thèse, nous utilisons
cette représentation pour calculer la nouvelle distance qui appartient
au domaine de la morphologie mathématique.
Les transformations de distance et les cartes de saillance qu’elles
induisent sont généralement utilisées
dans le traitement d’images, la vision par ordinateur et la
reconnaissance de formes. L’une des transformations
de distance les plus couramment utilisées est la géodésique.
Malheureusement, cette distance n’obtient pas
toujours des résultats satisfaisants sur des images bruitées ou floues.
Récemment, une nouvelle pseudo-distance,
appelée distance de barrière minimum (MBD), plus robuste aux variations
de pixels, a été introduite. Quelques
années plus tard, Géraud et al. ont proposé une bonne approximation
rapide de cette distance : la pseudo-distance
de Dahu. Puisque cette distance a été initialement développée pour les
images en niveaux de gris, nous proposons
ici une extension de cette transformation aux images multivariées ; nous
l’appelons la pseudo-distance vectorielle de Dahu.
Cette nouvelle distance est facilement calculée grâce à l’arbre
multivarié des formes (AdFM). Nous vous proposons
une méthode de calcul efficace de cette distance et de sa carte de
saillants dans cette thèse.
Nous étudions également les propriétés de cette distance dans le
traitement d’images floutées ou bruitées.
Pour valider cette nouvelle distance, nous fournissons des repères
démontrant à quel point la pseudo-distance
vectorielle de Dahu est plus robuste et compétitive par rapport aux
autres distances basées sur la
barrière minimum. Cette distance est prometteuse pour la détection des
objets saillants, la recherche du chemin
le plus court et la segmentation des objets. De plus, nous appliquons
cette distance pour détecter
des documents dans des vidéos. Notre méthode est une approche basée sur
les régions qui s’appuie sur
la saillance visuelle déduite de la pseudo-distance de Dahu. Nous
montrons que la performance de notre méthode
est compétitive par rapport aux méthodes de pointe utilisées dans le
concours Smartdoc 2015 ICDAR.
A bientôt,
Minh ON VU NGOC
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
13th European Lisp Symposium
Special Focus on Compilers
Call for papers
April 27 - April 28, 2020
GZ Riesbach
Zürich, Switzerland
http://www.european-lisp-symposium.org/2020
Sponsored by EPITA, Igalia S.L.
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Invited Speakers
~~~~~~~~~~~~~~~~
Andrew W. Keep (Cisco Systems, Inc.), on the Nanopass Framework.
Daniel Kochmański (Turtleware), on ECL, the Embeddable Common Lisp.
Important Dates
~~~~~~~~~~~~~~~
- Submission deadline: February 13, 2020
- Author notification: March 16, 2020
- Final papers due: April 6, 2020
- Symposium: April 27 - 28, 2020
Scope
~~~~~
The purpose of the European Lisp Symposium is to provide a forum for
the discussion and dissemination of all aspects of design,
implementation and application of any of the Lisp dialects, including
Common Lisp, Scheme, Emacs Lisp, Clojure, Racket, ACL2, AutoLisp,
ISLISP, Dylan, ECMAScript, SKILL and so on. We encourage everyone
interested in Lisp to participate.
The European Lisp Symposium 2020 invites high quality papers about
novel research results, insights and lessons learned from practical
applications, and educational perspectives. We also encourage
submissions about known ideas as long as they are presented in a new
setting and/or in a highly elegant way.
This year's focus will be directed towards "Compilers".
We especially invite submissions in the following areas:
- Compiler techniques
- Compiler passes
- Compiler compilers
- Showcasing of industrial or experimental compilers
- Code generation
- Compiler verification
- Compiler optimizations
- JIT compilers
Contributions are also welcome in other areas, including but not
limited to:
- Context-, aspect-, domain-oriented and generative programming
- Macro-, reflective-, meta- and/or rule-based development approaches
- Language design and implementation
- Language integration, inter-operation and deployment
- Development methodologies, support and environments
- Educational approaches and perspectives
- Experience reports and case studies
Technical Program
~~~~~~~~~~~~~~~~~
We invite submissions in the following forms:
* Papers: Technical papers of up to 15 pages that describe original
results or explain known ideas in new and elegant ways.
* Demonstrations: Abstracts of up to 4 pages for demonstrations of
tools, libraries, and applications.
* Tutorials: Abstracts of up to 4 pages for in-depth presentations
about topics of special interest for at least 90 minutes and up to
180 minutes.
All submissions should be formatted following the ACM SIGS guidelines
and include ACM Computing Classification System 2012 concepts and
terms. Submissions should be uploaded to Easy Chair, at the following
http://www.easychair.org/conferences/?conf=els2020
Note: to help us with the review process please indicate the type of
submission by entering either "paper", "demo", or "tutorial" in the
Keywords field.
Programme Chair
~~~~~~~~~~~~~~~
Ioanna M. Dimitriou H. - Igalia, Spain/Germany
Local Chair
~~~~~~~~~~~
Nicolas Hafner - Shinmera, Switzerland
Programme Committee
~~~~~~~~~~~~~~~~~~~
Andy Wingo - Igalia, Spain/France
Asumu Takikawa - Igalia, Spain/USA
Charlotte Herzeel - IMEC, Intel Exascience Lab, Belgium
Christophe Rhodes - Google, UK
Iréne Durand - Université Bordeaux 1, France
Jim Newton - EPITA Research Lab, France
Kent Pitman - HyperMeta, USA
Leonie Dreschler-Fischer - University of Hamburg, Germany
Marco Heisig - FAU Erlangen-Nürnberg, Germany
Mark Evenson - not.org, Austria
Max Rottenkolber - Interstellar Ventures, Germany
Paulo Matos - Igalia, Spain/Germany
Robert Goldman - SIFT, USA
Robert Strandh - Université Bordeaux 1, France
(more PC members to be announced)
--
Resistance is futile. You will be jazzimilated.
Lisp, Jazz, Aïkido: http://www.didierverna.info
Following yesterday's release of Declt, I'm happy to announce the
release of Quickref 3.0, our documentation aggregator for Quicklisp
libraries. This release is synchronized with that of Declt and the
latest Quicklisp distribution, resulting in 1792 reference manual for
Common Lisp libraries.
The official website has been updated yesterday:
https://quickref.common-lisp.net/
--
Resistance is futile. You will be jazzimilated.
Lisp, Jazz, Aïkido: http://www.didierverna.info
I'm pleased to announce the release of Declt 3.0 "Montgomery Scott", my
Texinfo reference manual generator for Common Lisp libraries.
This is a new major release of the library, although the changes may not
be so visible from the outside. The main concern in this release has
been to increase the robustness of the output, from three different
angles.
1. Many places from where Declt attempts to extract meaningful
information are underspecified (ASDF system slots notably).
2. The pretty-printing of Lisp items can be difficult, given the
liberalism and extensibility of the language.
3. Finally, several restrictions in the syntax of Texinfo itself (anchor
names notably) get in the way.
These issues were described in a paper presented at the TeX Users Group
conference, this summer in Palo Alto. This release goes a long way
toward fixing them.
The new generated reference manuals look mostly the same as before, but
some important things have changed under the hood, notably the names of
hyperlinks and cross-references (backward-incompatible, hence the
increase in the major version number). The output is now also much more
robust with respect to the final output format: the generation of HTML,
DVI, Postscript, PDF, and even plain Info should work fine and has been
tested on all Quicklisp libraries (in fact, Quickref will also be
upgraded in the near future to provide all those formats at once).
Those improvements do come at a cost. Unicode is now required (even for
Info readers). To be honest, many Lisp libraries already had that
implicit requirement. Also, this release depends on improvements and bug
fixes only available in Texinfo 6.7, now a requirement as well. I have
to thank Gavin Smith, from the Texinfo team, for his collaboration.
Apart from that, this release also has a number of bug fixes.
- Some method specializers were not handled properly.
- The manuals were missing documentation for some non-Lisp source files.
- There were some glitches in the pretty-printing of unreadable objects.
Declt's homepage:
https://www.lrde.epita.fr/~didier/software/lisp/typesetting.php#declt
--
Resistance is futile. You will be jazzimilated.
Lisp, Jazz, Aïkido: http://www.didierverna.info
Chers collègues,
La prochaine session du séminaire Performance et Généricité du LRDE
(Laboratoire de Recherche et Développement de l'EPITA) aura lieu le
Mardi 1 octobre 2019 (11h -- 12h), Amphi 4.
Vous trouverez sur le site du séminaire [1] les prochaines séances,
les résumés, captations vidéos et planches des exposés précédents [2],
le détail de cette séance [3] ainsi que le plan d'accès [4].
[1] http://seminaire.lrde.epita.fr
[2] http://seminaire.lrde.epita.fr/Archives
[3] http://seminaire.lrde.epita.fr/2019-10-01
[4] http://www.lrde.epita.fr/wiki/Contact
Au programme du Mardi 1 octobre 2019 :
* 11h -- 12h: The Loci Auto-Parallelizing Framework: An Overview and Future
Directions
-- Edward A. Luke, Professor, Department of Computer Science and Engineering,
Mississippi State University
http://web.cse.msstate.edu/~luke/loci/index.html
The Loci Auto-Parallelizing framework provides a Domain Specific
Language (DSL) for the creation of high performance numerical models.
The framework uses a logic-relation model to describe irregular
computations, provide guarantees of internal logical consistency, and
provides for automatic parallel execution. The framework has been used
to develop a number of advance computational models used in production
engineering processes. Currently Loci based tools form the backbone of
computational fluid dynamics tools used by NASA Marshall and Loci based
codes account for more than 20% of the computational workload on NASA’s
Pleiades supercomputer. This talk will provide an overview of the
framework, discuss its general approach, and provide comparisons to
other programming models through a mini-app benchmark. In addition,
future plans for developing efficient schedules of fine-grained parallel
and memory bandwidth constrained computations will be discussed.
Finally, some examples of the range of engineering simulations enabled
by the technology will be introduced and briefly discussed.
-- Dr. Ed Luke is a professor of computer science in the computer science
department of Mississippi State University. He received his Ph.D. in the
field of Computational Engineering in 1999 and conducts research at the
intersection between applied math, computer science. His research
focuses on creating systems to automatically parallelize numerical
algorithms, particularly those used to solve systems of partial
differential equations. Currently Dr. Luke is participating in active
collaborations with INRIA in Paris conducting research in the areas of
solver parallelization and mesh generation.
L'entrée du séminaire est libre. Merci de bien vouloir diffuser cette
information le plus largement possible. N'hésitez pas à nous faire
parvenir vos suggestions d'orateurs.
--
Edwin Carlinet
Laboratoire R&D de l'EPITA (LRDE)
_______________________________________________
Seminaire mailing list
Seminaire(a)lrde.epita.fr
https://lists.lrde.epita.fr/listinfo/seminaire
Greetings,
We are please to announce that the following paper has been accepted in 21st International Conference on Formal Engineering Methods (ICFEM'19). Please find the title and abstract below.
Title and corresponding authors:
"Combining Parallel Emptiness Checks with Partial Order Reductions"
Denis Poitrenaud (1,2) and Etienne Renault (3)
(1) Sorbonne Université, CNRS, LIP6, F-75005 Paris, France
(2) Université Paris Descartes, Paris, France
(3) LRDE, Epita, Paris, France
Abstract:
In explicit state model checking of concurrent systems, multicore emptiness checks and partial order reductions (POR) are two major techniques to handle large state spaces. The first one tries to take advantage of multi-core architectures while the second one may decrease exponentially the size of the state space to explore. For checking LTL properties, Bloemen and van de Pol [2] shown that the best performance is currently obtained using their multi-core SCC-based emptiness check. However, combining the latest SCC-based algorithm with POR is not trivial since a condition on cycles, the proviso, must be enforced on an algorithm which processes collaboratively cycles. In this paper, we suggest a pessimistic approach to tackle this problem for liveness properties. For safety ones, we propose an algorithm which takes benefit from the information computed by the SCC-based algorithm. We also present new parallel provisos for both safety and liveness prop- erties that relies on other multi-core emptiness checks. We observe that all presented algorithms maintain good reductions and scalability.
More information at :
https://www.lrde.epita.fr/wiki/Publications/renault.18.vecos <https://www.lrde.epita.fr/wiki/Publications/renault.18.vecos>
I'm happy to announce the the following article has been accepted in
to the 17th International Symposium on Automated Technology for
Verification and Analysis (ATVA'19), to be held on October 27-31, 2019
in Taipei, Taiwan.
Generic Emptiness Check for Fun and Profit
Christel Baier (1), František Blahoudek (2), Alexandre Duret-Lutz (3),
Joachim Klein (1), David Müller (1), and Jan Strejček (4)
(1) Technische Universität Dresden, Germany
(2) University of Mons, Belgium
(3) LRDE, EPITA, Le Kremlin-Bicêtre, France
(4) Masaryk University, Brno, Czech Republic
We present a new algorithm for checking the emptiness of ω-automata
with an Emerson-Lei acceptance condition (i.e., a positive Boolean
formula over sets of states or transitions that must be visited
infinitely or finitely often). The algorithm can also solve the
model checking problem of probabilistic positiveness of MDP under a
property given as a deterministic Emerson-Lei automaton. Although
both these problems are known to be NP-complete and our algorithm is
exponential in general, it runs in polynomial time for simpler
acceptance conditions like generalized Rabin, Streett, or parity. In
fact, the algorithm provides a unifying view on emptiness checks for
these simpler automata classes. We have implemented the algorithm in
Spot and PRISM and our experiments show improved performance over
previous solutions.
https://www.lrde.epita.fr/wiki/Publications/baier.19.atva
--
Alexandre Duret-Lutz
We are happy to announce the release of Spot 2.8
Spot is a library of algorithms for manipulating LTL formulas and
omega-automata objects (as commonly used in model checking).
This release contains code contributed by Clément Gillard and myself.
A detailed list of new features is given at the end of this email.
You can find the new release here:
http://www.lrde.epita.fr/dload/spot/spot-2.8.tar.gz
See https://spot.lrde.epita.fr/ for documentation and installation
instructions.
Note that since Debian 10 ("Buster") was released a few days ago, the
"stable" Debian packages we distribute are now built for Buster.
As always, please direct any feedback to <spot(a)lrde.epita.fr>.
New in spot 2.8 (2019-07-10)
Command-line tools:
- autfilt learned --highlight-accepting-run=NUM to highlight some
accepting run with color NUM.
- ltldo, ltlcross, and autcross are now preferring posix_spawn()
over fork()+exec() when available.
- ltlcross has new options --determinize-max-states=N and
--determinize-max-edges=M to restrict the use of
determinization-based complementation to cases where it produces
automata with at most N states and M edges. By default
determinization is now attempted up to 500 states and 5000
edges. This is an improvement over the previous default where
determinization-based complementation was not performed at all,
unless -D was specified.
- ltlcross will now skip unnecessary cross-checks and
consistency-checks (they are unnecessary when all automata
could be complemented and statistics were not required).
- genaut learned --m-nba=N to generate Max Michel's NBA familly.
(NBAs with N+1 states whose determinized have at least N! states.)
- Due to some new simplification of parity acceptance, the output of
"ltl2tgba -P -D" is now using a minimal number of colors. This
means that recurrence properties have an acceptance condition
among "Inf(0)", "t", or "f", and persistance properties have an
acceptance condition among "Fin(0)", "t", or "f".
- ltldo and ltlcross learned shorthands to call the tools ltl2na,
ltl2nba, and ltl2ngba from Owl 19.06. Similarly, autcross learned
a shorthand for Owl's dra2dpa.
Documentation:
- https://spot.lrde.epita.fr/tut90.html is a new file that explains
the purpose of the spot::bdd_dict object.
Library:
- Add generic_accepting_run() as a variant of generic_emptiness_check() that
returns an accepting run in an automaton with any acceptance condition.
- twa::accepting_run() and twa::intersecting_run() now work on
automata using Fin in their acceptance condition.
- simulation-based reductions have learned a trick that sometimes
improve transition-based output when the input is state-based.
(The automaton output by 'ltl2tgba -B GFa | autfilt --small' now
has 1 state instead of 2 in previous versions. Similarly,
'ltldo ltl2dstar -f 'GFa -> GFb' | autfilt --small' produces 1
state instead of 4.)
- simulation-based reductions hae learned another trick to better
merge states from transiant SCCs.
- acc_cond::top_disjuncts() and acc_cond::top_conjuncts() can be
used to split an acceptance condition on the top-level & or |.
These methods also exist in acc_cond::acc_code.
- minimize_obligation() learned to work on very weak automata even
if the formula or negated automaton are not supplied. (This
allows "autfilt [-D] --small" to minimize very-weak automata
whenever they are found to represent obligation properties.)
- There is a new spot::scc_and_mark_filter objet that simplifies the
creation of filters to restrict spot::scc_info to some particular
SCC while cutting new SCCs on given acceptance sets. This is used
by spot::generic_emptiness_check() when processing SCCs
recursively, and makes it easier to write similar code in Python.
- print_dot has a new option 'g', to hide edge labels. This is
helpful to display automata as "graphs", e.g., when illustrating
algorithms that do not care about labels.
- A new complement() function returns complemented automata with
unspecified acceptance condition, allowing different algorithms to
be used. The output can be alternating only if the input was
alternating.
- There is a new class output_aborter that is used to specify
upper bounds on the size of automata produced by some algorithms.
Several functions have been changed to accept an output_aborter.
This includes:
* tgba_determinize()
* tgba_powerset()
* minimize_obligation()
* minimize_wdba()
* remove_alternation()
* product()
* the new complement()
* the postprocessor class, via the "det-max-state" and
"det-max-edges" options.
- SVA's first_match operator can now be used in SERE formulas and
that is supported by the ltl_to_tgba_fm() translation. See
doc/tl/tl.pdf for the semantics. *WARNING* Because this adds a
new operator, any code that switches over the spot::op type may
need a new case for op::first_match. Furthermore, the output of
"randltl --psl" will be different from previous releases.
- The parser for SERE learned to recognize the ##n and ##[i:j]
operators from SVA. So {##2 a ##0 b[+] ##1 c ##2 e} is another
way to write {[*2];a:b[+];c;1;e}. The syntax {a ##[i:j] b} is
replaced in different ways depending on the values of i, a, and b.
The formula::sugar_delay() function implements this SVA operator in
terms of the existing PSL operators. ##[+] and ##[*] are sugar
for ##[1:$] and ##[0:$].
- The F[n:m] and G[n:m] operators introduced in Spot 2.7 now
support the case where m=$.
- spot::relabel_apply() makes it easier to reverse the effect
of spot::relabel() or spot::relabel_bse() on formula.
- The LTL simplifier learned the following rules:
F(G(a | Fb)) = FGa | GFb (if option "favor_event_univ")
G(F(a | Gb)) = GFa | FGb (if option "favor_event_univ")
F(G(a & Fb) = FGa & GFb (unless option "reduce_size_strictly")
G(F(a & Gb)) = GFa & FGb (unless option "reduce_size_strictly")
GF(f) = GF(dnf(f)) (unless option "reduce_size_strictly")
FG(f) = FG(cnf(f)) (unless option "reduce_size_strictly")
(f & g) R h = f R h if h implies g
(f & g) M h = f M h if h implies g
(f | g) W h = f W h if g implies h
(f | g) U h = f U h if g implies h
Gf | F(g & eventual) = f W (g & eventual) if !f implies g
Ff & G(g | universal) = f M (g | universal) if f implies !g
f U (g & eventual) = F(g & eventual) if !f implies g
f R (g | universal) = G(g | universal) if f implies !g
- cleanup_parity() and colorize_parity() were cleaned up a bit,
resulting in fewer colors used in some cases. In particular,
colorize_parity() learned that coloring transitiant edges does not
require the introduction of a new color.
- A new reduce_parity() function implements and generalizes the
algorithm for minimizing parity acceptance by Carton and Maceiras
(Computing the Rabin index of a parity automaton, 1999). This is
a better replacement for cleanup_parity() and colorize_parity().
See https://spot.lrde.epita.fr/ipynb/parity.html for examples.
- The postprocessor and translator classes are now using
reduce_parity() for further simplifications.
- The code for checking recurrence/persistence properties can also
use the fact the reduce_parity() with return "Inf(0)" (or "t" or
"f") for deterministic automata corresponding to recurrence
properties, and "Fin(0)" (or "t" or "f") for persistence
properties. This can be altered with the SPOT_PR_CHECK
environment variable.
Deprecation notices:
- The virtual function twa::intersecting_run() no longuer takes a
second "from_other" Boolean argument. This is a backward
incompatibility only for code that overrides this function in a
subclass. For backward compatibility with programs that simply
call this function with two argument, a non-virtual version of the
function has been introduced and marked as deprecated.
- The spot::acc_cond::format() methods have been deprecated. These
were used to display acceptance marks, but acceptance marks are
unrelated to acceptance conditions, so it's better to simply print
marks with operator<< without this extra step.
Bugs fixed:
- The gf_guarantee_to_ba() is relying on an inplace algorithm that
could produce a different number of edges for the same input in
two different transition order.
- A symmetry-based optimization of the LAR algorithm performed in
spot::to_parity() turned out to be incorrect. The optimization
has been removed. "ltlsynt --algo=lar" is the only code using
this function currently.
--
Alexandre Duret-Lutz